url
stringlengths 14
1.76k
| text
stringlengths 100
1.02M
| metadata
stringlengths 1.06k
1.1k
|
---|---|---|
https://www.lazymaths.com/smart-math/arithmetic-problem-8/
|
Categories
# [Smart Math] Arithmetic Problem 8
Here’s and example of a SMART MATH problem for ARITHMETIC.
### Problem
C’s age is 1.5 times the averages age of A, B and C. A’s age is half the average age of the three. What was their average age two years ago if the present age of B is 10 years?
1. 10 years
2. 8 years
3. 14 years
4. 7 years
5. 5 years
### The Usual Method
Let the present average age of the three be ‘x’ years.
C’s present age = 1.5x years
A’s present age = 0.5x years
B’s present age = 10 years
$\therefore$ Average age = $\frac{1.5x+0.5x+10}{3}=x$
$\therefore$x = 10 years
$\therefore$C’s present age = 1.5 x 10 = 15 years
$\therefore$A’s present age = 0.5 x 10 = 5 years
$\therefore$C’s age two years ago = 15 – 2 = 13 years
$\therefore$A’s age two years ago = 5 – 2 = 3 years
$\therefore$B’s age two years ago = 10 – 2 = 8 years
$\therefore$ Average age two years ago = $\frac{13+3+8}{3}=8$years
(Ans: 2)
Estimated Time to arrive at the answer = 75 seconds.
### Using Technique
Since, $\frac{1.5x+0.5x}{2}=x$
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 11, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6088786125183105, "perplexity": 2309.522396990317}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141184870.26/warc/CC-MAIN-20201125213038-20201126003038-00001.warc.gz"}
|
http://petebachant.me/
|
This post is a quick tutorial for preparing geometry in SolidWorks for meshing with OpenFOAM's snappyHexMesh. Note that these tips are mainly for external flows, but should generally carry over to internal geometries.
## 1. Clean it up
This step is common for all types of simulations. Unnecessary details should be removed from the CAD model such that the behavior can be simulated with minimal geometric complexity—use your judgement. For example, if simulating aerodynamics around a car body, bolt heads may be unnecessary. If working with an assembly, I prefer to save as single part first. Small features can then be removed, and the part can be simplified such that it is a single watertight object.
## 2. Get the coordinate system right
Make sure the SolidWorks model is located where you want it to be in the simulation, i.e., it is oriented properly with respect to the coordinate system of the OpenFOAM mesh. SolidWorks defines the positive z-direction normal to the front plane, whereas I prefer it normal to the top. Instead of using move/copy features, you can alternatively create a new coordinate system to select when exporting from SolidWorks.
## 3. Export to STL
Once the model is ready, select "File->Save As..." and pick "*.STL" under "Save as type:". Next, click the options button. Make sure the option to not move the model into positive space is checked, that the units are correct (OpenFOAM works in meters), and that you are saving as an ASCII, not binary, STL. Note the capital letters SolidWorks uses in the file extension by default, and that Unix file systems are case-sensitive.
You should now have an STL compatible with snappyHexMesh, and are ready to embark on the sometimes treacherous path towards its optimal meshing settings.
OpenFOAM cases are setup through a hierarchy of text files. Git is made to track changes in code, which is text. This makes Git a perfect candidate for tracking changes in simulation settings and parameters. In this post I will not really give a full picture of Git's operation and capabilities, but detail how I use Git with OpenFOAM cases.
## Why track changes?
Setting up an OpenFOAM case can be a real headache, since there are so many settings and parameters to change. In my experience, it's nice to be able to keep track of settings that worked, in case I change something and the simulation starts crashing. With Git it's easy to "check out" the last configuration that worked, and also keep track of what changes had favorable or negative effects, using commit messages.
## Branches
Git has the ability to create branches, that is, versions of files that one may switch between and edit independently. This is useful for OpenFOAM cases, because sometimes it's desirable to run multiple cases with some parameter variation. For example, one might want to run one simulation with a low Reynolds number mesh and one with a high Reynolds number mesh. One may also want a case that can be run with different versions of OpenFOAM. By putting these files on different branches, one can switch between these two sets of parameters with a single checkout command.
## Sharing, collaboration, and storage
Git has become the most popular way to share and collaborate on code, through the website GitHub. Using GitHub, another user may "fork" a set of case files, make some changes, and submit a "pull request" back to the main repository. This can be much more convenient than simply sharing zipped up folders of entire cases, as changes are more visible and reversible. Putting case files in a remote repository (private repositories can be hosted for free on BitBucket) is also a nice way to keep things backed up.
## How to set up a Git repository inside a case
The first important step is to create a .gitignore file in the top level case directory. Inside this file, you tell Git the file names or name patterns to not track. For OpenFOAM simulations, it is probably undesirable to track the actual simulation results, since this will grow the size of the repository significantly every time the simulation results are committed, i.e., all committed results would be saved for all time. My sample .gitignore file is shown below:
*.gz
log.*
postProcessing
*~
processor*
[0-9]*
[0-9]*.[0-9]*
!0.org
constant/extendedFeatureEdgeMesh
*.eMesh
constant/polyMesh
!constant/polyMesh/blockMeshDict
constant/cellToRegion
*.OpenFOAM
Note that I also ignore application logs, though these may be valuable troubleshooting tools. For example, one could checkout a previous commit, look at a log and compare with a newer version. However, these logs can get big, so I leave general performance information to the commit messages.
After a proper .gitignore file has been created, a repository can be initialized by running git init in a terminal (inside the top level case directory). Running git status will then list all the files that are not yet tracked. To add all of these files and create the first commit, run
git add -A
git commit -m "Initial commit message goes here"
Next, you may want to create a remote repository and set the remote location for the local repository. If using GitHub, create a new empty repository and follow these instructions for pointing your repository there.
Next, "push" your local changes to the remote by running
git push origin master
This pushes the changes to the master branch of the remote repo.
The case is now set to be tracked with Git. You can now continue on, and save your path towards, the perfect OpenFOAM case!
In order to perform turbine wake measurements in UNH's tow tank, I designed and built a 2-axis positioning system for our Nortek Vectrino acoustic Doppler velocimeter. Some specifications are shown in the table below:
Cross-stream (y-axis) travel 3.0 m Vertical (z-axis) travel 1.2 m Max drag force (estimated) 25 N at 2 m/s
## Components
The frame is constructed mostly from 80/20 15 series t-slot framing and hardware, with some custom brackets to mount to 1.25" pillow block linear bearings.
Attached to the frame are two Velmex BiSlide linear stages, with the y-axis driven by a stepper/belt and the z-axis driven by a stepper/ball screw. An additional igus DryLin polymer linear bearing, along with a second carriage on the z-stage BiSlide provide additional moment loading capacity.
The Vectrino probe is clamped in a cantilevered bar attached to a NACA 0020 strut. The foil is attached to an 80/20 extrusion, which mounts to the z-axis linear stage via two custom adapter blocks. Note that the adapter blocks' odd odd trapezoidal shape is due to the fact they're made from recycled material.
Two igus Energy Chain cable carriers span the horizontal and vertical axes to keep the Vectrino and motor cables from tangling during operation.
## Actuation and control
The stepper motors are driven by an ACS UDMlc drive controlled by an ACS NTM EtherCAT master controller, which provides synchronized operation with the tow and turbine axes through ACS' SPiiPlus interface. In addition to the ability to write motion control programs with their ACSPL+ language, COM, and C library, the ACSpy Python wrapper for the C library allows incorporating motion commands into Python programs, e.g., TurbineDAQ.
A full assembly of this design, minus some minor hardware, is available for download in (SolidWorks 2012 format) here. Contact me if you would like something similar designed and/or built.
Having used SolidWorks for a while, and likely entering the job market again in $O(1)$ years, I was curious as to which CAD software would be wisest to learn next, with respect to job opportunities and salaries. Looking to compile data from the current job market, I wrote Indeedy, a Python module for automatically searching Indeed.com. With this new tool I searched for "mechanical engineer" plus the names of various CAD systems, then plotted the number of results and average salary for each on the bar graph shown below.
From these results we can see demand for SolidWorks is dominant by a large margin. However, Catia and NX skills are compensated better on average, which makes sense considering their popularity in the aerospace and automotive industries. At first glance it seems there are more Catia than NX jobs, but when NX's results are combined with Unigraphics' (NX's ancestor), demand is comparable. However, this combination also comes with with a decrease in average salary, giving Catia a slight edge, and the title of Best CAD System to Learn (for me anyway).
Living in a studio apartment, storage space can be hard to come by. To have a place to store both my bike and hitch receiver bike rack, I designed and built a rolling stand that holds the rack upright, also serving as a nice work stand.
The frame is made from 80/20 t-slot extrusions and hardware, which are available at Amazon.com. On top is a 2 inch Reese hitch receiver, and below is a set of swivel casters, two of which are locking. One nice feature of this design is that it only requires cutting and tapping—no welding, drilling, or more complicated machining necessary before it's ready for assembly.
By default, the awesomely useful and fun IPython Notebook does not integrate with Windows so seamlessly. A console and then the notebook server must be opened in the proper directory in order to open a new or existing notebook. These extra steps make the IPython Notebook slightly less ideal for quickly jotting down ideas—one of its greatest uses!
Not to fear, however, as the Windows Registry can easily be modified to
1. Open notebooks directly from Windows Explorer.
2. Create a shortcut to launch the IPython Notebook server within the current directory.
These entries were put together from instructions located here. A similar solution can also be found here.
Windows Registry Editor Version 5.00
[HKEY_CLASSES_ROOT\Directory\Background\shell\ipynb]
@="Open IPython Notebook server here"
[HKEY_CLASSES_ROOT\Directory\Background\shell\ipynb\command]
@="\"C:\\Python27\\Scripts\\ipython.exe\" \"notebook\" \"%V\""
[HKEY_CLASSES_ROOT\Directory\shell\ipynb]
@="Open IPython Notebook server here"
[HKEY_CLASSES_ROOT\Directory\shell\ipynb\command]
@="\"C:\\Python27\\Scripts\\ipython.exe\" \"notebook\" \"%V\""
[HKEY_CLASSES_ROOT\ipynb_auto_file\shell\open\command]
@="\"C:\\Python27\\Scripts\\ipython.exe\" \"notebook\" \"%1\""
The code below shows how to use PyQt to create a status bar that shows the progress of an OpenFOAM simulation. Note that the arguments to re.findall may need to be tweaked depending on the specific case setup, and that the code assumes the simulation will be run in parallel, but it should be easy enough to tailor this sample to any case.
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
This program creates a progress bar for an OpenFOAM simulation
Based on http://acaciaecho.wordpress.com/2011/01/11/pyqtprogressbar/
Put this file in your case folder and run from a terminal
"""
import re
import os
import time
import numpy as np
from PyQt4 import QtCore, QtGui
def getParams():
"""Get run parameters"""
f = open("system/controlDict", "r")
if "endTime" in line:
endTime = re.findall("\d.\d+", line)
if endTime == []:
endTime = re.findall("\d+", line)
if "writeInterval" in line:
writeInterval = re.findall("\d.\d+", line)
f.close()
endTime = endTime[0]
writeInterval = writeInterval[0]
return endTime, writeInterval
procDone = QtCore.pyqtSignal(bool)
partDone = QtCore.pyqtSignal(int)
def run(self):
endTime, writeInterval = getParams()
done = False
while not done:
# Find highest valued folder in "processor0" folder
if os.path.isdir(endTime):
done = True
self.partDone.emit(100)
else:
dirs = os.listdir("processor0")
numdirs = np.array([])
for d in dirs:
try:
numdirs = np.append(numdirs, float(d))
except ValueError:
pass
self.partDone.emit(int(np.max(numdirs)/float(endTime)*100))
time.sleep(1)
self.procDone.emit(True)
def __init__(self, parent=None):
self.nameLine = QtGui.QLineEdit()
self.progressbar = QtGui.QProgressBar()
self.progressbar.setMinimum(1)
self.progressbar.setMaximum(100)
self.progressbar.setFixedWidth(400)
mainLayout = QtGui.QGridLayout()
self.setLayout(mainLayout)
self.setWindowTitle("Solving")
def updatePBar(self, val):
self.progressbar.setValue(val)
def fin(self):
sys.exit()
if __name__ == '__main__':
import sys
app = QtGui.QApplication(sys.path)
pbarwin.show()
app.exec_()
OpenFOAM runs can take a long time. Wouldn't it be nice to know when a simulation is done without having to keep checking the terminal? As it turns out, this is very easy to set up with Python (I got most of the code I used from here, which details how to send an SMS). Simply create a script called send_email.py in the OpenFOAM case directory that looks like this:
#!/usr/bin/python
import smtplib
from email.mime.text import MIMEText
server = smtplib.SMTP("smtp.gmail.com", 587)
server.starttls()
msg = MIMEText("The simulation is complete.")
msg["Subject"] = "Simulation finished"
msg["From"] = "Me"
msg["To"] = "Me"
Replace all the relevant info with your own. Note that this assumes you're using Gmail, but it can be adapted to any SMTP server. The script could also be expanded to email team members, extract information about the run, etc.
Change the permissions such that the file can be executed as a program, then at the bottom of your Allrun script add
./send_email.py
Voila. Now you'll get an email when your simulation finishes, and can go off and be productive elsewhere in the meantime.
The BeagleBone Black (BBB) is a $45 credit-card-sized computer that runs embedded Linux. I recently purchased a BBB along with an$8 4-phase stepper motor and driver to start playing around with slightly "closer-to-the-metal" motion control.
My first objective with the BBB was to simply get the stepper to move, and for this I started using the included Cloud9 IDE and BoneScript library. After a while messing around and learning a little JavaScript, along with some help from my programmer brother Tom, the stepper was moving in one direction using wave drive logic. However, the code felt awkward. I decided I'd rather use Python, since I'm a lot more comfortable with it, so I went out in search of Python packages or modules for working with the BeagleBone's general purpose I/O (GPIO) pins. I found two: Alexander Hiam's PyBBIO and Adafruit's Adafruit_BBIO. Adafruit_BBIO seemed to be more frequently maintained and better documented, so I installed it on the BeagleBone per their instructions, which was pretty quick and painless.
Next, I wrote a small, simple python module, BBpystepper, with a Stepper class for creating and working with a stepper control object. Currently the control uses full-step drive logic, with wave drive available by changing Stepper.drivemode. I may add half-stepping in the future, but for now full-stepping gets the job done.
## Usage
After installing Adafruit_BBIO and BBpystepper on the BeagleBone, the module can be imported into a Python script or run from a Python interpreter. For example:
>>> from bbpystepper import Stepper
>>> mystepper = Stepper()
>>> mystepper.rotate(180, 10) # Rotates motor 180 degrees at 10 RPM
>>> mystepper.rotate(-180, 5) # Rotates motor back 180 degrees at 5 RPM
>>> mystepper.angle
0.0
## Notes
• By default the GPIO pins used are P8_13, P8_14, P8_15, and P8_16. These can be changed by modifying the Stepper.pins list.
• By default the Stepper.steps_per_rev parameter is set to 2048 to match my motor (it has a built-in gearbox).
• The code doesn't keep track of where it ends in the sequence of pins. It simply sets all pins low after a move. This means there could be some additional error in the Stepper.angle variable if the amount of steps moved is not divisible by 4.
## Final Thoughts
As mentioned previously, half-stepping would be a nice future add-on, along with a more accurate way of keeping track of the motor shaft angle. Another logical next step would be to use one of the BBB's Programmable Real-Time Units (PRUs) to control the timing more precisely, therefore improving speed accuracy and allowing synchronization with other processes, e.g. data acquisition. However, for now this simple method gets the job done.
|
{"extraction_info": {"found_math": true, "script_math_tex": 1, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 1, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2654331922531128, "perplexity": 3109.218494985027}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440645330174.85/warc/CC-MAIN-20150827031530-00281-ip-10-171-96-226.ec2.internal.warc.gz"}
|
https://www.lettelevne.cz/08/MobileCrusher/14886.html
|
Contacts Us
Address: China Zhengzhou National High tech Industrial Development Zone
# fine grind ore mills
## Ore Grinding Mills - Mineral Processing Metallurgy
Ore Grinding Mills are used for the fine grinding as the last step in the reduction of an ore prior to concentration (gravity or flotation) or cyanidation.Practice varies, depending upon the type of ore and the amount of reduction required. In addition, some of the older properties continue with methods that perhaps are not considered the best in light of recent improvements but that cannot be ...
## Fine Grinding as Enabling Technology – The IsaMill
Three features of stirred mills that transform the economics of fine grinding are : - the high intensity attrition grinding environment - the ability to use fine grained media (eg 1 mm) to suit to the fine grained feed - the ability to use cheap natural products (local sand, slag, ore) as grinding media
## fine grind ore mills - apemonaco.org
XMQ Series Cone Ball Mill is a laboratory grinding equipment for wet fine grinding of ore (150 × 50 Cone Ball Mill also can be used for dry grinding).It is applied. Abstract “Fine grind attritional mills; can they or should they go coarser” to mill refractory sulphide gold ores to between 10 and 25 microns from a fairly standard ball.
## Mill (grinding) - Wikipedia
In materials processing a grinder is a machine for producing fine particle size reduction through attrition and compressive forces at the grain size level. See also crusher for mechanisms producing larger particles. In general, grinding processes require a relatively large amount of energy; for this reason, an experimental method to measure the energy used locally during milling with different machines was recently proposed. A typical type of fine grinder is the ball mill. A slightly inclined or horizontal rot
In materials processing a grinder is a machine for producing fine particle size reduction through attrition and compressive forces at the grain size level. See also crusher for mechanisms producing larger particles. In general, grinding processes require a relatively large amount of energy; for this reason, an experimental method to measure the energy used locally during milling with different machines was recently proposed. A typical type of fine grinder is the ball mill. A slightly inclined or horizontal rotating cylinder is partially filled with balls, usually stone or metal, which grind material to the necessary fineness by friction and impact with the tumbling balls. Ball mills normally operate with an approximate ball charge of 30%. Ball mills are characterized by their smaller (comparatively) diameter and longer length, and often have a length 1.5 to 2.5 times the diameter. The feed is at one end of the cylinder and the discharge is at the other. Ball mills are commonly used in the manufacture of Portland cement and finer grinding stages of mineral processing, one example being the Sepro tyre drive Grinding Mill. Industrial ball mills can be as large as 8.5 m (28 ft) in diameter with a 22 MW motor, drawing approximately 0.0011% of the total world's power (see List of countries by electricity consumption). However, small versions of ball mills can be found in laboratories where they are used for grinding sample material for quality assurance. The power predictions for ball mills typically use the following form of the Bond equation: $${\displaystyle E=10W\left({\frac {1}{\sqrt {P_{80}}}}-{\frac {1}{\sqrt {F_{80}}}}\right)\,}$$where • E is the energy (kilowatt-hours per metric or short ton)
Wikipedia CC-BY-SA 许可下的文字
## MicroGrinding Systems, Inc. Energy-Efficient Vibratory Mills
The VibroKinetic Energy (VKE) Mill is designed to fill the need for a fine to ultrafine grinding mill which is both low cost and economical to operate.Optional features include air classification while dry grinding, which greatly improves the efficiency of the mill. Wet grinding with hydrocyclone recycle for accurate sizing, and use of vibratory table for concentration of ore values is also ...
## HCH Ultra-fine Grinding Mill - Grinding Mills, Ultrafine ...
This mill is widely used to grind any non-metal minerals with Moh's hardness below 7 and moisture below 6%, such as talc, calcite, calciumcarbonate, dolomite, bentonite, kaolin, graphite, carbon black, etc. HCH Ultra-fine Grinding Mill features in space-saving, completeness, wide usage, easy operation, smooth performance, high performance-cost ...
## ISAMILL FINE GRINDING TECHNOLOGY AND ITS INDUSTRIAL ...
Tough economical conditions and high grade but fine-grained ore processing have forced the mining industry to look for more efficient processes. IsaMill Technology for ultra-fine and fine grinding is one such process employed at several mining operations in
## HCH Ultra-fine Grinding Mill - Grinding Mills, Ultrafine ...
This mill is widely used to grind any non-metal minerals with Moh's hardness below 7 and moisture below 6%, such as talc, calcite, calciumcarbonate, dolomite, bentonite, kaolin, graphite, carbon black, etc. HCH Ultra-fine Grinding Mill features in space-saving, completeness, wide usage, easy operation, smooth performance, high performance-cost ...
## Energy Use of Fine Grinding in Mineral Processing
B. Fine Grinding in Gravity-Induced Stirred Mills and Ball Mills Gravity-induced stirred (GIS) mills include the Tower mill, produced by Nippon Eirich, and the Vertimill, produced by Metso. Grinding to below 40 lm in GIS mills or ball mills is usually not recommended. In their product literature, Metso give 40 lm as the lower end of
## IsaMill - Wikipedia
In stirred-medium mills, the stirrers set the contents of the mixing chamber in motion, causing intensive collisions between the grinding medium and the ore particles and between the ore particles themselves. The grinding action is by attrition and abrasion, in which very fine particles are chipped from the surfaces of larger particles, rather than impact breakage.
## The Best Pepper Mills Serious Eats
Aug 23, 2017 The mill's grinding mechanism is over one and a quarter inches wide, the largest of all the mills tested, and must be responsible, at least in part, for the mill's impressive productivity: We reached a teaspoon of ground pepper in just six and a half rotations. Its major drawback, though, is fine grinding.
## Reducing Grinding Energy and Cost -Magnetite Iron Ore ...
Reducing Grinding Energy and Cost -Magnetite Iron Ore Design Case Study ... energy efficient stirred mills. Steel grinding media ... which is dominated by the cost of power required to fine grind ...
## Gold Mining Hammer Mill Crusher, Crushing Gold Ore To Fine ...
May 05, 2015 Video of our fine grinding hammer mill for gold mining and milling. We have easily replaceable hammers and armor in our mills and depending on the screen size installed in the mill it will produce ...
## The Best Countertop Grain Mills and Flour ... - Foodal
Jan 05, 2020 Every grain mill has a mechanism of some sort that will burst, crush, grind, shear, or shred grain into meal and flour, producing a variety of textures from coarse to very fine. Some burrs are more versatile than others, and are able to handle very hard or
## Ultra Fine Grinding - A Practical Alternative to Oxidative ...
Ultra Fine Grinding UFG mills overcome these limitations by the use of rotating stirrers inside a stationary mill shell. Ultra fine grinding mills have been in use for many years in a large number of every day applications such as pharmaceuticals, dyes, clays, paint and pigments before being used in the mineral processing industry.
## The Best Fine Grinding Mill Equipment For Potash Feldspar
The Best Fine Grinding Mill Equipment For Potash Feldspar . 2020-05-08 10:55; HCM; Where can we buy the potash feldspar fine grinding mill equipment? The superfine grinding mill machine produced by HCMilling(Guilin Hongcheng) company has the characteristics of high production, low consumption, environmental protection and noise reduction.
## Mineral Processing - KCGM
The ore is then ground into very fine particles less than one-fifth of one millimetre in diameter in large rotating S (Semi Autogenous Grinding) and Ball Mills which look like huge steel drums. Water is added to the process, creating mud-like slurry which is pumped to large tanks called flotation cells.
## Stirred mills - for wet grinding - Metso
The attrition grinding action, vertical arrangement, and the finer media size distribution contribute to make stirred mills energy efficient grinding machines. Vertimill® energy savings can range from 30% to greater than 50% compared with traditional mills, and the SMD energy savings can be far greater than 50% of ball mill energy in fine grinds.
## VXPmill for fine ultra fine grinding
There is a large gap between the tip speed and power intensity of the low-speed vertical mills and the high-speed horizontal mills. The VXP mill bridges the gap between high-speed and low-speed mills. The VXPmills are designed for fine and ultra fine grinding applications. In practice, the mill performs best when the feed has a normal distribution.
## Production. Copper. Interior of large ball mill showing ...
Production. Copper. Interior of large ball mill showing the iron balls which grind the ore to a fine size for subsequent treatment by flotation at one of the mills of the Utah Copper Company in Magna and Arthur in Utah. Salt Lake County Salt Lake County. United States Utah, None.
## Ultra Fine Grinding - A Practical Alternative to Oxidative ...
Ultra Fine Grinding UFG mills overcome these limitations by the use of rotating stirrers inside a stationary mill shell. Ultra fine grinding mills have been in use for many years in a large number of every day applications such as pharmaceuticals, dyes, clays, paint and pigments before being used in the mineral processing industry.
## The Best Fine Grinding Mill Equipment For Potash Feldspar
The Best Fine Grinding Mill Equipment For Potash Feldspar . 2020-05-08 10:55; HCM; Where can we buy the potash feldspar fine grinding mill equipment? The superfine grinding mill machine produced by HCMilling(Guilin Hongcheng) company has the characteristics of high production, low consumption, environmental protection and noise reduction.
## Mineral Processing - KCGM
The ore is then ground into very fine particles less than one-fifth of one millimetre in diameter in large rotating S (Semi Autogenous Grinding) and Ball Mills which look like huge steel drums. Water is added to the process, creating mud-like slurry which is pumped to large tanks called flotation cells.
## Stirred mills - for wet grinding - Metso
The attrition grinding action, vertical arrangement, and the finer media size distribution contribute to make stirred mills energy efficient grinding machines. Vertimill® energy savings can range from 30% to greater than 50% compared with traditional mills, and the SMD energy savings can be far greater than 50% of ball mill energy in fine grinds.
## VXPmill for fine ultra fine grinding
There is a large gap between the tip speed and power intensity of the low-speed vertical mills and the high-speed horizontal mills. The VXP mill bridges the gap between high-speed and low-speed mills. The VXPmills are designed for fine and ultra fine grinding applications. In practice, the mill performs best when the feed has a normal distribution.
## Production. Copper. Interior of large ball mill showing ...
Production. Copper. Interior of large ball mill showing the iron balls which grind the ore to a fine size for subsequent treatment by flotation at one of the mills of the Utah Copper Company in Magna and Arthur in Utah. Salt Lake County Salt Lake County. United States Utah, None.
## Ultra Fine Grinding Mills For Gold Ore
ultra fine grinding mills for gold ore - . Fine Grinding as Enabling Technology The IsaMill. Stirred milling was developed for fine grained ores that required an economic grind to sub 10 micron . Get Price And Support Online; Optimization of some parameters of stirred mill for ultra .
## Outotec HIGmill high intensity grinding mill - Outotec
Outotec HIGmill High Intensity Grinding Mill. With global ore grades declining and the demand for commodities continuing to increase, mining companies require new ways to cost-effectively liberate minerals and maximize recovery levels. The Outotec® HIGmill® is an advanced and energy-efficient fine and ultra-fine grinding solution that relies ...
## Raymond® Fine Grinding Roller Mill - Schenck Process
Raymond® Fine Grinding Roller Mill For finer products and increased production. The Raymond® fine grinding roller mill (US Patent Nos. 7665681 and 7963471) was specifically designed to achieve a product size distribution with d50 measurement of less than 10 microns. Available as new roller mills or as a retrofit to your existing mill.
## ore fine grinding impact mills - beblaromanina.it
Ore Grinding Mills Mineral Processing Metallurgy. Jul 11, 2016 Ore Grinding Mills are used for the fine grinding as the last step in the reduction of an ore prior to concentration (gravity or flotation) or cyanidation.Practice varies, depending upon the type of ore and the amount of reduction required.
## New and Used Ball Mills for Sale Ball Mill Supplier ...
New and Used Ball Mills for Sale Savona Equipment is a new and used Ball Mill supplier worldwide.A ball mill is a type of grinder used to grind materials into extremely fine powder for use in mineral dressing processes, paints, pyrotechnics, ceramics and selective laser sintering.
## New and Used Grinding Mills for Sale Savona Equipment
New and Used Ore Processing and Grinding Mills for Sale Savona Equipment is an ore processing and grinding mills supplier worldwide. We supply grinding mills for primary ore processing plants through to secondary and tertiary ore processing facilities. We buy, sell and consign grinding mills of all leading suppliers including Metso, , CITIC, Polysius, ABB Ltd, Polysius, Koppern and ...
## Ball Mill - an overview ScienceDirect Topics
The ball mill is a tumbling mill that uses steel balls as the grinding media. The length of the cylindrical shell is usually 1–1.5 times the shell diameter (Figure 8.11).The feed can be dry, with less than 3% moisture to minimize ball coating, or slurry containing 20–40% water by weight.
## ultra-fine grinding ... - International Mining
“This allows the VXP mill to be customised to a wide range of grinding applications,” he says. “Lower ore grades and complex minerology are driving demand for more efficient fine grinding. Typically, slurry enters the circuit at 40 to 70 microns in size, and leaves at about 15 to
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2249683290719986, "perplexity": 10795.81308568006}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107894175.55/warc/CC-MAIN-20201027111346-20201027141346-00317.warc.gz"}
|
https://www.codecogs.com/library/engineering/fluid_mechanics/floating_bodies/the-oscillation-of-floating-bodies.php
|
• https://me.yahoo.com
# The oscillation of Floating Bodies
The oscillation of floating bodies including the angle of heel and the period of oscillations
## Introduction
A single hulled ship subjected to any disturbing force will heel or roll. This page investigates the degree of roll and the periodic time of the movement.
## The Period Of Oscillation Of A Floating Body.
NOTE:
• 1) Torque = Moment of Inertia X Angular Acceleration.
• 2) The Solution of
$\frac{d^2y}{dx^2}\;+\;a^2y\;=\;0\;is\;&space;given\;&space;by\;y\;=\;A\,sin\,a\,x\;+\;B\,cos\,a\,x$
For small angles of heel, the Body can be regarded as Oscillating about it's Metacentre, in a manner similar to a Pendulum about it's point of suspension.
If
• W = The weight of the Ship
• M = The Metacentric Height
• k = The Radius of Gyration of the Ship about a horizontal Axis through the C.of.G.
• $\inline&space;\theta$ = The Rotation after a time t.
$Angular\;Acceleration&space;=&space;-\;\frac{d^2\theta&space;}{dt^2}\;(&space;Direction&space;\;towards\;&space;equilibrium\;&space;position)$
The Moment of Inertia of the ship about the C of G is $\inline&space;\frac{W}{g}k^2$ The Moment of Inertia of the ship about M is $\inline&space;&space;\frac{W}{g}\left(k^2+m^2&space;\right)$
Provided that m is small compared with k, these two can be regarded as equal
$\therefore\;\;\;\;\;I_G\approx&space;I_M$
For small angles of heel , the Righting Couple = $\inline&space;W\;m\;\theta$
$\therefore\;\;\;\;\;W\;m\;\theta&space;\;=\;I_m\times&space;-\frac{d^2\theta&space;}{dt^2}\;\approx&space;I_G\times-&space;\frac{d^2\theta&space;}{dt^2}$
$=\;-\frac{W}{g}\;k^2&space;\frac{d^2\theta&space;}{dt^2}$
$\therefore\;\;\;\;\;\frac{d^2\theta&space;}{dt^2}\;+\;\frac{gm}{k^2\theta&space;}&space;=&space;0$
The Solution of this Equation is given by:
$\theta&space;\;=\;A\;sin\sqrt{\frac{gm}{k^2}}\;t\;+\;b\;cos\sqrt{\frac{gm}{k^2}}\;t$
When t = 0 $\inline&space;\theta$= 0 and therefore B = 0
When $\inline&space;\theta$ = o and t = T/2 i.e. T = the time of a complete oscillation.
$A\;sin\sqrt{\frac{gm}{k^2}}\times\frac{T}{2}&space;=&space;0$
$Since\;A\;\neq&space;0\;\;\;\;sin\sqrt{\frac{gm}{k^2}}\times\frac{T}{2}&space;=&space;0$
The simplest solution;_
$\sqrt{\frac{gm}{k^2}}\times\frac{T}{2}\;=\;\pi$
$\therefore\;\;\;\;\;The\;Periodic\;Time\;T\;=\;2\pi&space;\sqrt{\frac{k^2}{mg}}$
### Example 1
A Solid Cylinder of Uniform material and with height equal to the Diameter is to float in water with it's axis vertical. Calculate the Metacentric hight and the specific gravity of the Cylinder so that it may have a Rolling Period of six seconds when the Diameter is 4 ft.
The Rolling Period is given by $\inline&space;T\;=\;2\pi&space;\sqrt{\frac{k^2}{hg}}\;Secs.$ Where h is the Metacentric Height and k is the Radius of Gyration of the Body about it's C of G.
$T\;=\;6Secs.\;=\;2\pi&space;\sqrt{\frac{k^2}{gh}}$
To Find H
$k^2\;=\;\frac{d^2}{16}\;+\;\frac{l^2}{12}\;=\;16\left(\frac{1}{16}\;+\;\frac{1}{12}&space;\right)\;=\;\frac{7}{3}\;ft^2$
$h\;=\;\frac{4\pi&space;^2k^2}{gT^2}\;=\;\frac{4\pi&space;^2\times&space;7/3}{32.2\times&space;36}\;=\;0.0795\;ft.$
To Find The Specific Gravity
$\frac{\pi&space;}{4}D^2\times&space;4\times&space;62.4\times&space;S\;=\;\frac{\pi&space;}{4}D^2x\times&space;62.4$
$\therefore\;\;\;\;\;\;\;s\;=\;\frac{x}{4}$
$OD\;=\;\frac{x}{2}\;\;\;\;and\;\;\;\;OG\;=\;2\;ft.$
$BM\;=\;\frac{I}{V}\;=\;\frac{\pi&space;\times&space;4^4}{64}\div&space;\frac{\pi&space;\times&space;h^2x}{4}\;=\;\frac{1}{x}$
MG (h) = BM + BO - BO - OG = $\inline&space;\frac{1}{x}\;+\;\frac{x}{2}\;-\;2\;=\;0.0795\;ft.$
$\therefore\;\;\;\;\;\;2\;+\;x^2\;-\;4.159\;=\;0$
Solving the quadratic X = 3.604 ft.
From which the Specific Gravity S = 0.901
## Worked Examples
The workings associated with these examples have been hidden. They can be seen by clicking on the red butttons.
### Example 2
A model ship weighs 80 lb. and floats in a tank of water. A ball of weight 1.6 oz. traveling horizontally at 15 ft./sec is arranged to strike the model broadside at a point 2 ft. vertically above its centre of gravity. The ball after striking, moves with the model and the amplitude of the resulting roll is $\inline&space;2^0$. The natural period of roll of the system is 1.5 sec. Find the metacentric height of the system assuming it to be constant over this amplitude. ( Assume that the vessel oscillates about its centre of gravity. (B.Sc. Part 2)
Let k be the radius of gyration of the model about the longitudinal axis through G and let the metacentric height which equals GM be h
When oscillating, the unbalanced couple, when $\inline&space;\theta$ which acts on the model is:-
$W\times&space;GM\times&space;\theta$
$\therefore&space;\;\;\;\;\;\;W\times&space;GM\times&space;\theta=W\times&space;h\times&space;\theta=\frac{W\times&space;k^2}{g}\;\;\;\frac{d^2\theta}{dt^2}$
Re-writing
$\frac{d^2\theta}{dt^2}=\frac{g\times&space;h}{k^2}\times&space;\theta$
This is Simple Harmonic motion with a periodic time of:-
$t=2\pi\sqrt{\frac{k^2}{g\times&space;h}}$
Note. For more details on Simple Harmonic Motion please see" Maths: Differential Equations: Equations of Motion"
It can be assumed that the ball, on striking the model looses its momentum and produces an impulsive torque which starts the oscillations of the model.
To consider the impact of the ball on the model, let:-
• $\inline&space;W_1$ be the weight of the ball. (Note in the question, the weight is given in ounces and therefore must be divided by 16 to convert it to lb.)
• $\inline&space;V_1$ be the velocity just before impact.
• $\inline&space;\omega_1$ be the angular velocity of the model after impact.
• "a" be the distance of the point of impact, measure from the centre of gravity (G) of the model.
• A be the maximum angular velocity of the model during its Simple Harmonic Motion.
• t be the periodic time.
$\frac{a\times&space;W_1\times&space;V_1}{g}=\frac{W\times&space;k^2\times&space;\omega_1}{g}$
Substituting in given values:-
$2\times&space;\frac{1.6}{16}\times&space;15=80\times&space;k^2\times&space;\omega_1$
$\therefore&space;\;\;\;\;\;\;k^2\times&space;\omega_1=\frac{3}{80}$
As the model starts at the centre of its harmonic motion, it can be seen that $\inline&space;omega_1$ is the maximum angular velocity.
$\text{Then}\;\;\;\;\;\;\omega_1=2\pi\frac{A}{t}\;\;=\;\;2\pi\times&space;\frac{2}{57.3\times&space;1.5}=0.146\;rad./sec.$
From equations (26) and (27)
$k^2=\frac{3}{80\times&space;0.146}$
Substituting in equation (23)
$1.5=2\pi\sqrt{\frac{3}{80\times&space;0.146\times&space;32.2\times&space;h}}$
$\text{From&space;which\;\;\;\;\;\;h=0.14\;ft.}$
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 44, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8570154309272766, "perplexity": 802.1913862104717}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662550298.31/warc/CC-MAIN-20220522220714-20220523010714-00516.warc.gz"}
|
https://scriptinghelpers.org/questions/113168/i-am-making-an-objective-system-but-the-ontouched-wont-work
|
Still have questions? Join our Discord server and get real time help.
0
# I am making an Objective system but the Ontouched wont work?
I am working on an Objective system for my horror game I am working on and after dozens of attempts, I can't seem to get it to work.
So the game starts off perfectly fine and working with it informing you on what is going on and what your goals are, and then you head off, so how it is supposed to work is when you exit the house it is supposed to activate an Ontouched script to change the objective to guide the player to walk to a building, but the Ontouched script won't work and the text wont change for some reason.
Objectiveblock1 is the block you touch when you exit the house, cancollide is false, anchored is true, transparency is 1, and I'm trying to compress the entire objective system into this one script too, but I don't know what is going wrong...
Here is the script, I would appreciate some help.
wait(8)
script.Parent.Text = ("Here is what is going on.")
wait(4)
script.Parent.Text = ("You just woke up")
wait(4)
script.Parent.Text = ("You had a very ominous dream, but you shrug it off.")
wait(5.5)
script.Parent.Text = ("And you cant wait to see your friends again!")
wait(5)
script.Parent.Text = ("So as you slowly woke up, you decided to head out.")
wait(4)
script.Parent.Text = ("Objective: Exit the house")
game.Workspace.Blocker1.Parent = game.ReplicatedStorage
function onTouched(part)
game.StarterGui.ScreenGui.Frame.TextButton.Text = ("Objective: ")
end
game.Workspace.Objectiveblock1.Touched:connect(onTouched)
Everything works until line 14.
0
Elyzzia 1023
4 days ago
assuming you're changing the text of the same button, you can change game.StarterGui.ScreenGui.Frame.TextButton.Text = ("Objective: ") to be script.Parent.Text = Objective: "
otherwise, if this is a localscript you can do
local player = game.Players.LocalPlayer -- at the top of the script
player.PlayerGui.ScreenGui.Frame.TextButton.Text = "Objective: "
for future reference, StarterGui is only cloned into the player's PlayerGui when they spawn
for updates to actually appear, you have to edit the player's PlayerGui instead
also if this is on the server then you're doing something very wrong, as messing with guis on the server is kind of incredibly nasty and bad practice
if the server needs a client to update a gui, or the server needs information from a gui, you should use a RemoteEvent instead
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.251373827457428, "perplexity": 2390.1331482792407}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400190270.10/warc/CC-MAIN-20200919044311-20200919074311-00553.warc.gz"}
|
http://www.bluegartr.com/archive/index.php/t-90889.html
|
PDA
View Full Version : Statistics halp
Julian
2010-03-25, 22:13
This is a practice test, not homework (I already know the answer, but I want to know how to find it)
An article in a journal reports that 34% of American fathers take no responsibility for child care. A researcher claims that the figure is higher for fathers in the town of Cheraw. A random sample of 225 fathers from Cheraw, yielded 97 who did not help with child care. Find the P-value for a test of the researcher's claim.
I don't really know this section that well :( I kinda get the null/alternate hypothesis stuff, for example, H0 here is that mu (would it be mu in this problem?) = .34, and H1: mu(?) > .34
Yeah I'm kinda lost, and this book is a piece of shit that isn't really helping, any help would be appreciated, thanks!
edit: Here's 2 more problems I don't really get:
In a poll of 278 voters in a certain city, 67% said that they backed a bill. The margin of error in the poll was reported as 6 percentage points (w/ a 95% degree of confidence) Which statement is correct?
A) The sample size is too small to achieve the stated margin of error
B) For the given sample size, the margin of error should be larger than stated
C) The reported margin of error is consistent with the sample size (correct answer)
D) There is not enough info to determine whether the margin of error is consistent w/ the sample size
E) For the given sample size, the margin of error should be smaller than stated.
Also another problem, except it's a poll of 390, 77% backed it, ME is 5% w/ 95% degree of confidence, and the answer is "The stated margin of error could be achieved with a smaller sample size"
Don't really get what's going on with those 2, any help would be appreciated :x thanks again
Heavencloud
2010-03-25, 23:00
H0 = u=.34
Ha = u>.34
Is a significance level given? I'm trying to remember how to solve this but it doesn't feel like I have enough information =p
Moss
2010-03-25, 23:48
Damn dude, I was doing these exact problems like 2 months ago. I'll look at my old stuff in class tomorrow.
Weiing
2010-03-26, 00:12
Ok... Let me see if I can recall stats looong time ago.
I don't know what formulae you have learned, but I was using this for my calculations:
N= 4p(1-p)/m²
Where N = sample size; p = percentage; m = margin of error
In a poll of 278 voters in a certain city, 67% said that they backed a bill. The margin of error in the poll was reported as 6 percentage points (w/ a 95% degree of confidence) Which statement is correct?
A) The sample size is too small to achieve the stated margin of error
B) For the given sample size, the margin of error should be larger than stated
C) The reported margin of error is consistent with the sample size (correct answer)
D) There is not enough info to determine whether the margin of error is consistent w/ the sample size
E) For the given sample size, the margin of error should be smaller than stated.
key: p=.67 m=.06
sticking it into the formula above:
N= (4)(.67)(1-.67)/(.06)²
= (4)(.67)(.33)/(.0036)
= .8844/.0036
= 245.667
thus, 245 is the minimum sample size needed for that margin of error. Since 278 is more than 245, then it is consistent (?????)
edit: punched in to see the sample size needed for a 5% margin of error, and you would need 353.76. so 278 falls within the range of 245-353
Also another problem, except it's a poll of 390, 77% backed it, ME is 5% w/ 95% degree of confidence, and the answer is "The stated margin of error could be achieved with a smaller sample size
key: p=.77 m=.05
N=(4)(.77)(1-.77)/(.05)²
= (4)(.77)(.23)/(.0025)
= .7084/.0025
= 283.36
Thus, the minimum sample size needed for that margin of error with that confidence interval is 283 people. 390 exceeds 283, so could have been done with fewer.
HOWEVER, I haven't done stats in over 3 years now. So I may be totally wrong and pulling stuff out of my butt. I sat here for a while punching in random numbers to try to figure it out so... I tried!
CDF
2010-03-26, 05:42
This is not a thorough explanation but may help:
1) Assuming you already covered the relevant material, the p-value to obtain is based off a test statistic that was already presented in lecture. Your textbook should have a section about "test for one proportion" or "test for a binomial proportion" where the normal approximation to the binomial distribution is applied and the corresponding test statistic given. The following is a brief summary of what is involved in the problem, implicitly or otherwise.
Stating the null and alternative hypotheses for a test of a single proportion:
http://latex.codecogs.com/gif.latex?H_0: \pi \leq .34
http://latex.codecogs.com/gif.latex?H_1: \pi > .34
Finding the sample proportion, which estimates the "true" proportion:
http://latex.codecogs.com/gif.latex?\hat{\pi} = 97/225
Obtaining the variance of the sample proportion given the null hypothesis (under which the proportion is .34):
http://latex.codecogs.com/gif.latex?Var(\hat{\pi}) = \frac{\pi(1-\pi)}{n}
You wouldn't use the standard error (based on 97/225) because the problem asks for a p-value, which is based on the null hypothesis.
Obtaining the Z-statistic:
http://latex.codecogs.com/gif.latex?Z = \frac{\hat{\pi} - \pi }{\sqrt{Var(\hat{\pi})}}
A statement of the p-value (you may use a standard normal table in the back of the book to obtain the actual value):
http://latex.codecogs.com/gif.latex?P\left ( Z > \frac{\hat{\pi} - \pi}{\sqrt{Var(\hat{\pi} )}} \right )
2) You should have been given the sample size formula for polling ("yes/no" response) somewhere. The general formula is:
http://latex.codecogs.com/gif.latex?n = \frac{Z^2\pi(1-\pi)}{(ME)^2}
where Z is the Z-score corresponding to your chosen significance level. Z is 1.96 given a 5% significance level, so Z^2 is often approximated as 4. Since you usually don't have a good idea what the true proportion really is, the margin of error is usually based on pi = .5 because that gives the maximum n required to achieve a target margin of error (you want to be sure you achieve the desired M.E.), so n is often approximated as
http://latex.codecogs.com/gif.latex?n \approx \frac{1}{(ME)^2}
Use this formula to answer the question.
3) Refer to the general sample size formula:
http://latex.codecogs.com/gif.latex?n = \frac{Z^2\pi(1-\pi)}{(ME)^2}
Note that Z^2 is actually 3.841, not 4. Also, the true proportion could be less or greater than .5. Either or both could be invoked to reach the given conclusion.
Julian
2010-03-26, 08:26
Thanks, much appreciated!
Here's another one:
Assume that a simple random sample has been selected from a normally distributed population. State the final conclusion.
Test the claim that the mean age of the prison population in one city is less than 26 years. Sample data are summarized as n=25, xbar=24.4 years, and s=9.2 years. Use a significance level of alpha = 0.05.
H0: mu=26 Ha: mu<26
(don't reject)
This problem is different than all the other problems of this type, so I'm not really sure how to go about it >_>. All the other problems have been just look at the p-value and alpha, and make a conclusion about it, which is easy, but I guess I gotta calculate the p-value for this one, and I have no idea how lol.
edit: If I did this right, I calculated t to be .8695, DF is 24, not sure how to go about calculating the p value.. :x Well, if I look at this crappy table I have, t for .1 is 1.318, so since it's less than that, it means the p-value is greater than .1 (so it's also greater than .05, the given alpha), so you don't reject?
Did I do that right? D:
Weiing
2010-03-26, 08:49
Yeah. I just calculated and got .8695 and DF of 24. Looking at my percentage points of t distribution, df of 24 with alpha of .05 shows 2.064. But since .8695 is less than that, can't reject the H0. So looks like you're right :D
Julian
2010-03-26, 09:24
Here's another:
In a survey of 5500 TV viewers, 20% said they watch network news programs. Find the standard error for the sample proportion.
(.0171)
I tried root(p(1-p)/n), which is (.2*.8/5500)^.5, but that gives .00539 <_<
EDIT: ok this one I have no clue how to do >_>
Scores on a certain test are normally distributed with a variance of 14. A researcher wishes to estimate the mean score achieved by all adults on the test. Find the sample size needed to assure with 98% confidence that the sample mean will not differ from the population mean by more than 2 units.
(20)
There's also a similar question to the above one:
The weekly earnings of students in one age group are normally distributed with a standard deviation of 81 dollars. A researcher wishes to estimate the mean weekly earnings of students in this age group. Find the sample size needed to assure with 98% confidence that the sample mean will not differ from the population mean by more than 5 dollars.
(1425)
These are the last questions in my practice exam, got everything else down, so this is it :D
Edit: trying to solve the above 2..
I basically have 5 +/- z * SE (or 2 for the first one) (z being 2.326 for 98% confidence)
Using SE = S/(n^.5) for standard deviation, and S=(variance/n)^.5 for the variance, I get 19 and 1420, which are... close enough? Especially compared to the other choices given. So I guess it's kinda right? idk <_<;
CDF
2010-03-26, 14:10
1) The reasoning on the lower bound of the p-value is fine. The t-statistic is actually -0.8695, though. Be careful if you actually have to report the p-value. To avoid making a "sign error," it could help to summarize what the p-value actually means (if this actually means anything to you):
http://latex.codecogs.com/gif.latex?P\left ( \bar{x} \leq 24.4 \right | H_0) = P(t\leq -0.8695 | H_0)
2) For n = 550, the standard error is 0.0171.
3) 20 and 1425 are based on the Z-score (2.32635) being rounded to the nearest hundredths place (2.33)
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9182250499725342, "perplexity": 762.628155116776}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394021338216/warc/CC-MAIN-20140305120858-00004-ip-10-183-142-35.ec2.internal.warc.gz"}
|
http://mathhelpforum.com/advanced-algebra/221347-systems-linear-equations-problem.html
|
# Thread: Systems of Linear Equations Problem
1. ## Systems of Linear Equations Problem
I'm having trouble solving a linear equations system with an unknown coefficient. I am trying to solve it using the reduce row echelon form, but to no avail.
x + y + 7z = -7
2x + 3y + 17z = -16
x + 2y + (k2 + 1)z = 3k
[ 1 1 7 | -7 ]
[ 2 3 17| -16]
[ 1 2 k2+1| 3k ]
Specifically, I am trying to simplify this equation to find:
a) no solution "inconsistent"
b) infinite number of solutions "consistent"
c) one solution "unique"
Right now I am just attempting a. I believe I have accidentally already solved c.
So following the procedure...
I subtract 2R1 from R2, giving in the second row:
[ 0 1 3 | -2 ]
I then subtract R1 from R3 giving in the third row:
[ 0 1 k2-6 | 3k+7 ]
Then subtracting R2 from R3 to give:
[ 0 0 k2-9 | 3k+9 ]
From here I am lost. My intuition tells me that 3rd row should simplify to something like this (x representing a non-zero value):
[ 0 0 0 | x ]
Any help would be greatly appreciated.
2. ## Re: Systems of Linear Equations Problem
From your last line simply divide through by k^2-9:
[ 0 0 1 | (3k+9)/k^2-9]
This tells you that z = (3k+9)/(k^2-9). As long as this value is defined you get a unique set of solutions for x, y, and z. For example if k=0 you get x=-1, y=1, and z = -1. If k= 4 you get x= -17, y= -11, z = 3. But notice that the (3k+0)/(k^2-9) is not defined for k= 3 or k=-3, so we need to see what happens for these two cases. For k=3 your last row becomes [0 0 0 | 15] which has no solution and for k = -3 it's [0 0 0| 0] which means the three original equations are not linearly indpendent and hence there are an infininite number of solutions.
3. ## Re: Systems of Linear Equations Problem
Hello,
I can help you with the first one. For a system of equations to be consistent (with either one or infinite solutions) it has to be satisfied that
$\det\left(\bf A\right)\neq 0$
For this reason, calculating the determinant of the coefficient matrix, you can determine the values of k where the system is not defined:
$\det\left(\bf A\right) = \left|\begin{matrix}1 & 1 & 7 \\ 2 & 3 & 17 \\ 1 & 2 & k^2 + 1 \end{matrix}\right|=k^2 -9=0$
Thus,
$k=\pm 3$ makes the determinant to be equal to zero, and thus, the system is not defined.
4. ## Re: Systems of Linear Equations Problem
Hello, brettm!
$\begin{array}{ccc} x + y +\qquad 7z &=& \text{-}7 \\ 2x + 3y +\qquad 17z &=& \text{-}16 \\ x + 2y + (k^2 + 1)z &=& 3k\end{array}$
Specifically, I am trying to simplify this equation to find:
. . (a) no solution, "inconsistent"
. . (b) infinite number of solutions, "consistent"
. . (c) one solution, "unique"
$\text{We have: }\:\left[\begin{array}{ccc|c} 1&1&7 & \text{-}7 \\ 2&3&17 & \text{-}16 \\ 1&2& k^2\!+\!1 & 3k \end{array}\right]$
$\begin{array}{c}\\ R_2-2R_1 \\ R_3-R_1 \end{array}\left[\begin{array}{ccc|c}1&1&7 & -7 \\ 0&1&3&-2 \\ 0&1&k^2\!-\!6 & 3k\!+\!7 \end{array}\right]$
$\begin{array}{c}R_1-R_2 \\ \\ R_3-R_2 \end{array}\;\left[\begin{array}{ccc|c}1&0&4&\text{-}5 \\ 0&1&3&\text{-}2 \\ 0&0&k^2\!-\!9 & 3k\!+\!9 \end{array}\right]$
$\text{If }k = 3\text{, we have: }\;\left[\begin{array}{ccc|c}1&0&4&\text{-}5 \\ 0&1&3&\text{-}2 \\ 0&0&0&18 \end{array}\right] \;\;\text{ . . . no solution}$
$\text{If }k = \text{-}3\text{, we have: }\;\left[\begin{array}{ccc|c}1&0&4&\text{-}5 \\ 0&1&3&\text{-}2 \\ 0&0&0&0 \end{array}\right]\;\;\text{ . . . infinite solutions}$
$\text{If }k \,\ne\,\pm3: \:$
$\begin{array}{c} \\ \\ R_3 \div (k^2-9)\end{array}\;\left[\begin{array}{ccc|c} 1&0&4 & \text{-}5 \\ 0&1&3&\text{-}2 \\ 0&0&1&\frac{3}{k-3}\end{array}\right]$
. . . $\begin{array}{c}R_1-4R_3 \\ R_2-3R_3 \\ \\ \end{array} \left[\begin{array}{ccc|c} 1&0&0 & \frac{\text{-}5k+3}{k-3} \\ 0&1&0& \frac{\text{-}2k-3}{k-3} \\ 0&0&1& \frac{3}{k-3} \end{array}\right]\;\;\text{ . . . one solution}$
5. ## Re: Systems of Linear Equations Problem
Thank you very much for your help everyone.
I've been trying to simplify the unique solution to the equation(s), i.e.:
$\frac{-5k+3}{k-3}$
But I'm beginning to think that I am meant to just leave this answer in terms of k, is this correct? I'm confused because the question states "Determine the value of the constant k such that the system..."
Also, what is the justification for arriving at $k = \pm3$ for no and infinite solutions? I am basically stating that for a system to have no solution, the bottom row of the matrix (excluding the augmented end) must have values of 0, with the augmented column having a non-zero value, $0x + 0y + 0z = 18$ (not possible therefore there is no solution. So I determine the value of k when $k^2 - 9 = 0$; then substituting in both values into $3k + 9$ seeing when it is a non-zero value, to find the answer for part a), no solutions. My justification for part b) is basically the same, just substituting in a value determined from $k^2 - 9 = 0$ that will give $3k + 9 = 0$ in the augmented column.
I am sure there has to be a better, more conventional way to explain this on my homework.
6. ## Re: Systems of Linear Equations Problem
Originally Posted by berni1984
Hello,
I can help you with the first one. For a system of equations to be consistent (with either one or infinite solutions) it has to be satisfied that
$\det\left(\bf A\right)\neq 0$
For this reason, calculating the determinant of the coefficient matrix, you can determine the values of k where the system is not defined:
$\det\left(\bf A\right) = \left|\begin{matrix}1 & 1 & 7 \\ 2 & 3 & 17 \\ 1 & 2 & k^2 + 1 \end{matrix}\right|=k^2 -9=0$
Thus,
$k=\pm 3$ makes the determinant to be equal to zero, and thus, the system is not defined.
This is a better solution, but we should note that a zero determinant could mean that the system either has no solutions or INFINITELY MANY solutions. So we can state that if \displaystyle \begin{align*} k \neq \pm 3 \end{align*} there is unique solution, and then the OP will need to substitute in each of k = 3 and k = -3 and reduce the system to echelon form to determine if there are no solutions or infinitely many solutions.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 23, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9233173131942749, "perplexity": 364.35193551385385}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988718034.77/warc/CC-MAIN-20161020183838-00048-ip-10-171-6-4.ec2.internal.warc.gz"}
|
http://www.physicsforums.com/showthread.php?s=6d80cb3f5caf01dd0b1c96705fafb29d&p=4542421
|
# Confinement potential
by Hluf
Tags: confinement, potential
P: 20 When we derive the Bethe-Salpeter equation, we have the relations of mass of constitute m, transverse momentum, q-hat and energy ω for equal mass constitute as ω^{2}=m^{2}+(q-hat)^{2}. What is the relations of ω,m and q-hat for low energy or long distance or low momentum? Can we say q-hat≈m? Thank you for taking time to help me!
Related Discussions High Energy, Nuclear, Particle Physics 39 High Energy, Nuclear, Particle Physics 2 High Energy, Nuclear, Particle Physics 4 Quantum Physics 3 High Energy, Nuclear, Particle Physics 17
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.960420548915863, "perplexity": 4547.9936586393105}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-15/segments/1398223211700.16/warc/CC-MAIN-20140423032011-00421-ip-10-147-4-33.ec2.internal.warc.gz"}
|
https://www.gamedev.net/blogs/entry/2259258-gamificator/
|
• entries
50
92
• views
66190
GAMIFICATOR
1264 views
Quick break from World making. Normal service will resume shortly.
So, as I looked at my life and noticed that I'm not getting ridiculously rich fast enough, I decided that I need to do something with it. I am going to gym, I study new things a lot (lately bioinformatics, astronomy, and cryptography, MOOCs are fun!), I help my friend with his project. But I wanted to do something to keep me motivated.
Enter gamification.
If you're not sure what that is, quick disclaimer, in my own words: It's using game stuff to make otherwise boring tasks fun. As you clean your dishes, you get experience points, when you get enough xp you level up! You may hate doing dishes, but being Level 41 Dirt Obliterator feels just cool, so you'll want to work just a bit more.
I decided to write a generic iOS/Android app that I could run on my phone, and level up in it. I took level up mechanics from Dungeons and Dragons, as they worked to get them solid way better than I ever could. This started as a really simple application just for myself with one screen, done in 30 minutes, but ended up as a bigger project. One that I'll upload to AppStore/Android Market even (for free), as soon as I'll make it pretty and robust enough (my girl is helping with art still, woo). So, without further ado, the mechanics:
1. Encounters
Each monster/group of monsters of strength equal roughly to yours that you meet while traveling is an encounter worth300 * level
experience points. I chose to assume each week counts as an encounter of your current level. At first you set your goals, say you want:
• 3 gym sessions
• 1 studying session
• 1 working on project session
each week. You add those, and then the experience points is divided equally between each session. If you get all or more of those, you get points assigned, for each one in each session type you miss, you get penalised. Each session type may have following kind:
• Sport
• Education
• Work/projects
• Food
• Art
• Entertainment
• Social
2. Levels
Levels in D&D (and thus in my app) are calculated with the following formula:
XPrequired = 1000 * (level + Combinatorial(level, 2));
So the required XP for levels are 0 for level 1, 1000 for level 2, 3000 for level 3, 6000 for level 4, etc. You look at your current XP and let the math figure out which level are you.
And the overview screen looks something like this:
It's not very pretty, but I need it done for tomorrow, when I'm at gym, so that I can start levelling up . According to my guesstimate (too lazy to calculate), it should be more than one year of training to reach level 20, which is when you're Epic Character in D&D standards, and after more than one year of sticking to the work, in mine as well.
Here's a quick video of the application working:
Obviously the app needs some work done before I can release it to people (because even free apps should be of SOME standard), but I really hope that you like the idea. Will be back this week with more World In 7 Days, because I've added that to my Gamificator, so now there's no escape ;).
See you then!
There are no comments to display.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.23856216669082642, "perplexity": 2049.1014495973095}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257647681.81/warc/CC-MAIN-20180321180325-20180321200325-00778.warc.gz"}
|
http://one-school.net/Malaysia/UniversityandCollege/SPM/revisioncard/physics/forceandmotion/impulse.html
|
# Impulse and Impulsive Force
Card 1: What is Impulse?
## Impulse
Impulse is defined as the product of the force (F) acting on an object and the time of action (t).
Impulse exerted on an object is equal to the momentum change of the object.
Impulse is a vector quantity.
Card 2: Formula of Impulse
## Formula of impulse
Impulse is the product of force and time.
Impulse = F × t
Impulse = momentum change
Impulse = mv - mu
Example 1
A car of mass 600kg is moving with velocity of 30m/s. A net force of 200N is applied on the car for 15s. Find the impulse exerted on the car and hence determine the final velocity of the car.
Impulse = F × t = (200) × (15) = 300oNs
Impulse = mv - mu
(3000) = 600v - 600(30)
600v = 3000 + 18000
v = 21000/600 = 35 m/s
[500,000N]
Card 3: What is impulsive force?
## Impulsive Force
Impulsive force is defined as the rate of change of momentum in a reaction.
It is a force which acts on an object for a very short interval during a collision or explosion.
Example 2
A car of mass 1000kg is traveling with a velocity of 25 m/s. The car hits a street lamp and is stopped in0.05 seconds. What is the impulsive force acting on the car during the crash?
\begin{gathered}
{\text{Impulsive Force,}} \hfill \\
F = \frac{{m(v - u)}}
{t} \hfill \\
F = \frac{{1000(0 - 25)}}
{{0.05}} \hfill \\
F = -500,000N \hfill \\
\end{gathered}
Card 4: Effects of Impulse vs Force
## Effects of impulse vs Force
A force determines the acceleration (rate of velocity change) of an object. A greater force produces a higher acceleration.
An impulse determines the velocity change of an object. A greater impulse yield a higher velocity change.
Card 5: Examples Involving Impulsive Force
## Examples Involving Impulsive Force
• Playing football
• Playing tennis
• Playing golf
• Playing baseball
Card 6: Effect of time on impulsive force 1 - Long Jump
## Long Jump
• The long jump pit is filled with sand to increase the reaction time when atlete land on it.
• This is to reduce the impulsive force acts on the leg of the atlete because impulsive force is inversely proportional to the reaction time.
Card 7: Effect of time on impulsive force 1 - High Jump
## High Jump
(This image is licenced under the GNU Free Document Licence. The original file is from the Wikipedia.org.)
• During a high jump, a high jumper will land on a thick, soft mattress after the jump.
• This is to increase the reaction time and hence reduces the impulsive force acting on the high jumper.
Card 8: Effect of time on impulsive force 1 - Jumping
## Jumping
A jumper bends his/her leg during landing. This is to increase the reaction time and hence reduce the impact of impulsive force acting on the leg of the jumper.
Card 9: Empty Card
Card 10: Empty Card
Table of Content
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8458269834518433, "perplexity": 2116.513758042424}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202131.54/warc/CC-MAIN-20190319203912-20190319225912-00101.warc.gz"}
|
http://www.physicsforums.com/showthread.php?t=20635
|
## [SOLVED] Re: "The other quaternions": how useful are 2x2 real matrices to
<jabberwocky><div class="vbmenu_control"><a href="jabberwocky:;" onClick="newWindow=window.open('','usenetCode','toolbar=no,location=no, scrollbars=yes,resizable=yes,status=no,width=650,height=400'); newWindow.document.write('<HTML><HEAD><TITLE>Usenet ASCII</TITLE></HEAD><BODY topmargin=0 leftmargin=0 BGCOLOR=#F1F1F1><table border=0 width=625><td bgcolor=midnightblue><font color=#F1F1F1>This Usenet message\'s original ASCII form: </font></td></tr><tr><td width=449><br><br><font face=courier><UL><PRE>\n\nDoug Sweetser wrote:\n\n>>The essential difference between the two is that, for the quaternions,\n>>the squares of all three of the non-real unit vectors i,j,k, are -1.\n>\n> If Hamilton would have chosen a righthanded system, this would be +1 I\n> believe.\n\nThis is like saying that if Gauss had chosen a different convention\nfor the complex numbers then i^2=1.\n\nYou\'d be _much_ more careful about what you post.\nIn a couple of years you\'ll probably be ashamed\nof what you posted in the last few months.\n\nArnold Neumaier\n\n</UL></PRE></font></td></tr></table></BODY><HTML>');"> <IMG SRC=/images/buttons/ip.gif BORDER=0 ALIGN=CENTER ALT="View this Usenet post in original ASCII form"> View this Usenet post in original ASCII form </a></div><P></jabberwocky>Doug Sweetser wrote:
>>The essential difference between the two is that, for the quaternions,
>>the squares of all three of the non-real unit vectors i,j,k, are -1.
>
> If Hamilton would have chosen a righthanded system, this would be $+1 I$
> believe.
This is like saying that if Gauss had chosen a different convention
for the complex numbers then $i^2=1$.
You'd be _much_ more careful about what you post.
In a couple of years you'll probably be ashamed
of what you posted in the last few months.
Arnold Neumaier
PhysOrg.com physics news on PhysOrg.com >> Iron-platinum alloys could be new-generation hard drives>> Lab sets a new record for creating heralded photons>> Breakthrough calls time on bootleg booze
Hello: Could someone remind me of how the difference in choice of handedness affects the algebra of quaternions? I cannot find the reference I had on the issue. Hamilton's way of writing them is only one representation. By the way, Gauss was the first to discover quaternions. It was in one of his notebooks. doug
Vaughan Pratt wrote: > While I'm familiar with some (but perhaps not all) of the applications > of the geometry of the quaternions to physics, I haven't encountered > any for the corresponding geometry of the 2x2 matrices. I'd therefore > greatly appreciate hearing about their applications to physics as a > noncommutative geometry. Commonplace physical phenomena especially > welcome. A similar issue arises with the Dirac algebra on spacetime - that is CL(3,1) has a real representation in terms of 4x4 real matrices (Majorana representation) (likewise CL(2,2)), while CL(1,3) is 2x2 quaternions. In the physical world, the Majorana representation and its spinors don't seem to correspond to real particles. So, while there may be some special configuration of, say, photons with a special polarization that have CL(1,1) as a basis, one shouldn't necessarily expect a general correspondence with physical systems. In the Dirac schema the Majorana particle is a neutral Fermion (massive or not) that is its own antiparticle, and this latter is what makes it unphysical (that is, it's not really a lepton). So the real physics (and mystery) is living in the idea of the unit pseudoscalar on whatever Clifford algebra comes under consideration. If Gauss were alive he'd say "The true metaphysics of $\gamma_5$ is hard" :) $$-drl$$
## [SOLVED] Re: "The other quaternions": how useful are 2x2 real matrices to
<jabberwocky><div class="vbmenu_control"><a href="jabberwocky:;" onClick="newWindow=window.open('','usenetCode','toolbar=no,location=no, scrollbars=yes,resizable=yes,status=no,width=650,height=400'); newWindow.document.write('<HTML><HEAD><TITLE>Usenet ASCII</TITLE></HEAD><BODY topmargin=0 leftmargin=0 BGCOLOR=#F1F1F1><table border=0 width=625><td bgcolor=midnightblue><font color=#F1F1F1>This Usenet message\'s original ASCII form: </font></td></tr><tr><td width=449><br><br><font face=courier><UL><PRE>Doug Sweetser wrote:\n\n> Could someone remind me of how the difference in choice of handedness\n> affects the algebra of quaternions? I cannot find the reference I had\n> on the issue. Hamilton\'s way of writing them is only one\n> representation.\n\nIt has nothing at all to do with it - quaternions as such are not tied\nto space geometry. But, if you make j correspond to the z-axis and k to\nthe y-axis, then you\'ve adopted a left-handed convention.\n\n> By the way, Gauss was the first to discover quaternions. It was in one\n> of his notebooks.\n\nI don\'t know where you "discovered" this fact but I\'d bet my last penny\non it being false. I suppose we must now add Hamilton to the growing\nlist of post-modernist bashees.\n\n-drl\n\n</UL></PRE></font></td></tr></table></BODY><HTML>');"> <IMG SRC=/images/buttons/ip.gif BORDER=0 ALIGN=CENTER ALT="View this Usenet post in original ASCII form"> View this Usenet post in original ASCII form </a></div><P></jabberwocky>Doug Sweetser wrote:
> Could someone remind me of how the difference in choice of handedness
> affects the algebra of quaternions? I cannot find the reference I had
> on the issue. Hamilton's way of writing them is only one
> representation.
It has nothing at all to do with $it -$ quaternions as such are not tied
to space geometry. But, if you make j correspond to the z-axis and k to
the y-axis, then you've adopted a left-handed convention.
> By the way, Gauss was the first to discover quaternions. It was in one
> of his notebooks.
I don't know where you "discovered" this fact but I'd bet my last penny
on it being false. I suppose we must now add Hamilton to the growing
list of post-modernist bashees.
$$-drl$$
Danny Ross Lunsford wrote: > It has nothing at all to do with $it -$ quaternions as such are not tied > to space geometry. But, if you make j correspond to the z-axis and k > to the y-axis, then you've adopted a left-handed convention. That is the issue I was referring too. >> By the way, Gauss was the first to discover quaternions. It was in >> one of his notebooks. > > I don't know where you "discovered" this fact but I'd bet my last > penny on it being false. I was surprised to read it too. Here is what some googling revealed: from http://www.infoplease.com/ce6/people/A0820346.html: Gauss was extremely careful and rigorous in all his work, insisting on a complete proof of any result before he would publish it. As a consequence, he made many discoveries that were not credited to him and had to be remade by others later; for example, he anticipated Bolyai and Lobachevsky in non-Euclidean geometry, Jacobi in the double periodicity of elliptic functions, Cauchy in the theory of functions of a complex variable, and Hamilton in quaternions. However, his published works were enough to establish his reputation as one of the greatest mathematicians of all time. It was only noticed many years after the fact. Hamilton should never be mentioned solo anyway, since Rodriguez discovered them independently, and with a purpose in mind, namely rotations. These are just historical tales. Hamilton gets so much more space because he wrote about the process of this particular discovery, which is not common. Few facts are even known about Rodriguez. He may have been a banker of Spanish decent, but my memory is vague here too. doug
In article , Doug Sweetser wrote: >Could someone remind me of how the difference in choice of handedness >affects the algebra of quaternions? I cannot find the reference I had >on the issue. Hamilton's way of writing them is only one >representation. A vector is a quaternion with a zero real part, and a unit vector x is one satisfying $x^2 = -1, i$.e. quaternion multiplication supplies a norm. Pick any two unit vectors x,y such that 1,x,y,xy are linearly independent. (So the multiplication supplies not just a norm but an inner product.) The set of coordinate systems of the form x,y,xy obtainable in this way is closed under rigid rotations. The set of those of the form $x,y,-xy (or y,x,xy, or x,y,yx)$ is also so closed, and consists of the reflections of the other set about an odd number of axes. Call these two sets respectively the right- and left-handed systems. The beauty of this definition is that it depends only on the operations of the algebra of quaternions, and is independent of any notational conventions Hamilton or anyone else might use. Is handedness defined $and/or$ useful for the enumeration i,1,j? (I'm asking, I don't know.) If either, then I believe there can be only two internally consistent nontrivial conventions, describable as follows. Take the 6 oriented necklaces containing beads 1,i,j,k (one of each) and take the handedness of the necklace obtained by deleting the bead opposite 1 to be always equal to that of the necklace obtained by deleting 1. The other convention is to make them always not equal. Reasoning: the position of 1 in (e.g.) i,1,j should not affect handedness (1 is not physical and should commute with everything), only the order of i and j. (Then there's the two trivial conventions: make all triples containing 1 right-handed, or its opposite convention, but I would become suddenly and violently ill if there were any mathematical reason for having to choose between those two conventions.) What I don't see is why the order of i and j should affect anything. Is there in fact a reason to define handedness of i,1,j? For example is there a natural homology of the quaternions that entails a handedness? >By the way, Gauss was the first to discover quaternions. It was in one >of his notebooks. Another area where I feel rather in the dark. I'd be interested in knowing a little more detail about exactly what was in the notebook. He wrote in an 1846 letter to Schumacher about Lobachevsky's geometry that he had "had the same conviction for 54 years," i.e. since the age of 15, but it is one thing to have a conviction and quite another to have the courage of that conviction to come out with it. Did he keep quiet about his ideas on nonstandard geometries because he thought they wouldn't be accepted, or because of a concern the authorities might treat him like Galileo for such spatial heresies, or because he didn't think the world was ready for it yet (he tended to act that way concerning his use of complex numbers in number theory, which he was in the habit of translating away), or because he was having difficulty getting the details to work out right? Hamilton famously had trouble figuring out the quaternions, until it hit him in a flash that the real axis should be a *separate* dimension from the three physical dimensions, commuting with everything while the physical dimensions had to anticommute. (Having that sudden insight on such a fundamental question as the algebra of 3-space must have been fun!) Vaughan Pratt -- Don't contact me at [email protected], substitute cs for boole instead.
Doug Sweetser wrote: [....] "It must be true, I found it on the internet!" If Gauss had acutally invented non-commutative algebra, he would get credit for it, insofar as it behind most all of particle physics and a lot of relativity. The true story of quaternions is well-known. The key idea only occured to Hamilton after 10 years of long hard thinking. $$-drl$$
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.776992678642273, "perplexity": 1514.7190886801175}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-20/segments/1368699881956/warc/CC-MAIN-20130516102441-00074-ip-10-60-113-184.ec2.internal.warc.gz"}
|
https://mathematica.stackexchange.com/questions/121466/weird-bug-with-esc-string-and-underscore
|
# Weird bug with Esc, string, and underscore? [closed]
There seems to be a bug with esc key, strings, and underscores.
For instance, see the following code:
a = f["1_2"];
If an esc character is typed anywhere between the start of the line and the letter f, the first quotation mark is no longer recognized.
It appears that this occurs only when:
• the string contains at least one underscore and at least one number (can have other characters too)
• the string needs to have some sort of brackets (parentheses, curly braces, etc.) around it.
Trying to evaluate the line when the phenomenon occurs leads to an error.
Typing anything other than esc before the second quotation mark fixes the line.
What is causing this?
## closed as off-topic by MarcoB, m_goldberg, C. E., corey979, FeyreFeb 6 '17 at 11:58
This question appears to be off-topic. The users who voted to close gave this specific reason:
• "The question is out of scope for this site. The answer to this question requires either advice from Wolfram support or the services of a professional consultant." – MarcoB, m_goldberg, C. E., corey979, Feyre
If this question can be reworded to fit the rules in the help center, please edit the question.
• I've seen similar kinds of problems. See, for instance, here. There, it's a problem with superscripts. – march Jul 25 '16 at 22:20
• Also, here and here. – march Jul 25 '16 at 22:37
• Also, double esc before a, then double backspace and highlighting is broken. – Kuba Feb 6 '17 at 8:19
• Probably related: undo after copying a string inside a string – Kuba Feb 6 '17 at 8:37
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.19195371866226196, "perplexity": 2100.8399757939915}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986697439.41/warc/CC-MAIN-20191019164943-20191019192443-00452.warc.gz"}
|
http://www.gamedev.net/topic/632378-how-to-manage-input-from-more-than-one-potential-source/?p=4987861
|
• Create Account
Banner advertising on our site currently available from just \$5!
# How to manage input from more than one potential source?
Old topic!
Guest, the last post of this topic is over 60 days old and at this point you may not reply in this topic. If you wish to continue this conversation start a new topic.
9 replies to this topic
### #1BUnzaga Members - Reputation: 267
Like
0Likes
Like
Posted 06 October 2012 - 05:37 AM
I am trying to put together a thorough input system, and I can't seem to work it through in my head (or in code).
Say for example, I want to use x number of keys, or mouse button(s) or joystick button(s) to perform a single continuous action, for example 'running forward'.
I can handle gathering the input, and I can handle associating the event with an action, but what I can't figure out, is the best way to cross-check for the current state of the 'other' inputs.
For discussion sake, lets say a key event would be something like... 'key_space', 'mouse_1', 'joy_1', followed by the appropriate data, such as 'true', 'false', .0121, etc.
So say I bound 'key_w', 'key_up', as well as 'mouse_2' to the action 'forward'.
So now if I press either of those keys, or the right mouse button, the 'forward' action will happen, meaning, the player will run 'forward'.
So now say I press all three, this works fine, because they all do the same function. But what happens when I let go of the right mouse button? In code, it would send through a 'mouse_2' event with the data 'false', which should cause the player to stop moving forward, unfortunately, both key_w, and key_up, are still down...
Is there any way to tackle something like this? Where I could have (n) number of inputs to handle a single thing like this? I know if I have a finite number of options, I could just check for the state of the keys when the 'up' event occurs... if(false == keys_w){/*stop*/}else{/*keep going*/}
I thought of using an integer for this, but that just seems sloppy... Like if I press key_w, held_amt++, if I press key_up, held_amt++, if I press mouse_2, held_amt++. Then when I release one of them, I do held_amt--, check if it is equal to zero, and then if it is, stop running...
I don't know, this just seems very sloppy to me.
On top of this, say I pressed forward, then back, then a different forward key. I would want it to respond properly to the currently held down keys, all of which could be reassigned and the amounts of them could be 1 to n.
By the way, I have already read through http://www.gamedev.n...stem-for-games/ which was informative and helpful, but it doesn't cover my issue. I have also tried looking up anything I can find in google, as well as on these boards, but most of them have to do with the lower level logic of how to just get multiple sources of input, rather than the higher functionality of how to orchestrate it.
EDIT: The only other method I was thinking of was polling the state of each potential input onUpdate, but I wanted use a callback system instead, onKeyChange... onKeyUp / onKeyDown type of thing.
Edited by BUnzaga, 06 October 2012 - 05:45 AM.
### #2MrDaaark Members - Reputation: 3555
Like
1Likes
Like
Posted 06 October 2012 - 11:57 AM
You don't seem to have a problem here as much as you just seem to have a bad design. Don't handle everything at once, just handle them separately.
Check the keyboard -- if a key is down, take the appropriate action in your program.
Check the mouse -- if a button is down, take the appropriate action in your program.
Check the gamepad -- of a button is down, take the appropriate action in your program.
Do this in a generic way, and just loop over all valid input methods. Better yet, if it's a game, let the player choose the preferred input method, and check only that. With the exception that you always want to check the keyboard for important keys like ESC or whatever else.
I have a Pen + Touch Wacom tablet plugged in, (like many people) and I don't expect any programs to account for it separably from my mouse. If I move my finger or pen on my tablet while moving my mouse, I'm going to get messed up cursor navigation. Especially since they both handle cursor movement very differently! There is no expectation of me (or any other user) that I should be able to make use of both at the same time.
### #3nobodynews Crossbones+ - Reputation: 2195
Like
1Likes
Like
Posted 06 October 2012 - 12:10 PM
I thought of using an integer for this, but that just seems sloppy... Like if I press key_w, held_amt++, if I press key_up, held_amt++, if I press mouse_2, held_amt++. Then when I release one of them, I do held_amt--, check if it is equal to zero, and then if it is, stop running...
I don't know, this just seems very sloppy to me.
Doesn't seem that sloppy to me given what you want to do.
What you could do is have an enum of possible actions (Action) mapped to an integer to count the number of inputs that are applied to that action like this:
std::map<Action, int> ActionBindings;
And for the callback system you want to associate a key with an action in another map like this:
std::map<Key, Action> KeyBindings;
When OnKeyUp/Down is called, map from the Key to the Action and then from the Action to the count for that Action. Then in the part of the code that performs actions just check for which actions are currently occurring and do actions based on that. Hope this helps.
C++: A Dialog | C++0x Features: Part1 (lambdas, auto, static_assert) , Part 2 (rvalue references) , Part 3 (decltype) | Write Games | Fix Your Timestep!
### #4nobodynews Crossbones+ - Reputation: 2195
Like
0Likes
Like
Posted 06 October 2012 - 12:24 PM
Oh if you need ALL keys to be applied then you'll need to keep track of that count as well. I'd do this by modifying the type of ActionBindings map to use a structure linstead like this:
struct KeyCount
{
int NecessaryCount;
int CurrentCount;
};
Then define ActionBindings like this instead:
std::map<Action, KeyCount> ActionBindings;
Then when you change the mapping of which keys go to which action you adjust NecessaryCount accordingly.
C++: A Dialog | C++0x Features: Part1 (lambdas, auto, static_assert) , Part 2 (rvalue references) , Part 3 (decltype) | Write Games | Fix Your Timestep!
### #5BUnzaga Members - Reputation: 267
Like
0Likes
Like
Posted 06 October 2012 - 01:43 PM
You had it right the first time nobodynews, it looks like the integer approach will be what I'll go with.
Daaark, you hit the nail on the head with your example. Say I do have both a pen and a mouse attached (for some crazy reason). I want to be able to take either a pen movement or a mouse movement, and apply both to the same action, but when one stops, and the other does not, I need a way to determine that:
1. The pen stopped.
2. The mouse is still moving.
3. Reflect this within the game.
By using nobodynews example, it would go something like...
1. Start moving mouse along the positive x axis, the mouse_px event will be triggered with a '1' value passed in (just as an example).
2. The px_axis_int will be incremented by +1.
3. Start moving pen or finger also along the positive x axis, the pen_px or finger_px event will be triggered with a '1' value passed in.
4. The px_axis_int will be incremented by +1 for the pen and or finger.
Since all three of these actions 'mouse_px', 'pen_px', and 'finger_px' are tied to the same event... uh... 'move_right', lets say... dVal.x = 1.0 So now the character starts moving along the x axis.
Now say I stop moving the mouse, I'll have to set px_axis_int -=1, then check if (px_axis_int == 0){dVal.x = 0.0}. I would do this for each input assigned to this action.
Unless someone has a better way to accomplish this? Of course these are just obscure examples, but I just like to cover all my bases... Most people wouldn't assign 'W', 'Up Arrow', 'Right Mouse Button', 'Joystick Button 8', all to the same action, and use them interchangeably, but 'what if'? I just want a system that works for whatever the user wants to end up doing...
### #6slmgc Members - Reputation: 168
Like
1Likes
Like
Posted 07 October 2012 - 12:12 PM
So say I bound 'key_w', 'key_up', as well as 'mouse_2' to the action 'forward'.
So now if I press either of those keys, or the right mouse button, the 'forward' action will happen, meaning, the player will run 'forward'.
So now say I press all three, this works fine, because they all do the same function. But what happens when I let go of the right mouse button? In code, it would send through a 'mouse_2' event with the data 'false', which should cause the player to stop moving forward, unfortunately, both key_w, and key_up, are still down...
Is there any way to tackle something like this?
Hey I myself usually do this:
I have an array of "actions" associated with keys, like "run forward", "jump", etc. And I do have several "keysets" which are bound to same "actions".
Something like:
{
"Jump": ['space', 'button 1'] // "action": [set_1, set_2]
"Run": ['shift', 'button 9'] // "another action": [set_1, set_2]
...
}
So, when I press any button, from any input, my engine sends an "action" event, associated with this button. And with the event I also send the index of the "set", which activated this event. For example, if I press "space" on my keyboard it sends something like: (action: "jump", keyset: 1) and when I press "button 1" on my gamepad, it sends the same "action-event", but with different keyset index (2 in this case). And now my InputManager can figure out which key was pressed and if I will press down "shift" AND "button 9", and release "shift" after that, InputManager would know, that the released button was from another set and will not switch off the current "action".
### #7BUnzaga Members - Reputation: 267
Like
0Likes
Like
Posted 07 October 2012 - 08:45 PM
slmgc, From what you are describing, it sounds like you are putting your actions into a queue or list of some sort, is that safe to assume? Then when you 'remove' or cancel the action, it removes only the one associated to that set or index.
It is an interesting approach, it sort of turns duplicate actions into unique actions via your indexing (set) system.
Thank you for sharing that with me.
### #8slmgc Members - Reputation: 168
Like
0Likes
Like
Posted 07 October 2012 - 10:44 PM
slmgc, From what you are describing, it sounds like you are putting your actions into a queue or list of some sort, is that safe to assume? Then when you 'remove' or cancel the action, it removes only the one associated to that set or index.
It is an interesting approach, it sort of turns duplicate actions into unique actions via your indexing (set) system.
Thank you for sharing that with me.
You can use bitwise operations on the "action" flag, for example: you have two keysets (1 and 2). Action flag will be 0 when no associated keys were pressed. When an action event comes (from keyset 2, for example), you set a bit flag with an index of the keyset for this action (so the action flag becomes 0010). When comes this action-event again, but with another keyset (1), you set a different bit (so now you flag is 0011). When a user releases a pressed key, you receive a reset event for this action for specific keyset (let it be 2), so you reset a specific bit of the action flag (it becomes 0001). When a user releases all pressed keys on all of the input devices, you just reset all of the bits of the action flag (so it becomes 0 again). The main plus of this approach is that you just have to compare if a specific action flag is zero or not, if not, then some key (associated with this action) from one of the input devices is pressed, if the flag becomes zero, then this action ends.
Hope this helps ;)
### #9zfvesoljc Members - Reputation: 450
Like
0Likes
Like
Posted 08 October 2012 - 01:20 AM
• Split inputs and actions, use "binds" to connect them
• Pool input states (for each input device), keep track of previous and current state, based on this you can handle the difference (press, holding, release)
• Based on input states, generate game actions, feed those to game logic
Then it becomes something like this:
// read device
for each input device
loop over every possible input (key, button, axis)
get current value (0.0-1.0), mark current time (useful for various things)
// input 2 action
for each action
for each bound input
action_value = max( inputdevice.currentvalue[bound_input], action_value );
// action logic
for each action
if action_value > prev_value && prev_value == 0 -> action was "pressed"
if action_value == 0 && prev_value > 0 -> action was "released"
if action_value > 0 && prev_value > 0 -> action is "held"
prev_value = action_value;
This works nice even for analog values; instead of making IsAction* queries, you can say GetActionValue and return the current action value (0.76 from gamepad analog stick). By tracking time on action, you can even detect difference between fast or long button press.
### #10haegarr Crossbones+ - Reputation: 4836
Like
0Likes
Like
Posted 08 October 2012 - 01:43 AM
It is IMHO wise to map input to actions with the sense to separate both subsystems. On the action side the things should be unique. Hence a way of making multiple equivalent input unique is needed.
When looking at different input methods it becomes clear that the first step would be in transforming input to meaningful values. E.g. do you need a trigger, then a key released-to-pressed transition may be the choice. If you need a temporary boolean state, then the key pressed state may be a choice. When you need a stable state, then you can use toggling by key presses of the same or even 2 different keys. As a strange example, you can transform a mouse move into a trigger when detecting a move by at least L units in horizontal direction. However, in the OP you want to have a temporary boolean state from key presses.
If all input sources for a given task are "normalized" in the above sense, functions may be used to mix different input signals. Typical functions would be min(...), max(...), and perhaps mean(...). E.g. applying min(...) to N state signals means that all of the signals need to be 1 to generate a 1 as output; similarly applying max(...) to N state signals means that any of the signals causes 1 as output. This may be implemented using an enum or using an approach with counting and comparing against a threshold. In the OP you want to use the min(...) function.
Old topic!
Guest, the last post of this topic is over 60 days old and at this point you may not reply in this topic. If you wish to continue this conversation start a new topic.
PARTNERS
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1995101422071457, "perplexity": 1799.1570641596302}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115858727.26/warc/CC-MAIN-20150124161058-00087-ip-10-180-212-252.ec2.internal.warc.gz"}
|
https://www.physicsforums.com/threads/standard-candles.129119/
|
# Standard candles
1. Aug 17, 2006
### wolram
http://arxiv.org/abs/astro-ph/0608324
Cosmological Implications of the Second Parameter of Type Ia Supernovae
Authors: Philipp Podsiadlowski (Oxford), Paolo A. Mazzali (MPA, Munich; Trieste), Pierre Lesaffre (Oxford, Cambridge, Paris), Christian Wolf (Oxford), Francisco Forster (Oxford)
Theoretical models predict that the initial metallicity of the progenitor of a Type Ia supernova (SN Ia) affects the peak of the supernova light curve. This can cause a deviation from the standard light curve calibration employed when using SNe Ia as standardizable distance candles and, if there is a systematic evolution of the metallicity of SN Ia progenitors, could affect the determination of cosmological parameters. Here we show that this metallicity effect can be substantially larger than has been estimated previously, when the neutronisation in the immediate pre-explosion phase in the CO white dwarf is taken into account, and quantitatively assess the importance of metallicity evolution for determining cosmological parameters. We show that, in principle, a moderate and plausible amount of metallicity evolution could mimic a lambda-dominated, flat Universe in an open, lambda-free Universe. However, the effect of metallicity evolution appears not large enough to explain the high-z SN Ia data in a flat Universe, for which there is strong independent evidence, without a cosmological constant. We also estimate the systematic uncertainties introduced by metallicity evolution in a lambda-dominated, flat Universe. We find that metallicity evolution may limit the precision with which Omega_m and w can be measured and that it will be difficult to distinguish evolution of the equation of state of dark energy from metallicity evolution, at least from SN Ia data alone.
May be not much, just another question mark.
2. Aug 18, 2006
### wolram
Double edged sword.
http://arxiv.org/abs/astro-ph/0608351
The impact of neutrino masses on the determination of dark energy properties
Authors: Axel De La Macorra, Alessandro Melchiorri, Paolo Serra, Rachel Bean
Recently, the Heidelberg-Moscow double beta decay experiment has claimed a detection for a neutrino mass with high significance. Here we consider the impact of this measurement on the determination of the dark energy equation of state. By combining the Heidelberg-Moscow result with the WMAP 3-years data and other cosmological datasets we constrain the equation of state to -1.67< w <-1.05 at 95% c.l., ruling out a cosmological constant at more than 95% c.l.. Interestingly enough, coupled neutrino-dark energy models may be consistent with such equation of state. While future data are certainly needed for a confirmation of the controversial Heildelberg-Moscow claim, our result shows that future laboratory searches for neutrino masses may play a crucial role in the determination of the dark energy properties.
3. Aug 21, 2006
### wolram
No acceleration
http://arxiv.org/abs/astro-ph/0608386
On the Absence of Cosmic Acceleration
Authors: John Middleditch
Report-no: LAUR 06-5685
Type Ia Supernovae (SNe) have been used by many to argue for an accelerated expansion of the universe. However, high velocity and polarized features in all nearby SNe Ia, show that the paradigm for Type Ia SNe is drastically and catastrophically invalid. By now it is also clear that an extreme version of the axisymmetry seen in SN 1987A is the correct paradigm for SNe Ia and Ic. A Ia/c is produced from the merger of two degenerate cores of common envelope WR stars, or of two CO white dwarfs. Its polar blowouts produce the observed high velocity and polarized spectral features in Ia's, and its equatorial bulge is much brighter in Ia's, due to the greater fraction of 56 Ni contained within it. These become classified as Ia's when viewed from the merger equator, and Ic's when viewed from the poles. Thus cosmology determined strictly from Ia's alone is flawed at its very foundation: the local sample is selectively biased. The problem arose with the more distant supernovae, when the high velocity polar blowout features, which initially obscure part of the Ia/c equatorial bulge, expose a greater fraction of it, particularly when viewed off the equator, during the interval when Delta m_15 is measured, leading to a smaller decrease in observed luminosity. The width-luminosity correction was thus too small, and the result was a distant SN Ia which appeared to be too faint for its redshift. When the errors introduced by this process and others are taken into account, there may be no cosmic acceleration effect in distant SNe.
4. Aug 21, 2006
### Garth
These papers may indicate the confidence placed in the standard
$\Lambda$CDM model may be misplaced.
Garth
5. Aug 21, 2006
### wolram
I am not going to say anything Garth, i will just post any thing that seems interesting in this thread, unless any one has objections.
6. Aug 21, 2006
### Garth
The point is: Are these 'SN Ia' standard candles?
Such a lot in the standard model hangs on the assumption that they are.
Yet the evidence that their luminosities are known depends only on those nearby, and therefore recent, ones that are able to be calibrated. As we go back into cosmological history metallicity may be expected to change, and as your links imply this may alter their expected luminosities and call into question the conclusions based on them.
Garth
7. Aug 21, 2006
### wolram
May be i should ask you what is standard in cosmology, is there a standard galaxy ,star, planet, moon, if not it is difficult to understand how there can be a (standard) body in the U.
8. Aug 21, 2006
### Garth
The concept of a standard candle is simply a luminous object of which we think we know the Absolute Magnitude. (Such as the Cepheid Variables)
The object's distance can then be calculated from the distance modulus.
Garth
9. Aug 21, 2006
### matt.o
I think Wolram was asking a deeper question here, Garth. He is pointing out the differences in other observed phenomena such as galaxies and then questioning how there can be standard candles if things are so different.
I would just like to say that as scientists, we generally like to have a few more cards up our sleeves, hence there are other tests which point to a dark energy density of ~.7 independant of the SN1a results. Such studies include the CMB, baryon acoustic fluctuations, galaxy cluster mass function etc.
10. Aug 21, 2006
### SpaceTiger
Staff Emeritus
That hasn't been true for some time. Pretty much everything we infer from Type Ia SNe (so far) has been checked with other methods (e.g. CMB, LSS). In fact, it would be very strange if the results were found to be invalid after having already been found consistent with these more reliable measurements.
11. Aug 22, 2006
### wolram
Thankyou, Matt.o,
i do not know about (baryon acoustic fluctuations) i will have to look up the
subject.
12. Aug 22, 2006
### Garth
If SN Ia are not standard candles and the universe is not accelerating would that not alter the interpretation of the CMB and LSS data?
Garth
13. Aug 22, 2006
### matt.o
Actually, I don't think baryon acoustic fluctuations have been used by themselves (yet) to constrain dark energy. The current galaxy redshift surveys don't cover enough volume. They have certainly been used in conjunction with WMAP data to constrain cosmological parameters http://arxiv.org/abs/astro-ph/0507583" [Broken] paper by Chris Blake and Karl Glazebrook.
I do know of galaxy redshift surveys that are going to attempt to estimate the equation of state of the dark energy, but they are a little while off yet. There are also planned missions to survey for clusters and measure the dark energy equation of state using the cluster mass function.
Last edited by a moderator: May 2, 2017
14. Aug 22, 2006
### SpaceTiger
Staff Emeritus
The paper in the OP says the following:
This doesn't suggest to me that the SN Ia results have been invalidated.
As for the Middleditch paper, I took a quick look at it and it sounds very crankish. I doubt it will make any waves.
If SNe Ia were not standard candles, why would you infer that the universe was not accelerating? The former statement would simply mean that they couldn't be used as distance indicators.
15. Aug 22, 2006
### Tzemach
There are still some big holes in our understanding, maybe if we collect enough details of anomolies we will be able to figure out just what is going on out there.
Here are a couple for the collection'
APM 8279+5255 the red shift of the quasar, a vibrant galaxy with a bright central region and massive central black hole, revealed that it contained much more iron than it should for its age. The amount of iron present is greater than the amount found in our own solar system. At a fraction of the age of the solar system this quasar should contain the lower percentage of iron.
Fossil cluster RX J1416.4+2315 Current projections state that the fossil group should not have had enough time to form given the age of the universe.
16. Aug 22, 2006
### Garth
I was simply reverting back to the pre-'cosmic acceleration' understanding.
The acceleration or non-acceleration of the universe determined from SNe Ia at varing high z is an important handle on the behaviour and therefore nature of DE.
The consequences of them not being standard candles would therefore be rather significant. Tzemach has highlighted again the Age problem question, I think there is a degeneracy in the interpretation of the data that may be resolved by attention to these anomalies.
Is there an age problem at high z?
Are SNe Ia standard candles?
Is there a genuine deficiency at low-l mode in the WMAP anisotropy power spectrum?
Just a few questions to chew over....
Garth
Last edited: Aug 22, 2006
17. Aug 22, 2006
### SpaceTiger
Staff Emeritus
It's not clear to me that your above statements really make that case. It would have made a big difference historically, but with the high-precision CMB and LSS measurements we have now, I don't think you'd see many people jumping ship from standard cosmology.
18. Aug 22, 2006
### Garth
Well ST, surely a project such as Destiny
depends on the assumption that SNe Ia are standard candles?
Are you saying that if, for the sake of argument, they prove to be intrinsically fainter (say) at high z, then the interpretation of the high precision WMAP and LSS data would not be affected?
Or do you think that WMAP, LSS, etc. have so proved cosmic acceleration independently of SNe Ia that the there is no question about it, and consequently SNe Ia have been verified to be standard candles?
Garth
Last edited: Aug 22, 2006
19. Aug 22, 2006
### SpaceTiger
Staff Emeritus
Again, this depends crucially on the distinction between them not being standard candles and them being renormalized. If they were renormalized and subsequently showed no acceleration, then that would be a serious problem for mainstream cosmology in general. That alone likely wouldn't be enough to convince folks (after all, it would also demonstrate the unreliability of supernovae). However, if subsequent measurements with other distance indicators gave the same result, the mainstream model would have to be changed.
If the supernovae were just shown to be unreliable as standard candles, then yes, I'm saying the interpretation would likely remain the same. Your choice of words in the above doesn't make it clear which case you're referring to.
Those data sets do not directly prove acceleration, but they do suggest it. Believe it or not, $\Lambda CDM$ was under serious consideration even prior to the supernova results because of COBE CMB measurements.
20. Aug 22, 2006
### matt.o
Exactly. From COBE, it was known we live in an $$\Omega$$ = 1 Universe, but the observed matter density only made up ~.3 of the total density. People were considering a cosmological constant to make up the rest of the energy density and along came the SN 1a results!
Similar Discussions: Standard candles
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.860864520072937, "perplexity": 1685.5117784188426}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-39/segments/1505818687837.85/warc/CC-MAIN-20170921191047-20170921211047-00246.warc.gz"}
|
http://poisson.phc.unipi.it/~maggiolo/index.php/2008/12/latex-class-for-lecture-notes/
|
# LaTeX class for lecture notes
In the last post I wrote about how to take notes with LaTeX; that post focused on speed, and there were a sty file with the macros defined for that purpose. Here I won’t speak about speed, but style.
To begin, LaTeX class file for lecture notes you can find the class file I wrote for my notes. Just put it in you LaTeX tree (or in the directory of your document), with the sty file, and write a document with the following structure.
\documentclass[OPTIONS]{Notes}
\title{TITLE}
\subject{SUBJECT}
\author{AUTHOR}
\email{EMAIL}
\speaker{SPEAKER}
\date{DD}{MM}{YYYY}
\dateend{DD}{MM}{YYYY}
\place{PLACE}
\begin{document}
\end{document}
Obviously you have to replace the values, or, when appropriate, delete some line. Note that “author” and “email” are name and e-mail of the notes taker. The options (see the first line) are these:
• english or italian: the document’s language;
• course, seminar or talk: a course is a medium-length document structured in sections, subsections and paragraphs; a seminar is a short document without structure or with subsections; a talk is the sheets you take with you when you give a seminar.
The second option defines the style of the document; if you want to tweak some parameters, there are also these fine-tuning options:
• headertitle, headersection, headersubsection or headerno: what to display on the top of the pages;
• theoremnosection, theoremsection or theoremsubsection: numbering of theorems, definitions, etc.;
• cleardoublepage or nocleardoublepage: whether we want empty double pages after a section ending;
• oneside or twoside: margins and headers optimized for one-sided or two-sided printing;
• onecolumn or twocolumn: how many columns per page.
The class defines also some commands:
• for course, the command \lecture{duration}{dd}{mm}{yyyy} which writes on the margin information about the lesson’s number, duration and date;
• for talk, the commands \separator, which simply draws a horizontal line, and \tosay{message} which prints the message inside a box (useful to remember things you don’t have to write on the blackboard);
• \mymarginpar{message}, which behaves like a regular \marginpar but with a custom style.
Two words on what you get with this class file: the main difference is the font (Palatino instead of Computer Modern); the second main difference is the use of small-caps instead of boldface. I have to say I needed these modifications because I was very tired of the standard, omnipresent LaTeX style. Other small notable differences are the centered titles—because I’m not writing a book—and the theorem numbers placed before the word “theorem”—to make easier searching for them. If you want to see an example, download a recent document from my lecture notes page.
Let’s give a look on the packages this class requires, at least the ones you should use in your everyday documents.
• hyperref: many PDF on the web doesn’t have hyperlinks, from the table of contents to the sections, to an equation from where they refer to it, etc. With this package you can have automatically all these links and also tweak PDF metadatas.
• mathtools: enhancements of the standard AMS-LaTeX packages; the documentation is an inspiring source of tips for beautiful typesetting.
• booktabs: standard LaTeX tables look poorer than they should; learn why reading booktabs’ documentation.
• multirow: tables with cells spanning multiple rows.
• fancyhdr: to customize the content of headers and footers.
• mparhack: workaround for a bug on LaTeX \marginpar.
• tikz: it’s a high-level drawing language, designed by Till Tantau (you can remember him for software like latex-beamer); you can see a gallery with examples here.
• mathdots: redefines \dots and its friends so that they change size when appropriate.
• xfrac: to typeset nice inline fractions, as in ½.
• faktor: same as before, but without shrinking denominator’s and numerator’s size; useful for quotients.
• cancel: to draw diagonal bars above terms in an equation.
Please note that my class is not licensed in any way. If you want to use it, you’re free to do anything you want with them. Anyway, I will be happy if you just drop me a line telling me you’re using it.
## 8 thoughts on “LaTeX class for lecture notes”
Seems like this doesn’t work. I get an error:
I changed the file to Notes.cls (from notes.cls) but then get more errors:
This is pdfTeX, Version 3.1415926-1.40.10 (TeX Live 2009/Debian)
entering extended mode
(/home/amg81/Documents/UW/Stat 581/.algebraic_stacks.tex.swp
LaTeX2e
Babel and hyphenation patterns for english, usenglishmax, dumylang, noh
(./Notes.cls
Document Class: Notes
(/usr/share/texmf-texlive/tex/latex/base/ifthen.sty)
(/usr/share/texmf-texlive/tex/latex/datetime/datetime.sty
(/usr/share/texmf-texlive/tex/latex/fmtcount/fmtcount.sty
(/usr/share/texmf-texlive/tex/latex/graphics/keyval.sty)
(/usr/share/texmf-texlive/tex/latex/amsmath/amsgen.sty)
(/usr/share/texmf-texlive/tex/latex/fmtcount/fc-english.def)
(/usr/share/texmf-texlive/tex/latex/fmtcount/fc-USenglish.def)
No configuration file fmtcount.cfg found.
)) (/usr/share/texmf-texlive/tex/latex/base/article.cls
Document Class: article 2007/10/19 v1.4h Standard LaTeX document class
(/usr/share/texmf-texlive/tex/latex/base/size10.clo))
(/usr/share/texmf-texlive/tex/generic/babel/babel.sty
(/usr/share/texmf-texlive/tex/generic/babel/english.ldf
(/usr/share/texmf-texlive/tex/generic/babel/babel.def)
/usr/share/texmf-texlive/tex/generic/babel/english.ldf:106: Missing number, tre
ated as zero.
\l@british
/usr/share/texmf-texlive/tex/generic/babel/english.ldf:106: ==> Fatal error oc
curred, no output PDF file produced!
Transcript written on /tmp/.algebraic_stacks.tex.log.
2. Stefano Post author
Adam: this is not a problem of my class (or, more precisely, probably I could avoid this problem but I haven’t time to do it). To look at more people having this problem, you can google it. An example is this.
To fix this (in a non-particularly-elegant way), you can comment out those lines 106-110 in /tex\generic\babel\english.ldf.
A more elegant, but maybe not working, way is to install the language UKenglish on your system (for example the package texlive-lang-ukenglish on Ubuntu).
3. Christopher
So, just letting you know that I have found this class file immensely useful in graduate school and use it daily for everything from lecture notes to preparing my own notes. Awesome design.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7135199904441833, "perplexity": 3271.2823081354404}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-23/segments/1406510270528.34/warc/CC-MAIN-20140728011750-00457-ip-10-146-231-18.ec2.internal.warc.gz"}
|
https://juanmarqz.wordpress.com/tag/3-variedad/
|
# Tag Archives: 3-variedad
## can you see a solid torus?
The image on the right is a solid torus with four lateral annulus in colours: orange, green, yellow and blue.
Solid tori are important elementary type of 3-manifolds. Also called orientable genus one handlebodies and can be fibered -in the sense of Seifert- by longitudinal circles, and in many different ways.
Among three dimensional technologeeks, they are customed to see a solid torus that can be fibered as the $M\ddot{o}\stackrel{\sim}\times I$, the twisted interval bundle over the Möbius strip. In constrast $M\ddot{o}\times I$ is the genus one non orientable handlebody: the solid Klein bottle. Isn’t difficult to prove that they are the only two $I$-bundles over $M\ddot{o}$
Leave a comment
Filed under fiber bundle, topology
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 4, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.49438682198524475, "perplexity": 1799.6025871485167}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496665726.39/warc/CC-MAIN-20191112175604-20191112203604-00282.warc.gz"}
|
http://www.talkstats.com/threads/differences-between-2-x1-and-x1-x2-under-normal-distribution.73857/
|
Differences between 2 X1 and X1 + X2 under normal distribution
shawnteh1711
New Member
The question is as below:
1. If X1 ~ N(μ, σ2) and X2 ~ N(μ, σ2) are normal random variables, then what is the difference between 2 X1 and X1 + X2? Discuss both their distributions and parameters.
Assume that X1 + X2 have same μ =1, σ = 2
2 X1 = 2(1,22) = 2(1,4) = (2,8)
X1 + X2 = (1 + 1, 22 + 22) = (2,8)
spunky
Doesn't actually exist
You're almost there, but you're forgetting how variances/standard deviations change under linear transformations.
Just go back to the definition of the variance and using the algebra of expectations see how Var(X1+X2) is different from Var (k*X1) (where k is a constant)
shawnteh1711
New Member
You're almost there, but you're forgetting how variances/standard deviations change under linear transformations.
Just go back to the definition of the variance and using the algebra of expectations see how Var(X1+X2) is different from Var (k*X1) (where k is a constant)
I look it up and got the following rule :
1) Var (k*X1) = k ^2 * Var (X)
2) If X and Y are independent,
V(X + Y) = V(X) + V(Y)
I try to apply to the question.
Given that X1 ~ N(µ, σ2) and X2 ~ N(µ, σ2)
Assume that X1 and X2 have same µ =1, σ =2
Var(2*X1) = 2^2*Var(1,2^2)
= 4*Var(1,4)
= Var(4,16)
Var(X1 + X2) = Var(X1) + Var (X2)
= Var(1, 2^2) + Var(1, 2^2)
= var (1,4) + var (1,4)
= var (2,8)
It means that Var (2*X1) has 2 times more variance value than Var(X1+X2). Is it correct?
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8528398275375366, "perplexity": 3257.288977711554}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027313803.9/warc/CC-MAIN-20190818104019-20190818130019-00197.warc.gz"}
|
https://worldwidescience.org/topicpages/v/volume-molar+mass+experiment.html
|
#### Sample records for volume-molar mass experiment
1. Oxygen from Hydrogen Peroxide. A Safe Molar Volume-Molar Mass Experiment.
Science.gov (United States)
Bedenbaugh, John H.; And Others
1988-01-01
Describes a molar volume-molar mass experiment for use in general chemistry laboratories. Gives background technical information, procedures for the titration of aqueous hydrogen peroxide with standard potassium permanganate and catalytic decomposition of hydrogen peroxide to produce oxygen, and a discussion of the results obtained in three…
2. Photon mass experiment
International Nuclear Information System (INIS)
Crandall, R.E.
1983-01-01
A Coulomb null experiment is described that enables physics students to obtain rigorous upper bounds on photon mass. The experimenter searches for subnanovolt signals that would escape a closed shell were photon mass to be positive. The approach can be adapted for several college levels. At the simplest level, a ''miniature'' low-cost experiment allows a student to verify the exponent ''-2'' in Coulomb's law to eight or more decimal places. An advanced student given a full-size apparatus (at greater cost) can obtain mass bounds very close to the established laboratory limit
3. Neutrino mass experiments
International Nuclear Information System (INIS)
Robertson, R.G.H.
1989-01-01
The current status of the experimental search for neutrino mass is reviewed, with emphasis on direct kinematic methods. Simpson and Hime report finding new evidence for a 17-keV neutrino in the β decay of 3 H and 35 S. The situation concerning the electron neutrino mass as measured in tritium beta decay has not changed significantly in the last two years. We discuss the ''model independent'' lower limit of 17 eV obtained by the ITEP group in light of existing data on the 3 H-- 3 He mass difference. 42 refs., 1 fig., 1 tab
4. Status of neutrino mass experiments
International Nuclear Information System (INIS)
Fackler, O.
1985-01-01
In 1980 two experiments ignited a fertile field of research - the determination of the neutrino masses. Subsequently, over 35 experiments using a variety of techniques have probed or are probing this question. Primarily the author discuss electron antineutrino (hereafter referred to as neutrino) mass experiments. Section I begins with a discussion of astronomical and terrestrial observations which motivated these experiments. In Section II, the author quote limits from muon and tau mass determinations. These limits are more thoroughly discussed in other paper. The author continues by describing the four approaches used to measure the electron neutrino mass. In Section III, tritium beta decay mass determinations are reviewed. This section includes a general summary of previous experimental results, and discussion of the major ongoing experiments. Section IV offers concluding remarks
5. Status of neutrino mass experiments
International Nuclear Information System (INIS)
Fackler, O.
1985-12-01
In 1980 two experiments ignited a fertile field of research the determination of the neutrino masses. Subsequently, over 35 experiments using a variety of techniques have probed or are probing this question. Primarily I will discuss electron antineutrino (hereafter referred to as neutrino) mass experiments. However, let me begin in Section I to discuss astronomical and terrestrial observations which motivated these experiments. In Section II, I will quote limits from muon and tau mass determinations. These limits are more thoroughly discussed in other papers. I will continue by describing the four approaches used to measure the electron neutrino mass. In Section III, tritium beta decay mass determinations will be reviewed. This section includes a general summary of previous experimental results, and discussion of the major ongoing experiments. Section IV offers concluding remarks. 24 refs., 24 figs
6. Current Direct Neutrino Mass Experiments
Directory of Open Access Journals (Sweden)
G. Drexlin
2013-01-01
Full Text Available In this contribution, we review the status and perspectives of direct neutrino mass experiments, which investigate the kinematics of β-decays of specific isotopes (3H, 187Re, 163Ho to derive model-independent information on the averaged electron (antineutrino mass. After discussing the kinematics of β-decay and the determination of the neutrino mass, we give a brief overview of past neutrino mass measurements (SN1987a-ToF studies, Mainz and Troitsk experiments for 3H, cryobolometers for 187Re. We then describe the Karlsruhe Tritium Neutrino (KATRIN experiment currently under construction at Karlsruhe Institute of Technology, which will use the MAC-E-Filter principle to push the sensitivity down to a value of 200 meV (90% C.L.. To do so, many technological challenges have to be solved related to source intensity and stability, as well as precision energy analysis and low background rate close to the kinematic endpoint of tritium β-decay at 18.6 keV. We then review new approaches such as the MARE, ECHO, and Project8 experiments, which offer the promise to perform an independent measurement of the neutrino mass in the sub-eV region. Altogether, the novel methods developed in direct neutrino mass experiments will provide vital information on the absolute mass scale of neutrinos.
7. The Mainz Neutrino Mass Experiment
Czech Academy of Sciences Publication Activity Database
Kraus, C.; Bornschein, L.; Bonn, J.; Bornschein, B.; Flatt, B.; Kovalík, Alojz; Müller, B.; Otten, EW; Schall, JP.; Thummler, T.; Weinheimer, C.
2005-01-01
Roč. 143, - (2005), s. 143 ISSN 0920-5632. [International Conference on Neutrino Physics and Astrophysics /21./. Paříž, 14.06.2004-19.06.2004] R&D Projects: GA MŠk 1P04LA213 Institutional research plan: CEZ:AV0Z10480505 Keywords : neutrino mass * tritium beta decay Subject RIV: BG - Nuclear, Atomic and Molecular Physics, Colliders Impact factor: 0.875, year: 2005
8. Pennsylvania's experience in mass screening
International Nuclear Information System (INIS)
Gerusky, T.M.
1975-01-01
A policy statement issued in 1972 by the Assistant Secretary for Health and Scientific Affairs of the Department of Health, Education, and Welfare recommended that community chest x-ray surveys should not be used as a screening procedure for the detection of cardiopulmonary disorders and that when chest x-ray screening programs are justified for selected population groups, the full size photograph, rather than the miniature film, is preferred. A survey conducted in 1974--75 revealed that chest x rays were required for prisoners, prison employees, school employees, food handlers, and students who wished to participate in sports. Meetings were held with medical associations in the hope of stopping the local mass-screening operations. Of 27 groups in Pennsylvania involved in tuberculosis screening, 12 groups refused or were unwilling to phase out their photofluorographic procedures. The problem will be resolved by regulation
9. Experiments for obtaining field influence mass particles.
CERN Document Server
Yahalomi, E
2010-01-01
Analyzing time dilation experiments the existence of a universal field interacting with moving mass particles is obtained. It is found that mass particle changes its properties depend on its velocity relative to this universal scalar field and not on its velocity relative to the laboratory. High energy proton momentum, energy and mass were calculated obtaining new results. Experiments in high energy accelerators are suggested as additional proofs for the existence of this universal field. This universal field may explain some results of other high energy experiments.
10. The IBM neutrino-mass experiment
International Nuclear Information System (INIS)
Clark, G.J.; Frisch, M.A.; Chaudhari, P.; Bregman, M.F.
1985-01-01
IBM is undertaking an experiment to measure the electron anti-neutrino mass. A high precision measurement of the tritium Β-decay spectrum near the end point is used to infer the neutrino mass. Electron energies are measured using a large spherical retarding grid analyzer. We are placing particular emphasis on understanding the complications introduced by solid state effects in the source
11. The IBM neutrino-mass experiment
International Nuclear Information System (INIS)
Clark, G.J.; Frisch, M.A.; Chaudhari, P.; Bregman, M.F.
1985-01-01
IBM is undertaking an experiment to measure the electron anti-neutrino mass. A high precision measurement of the tritium β-decay spectrum near the end point is used to infer the neutrino mass. Electron energies are measured using a large spherical retarding grid analyzer. They are placing particular emphasis on understanding the complications introduced by solid state effects in the source
12. Ozone mass transfer and kinetics experiments
International Nuclear Information System (INIS)
Bollyky, L.J.; Beary, M.M.
1981-12-01
Experiments were conducted at the Hanford Site to determine the most efficient pH and temperature levels for the destruction of complexants in Hanford high-level defense waste. These complexants enhance migration of radionuclides in the soil and inhibit the growth of crystals in the evaporator-crystallizer. Ozone mass transfer and kinetics tests have been outlined for the determination of critical mass transfer and kinetics parameters of the ozone-complexant reaction
13. Experiment for a precision neutrino mass measurement
International Nuclear Information System (INIS)
Fackler, O.; Mugge, M.; Sticker, H.; Woerner, R.
1984-04-01
We describe an experiment which is designed to determine the electron neutrino mass to better than 2 eV. Key features of the experiment are a high activity frozen tritium source and a high resolution electrostatic spectrometer designed to make a careful measurement of the tritium beta decay end point spectrum. The goal is to determine the neutrino mass to better than 1 eV statistically in a four day run. A series of these runs will allow study of potential systematics. The construction phase is nearly complete and preliminary data will be taken in late spring
Directory of Open Access Journals (Sweden)
Polat Dursun
2011-06-01
Full Text Available Background: Adnexal mass in pregnancy is a rare situation in daily clinical practice. Also, there is no consensus about the management of the adnexal mass diagnosed during pregnancy. Material&Methods: In this study,we retrospectively identified adnexal mass which was diagnosed during antenatal follow-up or cesarean section between 2000-2009 in Başkent University Hospital,Department of Obstetrics&Gynecology.Labor&delivery unit database, hospital records and pathology reports were evaluated in order to retrive the age of patients, number of gravida and parity, initial symptoms, the gestational age and the diameter of cyst, antenatal complications, time of delivery birth weight, indication of cesarean delivery, the type of surgical intervention during cearean delivery and pathology of the cyst. Results: We identified 27 pregnancy complicated with adnexal masses among the 2150 delivery ( 1.25%. Among these,25 of 27 pregnants were asymptomatic (92,6% while just 2 pregnant women came with the complaint of pain. In 2 of the patients (7,4% the cyst was known before the pregnancy while in 6 pregnant women (22,4% the cyst was diagnosed during antenatal care. Also, rest of the women (n=19, 70,4% were diagnosed during cesarean. The 3 of the cysts (11,1% was smaller than 6 cm while another 3 of the cysts (11,1% was greater than 6 cm. Cesarean and cystectomy was performed in 23 of this women. On the other hand, 2 of them had cesarean and unilateral ooferectomy.Pathologic examinations reported as; 6(22,2% dermoid cyst, 3 (11,1% endometrioma, 4(14,8% seros cystadenoma, 3(%11.1 Morgagni cyst, 4(%14.8 mucinous cyst, 3(%11.1 follicular cyst, 2(%7.4 siderophagic cyst, 1(%3.7 fibrom, 1 (%3.7 thecoma. Conclusion: Most of the adnexal masses diagnosed during antenatal period or cesarean section is benign. Therefore, if there is no sign of malignancy it can be conservatively managed during pregnancy and cesarean section.
15. Effects of local mass anomalies in Eoetvoes-like experiments
International Nuclear Information System (INIS)
Talmadge, C.; Aronson, S.H.; Fischbach, E.
1986-01-01
We consider in detail the effects of local mass anomalies in Eoetvoes-like experiments. It is shown that in the presence of an intermediate-range non-gravitational force, the dominant contributions to both the sign and magnitude of the Eoetvoes anomaly may come from nearby masses and not from the earth as a whole. This observation has important implications in the design and interpretation of future experiments, and in the formulation of unified theories incorporating new intermediate-range forces
16. A Shuttle Upper Atmosphere Mass Spectrometer /SUMS/ experiment
Science.gov (United States)
Blanchard, R. C.; Duckett, R. J.; Hinson, E. W.
1982-01-01
A magnetic mass spectrometer is currently being adapted to the Space Shuttle Orbiter to provide repeated high altitude atmosphere data to support in situ rarefied flow aerodynamics research, i.e., in the high velocity, low density flight regime. The experiment, called Shuttle Upper Atmosphere Mass Spectrometer (SUMS), is the first attempt to design mass spectrometer equipment for flight vehicle aerodynamic data extraction. The SUMS experiment will provide total freestream atmospheric quantitites, principally total mass density, above altitudes at which conventional pressure measurements are valid. Experiment concepts, the expected flight profile, tradeoffs in the design of the total system and flight data reduction plans are discussed. Development plans are based upon a SUMS first flight after the Orbiter initial development flights.
17. Evaluation of NO2 predictions by the plume volume molar ratio method (PVMRM) and ozone limiting method (OLM) in AERMOD using new field observations.
Science.gov (United States)
Hendrick, Elizabeth M; Tino, Vincent R; Hanna, Steven R; Egan, Bruce A
2013-07-01
The U.S. Environmental Protection Agency (EPA) plume volume molar ratio method (PVMRM) and the ozone limiting method (OLM) are in the AERMOD model to predict the 1-hr average NO2/NO(x) concentration ratio. These ratios are multiplied by the AERMOD predicted NO(x) concentration to predict the 1-hr average NO2 concentration. This paper first briefly reviews PVMRM and OLM and points out some scientific parameterizations that could be improved (such as specification of relative dispersion coefficients) and then discusses an evaluation of the PVMRM and OLM methods as implemented in AERMOD using a new data set. While AERMOD has undergone many model evaluation studies in its default mode, PVMRM and OLM are nondefault options, and to date only three NO2 field data sets have been used in their evaluations. Here AERMOD/PVMRM and AERMOD/OLM codes are evaluated with a new data set from a northern Alaskan village with a small power plant. Hourly pollutant concentrations (NO, NO2, ozone) as well as meteorological variables were measured at a single monitor 500 m from the plant. Power plant operating parameters and emissions were calculated based on hourly operator logs. Hourly observations covering 1 yr were considered, but the evaluations only used hours when the wind was in a 60 degrees sector including the monitor and when concentrations were above a threshold. PVMRM is found to have little bias in predictions of the C(NO2)/C(NO(x)) ratio, which mostly ranged from 0.2 to 0.4 at this site. OLM overpredicted the ratio. AERMOD overpredicts the maximum NO(x) concentration but has an underprediction bias for lower concentrations. AERMOD/PVMRM overpredicts the maximum C(NO2) by about 50%, while AERMOD/OLM overpredicts by a factor of 2. For 381 hours evaluated, there is a relative mean bias in C(NO2) predictions of near zero for AERMOD/PVMRM, while the relative mean bias reflects a factor of 2 overprediction for AERMOD/OLM. This study was initiated because the new stringent 1-hr NO2
18. Critical mass experiment using U-235 foils and lucite plates
International Nuclear Information System (INIS)
Sanchez, R.; Butterfield, K.; Kimpland, R.; Jaegers, P.
1998-01-01
The main objective of this experiment was to show how the multiplication of the system increases as moderated material is placed between highly enriched uranium foils. In addition, this experiment served to demonstrate the hand-stacking techniques, and approach to criticality by remote operation. This experiment was designed by Tom McLaughlin in the mid seventies as part of the criticality safety course that is taught at Los Alamos Critical Experiment Facility (LACEF). The W-U-235 ratio for this experiment was 215 which is where the minimum critical mass for this configuration occurs
19. Critical mass experiment using 235U foils and lucite plates
International Nuclear Information System (INIS)
Sanchez, R.; Butterfield, K.; Kimpland, R.; Jaegers, P.
1998-01-01
This experiment demonstrated how the neutron multiplication of a system increases as moderated material is placed between highly enriched uranium foils. In addition, this experiment served to demonstrate the hand-stacking technique and approach to criticality be remote operation. This experiment was designed by McLaughlin in the mid-seventies as part of the criticality safety course that is taught at the Los Alamos Critical Experiments Facility. The H/ 235 U ratio for this experiment was 215, which is the ratio at which the minimum critical mass for this configuration occurs
20. On the scalar electron mass limit from single photon experiments
International Nuclear Information System (INIS)
Grivaz, J.F.
1987-03-01
We discuss how the 90% C.L. lower limit on the mass of the scalar electron, as extracted from the single photon experiments, is affected by the way the background from radiative neutrino pair production is handled. We argue that some of the results presented at the Berkeley conference are overoptimistic, and that the mass lower limit is 65 GeV rather than the advertized value of 84 GeV, for the case of degenerate scalar electrons with massless photinos
1. A PMT mass testing facility for the JUNO experiment
Energy Technology Data Exchange (ETDEWEB)
Tietzsch, Alexander; Alsheimer, Isabell; Blum, David; Lachenmaier, Tobias; Sterr, Tobias [Physikalisches Institut, Universitaet Tuebingen (Germany); Bein, Bosse; Bick, Daniel; Ebert, Joachim; Hagner, Caren; Rebber, Henning; Steppat, Lisa; Wonsak, Bjoern [Institut fuer Experimentalphysik, Universitaet Hamburg (Germany)
2016-07-01
The JUNO (Jiangmen Underground Neutrino Observatory) experiment will be one of the big neutrino oscillation experiments starting in the next years. The main goal of JUNO is the determination of the neutrino mass hierarchy. To detect the sub-dominant effects in the oscillation pattern which depend on the mass hierarchy, the JUNO detector is planned with almost 20 kt fiducial volume, high light yield and energy resolution of better than 3%. In order to reach this, roughly 17000 newly developed high QE PMTs for the central detector, and additionally 2000 for the veto will be used. Each PMT has to be tested and characterized before it will be mounted in the experiment. This talk gives an overview on our plans and strategy for the mass test of all PMTs, and on the current status of the experimental test setup and next steps. The testing facility is developed in a cooperation between the Physical Institutes in Tuebingen and Hamburg within the JUNO collaboration.
2. Multifactorial Understanding of Ion Abundance in Tandem Mass Spectrometry Experiments.
Science.gov (United States)
Fazal, Zeeshan; Southey, Bruce R; Sweedler, Jonathan V; Rodriguez-Zas, Sandra L
2013-01-29
In a bottom-up shotgun approach, the proteins of a mixture are enzymatically digested, separated, and analyzed via tandem mass spectrometry. The mass spectra relating fragment ion intensities (abundance) to the mass-to-charge are used to deduce the amino acid sequence and identify the peptides and proteins. The variables that influence intensity were characterized using a multi-factorial mixed-effects model, a ten-fold cross-validation, and stepwise feature selection on 6,352,528 fragment ions from 61,543 peptide ions. Intensity was higher in fragment ions that did not have neutral mass loss relative to any mass loss or that had a +1 charge state. Peptide ions classified for proton mobility as non-mobile had lowest intensity of all mobility levels. Higher basic residue (arginine, lysine or histidine) counts in the peptide ion and low counts in the fragment ion were associated with lower fragment ion intensities. Higher counts of proline in peptide and fragment ions were associated with lower intensities. These results are consistent with the mobile proton theory. Opposite trends between peptide and fragment ion counts and intensity may be due to the different impact of factor under consideration at different stages of the MS/MS experiment or to the different distribution of observations across peptide and fragment ion levels. Presence of basic residues at all three positions next to the fragmentation site was associated with lower fragment ion intensity. The presence of proline proximal to the fragmentation site enhanced fragmentation and had the opposite trend when located distant from the site. A positive association between fragment ion intensity and presence of sulfur residues (cysteine and methionine) on the vicinity of the fragmentation site was identified. These results highlight the multi-factorial nature of fragment ion intensity and could improve the algorithms for peptide identification and the simulation in tandem mass spectrometry experiments.
3. Tuning the Mass of Chameleon Fields in Casimir Force Experiments
CERN Document Server
Brax, Ph; Davis, A C; Shaw, D J; Iannuzzi, D
2010-01-01
We have calculated the chameleon pressure between two parallel plates in the presence of an intervening medium that affects the mass of the chameleon field. As intuitively expected, the gas in the gap weakens the chameleon interaction mechanism with a screening effect that increases with the plate separation and with the density of the intervening medium. This phenomenon might open up new directions in the search of chameleon particles with future long range Casimir force experiments.
4. Status and perspectives of the Mainz neutrino mass experiment
International Nuclear Information System (INIS)
Backe, H.; Barth, H.; Bleile, A.
1997-01-01
New data from the decay of molecular tritium studied with the Mainz solenoid retarding spectrometer are presented. From a region close to the end-point we deduce an upper limit for the mass of the electron antineutrino of m ν c 2 ν 2 c 4 = - 22 ± 17 stat ± 14 sys eV 2 . Possible improvements and the perspectives of the experiment are discussed. (orig.)
5. TRIMS: Validating T2 Molecular Effects for Neutrino Mass Experiments
Science.gov (United States)
Lin, Ying-Ting; Trims Collaboration
2017-09-01
The Tritium Recoil-Ion Mass Spectrometer (TRIMS) experiment examines the branching ratio of the molecular tritium (T2) beta decay to the bound state (3HeT+). Measuring this branching ratio helps to validate the current molecular final-state theory applied in neutrino mass experiments such as KATRIN and Project 8. TRIMS consists of a magnet-guided time-of-flight mass spectrometer with a detector located on each end. By measuring the kinetic energy and time-of-flight difference of the ions and beta particles reaching the detectors, we will be able to distinguish molecular ions from atomic ones and hence derive the ratio in question. We will give an update on the apparatus, simulation software, and analysis tools, including efforts to improve the resolution of our detectors and to characterize the stability and uniformity of our field sources. We will also share our commissioning results and prospects for physics data. The TRIMS experiment is supported by U.S. Department of Energy Office of Science, Office of Nuclear Physics, Award Number DE-FG02-97ER41020.
6. Beams of mass-selected clusters: realization and first experiments
International Nuclear Information System (INIS)
Kamalou, O.
2007-04-01
The main objective of this work concerns the production of beams of mass-selected clusters of metallic and semiconductor materials. Clusters are produced in magnetron sputtering source combined with a gas aggregation chamber, cooled by liquid nitrogen circulation. Downstream of the cluster source, a Wiley-McLaren time-of-flight setup allows to select a given cluster size or a narrow size range. The pulsed mass-selected cluster ion beam is separated from the continuous neutral one by an electrostatic 90-quadrupole deflector. After the deflector, the density of the pulsed beam amounts to about 10 3 particles/cm 3 . Preliminary deposition experiments of mass-selected copper clusters with a deposition energy of about 0.5 eV/atom have ben performed on highly oriented pyrolytic graphite (HOPG) substrates, indicating that copper clusters are evidently mobile on the HOPG-surface until they reach cleavage steps, dislocation lines or other surface defects. In order to lower the cluster mobility on the HOPG-surface, we have first irradiated HOPG samples with slow highly charged ions (high dose) in order to create superficial defects. In a second step we have deposited mass-selected copper clusters on these pre-irradiated samples. The first analysis by AFM (Atomic Force Microscopy) techniques showed that the copper clusters are trapped on the defects produced by the highly charged ions. (author)
7. Statistical methods for quantitative mass spectrometry proteomic experiments with labeling
Directory of Open Access Journals (Sweden)
Oberg Ann L
2012-11-01
Full Text Available Abstract Mass Spectrometry utilizing labeling allows multiple specimens to be subjected to mass spectrometry simultaneously. As a result, between-experiment variability is reduced. Here we describe use of fundamental concepts of statistical experimental design in the labeling framework in order to minimize variability and avoid biases. We demonstrate how to export data in the format that is most efficient for statistical analysis. We demonstrate how to assess the need for normalization, perform normalization, and check whether it worked. We describe how to build a model explaining the observed values and test for differential protein abundance along with descriptive statistics and measures of reliability of the findings. Concepts are illustrated through the use of three case studies utilizing the iTRAQ 4-plex labeling protocol.
8. Statistical methods for quantitative mass spectrometry proteomic experiments with labeling.
Science.gov (United States)
Oberg, Ann L; Mahoney, Douglas W
2012-01-01
Mass Spectrometry utilizing labeling allows multiple specimens to be subjected to mass spectrometry simultaneously. As a result, between-experiment variability is reduced. Here we describe use of fundamental concepts of statistical experimental design in the labeling framework in order to minimize variability and avoid biases. We demonstrate how to export data in the format that is most efficient for statistical analysis. We demonstrate how to assess the need for normalization, perform normalization, and check whether it worked. We describe how to build a model explaining the observed values and test for differential protein abundance along with descriptive statistics and measures of reliability of the findings. Concepts are illustrated through the use of three case studies utilizing the iTRAQ 4-plex labeling protocol.
9. Mass Spectrometry Coupled Experiments and Protein Structure Modeling Methods
Directory of Open Access Journals (Sweden)
Lee Sael
2013-10-01
Full Text Available With the accumulation of next generation sequencing data, there is increasing interest in the study of intra-species difference in molecular biology, especially in relation to disease analysis. Furthermore, the dynamics of the protein is being identified as a critical factor in its function. Although accuracy of protein structure prediction methods is high, provided there are structural templates, most methods are still insensitive to amino-acid differences at critical points that may change the overall structure. Also, predicted structures are inherently static and do not provide information about structural change over time. It is challenging to address the sensitivity and the dynamics by computational structure predictions alone. However, with the fast development of diverse mass spectrometry coupled experiments, low-resolution but fast and sensitive structural information can be obtained. This information can then be integrated into the structure prediction process to further improve the sensitivity and address the dynamics of the protein structures. For this purpose, this article focuses on reviewing two aspects: the types of mass spectrometry coupled experiments and structural data that are obtainable through those experiments; and the structure prediction methods that can utilize these data as constraints. Also, short review of current efforts in integrating experimental data in the structural modeling is provided.
10. Newest results from the Mainz neutrino-mass experiment
International Nuclear Information System (INIS)
Bonn, J.; Bornschein, B.; Bornschein, L.; Fickinger, L.; Kraus, Ch.; Otten, E.W.; Ulrich, H.; Weinheimer, Ch.; Kazachenko, O.; Kovalik, A.
2000-01-01
The Mainz neutrino-mass experiment investigates the endpoint region of the tritium β-decay spectrum with a MAC-E spectrometer to determine the mass of the electron antineutrino. By the recent upgrade, the former problem of dewetting T 2 films has been solved, and the signal-to-background ratio was improved by a factor of 10. The latest measurement leads to m ν 2 -3.7 ± 5.3(stat.) ± 2.1(syst.) eV 2 /c 4 , from which an upper limit of m ν 2 (95% C.L.) is derived. Some indication for the anomaly, reported by the Troitsk group, was found, but its postulated half-year period is contradicted by our data. To push the sensitivity on the neutrino mass below 1 eV/c 2 , a new larger MAC-E spectrometer is proposed. Besides its integrating mode, it could run in a new nonintegration operation MAC-E-TOF mode
11. Neutrino mass and mixing: from theory to experiment
International Nuclear Information System (INIS)
King, Stephen F; Merle, Alexander; Morisi, Stefano; Shimizu, Yusuke; Tanimoto, Morimitsu
2014-01-01
The origin of fermion mass hierarchies and mixings is one of the unresolved and most difficult problems in high-energy physics. One possibility to address the flavour problems is by extending the standard model to include a family symmetry. In the recent years it has become very popular to use non-Abelian discrete flavour symmetries because of their power in the prediction of the large leptonic mixing angles relevant for neutrino oscillation experiments. Here we give an introduction to the flavour problem and to discrete groups that have been used to attempt a solution for it. We review the current status of models in light of the recent measurement of the reactor angle, and we consider different model-building directions taken. The use of the flavons or multi-Higgs scalars in model building is discussed as well as the direct versus indirect approaches. We also focus on the possibility of experimentally distinguishing flavour symmetry models by means of mixing sum rules and mass sum rules. In fact, we illustrate in this review the complete path from mathematics, via model building, to experiments, so that any reader interested in starting work in the field could use this text as a starting point in order to obtain a broad overview of the different subject areas
12. Guidelines for reporting quantitative mass spectrometry based experiments in proteomics.
Science.gov (United States)
Martínez-Bartolomé, Salvador; Deutsch, Eric W; Binz, Pierre-Alain; Jones, Andrew R; Eisenacher, Martin; Mayer, Gerhard; Campos, Alex; Canals, Francesc; Bech-Serra, Joan-Josep; Carrascal, Montserrat; Gay, Marina; Paradela, Alberto; Navajas, Rosana; Marcilla, Miguel; Hernáez, María Luisa; Gutiérrez-Blázquez, María Dolores; Velarde, Luis Felipe Clemente; Aloria, Kerman; Beaskoetxea, Jabier; Medina-Aunon, J Alberto; Albar, Juan P
2013-12-16
Mass spectrometry is already a well-established protein identification tool and recent methodological and technological developments have also made possible the extraction of quantitative data of protein abundance in large-scale studies. Several strategies for absolute and relative quantitative proteomics and the statistical assessment of quantifications are possible, each having specific measurements and therefore, different data analysis workflows. The guidelines for Mass Spectrometry Quantification allow the description of a wide range of quantitative approaches, including labeled and label-free techniques and also targeted approaches such as Selected Reaction Monitoring (SRM). The HUPO Proteomics Standards Initiative (HUPO-PSI) has invested considerable efforts to improve the standardization of proteomics data handling, representation and sharing through the development of data standards, reporting guidelines, controlled vocabularies and tooling. In this manuscript, we describe a key output from the HUPO-PSI-namely the MIAPE Quant guidelines, which have developed in parallel with the corresponding data exchange format mzQuantML [1]. The MIAPE Quant guidelines describe the HUPO-PSI proposal concerning the minimum information to be reported when a quantitative data set, derived from mass spectrometry (MS), is submitted to a database or as supplementary information to a journal. The guidelines have been developed with input from a broad spectrum of stakeholders in the proteomics field to represent a true consensus view of the most important data types and metadata, required for a quantitative experiment to be analyzed critically or a data analysis pipeline to be reproduced. It is anticipated that they will influence or be directly adopted as part of journal guidelines for publication and by public proteomics databases and thus may have an impact on proteomics laboratories across the world. This article is part of a Special Issue entitled: Standardization and
13. First low WIMP mass results in EDELWEISS III experiment
Energy Technology Data Exchange (ETDEWEB)
Scorza, Silvia [Karlsruher Institut fuer Technologie, Institut fuer Experimentelle Kernphysik, Postfach 3640, Karlsruhe (Germany); Collaboration: EDELWEISS-Collaboration
2016-07-01
The EDELWEISS-III collaboration is operating an experiment for the direct detection of Weakly Interacting Massive Particle (WIMPs) dark matter in the low radioactivity environment of the Modane Underground Laboratory. It consists of twenty-four advanced high purity germanium detectors operating at 18 mK in a dilution refrigerator in order to identify rare nuclear recoils induced by elastic scattering of WIMPs from our Galactic halo. The current EDELWEISS-III program, including improvements of the background, data-acquisition and the configuration is detailed. Sources of background along with the rejection techniques are discussed. Detector performances and a first low WIMP mass analysis of data acquired in a long-term campaign are presented as well.
14. A pilot mass-storage system for KEK belle experiment
International Nuclear Information System (INIS)
Fujii, Hirofumi; Itoh, Ryosuke; Manabe, Atsushi; Miyamoto, Akiya; Morita, Youhei; Nozaki, Tadao; Sasaki, Takashi; Watase, Yoshiyuko; Yamakasi, Tokuyuki
1996-01-01
A pilot mass-storage system for KEK Belle (B-meson Physics) experiment has been developed. This experiment requires a high speed and large data recording system. The required recording speed is about 3 MB/sec in average and 15 MB/sec at maximum. The required volume is more than 30 TB/year. We have developed a pilot system to study the high-speed and large volume data recording system which satisfies the above requirements. The system consists of (1) SONY-DIR 1000M data recorder with SCSI-2 Fast/Wide interface; the recording capability of the device is 16MB/sec art maximum. (2) SONY-DMS 24 tape robot (tape library) of which volume capacity is about 2 TB (3) high-speed TCP/IP network of HIPPI and (4) three workstations running under UNIX. For the system software, the CERN-SHIFT system has been installed for the study. Because that the tape device and the robot (tape library) system are completely part. The tape device and the robot (library) control path are directly connected to UNIX workstations. To achieve the required recording speed, we also developed an application interfaces for this tape server. We have made the user interface without using tape stating mechanism. This user interface reduces the overhead of the recording system has developed based on TCP/IP, so that the system is easy to expand and free from network media. (author)
15. Lab. experiments of mass transfer in the London clay
International Nuclear Information System (INIS)
Bourke, P.J.; Gilling, D.; Jefferies, N.L.; Lineham, T.R.; Lever, D.A.
1989-01-01
Aqueous phase mass transfer through the rocks surrounding a radioactive waste repository will take place by diffusion and convection. This paper presents a comprehensive set of measurements of the mass transfer characteristics for a single, naturally occurring, clay. These data are compared with the results predicted by mathematical models of mass transport in porous media, in order to build confidence in these models
16. Direct search for neutrino mass and anomaly in the tritium beta-spectrum: Status of 'Troitsk neutrino mass' experiment
International Nuclear Information System (INIS)
Lobashev, V.M.; Aseev, V.N.; Belesev, A.I.; Berlev, A.I.; Geraskin, E.V.; Golubev, A.A.; Kazachenko, O.V.; Kuznetsov, Yu.E.; Ostroumov, R.P.; Rivkis, L.A.; Stern, B.E.; Titov, N.A.; Zadoroghny, C.V.; Zakharov, Yu.I.
2000-01-01
Results of the 'Troitsk ν-mass' experiment on search for the neutrino rest mass in the tritium beta-decay are presented. New data on the time dependence of the anomalous, bump-like structure at the end of the beta spectrum reported earlier are discussed. Possible systematics is considered in view of contradiction of 'Troitsk nu-mass' observation with those of 'Mainz neutrino' set-up. An upper limit for electron antineutrino rest mass remains at m ν 2 at 95% C.L
17. Mass-dependent and non-mass-dependent isotope effects in ozone photolysis: Resolving theory and experiments
International Nuclear Information System (INIS)
Cole, Amanda S.; Boering, Kristie A.
2006-01-01
In addition to the anomalous 17 O and 18 O isotope effects in the three-body ozone formation reaction O+O 2 +M, isotope effects in the destruction of ozone by photolysis may also play a role in determining the isotopic composition of ozone and other trace gases in the atmosphere. While previous experiments on ozone photolysis at 254 nm were interpreted as evidence for preferential loss of light ozone that is anomalous (or 'non-mass-dependent'), recent semiempirical theoretical calculations predicted a preferential loss of heavy ozone at that wavelength that is mass dependent. Through photochemical modeling results presented here, we resolve this apparent contradiction between experiment and theory. Specifically, we show that the formation of ozone during the UV photolysis experiments is not negligible, as had been assumed, and that the well-known non-mass-dependent isotope effects in ozone formation can account for the non-mass-dependent enrichment of the heavy isotopologs of ozone observed in the experiment. Thus, no unusual non-mass-dependent fractionation in ozone photolysis must be invoked to explain the experimental results. Furthermore, we show that theoretical predictions of a mass-dependent preferential loss of the heavy isotopologs of ozone during UV photolysis are not inconsistent with the experimental data, particularly if mass-dependent isotope effects in the chemical loss reactions of ozone during the photolysis experiments or experimental artifacts enrich the remaining ozone in 17 O and 18 O. Before the calculated fractionation factors can be quantitatively evaluated, however, further investigation of possible mass-dependent isotope effects in the reactions of ozone with O( 1 D), O( 3 P), O 2 ( 1 Δ), and O 2 ( 1 Σ) is needed through experiments we suggest here
18. Bilateral adrenal masses: a single-centre experience
Directory of Open Access Journals (Sweden)
Nilesh Lomte
2016-05-01
19. Bilateral adrenal masses: a single-centre experience
Science.gov (United States)
Bandgar, Tushar; Khare, Shruti; Jadhav, Swati; Lila, Anurag; Goroshi, Manjunath; Kasaliwal, Rajeev; Khadilkar, Kranti; Shah, Nalini S
2016-01-01
20. Measurement of the W boson mass in the Delphi experiment
International Nuclear Information System (INIS)
Simard, L.
2000-01-01
After the Z 0 study during the first phase of LEP, the properties of the W boson, in particular its mass, are precisely measured at LEP2. After the implications of that measurement on the Higgs mass being explained, the analysis of the WW semileptonic events, where the two W decay into two quarks, a charged lepton and a neutrino, is described. It was carried out with the data sample collected at DELPHI in 1997 and 1998, corresponding to an integrated luminosity of 211.1 pb -1 . The measurement, based upon a likelihood fit applied both to simulation and data requires that all variables of simulation reproduce well the data. Comparisons between Monte Carlo and data are set out, as well as the selection of WW events and the kinematical fit used to improve the mass resolution. The method used to estimate the systematic errors on the measurement and the result of the measurement are presented. When combining these measurements with the measurements done in the hadronic channel, the mass and the width are measured. (author)
1. Bark beetle management after a mass attack - some Swiss experiences
Science.gov (United States)
B. Forster; F. Meier; R. Gall
2003-01-01
In 1990 and 1999, heavy storms accompanied by the worst gales ever recorded in Switzerland, struck Europe and left millions of cubic metres of windthrown Norway spruce trees; this provided breeding material for the eight-toothed spruce bark beetle (Ips typographus L.) and led to mass attacks in subsequent years which resulted in the additional loss...
2. Neutrino mass and mixing, and non-accelerator experiments
International Nuclear Information System (INIS)
Robertson, R.G.H.
1992-01-01
We review the current status of experimental knowledge about neutrinos derived from kinematic mass measurements, neutrino oscillation searches at reactors and accelerators, solar neutrinos, atmospheric neutrinos, and single and double beta decay. The solar neutrino results yield fairly strong and consistent indication that neutrino oscillations are occurring. Other evidence for new physics is less consistent and convincing
3. Guided fine needle aspiration cytology of retroperitoneal masses - Our experience
Directory of Open Access Journals (Sweden)
2011-01-01
Full Text Available Background : Early pathological classification of retroperitoneal masses is important for pin-point diagnosis and timely management. Aims : This study was done to evaluate the usefulness and drawbacks of guided fine needle aspiration cytology (FNAC of retroperitoneal masses covering a period of two years with an intention to distinguish between neoplastic and non-neoplastic lesions and to correlate with histologic findings. Materials and Methods : FNAC was done under radiological guidance in all cases using long needle fitted with disposable syringe. Appropriate staining was done and cytology was correlated with histology which was taken as the gold standard for comparison. Results : Fifty-one patients who presented with retroperitoneal masses were studied. Forty-four lesions were malignant cytologically and 7 were inflammatory (tuberculous. According to radiological and cytologic findings, we classified our cases into four groups: renal tumors, retroperitoneal lymphadenopathy, germ cell tumors, soft tissue tumors. Except for cases of non-Hodgkin lymphoma (NHL and metastatic lesions, we had sensitivity and specificity of 100%. In NHL the sensitivity and specificity were both 50%. In cases of metastatic adenocarcinoma, the sensitivity and specificity were 84.6% and 81.8%, respectively. Conclusions : Ignoring the pitfalls, guided FNAC is still an inexpensive and reliable method of early diagnosis of retroperitoneal lesions.
4. Accelerator mass spectrometry at Peking University: experiments and progress
International Nuclear Information System (INIS)
Chen Jiaer; Guo Zhiyu; Yan Shengqing; Li Renxing; Xiao Min; Li Kun; Liu Hongtao; Liu Kexin; Wang Jianjun; Li Bin; Lu Xiangyang; Yuan Sixun; Chen Tiemei; Gao Shijun; Zheng Shuhui; Chen Chengye; Liu Yan
1997-01-01
The Peking University Accelerator Mass Spectrometer (PKUAMS) has been put into routine operation. 14 C measurements of archaeological samples with fast cycling injection have shown good results. The new multi-target high-intensity sputtering ion source has been tested and 10 Be measurements were carried out with a new detector in which both the stopping of the intense flux of 10 B ions and the identification of 10 Be ions are performed. 26 Al samples were also measured. While various applications show good prospects for PKUAMS, further upgrade is desirable
5. Testicular sparing surgery in small testis masses: A multinstitutional experience
Directory of Open Access Journals (Sweden)
Andrea B. Galosi
2016-12-01
Full Text Available Introduction: The incidence of benign testicular tumors is increasing in particular in small lesion incidentally found at scrotal ultrasonography. Primary aim of this study was to perform radical surgery in malignant tumor. Secondary aim was to verify the efficacy of the diagnostic-therapeutic pathway recently adopted in management of small masses with testis sparing surgery in benign lesions. Materials and methods: In this multicenter study, we reviewed all patients with single testis lesion less than 15 mm at ultrasound as main diameter. We applied the diagnostic-therapeutic pathway described by Sbrollini et al. (Arch Ital Urol Androl 2014; 86:397 which comprises: 1 testicular tumor markers, 2 repeated scrotal ultrasound at the tertiary center, 3 surgical exploration with inguinal approach, intraoperative ultrasound, and intraoperative pathological examination. Definitive histology was reviewed by a dedicated uro-pathologist. Results: Twenty-eight patients completed this clinical flowchart. The mean lesion size was 9.3 mm (range 2.5-15. Testicular tumor markers were normal except in a case. Intraoperative ultrasound was necessary in 8/28 cases. We treated 11/28 (39.3% with immediate radical orchiectomy and 17/28 (60.7% with testis-sparing surgery. Definitive pathological results were: malignant tumor in 6 cases (seminoma, benign tumor in 10 cases (5 Leydig tumors, 2 Sertoli tumors, 1 epidermoid cyst, 1 adenomatoid tumor, 1 angiofibroma, benign disease in 11 (8 inflammation with haemorragic infiltration, 2 tubular atrophy, 1 fibrosis, and normal parenchyma in 1 case. We observed a good concordance between frozen section examination and definitive histology. Any malignant tumor was treated conservatively. Any delayed orchiectomy was necessary based on definitive histology. Conclusions: The incidence of benign lesions in 60% of small testis lesions with normal tumor markers makes orchiectomy an overtreatment. Testicular sparing surgery of single
6. Study of microstrip gas chambers for CMS experiment and measurement of the W boson mass in the DELPHI experiment
International Nuclear Information System (INIS)
Ripp-Baudot, I.
2004-06-01
In this document the author describes 3 fields of his research activities: first, the development and validation tests of micro-strip gas chambers for the CMS experiment; secondly, the measurements of the W boson mass and width by analysing the events: e + e - → W + W - → qq-bar qq-bar whose data have been collected in the DELPHI experiment (at the LEP-2 accelerator); and thirdly, the tagging of b-jets that is an essential tool for the study of the top quark. The last chapter is dedicated to what is expected from LHC experiments concerning the properties of the quark top: mass, spin, production and decay channels
7. The origin of mass and experiments on high-energy particle accelerators
International Nuclear Information System (INIS)
Ioffe, B.L.
2006-01-01
The visible world is one consisting of nucleons and electrons. The mass of nucleon arises from chiral symmetry breaking in quantum chromodynamics, so high energy accelerator experiments cannot give a clue to the nature of mass of matter in the visible world. The origin of the mass of the matter will be clarified when the mechanism of chiral symmetry breaking in quantum chromodynamics is established [ru
8. Are particle rest masses variable: Theory and constraints from solar system experiments
International Nuclear Information System (INIS)
Bekenstein, J.D.
1977-01-01
Particle rest mass variation in spacetime is considered. According to Dicke, if this is the case various null experiments indicate that all masses vary in the same way. Their variation relative to the Planck-Wheeler mass defines a universal scalar rest-mass field. We construct the relativistic dynamics for this field based on very general assumptions. In addition, we assume Einstein's equations to be valid in Planck-Wheeler units. A special case of the theory coincides with Dicke's reformulation of Brans-Dicke theory as general relativity with variable rest masses. In the general case the rest-mass field is some power r of a scalar field which obeys an ordinary scalar equation with coupling to the curvature of strength q. The r and q are the only parameters of the theory. Comparison with experiment is facilitated by recasting the theory into units in which rest masses are constant, the Planck-Wheeler mass varies, and the metric satisfies the equations of a small subset of the scalar-tensor theories of gravitation. The results of solar system experiments, usually used to test general relativity, are here used to delimit the acceptable values of r and q. We conclude that if cosmological considerations are not invoked, then the solar system experiments do not rule out the possibility of rest-mass variability. That is, there are theories which agree with all null and solar system experiments, and yet contradict the strong equivalence principle by allowing rest masses to vary relative to the Planck-Wheeler mass. We show that the field theory of the rest-mass field can be quantized and interpreted in terms of massless scalar quanta which interact very weakly with matter. This explains why they have not turned up in high-energy experiments. In future reports we shall investigate the implications of various cosmological and astrophysical data for the theory of variable rest masses. The ultimate goal is a firm decision on whether rest masses vary or not
9. A basis independent formulation of the connection between quark mass matrices, CP violation and experiment
International Nuclear Information System (INIS)
Jarlskog, C.; Stockholm Univ.; Bergen Univ.
1985-01-01
In the standard electroweak model, with three families, a one-to-one correspondence between certain determinants involving quark mass matrices (m and m' for charge 2/3 and -1/3 quarks respectively) and the presence/absence of CP violation is given. In an arbitrary basis for mass matrices, the quantity Im det[mm + , m'm' + ] appropriately normalized is introduced as a measure of CP violation. By this measure, CP is not maximally violated in any transition in Nature. Finally, constraints on quark mass matrices are derived from experiment. Any model of mass matrices, with the ambition to explain Nature, must satisfy these conditions. (orig.)
10. Status report on the Livermore-Rockefeller-Fermilab neutrino mass experiment
International Nuclear Information System (INIS)
Fackler, O.; Mugge, M.; Sticker, H.; White, R.M.; Woerner, R.
1986-03-01
An experiment is being performed to determine the electron neutrino mass with the precision of a few eV by measuring the tritium beta decay energy distribution near the endpoint. Key features of the experiment are a 2 eV resolution electrostatic spectrometer and a high-activity frozen tritium source
11. SUMS preliminary design and data analysis development. [shuttle upper atmosphere mass spectrometer experiment
Science.gov (United States)
Hinson, E. W.
1981-01-01
The preliminary analysis and data analysis system development for the shuttle upper atmosphere mass spectrometer (SUMS) experiment are discussed. The SUMS experiment is designed to provide free stream atmospheric density, pressure, temperature, and mean molecular weight for the high altitude, high Mach number region.
12. Mass
International Nuclear Information System (INIS)
Quigg, Chris
2007-01-01
In the classical physics we inherited from Isaac Newton, mass does not arise, it simply is. The mass of a classical object is the sum of the masses of its parts. Albert Einstein showed that the mass of a body is a measure of its energy content, inviting us to consider the origins of mass. The protons we accelerate at Fermilab are prime examples of Einsteinian matter: nearly all of their mass arises from stored energy. Missing mass led to the discovery of the noble gases, and a new form of missing mass leads us to the notion of dark matter. Starting with a brief guided tour of the meanings of mass, the colloquium will explore the multiple origins of mass. We will see how far we have come toward understanding mass, and survey the issues that guide our research today.
13. New experiments in organic, fast-atom-bomdardment, and secondary-ion mass spectrometry
International Nuclear Information System (INIS)
DiDonato, G.C.
1987-01-01
The goal of research presented in this dissertation is the creative use of new ionization and instrumental techniques in mass spectrometry. This goal manifests itself in three areas of mass spectrometry. In the first portion, modern, state-of-the-art instrumentation and new experiments were used to re-examine the mass spectra of transition-metal acetates and acetylacetonates. High resolution, chemical ionization, negative chemical ionization, and extended-mass-range mass spectrometry uncovered a wealth of new gas-phase ionic species. Energy-resolved mass spectrometry/mass spectrometry was applied to the characterization of molecular and fragment ion first-row transition-metal acetylacetonates, and comprises the second portion of the thesis. Studies in fast-atom-bombardment mass spectrometry are the subject of the third portion of the dissertation. Since fast-atom bombardment samples a liquid matrix, absolute and relative abundances of sputtered secondary ions are influenced by solution chemistry. The design and construction of an imaging secondary-ion mass spectrometer is the subject of the final portion of the thesis. This instrument provides for direct mass-spectrometric analysis of thin-layer and paper chromatograms and electrophoretograms
14. U-233 fuelled low critical mass solution reactor experiment PURNIMA II
International Nuclear Information System (INIS)
Srinivasan, M.; Chandramoleshwar, K.; Pasupathy, C.S.; Rasheed, K.K.; Subba Rao, K.
1987-01-01
A homogeneous U-233 uranyl nitrate solution fuelled BeO reflected, low critical mass reactor has been built at the Bhabha Atomic Research Centre, India. Christened PURNIMA II, the reactor was used for the study of the variation of critical mass as a function of fuel solution concentration to determine the minimum critical mass achievable for this geometry. Other experiments performed include the determination of temperature coefficient of reactivity, study of time behaviour of photoneutrons produced due to interaction between decaying U-233 fission product gammas and the beryllium reflector and reactor noise measurements. Besides being the only operational U-233 fuelled reactor at present, PURNIMA II also has the distinction of having attained the lowest critical mass of 397 g of fissile fuel for any operating reactor at the current time. The paper briefly describes the facility and gives an account of the experiments performed and results achieved. (author)
15. Prospects for cosmic neutrino detection in tritium experiments in the case of hierarchical neutrino masses
International Nuclear Information System (INIS)
Blennow, Mattias
2008-01-01
We discuss the effects of neutrino mixing and the neutrino mass hierarchy when considering the capture of the cosmic neutrino background (CNB) on radioactive nuclei. The implications of mixing and hierarchy at future generations of tritium decay experiments are considered. We find that the CNB should be detectable at these experiments provided that the resolution for the kinetic energy of the outgoing electron can be pushed to a few 0.01 eV for the scenario with inverted neutrino mass hierarchy, about an order of magnitude better than that of the upcoming KATRIN experiment. Another order of magnitude improvement is needed in the case of normal neutrino mass hierarchy. We also note that mixing effects generally make the prospects for CNB detection worse due to an increased maximum energy of the normal beta decay background
16. Probing the eV-mass range for solar axions with the CAST experiment
International Nuclear Information System (INIS)
Vogel, J.
2009-01-01
The CERN Axion Solar Telescope (CAST) is searching for solar axions, which could be produced in the core of the Sun via the so-called Primakoff effect. For this purpose, CAST uses a decommissioned LHC prototype magnet. In its magnetic field of 9 Tesla axions could be reconverted into X-ray photons. The magnet is mounted on a structure built to follow the Sun during sunrise and sunset for a total of about 3 hours per day. The analysis of the data acquired during the first phase of the experiment with vacuum in the magnetic field region yielded the most restrictive experimental upper limit on the axion-to-photon coupling constant for axion masses up to about 0.02 eV. In order to extend the sensitivity of the experiment to a wider mass range, the CAST experiment continued its search for axions with helium in the magnet bores. In this way it is possible to restore coherence for larger masses. Changing the pressure of the helium gas enables the experiment to scan different axion masses. In the first part of this second phase of CAST, helium-4 has been used and the axion mass region was extended up to 0.4 eV. Therefore the experiment enters the regions favored by axion models. In CAST's ongoing helium-3 phase the studied mass range is now further extended. We will present the final results of CAST's helium-4 phase. Furthermore the latest upgrades of the experiments will be shown and an outlook on CAST's status and prospects will be given. (author)
17. Analysis of high mass resolution PTR-TOF mass spectra from 1,3,5-trimethylbenzene (TMB) environmental chamber experiments
Science.gov (United States)
Müller, M.; Graus, M.; Wisthaler, A.; Hansel, A.; Metzger, A.; Dommen, J.; Baltensperger, U.
2012-01-01
A series of 1,3,5-trimethylbenzene (TMB) photo-oxidation experiments was performed in the 27-m3 Paul Scherrer Institute environmental chamber under various NOx conditions. A University of Innsbruck prototype high resolution Proton Transfer Reaction Time-of-Flight Mass Spectrometer (PTR-TOF) was used for measurements of gas and particulate phase organics. The gas phase mass spectrum displayed ~200 ion signals during the TMB photo-oxidation experiments. Molecular formulas CmHnNoOp were determined and ion signals were separated and grouped according to their C, O and N numbers. This allowed to determine the time evolution of the O:C ratio and of the average carbon oxidation state solid #000; color: #000;">OSC of the reaction mixture. Both quantities were compared with master chemical mechanism (MCMv3.1) simulations. The O:C ratio in the particle phase was about twice the O:C ratio in the gas phase. Average carbon oxidation states of secondary organic aerosol (SOA) samples solid #000; color: #000;">OSCSOA were in the range of -0.34 to -0.31, in agreement with expected average carbon oxidation states of fresh SOA (solid #000; color: #000;">OSC = -0.5-0).
18. Required momentum, heat, and mass transport experiments for liquid-metal blankets
International Nuclear Information System (INIS)
Tillack, M.S.; Sze, D.K.; Abdou, M.A.
1986-01-01
Through the effects on fluid flow, many aspects of blanket behavior are affected by magnetohydrodynamic (MHD) effects, including pressure drop, heat transfer, mass transfer, and structural behavior. In this paper, a set of experiments is examined that could be performed in order to reduce the uncertainties in the highly related set of issues dealing with momentum, heat, and mass transport under the influence of a strong magnetic field (i.e., magnetic transport phenomena). By improving our basic understanding and by providing direct experimental data on blanket behavior, these experiments will lead to improved designs and an accurate assessment of the attractiveness of liquid-metal blankets
19. Signal and noise in Gravity Recovery and Climate Experiment (GRACE) observed surface mass variations
NARCIS (Netherlands)
Schrama, E.J.O.; Wouters, B.; Lavallée, D.A.
2007-01-01
The Gravity Recovery and Climate Experiment (GRACE) product used for this study consists of 43 monthly potential coefficient sets released by the GRACE science team which are used to generate surface mass thickness grids expressed as equivalent water heights (EQWHs). We optimized both the smoothing
20. Changes in Kicking Pattern: Effect of Experience, Speed, Accuracy, and Effective Striking Mass
Science.gov (United States)
Southard, Dan L.
2014-01-01
Purpose: The purposes of this study were to: (a) examine the effect of experience and goal constraints (speed, accuracy) on kicking patterns; (b) determine if effective striking mass was independent of ankle velocity at impact; and (c) determine the accuracy of kicks relative to independent factors. Method: Twenty participants were recruited to…
1. Final scientific and technical report: New experiments to measure the neutrino mass scale
Energy Technology Data Exchange (ETDEWEB)
Monreal, Benjamin [Univ. of California, Santa Barbara, CA (United States)
2016-11-19
In this work, we made material progress towards future measurements of the mass of the neutrino. The neutrino is a fundamental particle, first observed in the 1950s and subjected to particularly intense study over the past 20 years. It is now known to have some, non-zero mass, but we are in an unusual situation of knowing the mass exists but not knowing what value it takes. The mass may be determined by precise measurements of certain radioactive decay distributions, particularly the beta decay of tritium. The KATRIN experiment is an international project which is nearing the beginning of a tritium measurement campaign using a large electrostatic spectrumeter. This research included participation in KATRIN, including construction and delivery of a key calibration subsystem, the Rear Section''. To obtain sensitivity beyond KATRIN's, new techniques are required; this work included R&D on a new technique we call CRES (Cyclotron Resonance Electron Spectroscopy) which has promise to enable even more sensitive tritium decay measurements. We successfully carried out CRES spectroscopy in a model system in 2014, making an important step towards the design of a next-generation tritium experiment with new neutrino mass measurement abilities.
2. Mass hierarchy sensitivity of medium baseline reactor neutrino experiments with multiple detectors
Directory of Open Access Journals (Sweden)
Hong-Xin Wang
2017-05-01
Full Text Available We report the neutrino mass hierarchy (MH determination of medium baseline reactor neutrino experiments with multiple detectors, where the sensitivity of measuring the MH can be significantly improved by adding a near detector. Then the impact of the baseline and target mass of the near detector on the combined MH sensitivity has been studied thoroughly. The optimal selections of the baseline and target mass of the near detector are ∼12.5 km and ∼4 kton respectively for a far detector with the target mass of 20 kton and the baseline of 52.5 km. As typical examples of future medium baseline reactor neutrino experiments, the optimal location and target mass of the near detector are selected for the specific configurations of JUNO and RENO-50. Finally, we discuss distinct effects of the reactor antineutrino energy spectrum uncertainty for setups of a single detector and double detectors, which indicate that the spectrum uncertainty can be well constrained in the presence of the near detector.
3. Mass hierarchy sensitivity of medium baseline reactor neutrino experiments with multiple detectors
Energy Technology Data Exchange (ETDEWEB)
Wang, Hong-Xin, E-mail: [email protected] [Department of Physics, Nanjing University, Nanjing 210093 (China); Zhan, Liang; Li, Yu-Feng; Cao, Guo-Fu [Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049 (China); Chen, Shen-Jian [Department of Physics, Nanjing University, Nanjing 210093 (China)
2017-05-15
We report the neutrino mass hierarchy (MH) determination of medium baseline reactor neutrino experiments with multiple detectors, where the sensitivity of measuring the MH can be significantly improved by adding a near detector. Then the impact of the baseline and target mass of the near detector on the combined MH sensitivity has been studied thoroughly. The optimal selections of the baseline and target mass of the near detector are ∼12.5 km and ∼4 kton respectively for a far detector with the target mass of 20 kton and the baseline of 52.5 km. As typical examples of future medium baseline reactor neutrino experiments, the optimal location and target mass of the near detector are selected for the specific configurations of JUNO and RENO-50. Finally, we discuss distinct effects of the reactor antineutrino energy spectrum uncertainty for setups of a single detector and double detectors, which indicate that the spectrum uncertainty can be well constrained in the presence of the near detector.
4. Determining the neutrino mass hierarchy with INO, T2K, NOvA and reactor experiments
International Nuclear Information System (INIS)
Ghosh, Anushree; Choubey, Sandhya; Thakore, Tarak
2013-01-01
The relatively large measured value of θ 13 has opened up the possibility of determining the neutrino mass hierarchy through earth matter effects. Amongst the current accelerator experiments only NOvA has a long enough baseline to observe earth matter effects. However, even NOvA is plagued with uncertainty on the knowledge of the true value of Δ CP which drastically reduces its sensitivity to the neutrino mass hierarchy. Earth matter effects in atmospheric neutrinos on the other hand is almost independent of δ CP . The 50 kton magnetized Iron CALorimeter at the India-based Neutrino Observatory (ICAL at the rate lNO) will be observing atmospheric neutrinos. The charge identification capability of this detector gives it an edge over others for mass hierarchy determination through observation of earth matter effects. We study in detail the neutrino mass hierarchy sensitivity of the data from this experiment simulated using the Nuance based generator developed for ICAL at the rate lNO and folded with the detector resolution and efficiencies obtained by the INO collaboration from a full detector Geant based simulation. The data from ICAL at the rate lNO is then combined with simulated of T2K, NOvA Double Chooz, RENO and Daya Bay experiments and a combined sensitivity study to the mass hierarchy performed. With 10 years of ICAL at the rate lNO data combined with T2K, NOvA and reactor data, one could get 2.8σ - 5σ discovery for the neutrino mass hierarchy depending on the true value of (θ23, θ13 and δ CP . (author)
5. The Majorana Experiment: a Straightforward Neutrino Mass Experiment Using The Double-Beta Decay of Ge-76
International Nuclear Information System (INIS)
Miley, Harry S.; Y Suzuki; M Nakahata; Y Itow; M Shiozawa; Y Obayashi
2004-01-01
The Majorana Experiment proposes to measure the effective mass of the electron neutrino to as low as 0.02 eV using well-tested technology. A half life of about 4E27 y, corresponding to a mass range of [0.02 - 0.07] eV can be reached by operating 500 kg of germanium enriched to 86% in Ge-76 deep underground. Radiological backgrounds of cosmogenic or primordial origin will be greatly reduced by ultra-low background screening of detector, structural, and shielding materials, by chemical processing of materials, and by electronic rejection of multi-site events in the detector. Electronic background reduction is achieved with pulse shape analysis, detector segmentation, and detector-to detector coincidence rejection
6. Measuring the mass of the W{sup {+-}} boson in the ALEPH experiment; Mesure de la masse du boson W{sup {+-}} dans l'experience ALEPH
Energy Technology Data Exchange (ETDEWEB)
Boumediene, D.E
2002-05-01
The W boson plays an important role in the standard model of weak interactions, its mass is predictable and its measurement is an efficient way to test the theory and to challenge experimental data. This thesis is dedicated to the measurement of the W mass (m{sub W}) through the ALEPH experiment. 3 important points are treated: measuring techniques, systematic effects and impact of the value measured on the standard model. As for measuring techniques: all the decay products of W have been reconstructed, this reconstruction is dependent on the decay way and has an impact on the determination of the mass. For the decay W{sup +}W{sup -} {yields} {tau}{nu}qq-bar, a specific reconstruction has been studied and used for every step of the measurement (selection, kinematic adjustment, adjustment of m{sub W}), this technique has improved the resolution of m{sub W}. For the leptonic decay W{sup +}W{sup -} {yields} l{nu}l{nu}, we have focused on the adjustment technique of m{sub W}, for the decays W{sup +}W{sup -} {yields} l{nu}qq-bar and W{sup +}W{sup -} {yields} qq-bar qq-bar, the standard measurement technique of ALEPH has been used. As for systematic effects, the statistical precision of the measurement campaign is high so it is necessary to understand the systematic effects that have an impact on the measured value. 2 effects have been thoroughly studied: the colour interconnection effect that concerns only the hadronic decay with 4 jets, the second effect appears in the simulation of showers in calorimeter-detectors. 2 values for m{sub W} are proposed, one that was obtained by only measuring the direction and the energy of the jet and the second one by discarding all the problematic events that were reconstructed in the detector. We obtain: m{sub W} = (80,392 {+-} 0,053) GeV/c{sup 2} and m{sub W} = (80,358 {+-} 0,050) GeV/c{sup 2}. By combining these values to the internationally agreed data, it has been possible to adjust the mass of the Higgs' boson and to give
7. Investigation of trailing mass in Z-pinch implosions and comparison to experiment
Science.gov (United States)
Yu, Edmund
2007-11-01
Wire-array Z pinches represent efficient, high-power x-ray sources with application to inertial confinement fusion, high energy density plasmas, and laboratory astrophysics. The first stage of a wire-array Z pinch is described by a mass ablation phase, during which stationary wires cook off material, which is then accelerated radially inwards by the JxB force. The mass injection rate varies axially and azimuthally, so that once the ablation phase concludes, the subsequent implosion is highly 3D in nature. In particular, a network of trailing mass and current is left behind the imploding plasma sheath, which can significantly affect pinch performance. In this work we focus on the implosion phase, electing to model the mass ablation via a mass injection scheme. Such a scheme has a number of injection parameters, but this freedom also allows us to gain understanding into the nature of the trailing mass network. For instance, a new result illustrates the role of azimuthal correlation. For an implosion which is 100% azimuthally correlated (corresponding to an azimuthally symmetric 2D r-z problem), current is forced to flow on the imploding plasma sheath, resulting in strong Rayleigh-Taylor (RT) growth. If, however, the implosion is not azimuthally symmetric, the additional azimuthal degree of freedom opens up new conducting paths of lower magnetic energy through the trailing mass network, effectively reducing RT growth. Consequently the 3D implosion experiences lower RT growth than the 2D r-z equivalent, and actually results in a more shell-like implosion. A second major goal of this work is to constrain the injection parameters by comparison to a well-diagnosed experimental data set, in which array mass was varied. In collaboration with R. Lemke, M. Desjarlais, M. Cuneo, C. Jennings, D. Sinars, E. Waisman
8. The interlaboratory experiment IDA-72 on mass spectrometric isotope dilution analysis. Vol. 2
International Nuclear Information System (INIS)
Beyrich, W.; Drosselmeyer, E.
1975-07-01
Volume II of the report on the IDA-72 experiment contains papers written by different authors on a number of special topics connected with the preparation, performance and evaluation of the interlaboratory test. In detail the sampling procedures for active samples of the reprocessing plant and the preparation of inactive reference and spike solution from standard material are described as well as new methods of sample conditioning by evaporation. An extra chapter is devoted to the chemical sample treatment as a preparation for mass spectrometric analysis of the U and Pu content of the solutions. Special topics are also methods for mass discrimination corrections, α-spectrometer measurements as a supplement for the determination of Pu-238 and the comparison of concentration determinations by mass spectrometric isotope dilution analysis with those performed by X-ray fluorescence spectrometry. The last part of this volume contains papers connected with the computerized statistical evaluation of the high number of data. (orig.) [de
9. The Colorado Ultraviolet Transit Experiment (CUTE): Observing Mass Loss on Short-Period Exoplanets
Science.gov (United States)
Egan, Arika; Fleming, Brian; France, Kevin
2018-06-01
The Colorado Ultraviolet Transit Experiment (CUTE) is an NUV spectrograph packaged into a 6U CubeSat, designed to characterize the interaction between exoplanetary atmospheres and their host stars. CUTE will conduct a transit spectroscopy survey, gathering data over multiple transits on more than 12 short-period exoplanets with a range of masses and radii. The instrument will characterize the spectral properties of the transit light curves to atomic and molecular absorption features predicted to exist in the upper atmospheres of these planets, including Mg I, Mg II, Fe II, and OH. The shape and evolution of these spectral light curves will be used to quantify mass loss rates, the stellar drives of that mass loss, and the possible existence of exoplanetary magnetic fiends. This poster presents the science motivation for CUTE, planned observation and data analysis methods, and expected results.
10. DESI and other Dark Energy experiments in the era of neutrino mass measurements
Energy Technology Data Exchange (ETDEWEB)
Font-Ribera, Andreu [Institute of Theoretical Physics, University of Zurich, Winterthurerstrasse 190, Zurich, 8057 (Switzerland); McDonald, Patrick; Mostek, Nick; Reid, Beth A.; Seo, Hee-Jong [Lawrence Berkeley National Laboratory, 1 Cyclotron Road, Berkeley, CA, 94720 (United States); Slosar, Anže, E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected] [Brookhaven National Laboratory, Upton, NY, 11973 (United States)
2014-05-01
We present Fisher matrix projections for future cosmological parameter measurements, including neutrino masses, Dark Energy, curvature, modified gravity, the inflationary perturbation spectrum, non-Gaussianity, and dark radiation. We focus on DESI and generally redshift surveys (BOSS, HETDEX, eBOSS, Euclid, and WFIRST), but also include CMB (Planck) and weak gravitational lensing (DES and LSST) constraints. The goal is to present a consistent set of projections, for concrete experiments, which are otherwise scattered throughout many papers and proposals. We include neutrino mass as a free parameter in most projections, as it will inevitably be relevant — DESI and other experiments can measure the sum of neutrino masses to ∼ 0.02 eV or better, while the minimum possible sum is ∼ 0.06 eV. We note that constraints on Dark Energy are significantly degraded by the presence of neutrino mass uncertainty, especially when using galaxy clustering only as a probe of the BAO distance scale (because this introduces additional uncertainty in the background evolution after the CMB epoch). Using broadband galaxy power becomes relatively more powerful, and bigger gains are achieved by combining lensing survey constraints with redshift survey constraints. We do not try to be especially innovative, e.g., with complex treatments of potential systematic errors — these projections are intended as a straightforward baseline for comparison to more detailed analyses.
11. Impact of the Spread of Mass Education on Married Women’s Experience with Domestic Violence
Science.gov (United States)
Ghimire, Dirgha J.; Axinn, William G.; Smith-Greenaway, Emily
2015-01-01
This paper investigates the association between mass education and married women’s experience with domestic violence in rural Nepal. Previous research on domestic violence in South Asian societies emphasizes patriarchal ideology and the widespread subordinate status of women within their communities and families. The recent spread of mass education is likely to shift these gendered dynamics, thereby lowering women’s likelihood of experiencing domestic violence. Using data from 1,775 currently married women from the Chitwan Valley Family Study in Nepal, we provide a thorough analysis of how the spread of mass education is associated with domestic violence among married women. The results show that women’s childhood access to school, their parents’ schooling, their own schooling, and their husbands’ schooling are each associated with their lower likelihood of experiencing domestic violence. Indeed, husbands’ education has a particularly strong, inverse association with women’s likelihood of experiencing domestic violence. These associations suggest that the proliferation of mass education will lead to a marked decline in women’s experience with domestic violence in Nepal. PMID:26463551
12. Impact of the spread of mass education on married women's experience with domestic violence.
Science.gov (United States)
Ghimire, Dirgha J; Axinn, William G; Smith-Greenaway, Emily
2015-11-01
13. Study of point defects in non crystalline alloys by high temperature mass transport experiments
International Nuclear Information System (INIS)
Limoge, Y.
1986-09-01
We present in this communication the results of new experiments designed to study the mass transport mechanism in non-crystalline metallic alloys. They are based on the isothermal measurement of the crystallization kinetics, either without constraint or under electron irradiation or hydrostatic pressure. These experiments show that in the alloys studied, (FeNi) 8 (Pb) 2 and Ni 6 Nb 4 ), irradiation enhances the diffusion on the one hand, and on the other that there exist an activation volume for diffusion, of the order of one atomic volume. We discuss then the atomic model of diffusion needed to explain our results
14. Consideration of possible mass and velocity corrections to magnetic cluster experiments
International Nuclear Information System (INIS)
Liu, Z.Y.; Dowben, P.A.; Popov, A.P.; Pappas, David P.
2003-01-01
Gadolinium occurs, in natural abundance, as several isotopes. The possible combinations of different gadolinium isotopes dictates that even for a fixed number of atoms in the cluster, clusters of gadolinium atoms will exhibit a range of masses. This and the expected consequence of the translation energy distributions are explored as possible corrections to Stern-Gerlach cluster beam-deflection experiments. Upon closer inspection of the experimental data, we find that the translation energy plus the vibrational temperature distribution may be inhomogeneous. This could be the origin of a long tail to high deflections in the experimental deflection profiles, at low cluster temperatures, in the magnetic cluster Stern-Gerlach experiments
15. Estudo experimental do volume molar excesso de soluções de dietil-1mina/acetonitrila e s-butil-amina/acetonitrila na faixa de 288,15 a 303,15k, a pressão atmosferica, com aplicação do modelo Eras
OpenAIRE
Ricardo Belchior Torres
1998-01-01
Resumo: Neste trabalho estudou-se o volume molar excesso vE de soluções de dietil-amina + acetonitrila e s-butil-amina + acetonitrila, à pressão atmosférica, em função da composição e temperatura. A metodologia utilizada foi a densitometria, utilizando um densímetro de oscilação mecânica fabricado pela Anton Paar (Modelo DMA 55). Os dados experimentais foram ajustados em função da composição e da temperatura, através de um polinômio de Redlich-Kister, e utilizados nos cálculos de grandezas as...
16. The Majorana experiment. A straightforward neutrino mass experiment using the double-beta decay of 76Ge
International Nuclear Information System (INIS)
Miley, H.S.
2004-01-01
The Majorana Experiment proposes to measure the effective mass of the electron neutrino to as low as 0.02 eV using well-tested technology. A half-life of about 4E27 y, corresponding to a mass range of [0.02 - 0.07] eV can be reached by operating 500 kg of germanium enriched to 86% in 76 Ge deep underground. Radiological backgrounds of cosmogenic or primordial origin will be greatly reduced by ultra-low-background screening of detector, structural, and shielding materials, by chemical processing of materials, and by electronic rejection of multi-site events in the detector. Electronic background reduction is achieved with pulse-shape analysis, detector segmentation, and detector-to-detector coincidence rejection. Sensitivity calculations assuming worst-case germanium cosmogenic activation predict rapid growth in mass sensitivity (T1/2 at 90%CL) after the beginning of detector production: [0.08-0.28] eV at ∼1 year, [0.04-0.14] eV at ∼2.5 years, [0.03-0.10] eV at ∼5 years, and [0.02-0.07] eV at ∼10 years. The impact of primordial backgrounds in structural and electronic components is being studied at the 1 μBq/kg level, and appears to be controllable to below levels needed to attain these results. (author)
17. The Majorana Experiment:. a Straightforward Neutrino Mass Experiment Using the Double-Beta Decay of 76GE
Science.gov (United States)
Miley, H. S.
2004-04-01
The Majorana Experiment proposes to measure the effective mass of the electron neutrino to as low as 0.02 eV using well-tested technology. A half-life of about 4E27 y, corresponding to a mass range of [0.02 - 0.07] eV can be reached by operating 500 kg of germanium enriched to 86% in 76Ge deep underground. Radiological backgrounds of cosmogenic or primordial origin will be greatly reduced by ultra-low-background screening of detector, structural, and shielding materials, by chemical processing of materials, and by electronic rejection of multi-site events in the detector. Electronic background reduction is achieved with pulse-shape analysis, detector segmentation, and detector-to-detector coincidence rejection. Sensitivity calculations assuming worst-case germanium cosmogenic activation predict rapid growth in mass sensitivity (T1/2 at 90%CL) after the beginning of detector production: [0.08-0.28] eV at ~1 year, [0.04-0.14] eV at ~2.5 years, [0.03-0.10] eV at ~5 years, and [0.02 - 0.07] eV at ~10 years. The impact of primordial backgrounds in structural and electronic components is being studied at the 1 μBq/kg level, and appears to be controllable to below levels needed to attain these results.
18. Determination of the Electron Neutrino Mass from Experiments on Electron-Capture Beta-Decay (EC)
CERN Multimedia
2002-01-01
The aim of the programme is to measure the electron-neutrino mass, for which at present an upper limit of 500~eV is known. \\\\ \\\\ The experiment studies the shape of the internal bremsstrahlung spectrum in electron-capture near its upper end-point and deduces a mass from small shape changes completely analogous to those in the well-known determination of the electron antineutrino mass in the tritium beta-minus decay. \\\\ \\\\ In a low-energy bremsstrahlung process, the capture takes place from a virtual S state associated with a radiative P~@A~S electromagnetic transition, and the resonant nature of the process leads to important enhancements of the photon intensities at low energy, in particular near the resonance energies co (X-rays). This effect gives this type of experiment a chance to compete with experiments on continuous beta spectra. \\\\ \\\\ The programme concentrates on two long-lived isotopes: \\\\ \\\\ 1)~~|1|6|3Ho. The Q value for this isotope has been found to be 2.6-2.7 keV. A detector specially construct...
19. A Search for New High-Mass Phenomena Producing Top Quarks with the ATLAS Experiment
CERN Document Server
The ATLAS collaboration
2011-01-01
A search for top quark pair resonances in the lepton plus jets final states has been performed with the ATLAS experiment at the LHC. The search uses a data sample corresponding to an integrated luminosity of 33 pb$^{-1}$, and was recorded at a proton-proton centre-of-mass energy of 7 TeV. No evidence of a resonance is found. Using the reconstructed $t\\bar{t}$ mass spectrum, limits are set on the production cross-section times branching ratio to $t\\bar{t}$ for narrow $Z'$ models. The observed 95\\% C.L. limits range from approximately 55~pb to 2.2~pb for masses going from $m=$ 500 GeV to $m=$ 1000 GeV. The analysis is also used to set limits on enhanced top quark production at high $t+X$ mass, using the production of quantum black holes to model the signal. In that context, enhanced $t+X$ production with a mass threshold below 2.35 TeV is excluded.
20. Interferometry with particles of non-zero rest mass: topological experiments
International Nuclear Information System (INIS)
Opat, G.I.
1994-01-01
Interferometry as a space-time process is described, together with its topology. Starting from this viewpoint, a convenient unified formalism for the phase shifts which arise in particle interferometry is developed. This formalism is based on a covariant form of Hamilton's action principle and Lagrange's equations of motion. It will be shown that this Lorentz invariant formalism yields a simple perturbation theoretic expression for the general phase shift that arises in matter-wave interferometry. The Lagrangian formalism is compared with the more usual formalism based on the wave propagation vector and frequency. The resulting formalism will be used to analyse the Sagnac effect, gravitational field measurements, and several Aharonov-Bohm-like topological phase shifts. Several topological interferometric experiments using particles of non-zero rest mass are discussed. These experiments involve the use of electrons, neutrons and neutral atoms. Neutron experiments will be emphasised. 45 refs., 15 figs
1. The influence of mass transfer on solute transport in column experiments with an aggregated soil
Science.gov (United States)
Roberts, Paul V.; Goltz, Mark N.; Summers, R. Scott; Crittenden, John C.; Nkedi-Kizza, Peter
1987-06-01
The spreading of concentration fronts in dynamic column experiments conducted with a porous, aggregated soil is analyzed by means of a previously documented transport model (DFPSDM) that accounts for longitudinal dispersion, external mass transfer in the boundary layer surrounding the aggregate particles, and diffusion in the intra-aggregate pores. The data are drawn from a previous report on the transport of tritiated water, chloride, and calcium ion in a column filled with Ione soil having an average aggregate particle diameter of 0.34 cm, at pore water velocities from 3 to 143 cm/h. The parameters for dispersion, external mass transfer, and internal diffusion were predicted for the experimental conditions by means of generalized correlations, independent of the column data. The predicted degree of solute front-spreading agreed well with the experimental observations. Consistent with the aggregate porosity of 45%, the tortuosity factor for internal pore diffusion was approximately equal to 2. Quantitative criteria for the spreading influence of the three mechanisms are evaluated with respect to the column data. Hydrodynamic dispersion is thought to have governed the front shape in the experiments at low velocity, and internal pore diffusion is believed to have dominated at high velocity; the external mass transfer resistance played a minor role under all conditions. A transport model such as DFPSDM is useful for interpreting column data with regard to the mechanisms controlling concentration front dynamics, but care must be exercised to avoid confounding the effects of the relevant processes.
2. Optimizing Mass Spectrometry Analyses: A Tailored Review on the Utility of Design of Experiments.
Science.gov (United States)
Hecht, Elizabeth S; Oberg, Ann L; Muddiman, David C
2016-05-01
Mass spectrometry (MS) has emerged as a tool that can analyze nearly all classes of molecules, with its scope rapidly expanding in the areas of post-translational modifications, MS instrumentation, and many others. Yet integration of novel analyte preparatory and purification methods with existing or novel mass spectrometers can introduce new challenges for MS sensitivity. The mechanisms that govern detection by MS are particularly complex and interdependent, including ionization efficiency, ion suppression, and transmission. Performance of both off-line and MS methods can be optimized separately or, when appropriate, simultaneously through statistical designs, broadly referred to as "design of experiments" (DOE). The following review provides a tutorial-like guide into the selection of DOE for MS experiments, the practices for modeling and optimization of response variables, and the available software tools that support DOE implementation in any laboratory. This review comes 3 years after the latest DOE review (Hibbert DB, 2012), which provided a comprehensive overview on the types of designs available and their statistical construction. Since that time, new classes of DOE, such as the definitive screening design, have emerged and new calls have been made for mass spectrometrists to adopt the practice. Rather than exhaustively cover all possible designs, we have highlighted the three most practical DOE classes available to mass spectrometrists. This review further differentiates itself by providing expert recommendations for experimental setup and defining DOE entirely in the context of three case-studies that highlight the utility of different designs to achieve different goals. A step-by-step tutorial is also provided.
3. A low-mass faraday cup experiment for the solar wind
Science.gov (United States)
Lazarus, A. J.; Steinberg, J. T.; Mcnutt, R. L., Jr.
1993-01-01
Faraday cups have proven to be very reliable and accurate instruments capable of making 3-D velocity distribution measurements on spinning or 3-axis stabilized spacecraft. Faraday cup instrumentation continues to be appropriate for heliospheric missions. As an example, the reductions in mass possible relative to the solar wind detection system about to be flown on the WIND spacecraft were estimated. Through the use of technology developed or used at the MIT Center for Space Research but were not able to utilize for WIND: surface-mount packaging, field-programmable gate arrays, an optically-switched high voltage supply, and an integrated-circuit power converter, it was estimated that the mass of the Faraday Cup system could be reduced from 5 kg to 1.8 kg. Further redesign of the electronics incorporating hybrid integrated circuits as well as a decrease in the sensor size, with a corresponding increase in measurement cycle time, could lead to a significantly lower mass for other mission applications. Reduction in mass of the entire spacecraft-experiment system is critically dependent on early and continual collaborative efforts between the spacecraft engineers and the experimenters. Those efforts concern a range of issues from spacecraft structure to data systems to the spacecraft power voltage levels. Requirements for flight qualification affect use of newer, lighter electronics packaging and its implementation; the issue of quality assurance needs to be specifically addressed. Lower cost and reduced mass can best be achieved through the efforts of a relatively small group dedicated to the success of the mission. Such a group needs a fixed budget and greater control over quality assurance requirements, together with a reasonable oversight mechanism.
4. Determination of the W boson mass in the hadronic decay channel in the DELPHI experiment
International Nuclear Information System (INIS)
Zoller, Ph.
2001-04-01
The subject of this thesis is the determination of the boson W mass in the fully hadronic channel within DELPHI experiment. Data are collected by the detector DELPHI during 1998 and 1999 with the e + e - collider LEP at CERN. W bosons collected were produced by pair in the process e + e - → W + W - with a center of mass energy between 189 and 202 GeV. A significant part of work in the thesis was devoted to the pairing algorithm creation for the four jets and to the final mass determination of the W boson. An event per event log-likelihood method was used for the constrained fit. We also compare for each event selected the reconstructed mass for W + and W - bosons with simulated histograms corrected by a ratio of a modified Breit-Wigner function. Finally, the last part of my work has consisted in studying the influence, on one hand, of some parameters resulting from theoretical models and, on other hand, of the experimental observables responsible of a mass shift. So, I was able to study the systematics effects owing to the final state interactions, to parameters used by models of fragmentation and to their energy and angular resolution. The final result we get is M W = (80.430 ± 0.081 (stat) ± 0.037 (syst) ± 0.050 (FSI) ± 0.017 (LEP)) GeV/c 2 where 'stat' means statistical error, 'LEP' means systematic error due to the LEP beam energy, 'FSI' means systematic error due to the interconnection in the final state, and 'syst' means other systematic errors
5. Late time neutrino masses, the LSND experiment, and the cosmic microwave background.
Science.gov (United States)
Chacko, Z; Hall, Lawrence J; Oliver, Steven J; Perelstein, Maxim
2005-03-25
Models with low-scale breaking of global symmetries in the neutrino sector provide an alternative to the seesaw mechanism for understanding why neutrinos are light. Such models can easily incorporate light sterile neutrinos required by the Liquid Scintillator Neutrino Detector experiment. Furthermore, the constraints on the sterile neutrino properties from nucleosynthesis and large-scale structure can be removed due to the nonconventional cosmological evolution of neutrino masses and densities. We present explicit, fully realistic supersymmetric models, and discuss the characteristic signatures predicted in the angular distributions of the cosmic microwave background.
6. Parameterization experiments performed via synthetic mass movements prototypes generated by 3D slope stability simulator
Science.gov (United States)
Colangelo, Antonio C.
2010-05-01
each cell in synthetic slope systems performed by relief unity emulator. The central methodological strategy is to locate the potential rupture surfaces (prs), main material discontinuities, like soil-regolith or regolith-rock transitions. Inner these "prs", we would to outline the effective potential rupture surfaces (eprs). This surface is a sub-set of the "prs" that presents safety factor less than unity (fwalls, the "slope stability simulator" generates a synthetic mass movement. The overlay material will slide until that a new equilibrium be attained at residual shear strength. These devices generate graphic 3D cinematic sequences of experiments in synthetic slope systems and numerical results about physical and morphological data about scars and deposits. Thus, we have a detailed geotechnical, morphological, topographic and morphometric description of these mass movements prototypes, for deal with effective mass movements found in the real environments.
7. An experiment to measure the electron neutrino mass using a cryogenic tritium source
International Nuclear Information System (INIS)
Fackler, O.; Jeziorski, B.; Kolos, W.; Monkhorst, H.; Mugge, M.; Sticker, H.; Szalewicz, K.; White, R.M.; Woerner, R.
1985-01-01
An experiment has been performed to determine the electron neutrino mass with the precision of a few eV by measuring the tritium beta decay energy distribution near the endpoint. Key features of the experiment are a 2 eV resolution electrostatic spectrometer and a high-activity frozen tritium source. It is important that the source have electronic wavefunctions which can be accurately calculated. These calculations have been made for tritium and the HeT + daughter ion and allow determination of branching fractions to 0.1% and energy of the excited states to 0.1 eV. The excited final molecular state calculations and the experimental apparatus are discussed. 4 refs., 5 figs
8. FIX-II/2032, BWR Pump Trip Experiment 2032, Simulation Mass Flow and Power Transients
International Nuclear Information System (INIS)
1988-01-01
1 - Description of test facility: In the FIX-II pump trip experiments, mass flow and power transients were simulated subsequent to a total loss of power to the recirculation pumps in an internal pump boiling water reactor. The aim was to determine the initial power limit to give dryout in the fuel bundle for the specified transient. In addition, the peak cladding temperature was measured and the rewetting was studied. 2 - Description of test: Pump trip experiment 2032 was a part of test group 2, i.e. the mass flow transient was to simulate the pump coast down with a pump inertia of 11.3 kg.m -2 . The initial power in the 36-rod bundle was 4.44 MW which gave dryout after 1.4 s from the start of the flow transient. A maximum rod cladding temperature of 457 degrees C was measured. Rewetting was obtained after 7.6 s. 3 - Experimental limitations or shortcomings: No ECCS injection systems
9. Further development of drag bodies for the measurement of mass flow rates during blowdown experiments
International Nuclear Information System (INIS)
Brockmann, E.; John, H.; Reimann, J.
1983-01-01
Drag bodies have already been used for sometime for the measurement of mass flow rates in blowdown experiments. Former research concerning the drag body behaviour in non-homogeneous two-phase flows frequently dealt with special effects by means of theoretical models only. For pipe flows most investigations were conducted for ratios of drag plate area to pipe cross section smaller 0.02. The present paper gives the results of experiments with drag bodies in a horizontal, non-homogeneous two-phase pipe flow with slip, which were carried through under the sponsorship of the German Ministry for Research and Technology (BMFT). Special interest was layed on the behaviour of the drag coefficient in stationary flows and at various cross sectional ratios. Both design and response of various drag bodies, which were developed at the Battelle-Institut, were tested in stationary and instationary two-phase flows. The influences of density and velocity profiles as well as the drag body position were studied. The results demonstrate, that the drag body is capable of measuring mass flow rates in connection with a gamma densitometer also in non-homogeneous two-phase flows. Satisfying results could be obtained, using simply the drag coefficient which was determined from single-phase flow calibrations
10. Estimation of relative permeability and capillary pressure from mass imbibition experiments
Science.gov (United States)
Alyafei, Nayef; Blunt, Martin J.
2018-05-01
We perform spontaneous imbibition experiments on three carbonates - Estaillades, Ketton, and Portland - which are three quarry limestones that have very different pore structures and span wide range of permeability. We measure the mass of water imbibed in air saturated cores as a function of time under strongly water-wet conditions. Specifically, we perform co-current spontaneous experiments using a highly sensitive balance to measure the mass imbibed as a function of time for the three rocks. We use cores measuring 37 mm in diameter and three lengths of approximately 76 mm, 204 mm, and 290 mm. We show that the amount imbibed scales as the square root of time and find the parameter C, where the volume imbibed per unit cross-sectional area at time t is Ct1/2. We find higher C values for higher permeability rocks. Employing semi-analytical solutions for one-dimensional flow and using reasonable estimates of relative permeability and capillary pressure, we can match the experimental data. We finally discuss how, in combination with conventional measurements, we can use theoretical solutions and imbibition measurements to find or constrain relative permeability and capillary pressure.
11. Safety analysis report for the Hanford Critical Mass Laboratory: Supplement No. 2. Experiments with heterogeneous assemblies
International Nuclear Information System (INIS)
Gore, B.F.; Davenport, L.C.
1981-04-01
Factors affecting the safety of criticality experiments using heterogeneous assemblies are described and assessed. It is concluded that there is no substantial change in safety from experiments already being routinely performed at the Critical Mass Laboratory (CML), and that laboratory and personnel safety are adequately provided by the combination of engineered and administrative safety limits enforced at the CML. This conclusion is based on the analysis of operational controls, potential hazards, and the consequences of accidents. Contingencies considered that could affect nuclear criticality include manual changes in fuel loadings, water flooding, fire, explosion, loss of services, earthquake, windstorm, and flood. Other potential hazards considered include radiation exposure to personnel, and potential releases within the Assembly Room and outside to the environment. It is concluded that the Maximum Credible Nuclear Burst of 3 x 10 18 fissions (which served as the design basis for the CML) is valid for heterogeneous assemblies as well as homogeneous assemblies. This is based upon examination of the results of reactor destructive tests and the results of the SL-1 reactor destructive accident. The production of blast effects which might jeopardize the CML critical assembly room (of thick reinforced concrete) is not considered credible due to the extreme circumstances required to produce blast effects in reactor destructive tests. Consequently, it is concluded that, for experiments with heterogeneous assemblies, the consequences of the Maximum Credible Burst are unchanged from those previously estimated for experiments with homogeneous systems
12. Experience in the management of the mass casualty from the January 2010 Jos Crisis.
Science.gov (United States)
Ozoilo, K N; Amupitan, I; Peter, S D; Ojo, E O; Ismaila, B O; Ode, M; Adoga, A A; Adoga, A S
2016-01-01
On the 17 of January 2010, a sectarian crisis broke out in Jos the capital of Plateau state, Nigeria. It created a mass casualty situation in the Jos University Teaching Hospital. We present the result of the hospital management of that mass casualty incident. To share our experience in the management of the mass casualty situation arising from the sectarian crisis of Jos in January 2010. We retrospectively reviewed the hospital records of patients who were treated in our hospital with injuries sustained in the Jos crisis of January 2010. A total of 168 patients presented over a four day period. There were 108 males (64.3%) and 60 females (35.7%). The mean age was 26 ± 16 years. Injury was caused by gunshots in 68 patients (40.5%), machete in 56 (33.3%), falls in 22 (13.1%) and burning in 21 (13.1%). The body parts injured were the upper limbs in 61(36.3%) patients, lower limbs 44 (26.2%) and scalp 43 (25.6%). Majority, 125 (74.4%) did not require formal operative care. Fourteen (8.3%) patients had complications out of which 10 (6.0%) were related to infections. There were 5 (3.1%) hospital mortalities and the mean duration of hospital stay was 4.2 days. The hospital operations returned to routine 24 hours after the last patient was brought in. As a result of changes made to our protocol, management proceeded smoothly and there was no stoppage of the hospital response at any point. This civil crisis involved mostly young males. Injuries were mainly lacerations from machete and gunshot injuries. Majority of the victims did not require formal surgical operations beyond initial care. Maintaining continuity in the positions of the Incident commander and the mass casualty commander ensure a smooth disaster response with fewer challenges.
13. Fundamental Drop Dynamics and Mass Transfer Experiments to Support Solvent Extraction Modeling Efforts
International Nuclear Information System (INIS)
Christensen, Kristi; Rutledge, Veronica; Garn, Troy
2011-01-01
In support of the Nuclear Energy Advanced Modeling Simulation Safeguards and Separations (NEAMS SafeSep) program, the Idaho National Laboratory (INL) worked in collaboration with Los Alamos National Laboratory (LANL) to further a modeling effort designed to predict mass transfer behavior for selected metal species between individual dispersed drops and a continuous phase in a two phase liquid-liquid extraction (LLE) system. The purpose of the model is to understand the fundamental processes of mass transfer that occur at the drop interface. This fundamental understanding can be extended to support modeling of larger LLE equipment such as mixer settlers, pulse columns, and centrifugal contactors. The work performed at the INL involved gathering the necessary experimental data to support the modeling effort. A custom experimental apparatus was designed and built for performing drop contact experiments to measure mass transfer coefficients as a function of contact time. A high speed digital camera was used in conjunction with the apparatus to measure size, shape, and velocity of the drops. In addition to drop data, the physical properties of the experimental fluids were measured to be used as input data for the model. Physical properties measurements included density, viscosity, surface tension and interfacial tension. Additionally, self diffusion coefficients for the selected metal species in each experimental solution were measured, and the distribution coefficient for the metal partitioning between phases was determined. At the completion of this work, the INL has determined the mass transfer coefficient and a velocity profile for drops rising by buoyancy through a continuous medium under a specific set of experimental conditions. Additionally, a complete set of experimentally determined fluid properties has been obtained. All data will be provided to LANL to support the modeling effort.
14. Single center experience with percutaneous and laparoscopic cryoablation of small renal masses.
Science.gov (United States)
Malcolm, John B; Berry, Tristan T; Williams, Michael B; Logan, Joshua E; Given, Robert W; Lance, Raymond S; Barone, Bethany; Shaves, Sarah; Vingan, Harlan; Fabrizio, Michael D
2009-06-01
While partial nephrectomy remains the gold standard for the management of most small renal masses, increasing experience with renal cryoablation has suggested a viable alternative with a favorable morbidity profile and good efficacy. We report intermediate-term oncologic outcomes from a single-center experience with laparoscopic and percutaneous renal cryoablation. We performed a retrospective review of our laparoscopic renal cryoablation (LRC) and percutaneous renal cryoablation (PRC) experience between January 2003 and April 2007. Patients with at least 12 months of follow-up were included in the analysis. Follow-up consisted of imaging and laboratory studies at regular intervals. Persistent mass enhancement or interval tumor growth was considered a treatment failure. Sixty-six patients (44% women/56% men; 42% African-American/58% Caucasian/other; mean body mass index, 29.7) with 72 tumors underwent either LRC (n = 52) or PRC (n = 20) with a mean follow-up of 30 months (median 25.1 mos; range 13-63 mos). Average patient age was 66.5 years (range 34-82 yrs). Mean tumor size was 2.33 cm (range 1-4.6 cm). Comorbid conditions were prevalent: 76% hypertension, 36% hyperlipidemia, 24% chronic kidney disease, 29% diabetes mellitus, 36% tobacco use, and 32% heart disease. RESULTS of pretreatment biopsy were 62% renal-cell carcinoma and 38% benign or nondiagnostic. Overall cancer-specific and cancer-free survival were 100% and 97%, respectively. There were two treatment failures (3.8%) in the LRC group and five primary failures in the PRC group (25%) (P = 0.015), four of which were salvaged with repeated PRC with no evidence of recurrence at 6 to 36 months of follow-up. There has been no significant local or metastatic progression. LRC and PRC achieved good oncologic control with minimal morbidity at a mean follow-up of 30 months in a patient cohort characterized by numerous comorbid conditions. PRC had a significantly higher primary treatment failure rate than LRC, but
15. Measuring the mass of the W± boson in the ALEPH experiment
International Nuclear Information System (INIS)
Boumediene, D.E.
2002-05-01
The W boson plays an important role in the standard model of weak interactions, its mass is predictable and its measurement is an efficient way to test the theory and to challenge experimental data. This thesis is dedicated to the measurement of the W mass (m W ) through the ALEPH experiment. 3 important points are treated: measuring techniques, systematic effects and impact of the value measured on the standard model. As for measuring techniques: all the decay products of W have been reconstructed, this reconstruction is dependent on the decay way and has an impact on the determination of the mass. For the decay W + W - → τνqq-bar, a specific reconstruction has been studied and used for every step of the measurement (selection, kinematic adjustment, adjustment of m W ), this technique has improved the resolution of m W . For the leptonic decay W + W - → lνlν, we have focused on the adjustment technique of m W , for the decays W + W - → lνqq-bar and W + W - → qq-bar qq-bar, the standard measurement technique of ALEPH has been used. As for systematic effects, the statistical precision of the measurement campaign is high so it is necessary to understand the systematic effects that have an impact on the measured value. 2 effects have been thoroughly studied: the colour interconnection effect that concerns only the hadronic decay with 4 jets, the second effect appears in the simulation of showers in calorimeter-detectors. 2 values for m W are proposed, one that was obtained by only measuring the direction and the energy of the jet and the second one by discarding all the problematic events that were reconstructed in the detector. We obtain: m W = (80,392 ± 0,053) GeV/c 2 and m W = (80,358 ± 0,050) GeV/c 2 . By combining these values to the internationally agreed data, it has been possible to adjust the mass of the Higgs' boson and to give an upper limit for its mass: m H 2 . (A.C.)
16. Experiences with gStore, a scalable mass storage system with tape backend
International Nuclear Information System (INIS)
Goeringer, H; Feyerabend, M; Sedykh, S
2008-01-01
GSI is a center for heavy ion research and host of an Alice Tier2 center. The GSI Mass Storage System gStore manages ∼200 TB experiment data currently with different life times and access patterns. The data are available 24 hours per day and seven days per week for fast and highly parallel access. For Alice users all gStore data are worldwide accessible via Alice grid software, and for the end of 2007 it is planned to provide ∼200TB via xrootd backed with gStore. Successfully in operation for more than 10 years gStore has been developed in parallel continuously by only two FTEs mastering a growth of nearly two orders of magnitude. In 2014 the future FAIR experiments at GSI, CBM and Panda, will have requirements for data capacity and I/O bandwidth reaching those of the current LHC experiments at CERN. This needs another growth of gStore of three orders of magnitude. This paper describes gStore and its potential to master also the challenges coming with the FAIR project
17. Experiences with gStore, a scalable mass storage system with tape backend
Energy Technology Data Exchange (ETDEWEB)
Goeringer, H; Feyerabend, M; Sedykh, S [GSI, Gesellschaft fuer Schwerionenforschung Darmstadt (Germany)
2008-07-15
GSI is a center for heavy ion research and host of an Alice Tier2 center. The GSI Mass Storage System gStore manages {approx}200 TB experiment data currently with different life times and access patterns. The data are available 24 hours per day and seven days per week for fast and highly parallel access. For Alice users all gStore data are worldwide accessible via Alice grid software, and for the end of 2007 it is planned to provide {approx}200TB via xrootd backed with gStore. Successfully in operation for more than 10 years gStore has been developed in parallel continuously by only two FTEs mastering a growth of nearly two orders of magnitude. In 2014 the future FAIR experiments at GSI, CBM and Panda, will have requirements for data capacity and I/O bandwidth reaching those of the current LHC experiments at CERN. This needs another growth of gStore of three orders of magnitude. This paper describes gStore and its potential to master also the challenges coming with the FAIR project.
18. Measurement of the W{+-} boson mass in the ALEPH experiment at LEP; Mesure de la masse du boson W{+-} dans l`experience ALEPH au LEP
Energy Technology Data Exchange (ETDEWEB)
Trabelsi, A.
1996-12-31
A precise measurement of the W boson mass is a powerful test of the standard model (SM) and its possible extensions. The predictions for the value of W mass, including radiative corrections, from the SM and from the minimal supersymmetric model are different with a small overlap. A measurement of the W mass up to 30 MeV can favour one of these models. Such a precision constraint also the Higgs boson mass to 30 percent. In this study an invariant mass measurement method is developed, using the decay products of the W boson pairs produced in the electron-positron collider, LEP. The two main channels, hadronic and semileptonic, are used. By analyzing different jet algorithms, the on from DURHAM was found to give the highest correct assignment efficiency. To improve the jet momentum reconstruction, a constrained fit method is applied, taking into account the ALEPH detector resolution. The momentum resolution is improved by a factor. Combining the results from the two channels, a statistical error of 44 MeV on the W mass is achieved, the corresponding systematic error is 20 MeV. (author).
19. Testing Suitability of Cell Cultures for SILAC-Experiments Using SWATH-Mass Spectrometry.
Science.gov (United States)
Reinders, Yvonne; Völler, Daniel; Bosserhoff, Anja-K; Oefner, Peter J; Reinders, Jörg
2016-01-01
Precise quantification is a major issue in contemporary proteomics. Both stable-isotope-labeling and label-free methods have been established for differential protein quantification and both approaches have different advantages and disadvantages. The present protocol uses the superior precision of label-free SWATH-mass spectrometry to test for suitability of cell lines for a SILAC-labeling approach as systematic regulations may be introduced upon incorporation of the "heavy" amino acids. The SILAC-labeled cell cultures can afterwards be used for further analyses where stable-isotope-labeling is mandatory or has substantial advantages over label-free approaches such as pulse-chase-experiments and differential protein interaction analyses based on co-immunoprecipitation. As SWATH-mass spectrometry avoids the missing-value-problem typically caused by undersampling in highly complex samples and shows superior precision for the quantification, it is better suited for the detection of systematic changes caused by the SILAC-labeling and thus, can serve as a useful tool to test cell lines for changes upon SILAC-labeling.
20. Mass transfer experiments for the heat load during in-vessel retention of core melt
Energy Technology Data Exchange (ETDEWEB)
Park, Hae Kyun; Chung, Bum Jin [Dept. of Nuclear Engineering, Kyung Hee University, Seoul (Korea, Republic of)
2016-08-15
We investigated the heat load imposed on the lower head of a reactor vessel by the natural convection of the oxide pool in a severe accident. Mass transfer experiments using a CuSO{sub 4}–H{sub 2}SO{sub 4} electroplating system were performed based on the analogy between heat and mass transfer. The Ra′{sub H} of 10{sup 14} order was achieved with a facility height of only 0.1 m. Three different volumetric heat sources were compared; two had identical configurations to those previously reported, and the other was designed by the authors. The measured Nu's of the lower head were about 30% lower than those previously reported. The measured angular heat flux ratios were similar to those reported in existing studies except for the peaks appearing near the top. The volumetric heat sources did not affect the Nu of the lower head but affected the Nu of the top plate by obstructing the rising flow from the bottom.
1. [The impact of experience in bearing child on the body mass index and obesity in women].
Science.gov (United States)
Lai, Jian-qiang; Yin, Shi-an
2009-02-01
To analyze the relations of body mass index(BMI)and obese prevalence in differently aged women and explore the effective strategy for preventing obesity among adult Chinese women. This study was based on the data from 2002 National Nutrition and Health Survey. The method of multi-steps cluster sampling was adopted. Total subjects including unmarried women (n = 2474), married women without the experience of childbearing (n = 10,816), and married and bearing-child women (n = 4103), were 17,393. In urban areas, the average body weights of unmarried, married and without childbearing experience, and the married with born-child were (53.7 +/- 9.0) kg, (57.6 +/- 9.4) kg and (54.5 +/- 8.5) kg respectively; the body weights of unmarried, married and without childbearing experience were significantly higher than that of the married with born-child women (t = 12.25, P unmarried, married without childbearing experience, and the married with born-child women were (21.1 +/- 3.3) kg/m(2), (22.8 +/- 3.4) kg/m(2) and (22.0 +/- 2.9) kg/m(2) respectively; the BMIs of married without childbearing experience and married with born-child women were significantly higher than that of unmarried women (t = 14.88, P unmarried, married without childbearing experience, and the married with born-child women were (52.3 +/- 7.8) kg, (55.3 +/- 8.6) kg and (52.8 +/- 8.1) kg respectively; the body weights of unmarried, the married with born-child women were significantly higher than that of married without childbearing experience (t = 11.67, P unmarried, married without childbearing experience, and the married with born-child women were (21.2 +/- 2.8) kg/m(2), (22.5 +/- 3.1) kg/m(2), and (21.8 +/- 3.0) kg/m(2) respectively; the BMIs of married and the married with born-child were significantly higher than that of unmarried women (t = 13.80, P unmarried women (18.1%) was higher than that of married without childbearing experience and married with born-child group (7.3% vs. 9.1%; comparing with married
2. Single Particle Laser Mass Spectrometry Applied to Differential Ice Nucleation Experiments at the AIDA Chamber
International Nuclear Information System (INIS)
Gallavardin, S. J.; Froyd, Karl D.; Lohmann, U.; Moehler, Ottmar; Murphy, Daniel M.; Cziczo, Dan
2008-01-01
Experiments conducted at the Aerosol Interactions and Dynamics in the Atmosphere (AIDA) chamber located in Karlsruhe, Germany permit investigation of particle properties that affect the nucleation of ice at temperature and water vapor conditions relevant to cloud microphysics and climate issues. Ice clouds were generated by heterogeneous nucleation of Arizona test dust (ATD), illite, and hematite and homogeneous nucleation of sulfuric acid. Ice crystals formed in the chamber were inertially separated from unactivated, or 'interstitial' aerosol particles with a pumped counterflow virtual impactor (PCVI), then evaporated. The ice residue (i.e., the aerosol which initiated ice nucleation plus any material which was scavenged from the gas- and/or particle-phase), was chemically characterized at the single particle level using a laser ionization mass spectrometer. In this manner the species that first nucleated ice could be identified out of a mixed aerosol population in the chamber. Bare mineral dust particles were more effective ice nuclei (IN) than similar particles with a coating. Metallic particles from contamination in the chamber initiated ice nucleation before other species but there were few enough that they did not compromise the experiments. Nitrate, sulfate, and organics were often detected on particles and ice residue, evidently from scavenging of trace gas-phase species in the chamber. Hematite was a more effective ice nucleus than illite. Ice residue was frequently larger than unactivated test aerosol due to the formation of aggregates due to scavenging, condensation of contaminant gases, and the predominance of larger aerosol in nucleation
3. In silico proteome analysis to facilitate proteomics experiments using mass spectrometry
Directory of Open Access Journals (Sweden)
Lindo Micheal
2003-08-01
Full Text Available Abstract Proteomics experiments typically involve protein or peptide separation steps coupled to the identification of many hundreds to thousands of peptides by mass spectrometry. Development of methodology and instrumentation in this field is proceeding rapidly, and effective software is needed to link the different stages of proteomic analysis. We have developed an application, proteogest, written in Perl that generates descriptive and statistical analyses of the biophysical properties of multiple (e.g. thousands protein sequences submitted by the user, for instance protein sequences inferred from the complete genome sequence of a model organism. The application also carries out in silico proteolytic digestion of the submitted proteomes, or subsets thereof, and the distribution of biophysical properties of the resulting peptides is presented. proteogest is customizable, the user being able to select many options, for instance the cleavage pattern of the digestion treatment or the presence of modifications to specific amino acid residues. We show how proteogest can be used to compare the proteomes and digested proteome products of model organisms, to examine the added complexity generated by modification of residues, and to facilitate the design of proteomics experiments for optimal representation of component proteins.
4. Experiments and Modelling Techniques for Heat and Mass Transfer in Light Water Reactors
International Nuclear Information System (INIS)
Ambrosini, W.; Bucci, M.; Forgione, N.; Manfredini, A.; Oriolo, F.
2009-01-01
The paper summarizes the lesson learned from theoretical and experimental activities performed at the University of Pisa, Pisa, Italy, in past decades in order to develop a general methodology of analysis of heat and mass transfer phenomena of interest for nuclear reactor applications. An overview of previously published results is proposed, highlighting the rationale at the basis of the performed work and its relevant conclusions. Experimental data from different sources provided information for model development and assessment. They include condensation experiments performed at SIET (Piacenza, Italy) on the PANTHERS prototypical PCCS module, falling film evaporation tests for simulating AP600-like outer shell spraying conditions, performed at the University of Pisa, experimental data concerning condensation on finned tubes, collected by CISE (Piacenza, Italy) in the frame of the INCON EU Project, and experimental tests performed in the CONAN experimental facility installed at the University of Pisa. The experience gained in these activities is critically reviewed and discussed to highlight the relevant obtained conclusions and the perspectives for future work
5. The interlaboratory experiment IDA-72 on mass spectrometric isotope dilution analysis. Vol. 1
International Nuclear Information System (INIS)
Beyrich, W.; Drosselmeyer, E.
1975-07-01
Within the framework of the Safeguards Project of the Federal Republic of Germany at the Nuclear Research Center Karlsruhe an analytical intercomparison program was carried out in cooperation with 22 laboratories of 13 countries or international organizations. The main objective was the acquisition of basic data on the errors involved in the mass spectrometric isotope dilution analysis if it is applied to the determination of uranium and plutonium in diluted active feed solutions of reprocessing plants in routine operation. The results were evaluated by statistical methods mainly in regard to the calculation of the estimates of the variances for the different error components contributing to the total error of this analytical technique. Furthermore, the performance of two new methods for sample conditioning suggested by the International Atomic Energy Agency, Vienna, and the European Institute for Transuranium Elements (EURATOM), Karlsruhe, was successfully tested. The results of some investigations on the stability of diluted high active feed solutions and on comparison analysis by X-ray fluorescence spectrometry are also included. Data on the analytical efforts (manhours) invested in this study are reported as well as general experiences made in the organization and performance of an experiment on such an extended international level. (orig.) [de
6. $W$ mass measurement and simulation of the transition radiation tracker at the ATLAS experiment
CERN Document Server
2008-01-01
At the time of writing, the final preparation toward LHC startup is ongoing. All the magnets of the machine have been installed and are currently being cooled. Most sub-detectors of the four experiments situated at the LHC ring are installed in their final positions and are being integrated into their respective data acquisition systems. This thesis concerns itself with the ATLAS experiment, focusing on a sub-detector called the Transition Radiation Tracker (TRT). Some attention is given to the hardware testing of the detector modules, but the main focus lies on the simulation of the detector and the comparison of the simulation with test-beam data, as well as with data collected during the commissioning phase using cosmic muons. There is little doubt that LHC will bring insight with respect to the understanding of the universe on the fundamental level. In particular, it is anticipated that light will be shed on the origin of mass which according to our current understanding proceeds via the Higgs mechanism. ...
7. Simulation Experiments with the Model of Information-Psychological Influences on Mass Consciousness
Directory of Open Access Journals (Sweden)
2017-06-01
which system- dynamic models are based are systems of differential equations. The authors of the article constructed an imitation model that makes it possible to identify the interrelationships between factors that effect the dynamics of information and psychological influences on mass consciousness, and its implementation has been carried out in the Anylogic software environment. This environment is one of the most promising environments for simulation. Temporal graphs reflecting the dynamics of the number of persons capable of adopting the idea of information and psychological influence, the number of persons who adopted the idea of information and psychological influences, and graphs reflecting the distribution in time of the number of persons who adopted the idea of information and psychological influences under certain initial conditions are given. Simulation experiments related to variations of the model parameters in which the dependence of increasing rate of person’s number who adopted the idea of information and psychological effects on the frequency of person’s contacts sharing these ideas is studied.
8. Primary care nurses' experiences of how the mass media influence frontline healthcare in the UK.
Science.gov (United States)
van Bekkum, Jennifer E; Hilton, Shona
2013-11-24
Mass media plays an important role in communicating about health research and services to patients, and in shaping public perceptions and decisions about health. Healthcare professionals also play an important role in providing patients with credible, evidence-based and up-to-date information on a wide range of health issues. This study aims to explore primary care nurses' experiences of how mass media influences frontline healthcare. In-depth telephone interviews were carried out with 18 primary care nurses (nine health visitors and nine practice nurses) working in the United Kingdom (UK). Interviews were recorded and transcribed. The data was analysed using thematic analysis, with a focus on constant comparative analysis. Three themes emerged from the data. First, participants reported that their patients were frequently influenced by controversial health stories reported in the media, which affected their perceptions of, and decisions about, care. This, in turn, impinged upon participants' workloads as they had to spend additional time discussing information and reassuring patients. Second, participants also recalled times in their own careers when media reports had contributed to a decline in their confidence in current healthcare practices and treatments. Third, the participants in this study suggested a real need for additional resources to support and expand their own media literacy skills, which could be shared with patients. In an ever expanding media landscape with greater reporting on health, nurses working in the primary care setting face increasing pressure to effectively manage media stories that dispute current health policies and practices. These primary care nurses were keen to expand their media literacy skills to develop critical autonomy in relation to all media, and to facilitate more meaningful conversations with their patients about their health concerns and choices.
9. Estimation of regional mass anomalies from Gravity Recovery and Climate Experiment (GRACE) over Himalayan region
Science.gov (United States)
Agrawal, R.; Singh, S. K.; Rajawat, A. S.; Ajai
2014-11-01
Time-variable gravity changes are caused by a combination of postglacial rebound, redistribution of water and snow/ice on land and as well as in the ocean. The Gravity Recovery and Climate Experiment (GRACE) satellite mission, launched in 2002, provides monthly average of the spherical harmonic co-efficient. These spherical harmonic co-efficient describe earth's gravity field with a resolution of few hundred kilometers. Time-variability of gravity field represents the change in mass over regional level with accuracies in cm in terms of Water Equivalent Height (WEH). The WEH reflects the changes in the integrated vertically store water including snow cover, surface water, ground water and soil moisture at regional scale. GRACE data are also sensitive towards interior strain variation, surface uplift and surface subsidence cover over a large area. GRACE data was extracted over the three major Indian River basins, Indus, Ganga and Brahmaputra, in the Himalayas which are perennial source of fresh water throughout the year in Northern Indian Plain. Time series analysis of the GRACE data was carried out from 2003-2012 over the study area. Trends and amplitudes of the regional mass anomalies in the region were estimated using level 3 GRACE data product with a spatial resolution at 10 by 10 grid provided by Center for Space Research (CSR), University of Texas at Austin. Indus basin has shown a subtle decreasing trend from 2003-2012 however it was observed to be statistically insignificant at 95 % confidence level. Ganga and Brahmaputra basins have shown a clear decreasing trend in WEH which was also observed to be statistically significant. The trend analysis over Ganga and Brahamputra basins have shown an average annual change of -1.28 cm and -1.06 cm in terms of WEH whereas Indus basin has shown a slight annual change of -0.07 cm. This analysis will be helpful to understand the loss of mass in terms of WEH over Indian Himalayas and will be crucial for hydrological and
10. Position sensitive detection coupled to high-resolution time-of-flight mass spectrometry: Imaging for molecular beam deflection experiments
International Nuclear Information System (INIS)
Abd El Rahim, M.; Antoine, R.; Arnaud, L.; Barbaire, M.; Broyer, M.; Clavier, Ch.; Compagnon, I.; Dugourd, Ph.; Maurelli, J.; Rayane, D.
2004-01-01
We have developed and tested a high-resolution time-of-flight mass spectrometer coupled to a position sensitive detector for molecular beam deflection experiments. The major achievement of this new spectrometer is to provide a three-dimensional imaging (X and Y positions and time-of-flight) of the ion packet on the detector, with a high acquisition rate and a high resolution on both the mass and the position. The calibration of the experimental setup and its application to molecular beam deflection experiments are discussed
11. A population balance approach considering heat and mass transfer-Experiments and CFD simulations
International Nuclear Information System (INIS)
Krepper, Eckhard; Beyer, Matthias; Lucas, Dirk; Schmidtke, Martin
2011-01-01
Highlights: → The MUSIG approach was extended by mass transfer between the size groups to describe condensation or re-evaporation. → Experiments on steam bubble condensation in vertical co-current steam/water flows have been carried out. The cross sectional gas fraction distribution, the bubble size distribution ad the gas velocity profiles were measured. → The following phenomena could be reproduced with good agreement to the experiments: (a) Dependence of the condensation rate on the initial bubble size distribution and (b) re-evaporation over the height in tests with low inlet temperature subcooling. - Abstract: Bubble condensation in sub-cooled water is a complex process, to which various phenomena contribute. Since the condensation rate depends on the interfacial area density, bubble size distribution changes caused by breakup and coalescence play a crucial role. Experiments on steam bubble condensation in vertical co-current steam/water flows have been carried out in an 8 m long vertical DN200 pipe. Steam is injected into the pipe and the development of the bubbly flow is measured at different distances to the injection using a pair of wire mesh sensors. By varying the steam nozzle diameter the initial bubble size can be influenced. Larger bubbles come along with a lower interfacial area density and therefore condensate slower. Steam pressures between 1 and 6.5 MPa and sub-cooling temperatures from 2 to 12 K were applied. Due to the pressure drop along the pipe, the saturation temperature falls towards the upper pipe end. This affects the sub-cooling temperature and can even cause re-evaporation in the upper part of the test section. The experimental configurations are simulated with the CFD code CFX using an extended MUSIG approach, which includes the bubble shrinking or growth due to condensation or re-evaporation. The development of the vapour phase along the pipe with respect to vapour void fractions and bubble sizes is qualitatively well reproduced
12. Measuring galaxy cluster masses with CMB lensing using a Maximum Likelihood estimator: statistical and systematic error budgets for future experiments
Energy Technology Data Exchange (ETDEWEB)
Raghunathan, Srinivasan; Patil, Sanjaykumar; Bianchini, Federico; Reichardt, Christian L. [School of Physics, University of Melbourne, 313 David Caro building, Swanston St and Tin Alley, Parkville VIC 3010 (Australia); Baxter, Eric J. [Department of Physics and Astronomy, University of Pennsylvania, 209 S. 33rd Street, Philadelphia, PA 19104 (United States); Bleem, Lindsey E. [Argonne National Laboratory, High-Energy Physics Division, 9700 S. Cass Avenue, Argonne, IL 60439 (United States); Crawford, Thomas M. [Kavli Institute for Cosmological Physics, University of Chicago, 5640 South Ellis Avenue, Chicago, IL 60637 (United States); Holder, Gilbert P. [Department of Astronomy and Department of Physics, University of Illinois, 1002 West Green St., Urbana, IL 61801 (United States); Manzotti, Alessandro, E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected], E-mail: [email protected] [Department of Astronomy and Astrophysics, University of Chicago, 5640 South Ellis Avenue, Chicago, IL 60637 (United States)
2017-08-01
We develop a Maximum Likelihood estimator (MLE) to measure the masses of galaxy clusters through the impact of gravitational lensing on the temperature and polarization anisotropies of the cosmic microwave background (CMB). We show that, at low noise levels in temperature, this optimal estimator outperforms the standard quadratic estimator by a factor of two. For polarization, we show that the Stokes Q/U maps can be used instead of the traditional E- and B-mode maps without losing information. We test and quantify the bias in the recovered lensing mass for a comprehensive list of potential systematic errors. Using realistic simulations, we examine the cluster mass uncertainties from CMB-cluster lensing as a function of an experiment's beam size and noise level. We predict the cluster mass uncertainties will be 3 - 6% for SPT-3G, AdvACT, and Simons Array experiments with 10,000 clusters and less than 1% for the CMB-S4 experiment with a sample containing 100,000 clusters. The mass constraints from CMB polarization are very sensitive to the experimental beam size and map noise level: for a factor of three reduction in either the beam size or noise level, the lensing signal-to-noise improves by roughly a factor of two.
13. Mass transfer in electromembrane extraction - The link between theory and experiments
DEFF Research Database (Denmark)
Huang, Chuixiu; Jensen, Henrik; Seip, Knut Fredrik
2016-01-01
typically been combined with chromatography, mass spectrometry, and electrophoresis for analyte separation and detection. At the moment, close to 125 research papers have been published with focus on electromembrane extraction. Electromembrane extraction is a hybrid technique between electrophoresis....... This review summarizes recent efforts to describe the fundamentals of mass transfer in electromembrane extraction, and aim to give an up-to-date understanding of the processes involved....... and liquid–liquid extraction, and the fundamental principles for mass transfer have only partly been investigated. Thus, although there is great interest in electromembrane extraction, the fundamental principle for mass transfer has to be described in more detail for the scientific acceptance of the concept...
14. Minimizing Back Exchange in the Hydrogen Exchange-Mass Spectrometry Experiment
Science.gov (United States)
Walters, Benjamin T.; Ricciuti, Alec; Mayne, Leland; Englander, S. Walter
2012-12-01
The addition of mass spectrometry (MS) analysis to the hydrogen exchange (HX) proteolytic fragmentation experiment extends powerful HX methodology to the study of large biologically important proteins. A persistent problem is the degradation of HX information due to back exchange of deuterium label during the fragmentation-separation process needed to prepare samples for MS measurement. This paper reports a systematic analysis of the factors that influence back exchange (solution pH, ionic strength, desolvation temperature, LC column interaction, flow rates, system volume). The many peptides exhibit a range of back exchange due to intrinsic amino acid HX rate differences. Accordingly, large back exchange leads to large variability in D-recovery from one residue to another as well as one peptide to another that cannot be corrected for by reference to any single peptide-level measurement. The usual effort to limit back exchange by limiting LC time provides little gain. Shortening the LC elution gradient by 3-fold only reduced back exchange by ~2 %, while sacrificing S/N and peptide count. An unexpected dependence of back exchange on ionic strength as well as pH suggests a strategy in which solution conditions are changed during sample preparation. Higher salt should be used in the first stage of sample preparation (proteolysis and trapping) and lower salt (<20 mM) and pH in the second stage before electrospray injection. Adjustment of these and other factors together with recent advances in peptide fragment detection yields hundreds of peptide fragments with D-label recovery of 90 % ± 5 %.
15. Minimizing back exchange in the hydrogen exchange-mass spectrometry experiment.
Science.gov (United States)
Walters, Benjamin T; Ricciuti, Alec; Mayne, Leland; Englander, S Walter
2012-12-01
The addition of mass spectrometry (MS) analysis to the hydrogen exchange (HX) proteolytic fragmentation experiment extends powerful HX methodology to the study of large biologically important proteins. A persistent problem is the degradation of HX information due to back exchange of deuterium label during the fragmentation-separation process needed to prepare samples for MS measurement. This paper reports a systematic analysis of the factors that influence back exchange (solution pH, ionic strength, desolvation temperature, LC column interaction, flow rates, system volume). The many peptides exhibit a range of back exchange due to intrinsic amino acid HX rate differences. Accordingly, large back exchange leads to large variability in D-recovery from one residue to another as well as one peptide to another that cannot be corrected for by reference to any single peptide-level measurement. The usual effort to limit back exchange by limiting LC time provides little gain. Shortening the LC elution gradient by 3-fold only reduced back exchange by ~2%, while sacrificing S/N and peptide count. An unexpected dependence of back exchange on ionic strength as well as pH suggests a strategy in which solution conditions are changed during sample preparation. Higher salt should be used in the first stage of sample preparation (proteolysis and trapping) and lower salt (<20 mM) and pH in the second stage before electrospray injection. Adjustment of these and other factors together with recent advances in peptide fragment detection yields hundreds of peptide fragments with D-label recovery of 90% ± 5%.
16. Search of low-mass WIMPs with a p -type point contact germanium detector in the CDEX-1 experiment
Science.gov (United States)
Zhao, W.; Yue, Q.; Kang, K. J.; Cheng, J. P.; Li, Y. J.; Wong, H. T.; Lin, S. T.; Chang, J. P.; Chen, J. H.; Chen, Q. H.; Chen, Y. H.; Deng, Z.; Du, Q.; Gong, H.; Hao, X. Q.; He, H. J.; He, Q. J.; Huang, H. X.; Huang, T. R.; Jiang, H.; Li, H. B.; Li, J.; Li, J.; Li, J. M.; Li, X.; Li, X. Y.; Li, Y. L.; Lin, F. K.; Liu, S. K.; Lü, L. C.; Ma, H.; Ma, J. L.; Mao, S. J.; Qin, J. Q.; Ren, J.; Ren, J.; Ruan, X. C.; Sharma, V.; Shen, M. B.; Singh, L.; Singh, M. K.; Soma, A. K.; Su, J.; Tang, C. J.; Wang, J. M.; Wang, L.; Wang, Q.; Wu, S. Y.; Wu, Y. C.; Xianyu, Z. Z.; Xiao, R. Q.; Xing, H. Y.; Xu, F. Z.; Xu, Y.; Xu, X. J.; Xue, T.; Yang, L. T.; Yang, S. W.; Yi, N.; Yu, C. X.; Yu, H.; Yu, X. Z.; Zeng, M.; Zeng, X. H.; Zeng, Z.; Zhang, L.; Zhang, Y. H.; Zhao, M. G.; Zhou, Z. Y.; Zhu, J. J.; Zhu, W. B.; Zhu, X. Z.; Zhu, Z. H.; CDEX Collaboration
2016-05-01
The CDEX-1 experiment conducted a search of low-mass (events is observed after the subtraction of the known background. Using 335.6 kg-days of data, exclusion constraints on the weakly interacting massive particle-nucleon spin-independent and spin-dependent couplings are derived.
17. Surface mass redistribution inversion from global GPS deformation and Gravity Recovery and Climate Experiment (GRACE) gravity data
NARCIS (Netherlands)
Kusche, J.; Schrama, E.J.O.
2005-01-01
Monitoring hydrological redistributions through their integrated gravitational effect is the primary aim of the Gravity Recovery and Climate Experiment (GRACE) mission. Time?variable gravity data from GRACE can be uniquely inverted to hydrology, since mass transfers located at or near the Earth's
18. An Advanced Analytical Chemistry Experiment Using Gas Chromatography-Mass Spectrometry, MATLAB, and Chemometrics to Predict Biodiesel Blend Percent Composition
Science.gov (United States)
Pierce, Karisa M.; Schale, Stephen P.; Le, Trang M.; Larson, Joel C.
2011-01-01
We present a laboratory experiment for an advanced analytical chemistry course where we first focus on the chemometric technique partial least-squares (PLS) analysis applied to one-dimensional (1D) total-ion-current gas chromatography-mass spectrometry (GC-TIC) separations of biodiesel blends. Then, we focus on n-way PLS (n-PLS) applied to…
19. Reinterpreting Higher Education Quality in Response to Policies of Mass Education: The Australian Experience
Science.gov (United States)
Pitman, Tim
2014-01-01
This article explores the relationship between mass education, higher education quality and policy development in Australia in the period 2008-2014, during which access to higher education was significantly increased. Over this time, which included a change of national government, the discursive relationship between mass higher education and…
20. Note: A versatile mass spectrometer chamber for molecular beam and temperature programmed desorption experiments
Energy Technology Data Exchange (ETDEWEB)
Tonks, James P., E-mail: [email protected] [Department of Mechanical Engineering Sciences, University of Surrey, Guildford, Surrey GU2 7XH (United Kingdom); AWE Plc, Aldermaston, Reading, Berkshire RG7 4PR (United Kingdom); Galloway, Ewan C., E-mail: [email protected]; King, Martin O. [AWE Plc, Aldermaston, Reading, Berkshire RG7 4PR (United Kingdom); Kerherve, Gwilherm [VACGEN Ltd, St. Leonards-On-Sea, East Sussex TN38 9NN (United Kingdom); Watts, John F. [Department of Mechanical Engineering Sciences, University of Surrey, Guildford, Surrey GU2 7XH (United Kingdom)
2016-08-15
A dual purpose mass spectrometer chamber capable of performing molecular beam scattering (MBS) and temperature programmed desorption (TPD) is detailed. Two simple features of this design allow it to perform these techniques. First, the diameter of entrance aperture to the mass spectrometer can be varied to maximize signal for TPD or to maximize angular resolution for MBS. Second, the mass spectrometer chamber can be radially translated so that it can be positioned close to the sample to maximize signal or far from the sample to maximize angular resolution. The performance of this system is described and compares well with systems designed for only one of these techniques.
1. Quantitative shear wave ultrasound elastography: initial experience in solid breast masses.
Science.gov (United States)
Evans, Andrew; Whelehan, Patsy; Thomson, Kim; McLean, Denis; Brauer, Katrin; Purdie, Colin; Jordan, Lee; Baker, Lee; Thompson, Alastair
2010-01-01
Shear wave elastography is a new method of obtaining quantitative tissue elasticity data during breast ultrasound examinations. The aims of this study were (1) to determine the reproducibility of shear wave elastography (2) to correlate the elasticity values of a series of solid breast masses with histological findings and (3) to compare shear wave elastography with greyscale ultrasound for benign/malignant classification. Using the Aixplorer® ultrasound system (SuperSonic Imagine, Aix en Provence, France), 53 solid breast lesions were identified in 52 consecutive patients. Two orthogonal elastography images were obtained of each lesion. Observers noted the mean elasticity values in regions of interest (ROI) placed over the stiffest areas on the two elastography images and a mean value was calculated for each lesion. A sub-set of 15 patients had two elastography images obtained by an additional operator. Reproducibility of observations was assessed between (1) two observers analysing the same pair of images and (2) findings from two pairs of images of the same lesion taken by two different operators. All lesions were subjected to percutaneous biopsy. Elastography measurements were correlated with histology results. After preliminary experience with 10 patients a mean elasticity cut off value of 50 kilopascals (kPa) was selected for benign/malignant differentiation. Greyscale images were classified according to the American College of Radiology (ACR) Breast Imaging Reporting and Data System (BI-RADS). BI-RADS categories 1-3 were taken as benign while BI-RADS categories 4 and 5 were classified as malignant. Twenty-three benign lesions and 30 cancers were diagnosed on histology. Measurement of mean elasticity yielded an intraclass correlation coefficient of 0.99 for two observers assessing the same pairs of elastography images. Analysis of images taken by two independent operators gave an intraclass correlation coefficient of 0.80. Shear wave elastography versus
2. Measurement of the W-boson mass at the ATLAS experiment
CERN Document Server
Kivernyk, Oleh; The ATLAS collaboration
2018-01-01
We present the results of $W$-boson mass measurements with the ATLAS detector at the LHC based on the 2011 data-set recorded at a centre-of-mass energy of $\\sqrt{s} =7$~\\TeV, and corresponding to 4.6~fb$^{-1}$ of integrated luminosity. The selected data sample consists of 7.8$\\times 10^6$ $W\\rightarrow \\mu\ 3. Precise measurement of the top-quark mass at the CMS experiment using the ideogram method International Nuclear Information System (INIS) Seidel, Markus 2015-08-01 The mass of the top quark is measured using a sample of t anti t candidate events with one electron or muon and at least four jets in the final state, collected by CMS in proton-proton collisions at √(s)=8 TeV at the LHC. The candidate events are selected from data corresponding to an integrated luminosity of 19.7 fb -1 . For each event the top-quark mass is reconstructed from a kinematic fit of the decay products to a t anti t hypothesis. In order to minimize the uncertainties from jet energy corrections, the top-quark mass is determined simultaneously with a jet energy scale factor (JSF), constrained by the known mass of the W boson decaying to quark-antiquark pairs. A joint likelihood fit taking into account multiple interpretations per event - the ideogram method - is used. From the simultaneous fit, a top-quark mass of 172.15±0.19(stat.+JSF)±0.61(syst.) GeV is obtained. Using an additional constraint from the determination of the jet energy scale in γ/Z+jet events yields m t =172.38±0.16(stat.+JSF)±0.49(syst.) GeV. The results are discussed in the context of different event generator implementations. Possible kinematic biases are studied by performing the measurement in different regions of the phase space. 4. Reconstructing neutrino properties from collider experiments in a Higgs triplet neutrino mass model International Nuclear Information System (INIS) Aristizabal Sierra, D.; Hirsch, M.; Valle, J. W. F.; Villanova del Moral, A. 2003-01-01 We extend the minimal supersymmetric standard model with bilinear R-parity violation to include a pair of Higgs triplet superfields. The neutral components of the Higgs triplets develop small vacuum expectation values (VEVs) quadratic in the bilinear R-parity breaking parameters. In this scheme the atmospheric neutrino mass scale arises from bilinear R-parity breaking while for reasonable values of parameters the solar neutrino mass scale is generated from the small Higgs triplet VEVs. We calculate neutrino masses and mixing angles in this model and show how the model can be tested at future colliders. The branching ratios of the doubly charged triplet decays are related to the solar neutrino angle via a simple formula 5. Beta Decay in the Field of an Electromagnetic Wave and Experiments on Measuring the Neutrino Mass International Nuclear Information System (INIS) Dorofeev, O.F.; Lobanov, A.E. 2005-01-01 Investigations of the effect of an electromagnetic wave field on the beta-decay process are used to analyze the tritium-decay experimental data on the neutrino mass. It is shown that the electromagnetic wave can distort the beta spectrum, shifting the end point to the higher energy region. This phenomenon is purely classical and it is associated with the electron acceleration in the radiation field. Since strong magnetic fields exist in setups for precise measurement of the neutrino mass, the indicated field can appear owing to the synchrotron radiation mechanism. The phenomenon under consideration can explain the experimentally observed anomalies in the spectrum of the decay electrons; in particular, the effect of the 'negative square of the neutrino mass' 6. Negative ion mass spectra and particulate formation in rf silane plasma deposition experiments International Nuclear Information System (INIS) Howling, A.A.; Dorier, J.L.; Hollenstein, C. 1992-09-01 Negative ions have been clearly identified in silane rf plasmas used for the deposition of amorphous silicon. Mass spectra were measured for monosilicon up to pentasilicon negative ion radical groups in power-modulated plasmas by means of a mass spectrometer mounted just outside the glow region. Negative ions were only observed over a limited range of power modulation frequency which corresponds to particle-free conditions. The importance of negative ions regarding particulate formation is demonstrated and commented upon. (author) 3 figs., 19 refs 7. Gas-phase fragmentation of peptides to increase the spatial resolution of the Hydrogen Exchange Mass Spectrometry experiment DEFF Research Database (Denmark) Jensen, Pernille Foged; Rand, Kasper Dyrberg 2016-01-01 are produced after precursor ion selection and thus do not add complexity to the LC-MS analysis. The key to obtaining optimal spatial resolution in a hydrogen exchange mass spectrometry (HX-MS) experiment is the fragmentation efficiency. This chapter discusses common fragmentation techniques like collision....../D scrambling, thus making them suitable for HX applications. By combining the classic bottom-up HX-MS workflow with gas-phase fragmentation by ETD, detailed information on protein HX can be obtained.... 8. Precise measurement of the top-quark mass at the CMS experiment using the ideogram method CERN Document Server Seidel, Markus; Stadie, Hartmut 2015-01-01 The mass of the top quark is measured using a sample of tt candidate events with oneelectron or muon and at least four jets in the final state, collected by CMS in proton√proton collisions at s = 8 TeV at the LHC. The candidate events are selected fromdata corresponding to an integrated luminosity of 19.7 fb−1 . For each event thetop-quark mass is reconstructed from a kinematic fit of the decay products to att hypothesis. In order to minimize the uncertainties from jet energy corrections,the top-quark mass is determined simultaneously with a jet energy scale factor(JSF), constrained by the known mass of the W boson decaying to quark-antiquarkpairs. A joint likelihood fit taking into account multiple interpretations per event– the ideogram method – is used. From the simultaneous fit, a top-quark massof 172.15 ± 0.19 (stat.+JSF) ± 0.61 (syst.) GeV is obtained. Using an additionalconstraint from the determination of the jet energy scale in γ/Z+jet events yieldsmt = 172.38 ± 0.16 (stat.+JSF) ± 0.... 9. qcML : an exchange format for quality control metrics from mass spectrometry experiments NARCIS (Netherlands) Walzer, Mathias; Pernas, Lucia Espona; Nasso, Sara; Bittremieux, Wout; Nahnsen, Sven; Kelchtermans, Pieter; Pichler, Peter; van den Toorn, Henk W P|info:eu-repo/dai/nl/31093205X; Staes, An; Vandenbussche, Jonathan; Mazanek, Michael; Taus, Thomas; Scheltema, Richard A; Kelstrup, Christian D; Gatto, Laurent; van Breukelen, Bas|info:eu-repo/dai/nl/244219087; Aiche, Stephan; Valkenborg, Dirk; Laukens, Kris; Lilley, Kathryn S; Olsen, Jesper V; Heck, Albert J R|info:eu-repo/dai/nl/105189332; Mechtler, Karl; Aebersold, Ruedi; Gevaert, Kris; Vizcaíno, Juan Antonio; Hermjakob, Henning; Kohlbacher, Oliver; Martens, Lennart Quality control is increasingly recognized as a crucial aspect of mass spectrometry based proteomics. Several recent papers discuss relevant parameters for quality control and present applications to extract these from the instrumental raw data. What has been missing, however, is a standard data 10. Vibration induced sliding: theory and experiment for a beam with a spring-loaded mass DEFF Research Database (Denmark) Miranda, Erik; Thomsen, Jon Juel 1998-01-01 The study sets up a simple model for predicting vibration induced sliding of mass, and provides quantitative experimental evidence for the validity of the model. The results lend confidence to recent theoretical developments on using vibration induced sliding for passive vibration damping, and co... 11. KATRIN :a New Beta-Spectroscopic Experiment to Determine the Neutrino Mass Czech Academy of Sciences Publication Activity Database Dragoun, Otokar 2007-01-01 Roč. 958, - (2007), s. 193-196. ISBN 978-0-7354-0472-4. ISSN N R&D Projects: GA MŠk LC07050 Institutional research plan: CEZ:AV0Z10480505 Keywords : neutrino mass * beta ray spectroscopy Subject RIV: BE - Theoretical Physics 12. Combined Measurement of the Higgs Boson Mass in pp Collisions at sqrt[s]=7 and 8 TeV with the ATLAS and CMS Experiments. Science.gov (United States) Aad, G; Abbott, B; Abdallah, J; Abdinov, O; Aben, R; Abolins, M; AbouZeid, O S; Abramowicz, H; Abreu, H; Abreu, R; Abulaiti, Y; Acharya, B S; Adamczyk, L; Adams, D L; Adelman, J; Adomeit, S; Adye, T; Affolder, A A; Agatonovic-Jovin, T; Aguilar-Saavedra, J A; Ahlen, S P; Ahmadov, F; Aielli, G; Akerstedt, H; Åkesson, T P A; Akimoto, G; Akimov, A V; Alberghi, G L; Albert, J; Albrand, S; Alconada Verzini, M J; Aleksa, M; Aleksandrov, I N; Alexa, C; Alexander, G; Alexopoulos, T; Alhroob, M; Alimonti, G; Alio, L; Alison, J; Alkire, S P; Allbrooke, B M M; Allport, P P; Aloisio, A; Alonso, A; Alonso, F; Alpigiani, C; Altheimer, A; Alvarez Gonzalez, B; Álvarez Piqueras, D; Alviggi, M G; Amadio, B T; Amako, K; Amaral Coutinho, Y; Amelung, C; Amidei, D; Amor Dos Santos, S P; Amorim, A; Amoroso, S; Amram, N; Amundsen, G; Anastopoulos, C; Ancu, L S; Andari, N; Andeen, T; Anders, C F; Anders, G; Anders, J K; Anderson, K J; Andreazza, A; Andrei, V; Angelidakis, S; Angelozzi, I; Anger, P; Angerami, A; Anghinolfi, F; Anisenkov, A V; Anjos, N; Annovi, A; Antonelli, M; Antonov, A; Antos, J; Anulli, F; Aoki, M; Aperio Bella, L; Arabidze, G; Arai, Y; Araque, J P; Arce, A T H; Arduh, F A; Arguin, J-F; Argyropoulos, S; Arik, M; Armbruster, A J; Arnaez, O; Arnal, V; Arnold, H; Arratia, M; Arslan, O; Artamonov, A; Artoni, G; Asai, S; Asbah, N; Ashkenazi, A; Åsman, B; Asquith, L; Assamagan, K; Astalos, R; Atkinson, M; Atlay, N B; Auerbach, B; Augsten, K; Aurousseau, M; Avolio, G; Axen, B; Ayoub, M K; Azuelos, G; Baak, M A; Baas, A E; Bacci, C; Bachacou, H; Bachas, K; Backes, M; Backhaus, M; Badescu, E; Bagiacchi, P; Bagnaia, P; Bai, Y; Bain, T; Baines, J T; Baker, O K; Balek, P; Balestri, T; Balli, F; Banas, E; Banerjee, Sw; Bannoura, A A E; Bansil, H S; Barak, L; Baranov, S P; Barberio, E L; Barberis, D; Barbero, M; Barillari, T; Barisonzi, M; Barklow, T; Barlow, N; Barnes, S L; Barnett, B M; Barnett, R M; Barnovska, Z; Baroncelli, A; Barone, G; Barr, A J; Barreiro, F; Barreiro Guimarães da Costa, J; Bartoldus, R; Barton, A E; Bartos, P; Bassalat, A; Basye, A; Bates, R L; Batista, S J; Batley, J R; Battaglia, M; Bauce, M; Bauer, F; Bawa, H S; Beacham, J B; Beattie, M D; Beau, T; Beauchemin, P H; Beccherle, R; Bechtle, P; Beck, H P; Becker, K; Becker, M; Becker, S; Beckingham, M; Becot, C; Beddall, A J; Beddall, A; Bednyakov, V A; Bee, C P; Beemster, L J; Beermann, T A; Begel, M; Behr, J K; Belanger-Champagne, C; Bell, W H; Bella, G; Bellagamba, L; Bellerive, A; Bellomo, M; Belotskiy, K; Beltramello, O; Benary, O; Benchekroun, D; Bender, M; Bendtz, K; Benekos, N; Benhammou, Y; Benhar Noccioli, E; Benitez Garcia, J A; Benjamin, D P; Bensinger, J R; Bentvelsen, S; Beresford, L; Beretta, M; Berge, D; Bergeaas Kuutmann, E; Berger, N; Berghaus, F; Beringer, J; Bernard, C; Bernard, N R; Bernius, C; Bernlochner, F U; Berry, T; Berta, P; Bertella, C; Bertoli, G; Bertolucci, F; Bertsche, C; Bertsche, D; Besana, M I; Besjes, G J; Bessidskaia Bylund, O; Bessner, M; Besson, N; Betancourt, C; Bethke, S; Bevan, A J; Bhimji, W; Bianchi, R M; Bianchini, L; Bianco, M; Biebel, O; Bieniek, S P; Biglietti, M; Bilbao De Mendizabal, J; Bilokon, H; Bindi, M; Binet, S; Bingul, A; Bini, C; Black, C W; Black, J E; Black, K M; Blackburn, D; Blair, R E; Blanchard, J-B; Blanco, J E; Blazek, T; Bloch, I; Blocker, C; Blum, W; Blumenschein, U; Bobbink, G J; Bobrovnikov, V S; Bocchetta, S S; Bocci, A; Bock, C; Boehler, M; Bogaerts, J A; Bogdanchikov, A G; Bohm, C; Boisvert, V; Bold, T; Boldea, V; Boldyrev, A S; Bomben, M; Bona, M; Boonekamp, M; Borisov, A; Borissov, G; Borroni, S; Bortfeldt, J; Bortolotto, V; Bos, K; Boscherini, D; Bosman, M; Boudreau, J; Bouffard, J; Bouhova-Thacker, E V; Boumediene, D; Bourdarios, C; Bousson, N; Boveia, A; Boyd, J; Boyko, I R; Bozic, I; Bracinik, J; Brandt, A; Brandt, G; Brandt, O; Bratzler, U; Brau, B; Brau, J E; Braun, H M; Brazzale, S F; Brendlinger, K; Brennan, A J; Brenner, L; Brenner, R; Bressler, S; Bristow, K; Bristow, T M; Britton, D; Britzger, D; Brochu, F M; Brock, I; Brock, R; Bronner, J; Brooijmans, G; Brooks, T; Brooks, W K; Brosamer, J; Brost, E; Brown, J; Bruckman de Renstrom, P A; Bruncko, D; Bruneliere, R; Bruni, A; Bruni, G; Bruschi, M; Bryngemark, L; Buanes, T; Buat, Q; Buchholz, P; Buckley, A G; Buda, S I; Budagov, I A; Buehrer, F; Bugge, L; Bugge, M K; Bulekov, O; Bullock, D; Burckhart, H; Burdin, S; Burghgrave, B; Burke, S; Burmeister, I; Busato, E; Büscher, D; Büscher, V; Bussey, P; Buszello, C P; Butler, J M; Butt, A I; Buttar, C M; Butterworth, J M; Butti, P; Buttinger, W; Buzatu, A; Buzykaev, R; Cabrera Urbán, S; Caforio, D; Cairo, V M; Cakir, O; Calafiura, P; Calandri, A; Calderini, G; Calfayan, P; Caloba, L P; Calvet, D; Calvet, S; Camacho Toro, R; Camarda, S; Camarri, P; Cameron, D; Caminada, L M; Caminal Armadans, R; Campana, S; Campanelli, M; Campoverde, A; Canale, V; Canepa, A; Cano Bret, M; Cantero, J; Cantrill, R; Cao, T; Capeans Garrido, M D M; Caprini, I; Caprini, M; Capua, M; Caputo, R; Cardarelli, R; Carli, T; Carlino, G; Carminati, L; Caron, S; Carquin, E; Carrillo-Montoya, G D; Carter, J R; Carvalho, J; Casadei, D; Casado, M P; Casolino, M; Castaneda-Miranda, E; Castelli, A; Castillo Gimenez, V; Castro, N F; Catastini, P; Catinaccio, A; Catmore, J R; Cattai, A; Caudron, J; Cavaliere, V; Cavalli, D; Cavalli-Sforza, M; Cavasinni, V; Ceradini, F; Cerio, B C; Cerny, K; Cerqueira, A S; Cerri, A; Cerrito, L; Cerutti, F; Cerv, M; Cervelli, A; Cetin, S A; Chafaq, A; Chakraborty, D; Chalupkova, I; Chang, P; Chapleau, B; Chapman, J D; Charlton, D G; Chau, C C; Chavez Barajas, C A; Cheatham, S; Chegwidden, A; Chekanov, S; Chekulaev, S V; Chelkov, G A; Chelstowska, M A; Chen, C; Chen, H; Chen, K; Chen, L; Chen, S; Chen, X; Chen, Y; Cheng, H C; Cheng, Y; Cheplakov, A; Cheremushkina, E; Cherkaoui El Moursli, R; Chernyatin, V; Cheu, E; Chevalier, L; Chiarella, V; Childers, J T; Chiodini, G; Chisholm, A S; Chislett, R T; Chitan, A; Chizhov, M V; Choi, K; Chouridou, S; Chow, B K B; Christodoulou, V; Chromek-Burckhart, D; Chu, M L; Chudoba, J; Chuinard, A J; Chwastowski, J J; Chytka, L; Ciapetti, G; Ciftci, A K; Cinca, D; Cindro, V; Cioara, I A; Ciocio, A; Citron, Z H; Ciubancan, M; Clark, A; Clark, B L; Clark, P J; Clarke, R N; Cleland, W; Clement, C; Coadou, Y; Cobal, M; Coccaro, A; Cochran, J; Coffey, L; Cogan, J G; Cole, B; Cole, S; Colijn, A P; Collot, J; Colombo, T; Compostella, G; Conde Muiño, P; Coniavitis, E; Connell, S H; Connelly, I A; Consonni, S M; Consorti, V; Constantinescu, S; Conta, C; Conti, G; Conventi, F; Cooke, M; Cooper, B D; Cooper-Sarkar, A M; Cornelissen, T; Corradi, M; Corriveau, F; Corso-Radu, A; Cortes-Gonzalez, A; Cortiana, G; Costa, G; Costa, M J; Costanzo, D; Côté, D; Cottin, G; Cowan, G; Cox, B E; Cranmer, K; Cree, G; Crépé-Renaudin, S; Crescioli, F; Cribbs, W A; Crispin Ortuzar, M; Cristinziani, M; Croft, V; Crosetti, G; Cuhadar Donszelmann, T; Cummings, J; Curatolo, M; Cuthbert, C; Czirr, H; Czodrowski, P; D'Auria, S; D'Onofrio, M; Da Cunha Sargedas De Sousa, M J; Da Via, C; Dabrowski, W; Dafinca, A; Dai, T; Dale, O; Dallaire, F; Dallapiccola, C; Dam, M; Dandoy, J R; Dang, N P; Daniells, A C; Danninger, M; Dano Hoffmann, M; Dao, V; Darbo, G; Darmora, S; Dassoulas, J; Dattagupta, A; Davey, W; David, C; Davidek, T; Davies, E; Davies, M; Davison, P; Davygora, Y; Dawe, E; Dawson, I; Daya-Ishmukhametova, R K; De, K; de Asmundis, R; De Castro, S; De Cecco, S; De Groot, N; de Jong, P; De la Torre, H; De Lorenzi, F; De Nooij, L; De Pedis, D; De Salvo, A; De Sanctis, U; De Santo, A; De Vivie De Regie, J B; Dearnaley, W J; Debbe, R; Debenedetti, C; Dedovich, D V; Deigaard, I; Del Peso, J; Del Prete, T; Delgove, D; Deliot, F; Delitzsch, C M; Deliyergiyev, M; Dell'Acqua, A; Dell'Asta, L; Dell'Orso, M; Della Pietra, M; Della Volpe, D; Delmastro, M; Delsart, P A; Deluca, C; DeMarco, D A; Demers, S; Demichev, M; Demilly, A; Denisov, S P; Derendarz, D; Derkaoui, J E; Derue, F; Dervan, P; Desch, K; Deterre, C; Deviveiros, P O; Dewhurst, A; Dhaliwal, S; Di Ciaccio, A; Di Ciaccio, L; Di Domenico, A; Di Donato, C; Di Girolamo, A; Di Girolamo, B; Di Mattia, A; Di Micco, B; Di Nardo, R; Di Simone, A; Di Sipio, R; Di Valentino, D; Diaconu, C; Diamond, M; Dias, F A; Diaz, M A; Diehl, E B; Dietrich, J; Diglio, S; Dimitrievska, A; Dingfelder, J; Dita, P; Dita, S; Dittus, F; Djama, F; Djobava, T; Djuvsland, J I; do Vale, M A B; Dobos, D; Dobre, M; Doglioni, C; Dohmae, T; Dolejsi, J; Dolezal, Z; Dolgoshein, B A; Donadelli, M; Donati, S; Dondero, P; Donini, J; Dopke, J; Doria, A; Dova, M T; Doyle, A T; Drechsler, E; Dris, M; Dubreuil, E; Duchovni, E; Duckeck, G; Ducu, O A; Duda, D; Dudarev, A; Duflot, L; Duguid, L; Dührssen, M; Dunford, M; Duran Yildiz, H; Düren, M; Durglishvili, A; Duschinger, D; Dyndal, M; Eckardt, C; Ecker, K M; Edgar, R C; Edson, W; Edwards, N C; Ehrenfeld, W; Eifert, T; Eigen, G; Einsweiler, K; Ekelof, T; El Kacimi, M; Ellert, M; Elles, S; Ellinghaus, F; Elliot, A A; Ellis, N; Elmsheuser, J; Elsing, M; Emeliyanov, D; Enari, Y; Endner, O C; Endo, M; Engelmann, R; Erdmann, J; Ereditato, A; Ernis, G; Ernst, J; Ernst, M; Errede, S; Ertel, E; Escalier, M; Esch, H; Escobar, C; Esposito, B; Etienvre, A I; Etzion, E; Evans, H; Ezhilov, A; Fabbri, L; Facini, G; Fakhrutdinov, R M; Falciano, S; Falla, R J; Faltova, J; Fang, Y; Fanti, M; Farbin, A; Farilla, A; Farooque, T; Farrell, S; Farrington, S M; Farthouat, P; Fassi, F; Fassnacht, P; Fassouliotis, D; Faucci Giannelli, M; Favareto, A; Fayard, L; Federic, P; Fedin, O L; Fedorko, W; Feigl, S; Feligioni, L; Feng, C; Feng, E J; Feng, H; Fenyuk, A B; Fernandez Martinez, P; Fernandez Perez, S; Ferrag, S; Ferrando, J; Ferrari, A; Ferrari, P; Ferrari, R; Ferreira de Lima, D E; Ferrer, A; Ferrere, D; Ferretti, C; Ferretto Parodi, A; Fiascaris, M; Fiedler, F; Filipčič, A; Filipuzzi, M; Filthaut, F; Fincke-Keeler, M; Finelli, K D; Fiolhais, M C N; Fiorini, L; Firan, A; Fischer, A; Fischer, C; Fischer, J; Fisher, W C; Fitzgerald, E A; Flechl, M; Fleck, I; Fleischmann, P; Fleischmann, S; Fletcher, G T; Fletcher, G; Flick, T; Floderus, A; Flores Castillo, L R; Flowerdew, M J; Formica, A; Forti, A; Fournier, D; Fox, H; Fracchia, S; Francavilla, P; Franchini, M; Francis, D; Franconi, L; Franklin, M; Fraternali, M; Freeborn, D; French, S T; Friedrich, F; Froidevaux, D; Frost, J A; Fukunaga, C; Fullana Torregrosa, E; Fulsom, B G; Fuster, J; Gabaldon, C; Gabizon, O; Gabrielli, A; Gabrielli, A; Gadatsch, S; Gadomski, S; Gagliardi, G; Gagnon, P; Galea, C; Galhardo, B; Gallas, E J; Gallop, B J; Gallus, P; Galster, G; Gan, K K; Gao, J; Gao, Y; Gao, Y S; Garay Walls, F M; Garberson, F; García, C; García Navarro, J E; Garcia-Sciveres, M; Gardner, R W; Garelli, N; Garonne, V; Gatti, C; Gaudiello, A; Gaudio, G; Gaur, B; Gauthier, L; Gauzzi, P; Gavrilenko, I L; Gay, C; Gaycken, G; Gazis, E N; Ge, P; Gecse, Z; Gee, C N P; Geerts, D A A; Geich-Gimbel, Ch; Geisler, M P; Gemme, C; Genest, M H; Gentile, S; George, M; George, S; Gerbaudo, D; Gershon, A; Ghazlane, H; Giacobbe, B; Giagu, S; Giangiobbe, V; Giannetti, P; Gibbard, B; Gibson, S M; Gilchriese, M; Gillam, T P S; Gillberg, D; Gilles, G; Gingrich, D M; Giokaris, N; Giordani, M P; Giorgi, F M; Giorgi, F M; Giraud, P F; Giromini, P; Giugni, D; Giuliani, C; Giulini, M; Gjelsten, B K; Gkaitatzis, S; Gkialas, I; Gkougkousis, E L; Gladilin, L K; Glasman, C; Glatzer, J; Glaysher, P C F; Glazov, A; Goblirsch-Kolb, M; Goddard, J R; Godlewski, J; Goldfarb, S; Golling, T; Golubkov, D; Gomes, A; Gonçalo, R; Goncalves Pinto Firmino Da Costa, J; Gonella, L; González de la Hoz, S; Gonzalez Parra, G; Gonzalez-Sevilla, S; Goossens, L; Gorbounov, P A; Gordon, H A; Gorelov, I; Gorini, B; Gorini, E; Gorišek, A; Gornicki, E; Goshaw, A T; Gössling, C; Gostkin, M I; Goujdami, D; Goussiou, A G; Govender, N; Grabas, H M X; Graber, L; Grabowska-Bold, I; Grafström, P; Grahn, K-J; Gramling, J; Gramstad, E; Grancagnolo, S; Grassi, V; Gratchev, V; Gray, H M; Graziani, E; Greenwood, Z D; Gregersen, K; Gregor, I M; Grenier, P; Griffiths, J; Grillo, A A; Grimm, K; Grinstein, S; Gris, Ph; Grivaz, J-F; Grohs, J P; Grohsjean, A; Gross, E; Grosse-Knetter, J; Grossi, G C; Grout, Z J; Guan, L; Guenther, J; Guescini, F; Guest, D; Gueta, O; Guido, E; Guillemin, T; Guindon, S; Gul, U; Gumpert, C; Guo, J; Gupta, S; Gutierrez, P; Gutierrez Ortiz, N G; Gutschow, C; Guyot, C; Gwenlan, C; Gwilliam, C B; Haas, A; Haber, C; Hadavand, H K; Haddad, N; Haefner, P; Hageböck, S; Hajduk, Z; Hakobyan, H; Haleem, M; Haley, J; Hall, D; Halladjian, G; Hallewell, G D; Hamacher, K; Hamal, P; Hamano, K; Hamer, M; Hamilton, A; Hamity, G N; Hamnett, P G; Han, L; Hanagaki, K; Hanawa, K; Hance, M; Hanke, P; Hanna, R; Hansen, J B; Hansen, J D; Hansen, M C; Hansen, P H; Hara, K; Hard, A S; Harenberg, T; Hariri, F; Harkusha, S; Harrington, R D; Harrison, P F; Hartjes, F; Hasegawa, M; Hasegawa, S; Hasegawa, Y; Hasib, A; Hassani, S; Haug, S; Hauser, R; Hauswald, L; Havranek, M; Hawkes, C M; Hawkings, R J; Hawkins, A D; Hayashi, T; Hayden, D; Hays, C P; Hays, J M; Hayward, H S; Haywood, S J; Head, S J; Heck, T; Hedberg, V; Heelan, L; Heim, S; Heim, T; Heinemann, B; Heinrich, L; Hejbal, J; Helary, L; Hellman, S; Hellmich, D; Helsens, C; Henderson, J; Henderson, R C W; Heng, Y; Hengler, C; Henrichs, A; Henriques Correia, A M; Henrot-Versille, S; Herbert, G H; Hernández Jiménez, Y; Herrberg-Schubert, R; Herten, G; Hertenberger, R; Hervas, L; Hesketh, G G; Hessey, N P; Hetherly, J W; Hickling, R; Higón-Rodriguez, E; Hill, E; Hill, J C; Hiller, K H; Hillier, S J; Hinchliffe, I; Hines, E; Hinman, R R; Hirose, M; Hirschbuehl, D; Hobbs, J; Hod, N; Hodgkinson, M C; Hodgson, P; Hoecker, A; Hoeferkamp, M R; Hoenig, F; Hohlfeld, M; Hohn, D; Holmes, T R; Hong, T M; Hooft van Huysduynen, L; Hopkins, W H; Horii, Y; Horton, A J; Hostachy, J-Y; Hou, S; Hoummada, A; Howard, J; Howarth, J; Hrabovsky, M; Hristova, I; Hrivnac, J; Hryn'ova, T; Hrynevich, A; Hsu, C; Hsu, P J; Hsu, S-C; Hu, D; Hu, Q; Hu, X; Huang, Y; Hubacek, Z; Hubaut, F; Huegging, F; Huffman, T B; Hughes, E W; Hughes, G; Huhtinen, M; Hülsing, T A; Huseynov, N; Huston, J; Huth, J; Iacobucci, G; Iakovidis, G; Ibragimov, I; Iconomidou-Fayard, L; Ideal, E; Idrissi, Z; Iengo, P; Igonkina, O; Iizawa, T; Ikegami, Y; Ikematsu, K; Ikeno, M; Ilchenko, Y; Iliadis, D; Ilic, N; Inamaru, Y; Ince, T; Ioannou, P; Iodice, M; Iordanidou, K; Ippolito, V; Irles Quiles, A; Isaksson, C; Ishino, M; Ishitsuka, M; Ishmukhametov, R; Issever, C; Istin, S; Iturbe Ponce, J M; Iuppa, R; Ivarsson, J; Iwanski, W; Iwasaki, H; Izen, J M; Izzo, V; Jabbar, S; Jackson, B; Jackson, M; Jackson, P; Jaekel, M R; Jain, V; Jakobs, K; Jakobsen, S; Jakoubek, T; Jakubek, J; Jamin, D O; Jana, D K; Jansen, E; Jansky, R W; Janssen, J; Janus, M; Jarlskog, G; Javadov, N; Javůrek, T; Jeanty, L; Jejelava, J; Jeng, G-Y; Jennens, D; Jenni, P; Jentzsch, J; Jeske, C; Jézéquel, S; Ji, H; Jia, J; Jiang, Y; Jiggins, S; Jimenez Pena, J; Jin, S; Jinaru, A; Jinnouchi, O; Joergensen, M D; Johansson, P; Johns, K A; Jon-And, K; Jones, G; Jones, R W L; Jones, T J; Jongmanns, J; Jorge, P M; Joshi, K D; Jovicevic, J; Ju, X; Jung, C A; Jussel, P; Juste Rozas, A; Kaci, M; Kaczmarska, A; Kado, M; Kagan, H; Kagan, M; Kahn, S J; Kajomovitz, E; Kalderon, C W; Kama, S; Kamenshchikov, A; Kanaya, N; Kaneda, M; Kaneti, S; Kantserov, V A; Kanzaki, J; Kaplan, B; Kapliy, A; Kar, D; Karakostas, K; Karamaoun, A; Karastathis, N; Kareem, M J; Karnevskiy, M; Karpov, S N; Karpova, Z M; Karthik, K; Kartvelishvili, V; Karyukhin, A N; Kashif, L; Kass, R D; Kastanas, A; Kataoka, Y; Katre, A; Katzy, J; Kawagoe, K; Kawamoto, T; Kawamura, G; Kazama, S; Kazanin, V F; Kazarinov, M Y; Keeler, R; Kehoe, R; Keller, J S; Kempster, J J; Keoshkerian, H; Kepka, O; Kerševan, B P; Kersten, S; Keyes, R A; Khalil-Zada, F; Khandanyan, H; Khanov, A; Kharlamov, A G; Khoo, T J; Khovanskiy, V; Khramov, E; Khubua, J; Kim, H Y; Kim, H; Kim, S H; Kim, Y; Kimura, N; Kind, O M; King, B T; King, M; King, R S B; King, S B; Kirk, J; Kiryunin, A E; Kishimoto, T; Kisielewska, D; Kiss, F; Kiuchi, K; Kivernyk, O; Kladiva, E; Klein, M H; Klein, M; Klein, U; Kleinknecht, K; Klimek, P; Klimentov, A; Klingenberg, R; Klinger, J A; Klioutchnikova, T; Kluge, E-E; Kluit, P; Kluth, S; Kneringer, E; Knoops, E B F G; Knue, A; Kobayashi, A; Kobayashi, D; Kobayashi, T; Kobel, M; Kocian, M; Kodys, P; Koffas, T; Koffeman, E; Kogan, L A; Kohlmann, S; Kohout, Z; Kohriki, T; Koi, T; Kolanoski, H; Koletsou, I; Komar, A A; Komori, Y; Kondo, T; Kondrashova, N; Köneke, K; König, A C; König, S; Kono, T; Konoplich, R; Konstantinidis, N; Kopeliansky, R; Koperny, S; Köpke, L; Kopp, A K; Korcyl, K; Kordas, K; Korn, A; Korol, A A; Korolkov, I; Korolkova, E V; Kortner, O; Kortner, S; Kosek, T; Kostyukhin, V V; Kotov, V M; Kotwal, A; Kourkoumeli-Charalampidi, A; Kourkoumelis, C; Kouskoura, V; Koutsman, A; Kowalewski, R; Kowalski, T Z; Kozanecki, W; Kozhin, A S; Kramarenko, V A; Kramberger, G; Krasnopevtsev, D; Krasznahorkay, A; Kraus, J K; Kravchenko, A; Kreiss, S; Kretz, M; Kretzschmar, J; Kreutzfeldt, K; Krieger, P; Krizka, K; Kroeninger, K; Kroha, H; Kroll, J; Kroseberg, J; Krstic, J; Kruchonak, U; Krüger, H; Krumnack, N; Krumshteyn, Z V; Kruse, A; Kruse, M C; Kruskal, M; Kubota, T; Kucuk, H; Kuday, S; Kuehn, S; Kugel, A; Kuger, F; Kuhl, A; Kuhl, T; Kukhtin, V; Kulchitsky, Y; Kuleshov, S; Kuna, M; Kunigo, T; Kupco, A; Kurashige, H; Kurochkin, Y A; Kurumida, R; Kus, V; Kuwertz, E S; Kuze, M; Kvita, J; Kwan, T; Kyriazopoulos, D; La Rosa, A; La Rosa Navarro, J L; La Rotonda, L; Lacasta, C; Lacava, F; Lacey, J; Lacker, H; Lacour, D; Lacuesta, V R; Ladygin, E; Lafaye, R; Laforge, B; Lagouri, T; Lai, S; Lambourne, L; Lammers, S; Lampen, C L; Lampl, W; Lançon, E; Landgraf, U; Landon, M P J; Lang, V S; Lange, J C; Lankford, A J; Lanni, F; Lantzsch, K; Laplace, S; Lapoire, C; Laporte, J F; Lari, T; Lasagni Manghi, F; Lassnig, M; Laurelli, P; Lavrijsen, W; Law, A T; Laycock, P; Le Dortz, O; Le Guirriec, E; Le Menedeu, E; LeBlanc, M; LeCompte, T; Ledroit-Guillon, F; Lee, C A; Lee, S C; Lee, L; Lefebvre, G; Lefebvre, M; Legger, F; Leggett, C; Lehan, A; Lehmann Miotto, G; Lei, X; Leight, W A; Leisos, A; Leister, A G; Leite, M A L; Leitner, R; Lellouch, D; Lemmer, B; Leney, K J C; Lenz, T; Lenzi, B; Leone, R; Leone, S; Leonidopoulos, C; Leontsinis, S; Leroy, C; Lester, C G; Levchenko, M; Levêque, J; Levin, D; Levinson, L J; Levy, M; Lewis, A; Leyko, A M; Leyton, M; Li, B; Li, H; Li, H L; Li, L; Li, L; Li, S; Li, Y; Liang, Z; Liao, H; Liberti, B; Liblong, A; Lichard, P; Lie, K; Liebal, J; Liebig, W; Limbach, C; Limosani, A; Lin, S C; Lin, T H; Linde, F; Lindquist, B E; Linnemann, J T; Lipeles, E; Lipniacka, A; Lisovyi, M; Liss, T M; Lissauer, D; Lister, A; Litke, A M; Liu, B; Liu, D; Liu, J; Liu, J B; Liu, K; Liu, L; Liu, M; Liu, M; Liu, Y; Livan, M; Lleres, A; Llorente Merino, J; Lloyd, S L; Lo Sterzo, F; Lobodzinska, E; Loch, P; Lockman, W S; Loebinger, F K; Loevschall-Jensen, A E; Loginov, A; Lohse, T; Lohwasser, K; Lokajicek, M; Long, B A; Long, J D; Long, R E; Looper, K A; Lopes, L; Lopez Mateos, D; Lopez Paredes, B; Lopez Paz, I; Lorenz, J; Lorenzo Martinez, N; Losada, M; Loscutoff, P; Lösel, P J; Lou, X; Lounis, A; Love, J; Love, P A; Lu, N; Lubatti, H J; Luci, C; Lucotte, A; Luehring, F; Lukas, W; Luminari, L; Lundberg, O; Lund-Jensen, B; Lynn, D; Lysak, R; Lytken, E; Ma, H; Ma, L L; Maccarrone, G; Macchiolo, A; Macdonald, C M; Machado Miguens, J; Macina, D; Madaffari, D; Madar, R; Maddocks, H J; Mader, W F; Madsen, A; Maeland, S; Maeno, T; Maevskiy, A; Magradze, E; Mahboubi, K; Mahlstedt, J; Maiani, C; Maidantchik, C; Maier, A A; Maier, T; Maio, A; Majewski, S; Makida, Y; Makovec, N; Malaescu, B; Malecki, Pa; Maleev, V P; Malek, F; Mallik, U; Malon, D; Malone, C; Maltezos, S; Malyshev, V M; Malyukov, S; Mamuzic, J; Mancini, G; Mandelli, B; Mandelli, L; Mandić, I; Mandrysch, R; Maneira, J; Manfredini, A; Manhaes de Andrade Filho, L; Manjarres Ramos, J; Mann, A; Manning, P M; Manousakis-Katsikakis, A; Mansoulie, B; Mantifel, R; Mantoani, M; Mapelli, L; March, L; Marchiori, G; Marcisovsky, M; Marino, C P; Marjanovic, M; Marroquim, F; Marsden, S P; Marshall, Z; Marti, L F; Marti-Garcia, S; Martin, B; Martin, T A; Martin, V J; Martin Dit Latour, B; Martinez, M; Martin-Haugh, S; Martoiu, V S; Martyniuk, A C; Marx, M; Marzano, F; Marzin, A; Masetti, L; Mashimo, T; Mashinistov, R; Masik, J; Maslennikov, A L; Massa, I; Massa, L; Massol, N; Mastrandrea, P; Mastroberardino, A; Masubuchi, T; Mättig, P; Mattmann, J; Maurer, J; Maxfield, S J; Maximov, D A; Mazini, R; Mazza, S M; Mazzaferro, L; Mc Goldrick, G; Mc Kee, S P; McCarn, A; McCarthy, R L; McCarthy, T G; McCubbin, N A; McFarlane, K W; Mcfayden, J A; Mchedlidze, G; McMahon, S J; McPherson, R A; Medinnis, M; Meehan, S; Mehlhase, S; Mehta, A; Meier, K; Meineck, C; Meirose, B; Mellado Garcia, B R; Meloni, F; Mengarelli, A; Menke, S; Meoni, E; Mercurio, K M; Mergelmeyer, S; Mermod, P; Merola, L; Meroni, C; Merritt, F S; Messina, A; Metcalfe, J; Mete, A S; Meyer, C; Meyer, C; Meyer, J-P; Meyer, J; Middleton, R P; Miglioranzi, S; Mijović, L; Mikenberg, G; Mikestikova, M; Mikuž, M; Milesi, M; Milic, A; Miller, D W; Mills, C; Milov, A; Milstead, D A; Minaenko, A A; Minami, Y; Minashvili, I A; Mincer, A I; Mindur, B; Mineev, M; Ming, Y; Mir, L M; Mitani, T; Mitrevski, J; Mitsou, V A; Miucci, A; Miyagawa, P S; Mjörnmark, J U; Moa, T; Mochizuki, K; Mohapatra, S; Mohr, W; Molander, S; Moles-Valls, R; Mönig, K; Monini, C; Monk, J; Monnier, E; Montejo Berlingen, J; Monticelli, F; Monzani, S; Moore, R W; Morange, N; Moreno, D; Moreno Llácer, M; Morettini, P; Morgenstern, M; Morii, M; Morinaga, M; Morisbak, V; Moritz, S; Morley, A K; Mornacchi, G; Morris, J D; Mortensen, S S; Morton, A; Morvaj, L; Mosidze, M; Moss, J; Motohashi, K; Mount, R; Mountricha, E; Mouraviev, S V; Moyse, E J W; Muanza, S; Mudd, R D; Mueller, F; Mueller, J; Mueller, K; Mueller, R S P; Mueller, T; Muenstermann, D; Mullen, P; Munwes, Y; Murillo Quijada, J A; Murray, W J; Musheghyan, H; Musto, E; Myagkov, A G; Myska, M; Nackenhorst, O; Nadal, J; Nagai, K; Nagai, R; Nagai, Y; Nagano, K; Nagarkar, A; Nagasaka, Y; Nagata, K; Nagel, M; Nagy, E; Nairz, A M; Nakahama, Y; Nakamura, K; Nakamura, T; Nakano, I; Namasivayam, H; Naranjo Garcia, R F; Narayan, R; Naumann, T; Navarro, G; Nayyar, R; Neal, H A; Nechaeva, P Yu; Neep, T J; Nef, P D; Negri, A; Negrini, M; Nektarijevic, S; Nellist, C; Nelson, A; Nemecek, S; Nemethy, P; Nepomuceno, A A; Nessi, M; Neubauer, M S; Neumann, M; Neves, R M; Nevski, P; Newman, P R; Nguyen, D H; Nickerson, R B; Nicolaidou, R; Nicquevert, B; Nielsen, J; Nikiforou, N; Nikiforov, A; Nikolaenko, V; Nikolic-Audit, I; Nikolopoulos, K; Nilsen, J K; Nilsson, P; Ninomiya, Y; Nisati, A; Nisius, R; Nobe, T; Nomachi, M; Nomidis, I; Nooney, T; Norberg, S; Nordberg, M; Novgorodova, O; Nowak, S; Nozaki, M; Nozka, L; Ntekas, K; Nunes Hanninger, G; Nunnemann, T; Nurse, E; Nuti, F; O'Brien, B J; O'grady, F; O'Neil, D C; O'Shea, V; Oakham, F G; Oberlack, H; Obermann, T; Ocariz, J; Ochi, A; Ochoa, I; Ochoa-Ricoux, J P; Oda, S; Odaka, S; Ogren, H; Oh, A; Oh, S H; Ohm, C C; Ohman, H; Oide, H; Okamura, W; Okawa, H; Okumura, Y; Okuyama, T; Olariu, A; Olivares Pino, S A; Oliveira Damazio, D; Oliver Garcia, E; Olszewski, A; Olszowska, J; Onofre, A; Onyisi, P U E; Oram, C J; Oreglia, M J; Oren, Y; Orestano, D; Orlando, N; Oropeza Barrera, C; Orr, R S; Osculati, B; Ospanov, R; Otero Y Garzon, G; Otono, H; Ouchrif, M; Ouellette, E A; Ould-Saada, F; Ouraou, A; Oussoren, K P; Ouyang, Q; Ovcharova, A; Owen, M; Owen, R E; Ozcan, V E; Ozturk, N; Pachal, K; Pacheco Pages, A; Padilla Aranda, C; Pagáčová, M; Pagan Griso, S; Paganis, E; Pahl, C; Paige, F; Pais, P; Pajchel, K; Palacino, G; Palestini, S; Palka, M; Pallin, D; Palma, A; Pan, Y B; Panagiotopoulou, E; Pandini, C E; Panduro Vazquez, J G; Pani, P; Panitkin, S; Pantea, D; Paolozzi, L; Papadopoulou, Th D; Papageorgiou, K; Paramonov, A; Paredes Hernandez, D; Parker, M A; Parker, K A; Parodi, F; Parsons, J A; Parzefall, U; Pasqualucci, E; Passaggio, S; Pastore, F; Pastore, Fr; Pásztor, G; Pataraia, S; Patel, N D; Pater, J R; Pauly, T; Pearce, J; Pearson, B; Pedersen, L E; Pedersen, M; Pedraza Lopez, S; Pedro, R; Peleganchuk, S V; Pelikan, D; Peng, H; Penning, B; Penwell, J; Perepelitsa, D V; Perez Codina, E; Pérez García-Estañ, M T; Perini, L; Pernegger, H; Perrella, S; Peschke, R; Peshekhonov, V D; Peters, K; Peters, R F Y; Petersen, B A; Petersen, T C; Petit, E; Petridis, A; Petridou, C; Petrolo, E; Petrucci, F; Pettersson, N E; Pezoa, R; Phillips, P W; Piacquadio, G; Pianori, E; Picazio, A; Piccaro, E; Piccinini, M; Pickering, M A; Piegaia, R; Pignotti, D T; Pilcher, J E; Pilkington, A D; Pina, J; Pinamonti, M; Pinfold, J L; Pingel, A; Pinto, B; Pires, S; Pitt, M; Pizio, C; Plazak, L; Pleier, M-A; Pleskot, V; Plotnikova, E; Plucinski, P; Pluth, D; Poettgen, R; Poggioli, L; Pohl, D; Polesello, G; Policicchio, A; Polifka, R; Polini, A; Pollard, C S; Polychronakos, V; Pommès, K; Pontecorvo, L; Pope, B G; Popeneciu, G A; Popovic, D S; Poppleton, A; Pospisil, S; Potamianos, K; Potrap, I N; Potter, C J; Potter, C T; Poulard, G; Poveda, J; Pozdnyakov, V; Pralavorio, P; Pranko, A; Prasad, S; Prell, S; Price, D; Price, L E; Primavera, M; Prince, S; Proissl, M; Prokofiev, K; Prokoshin, F; Protopapadaki, E; Protopopescu, S; Proudfoot, J; Przybycien, M; Ptacek, E; Puddu, D; Pueschel, E; Puldon, D; Purohit, M; Puzo, P; Qian, J; Qin, G; Qin, Y; Quadt, A; Quarrie, D R; Quayle, W B; Queitsch-Maitland, M; Quilty, D; Raddum, S; Radeka, V; Radescu, V; Radhakrishnan, S K; Radloff, P; Rados, P; Ragusa, F; Rahal, G; Rajagopalan, S; Rammensee, M; Rangel-Smith, C; Rauscher, F; Rave, S; Ravenscroft, T; Raymond, M; Read, A L; Readioff, N P; Rebuzzi, D M; Redelbach, A; Redlinger, G; Reece, R; Reeves, K; Rehnisch, L; Reisin, H; Relich, M; Rembser, C; Ren, H; Renaud, A; Rescigno, M; Resconi, S; Rezanova, O L; Reznicek, P; Rezvani, R; Richter, R; Richter, S; Richter-Was, E; Ricken, O; Ridel, M; Rieck, P; Riegel, C J; Rieger, J; Rijssenbeek, M; Rimoldi, A; Rinaldi, L; Ristić, B; Ritsch, E; Riu, I; Rizatdinova, F; Rizvi, E; Robertson, S H; Robichaud-Veronneau, A; Robinson, D; Robinson, J E M; Robson, A; Roda, C; Roe, S; Røhne, O; Rolli, S; Romaniouk, A; Romano, M; Romano Saez, S M; Romero Adam, E; Rompotis, N; Ronzani, M; Roos, L; Ros, E; Rosati, S; Rosbach, K; Rose, P; Rosendahl, P L; Rosenthal, O; Rossetti, V; Rossi, E; Rossi, L P; Rosten, R; Rotaru, M; Roth, I; Rothberg, J; Rousseau, D; Royon, C R; Rozanov, A; Rozen, Y; Ruan, X; Rubbo, F; Rubinskiy, I; Rud, V I; Rudolph, C; Rudolph, M S; Rühr, F; Ruiz-Martinez, A; Rurikova, Z; Rusakovich, N A; Ruschke, A; Russell, H L; Rutherfoord, J P; Ruthmann, N; Ryabov, Y F; Rybar, M; Rybkin, G; Ryder, N C; Saavedra, A F; Sabato, G; Sacerdoti, S; Saddique, A; Sadrozinski, H F-W; Sadykov, R; Safai Tehrani, F; Saimpert, M; Sakamoto, H; Sakurai, Y; Salamanna, G; Salamon, A; Saleem, M; Salek, D; Sales De Bruin, P H; Salihagic, D; Salnikov, A; Salt, J; Salvatore, D; Salvatore, F; Salvucci, A; Salzburger, A; Sampsonidis, D; Sanchez, A; Sánchez, J; Sanchez Martinez, V; Sandaker, H; Sandbach, R L; Sander, H G; Sanders, M P; Sandhoff, M; Sandoval, C; Sandstroem, R; Sankey, D P C; Sannino, M; Sansoni, A; Santoni, C; Santonico, R; Santos, H; Santoyo Castillo, I; Sapp, K; Sapronov, A; Saraiva, J G; Sarrazin, B; Sasaki, O; Sasaki, Y; Sato, K; Sauvage, G; Sauvan, E; Savage, G; Savard, P; Sawyer, C; Sawyer, L; Saxon, J; Sbarra, C; Sbrizzi, A; Scanlon, T; Scannicchio, D A; Scarcella, M; Scarfone, V; Schaarschmidt, J; Schacht, P; Schaefer, D; Schaefer, R; Schaeffer, J; Schaepe, S; Schaetzel, S; Schäfer, U; Schaffer, A C; Schaile, D; Schamberger, R D; Scharf, V; Schegelsky, V A; Scheirich, D; Schernau, M; Schiavi, C; Schillo, C; Schioppa, M; Schlenker, S; Schmidt, E; Schmieden, K; Schmitt, C; Schmitt, S; Schmitt, S; Schneider, B; Schnellbach, Y J; Schnoor, U; Schoeffel, L; Schoening, A; Schoenrock, B D; Schopf, E; Schorlemmer, A L S; Schott, M; Schouten, D; Schovancova, J; Schramm, S; Schreyer, M; Schroeder, C; Schuh, N; Schultens, M J; Schultz-Coulon, H-C; Schulz, H; Schumacher, M; Schumm, B A; Schune, Ph; Schwanenberger, C; Schwartzman, A; Schwarz, T A; Schwegler, Ph; Schwemling, Ph; Schwienhorst, R; Schwindling, J; Schwindt, T; Schwoerer, M; Sciacca, F G; Scifo, E; Sciolla, G; Scuri, F; Scutti, F; Searcy, J; Sedov, G; Sedykh, E; Seema, P; Seidel, S C; Seiden, A; Seifert, F; Seixas, J M; Sekhniaidze, G; Sekhon, K; Sekula, S J; Selbach, K E; Seliverstov, D M; Semprini-Cesari, N; Serfon, C; Serin, L; Serkin, L; Serre, T; Sessa, M; Seuster, R; Severini, H; Sfiligoj, T; Sforza, F; Sfyrla, A; Shabalina, E; Shamim, M; Shan, L Y; Shang, R; Shank, J T; Shapiro, M; Shatalov, P B; Shaw, K; Shaw, S M; Shcherbakova, A; Shehu, C Y; Sherwood, P; Shi, L; Shimizu, S; Shimmin, C O; Shimojima, M; Shiyakova, M; Shmeleva, A; Shoaleh Saadi, D; Shochet, M J; Shojaii, S; Shrestha, S; Shulga, E; Shupe, M A; Shushkevich, S; Sicho, P; Sidiropoulou, O; Sidorov, D; Sidoti, A; Siegert, F; Sijacki, Dj; Silva, J; Silver, Y; Silverstein, S B; Simak, V; Simard, O; Simic, Lj; Simion, S; Simioni, E; Simmons, B; Simon, D; Simoniello, R; Sinervo, P; Sinev, N B; Siragusa, G; Sisakyan, A N; Sivoklokov, S Yu; Sjölin, J; Sjursen, T B; Skinner, M B; Skottowe, H P; Skubic, P; Slater, M; Slavicek, T; Slawinska, M; Sliwa, K; Smakhtin, V; Smart, B H; Smestad, L; Smirnov, S Yu; Smirnov, Y; Smirnova, L N; Smirnova, O; Smith, M N K; Smith, R W; Smizanska, M; Smolek, K; Snesarev, A A; Snidero, G; Snyder, S; Sobie, R; Socher, F; Soffer, A; Soh, D A; Solans, C A; Solar, M; Solc, J; Soldatov, E Yu; Soldevila, U; Solodkov, A A; Soloshenko, A; Solovyanov, O V; Solovyev, V; Sommer, P; Song, H Y; Soni, N; Sood, A; Sopczak, A; Sopko, B; Sopko, V; Sorin, V; Sosa, D; Sosebee, M; Sotiropoulou, C L; Soualah, R; Soueid, P; Soukharev, A M; South, D; Sowden, B C; Spagnolo, S; Spalla, M; Spanò, F; Spearman, W R; Spettel, F; Spighi, R; Spigo, G; Spiller, L A; Spousta, M; Spreitzer, T; St Denis, R D; Staerz, S; Stahlman, J; Stamen, R; Stamm, S; Stanecka, E; Stanescu, C; Stanescu-Bellu, M; Stanitzki, M M; Stapnes, S; Starchenko, E A; Stark, J; Staroba, P; Starovoitov, P; Staszewski, R; Stavina, P; Steinberg, P; Stelzer, B; Stelzer, H J; Stelzer-Chilton, O; Stenzel, H; Stern, S; Stewart, G A; Stillings, J A; Stockton, M C; Stoebe, M; Stoicea, G; Stolte, P; Stonjek, S; Stradling, A R; Straessner, A; Stramaglia, M E; Strandberg, J; Strandberg, S; Strandlie, A; Strauss, E; Strauss, M; Strizenec, P; Ströhmer, R; Strom, D M; Stroynowski, R; Strubig, A; Stucci, S A; Stugu, B; Styles, N A; Su, D; Su, J; Subramaniam, R; Succurro, A; Sugaya, Y; Suhr, C; Suk, M; Sulin, V V; Sultansoy, S; Sumida, T; Sun, S; Sun, X; Sundermann, J E; Suruliz, K; Susinno, G; Sutton, M R; Suzuki, S; Suzuki, Y; Svatos, M; Swedish, S; Swiatlowski, M; Sykora, I; Sykora, T; Ta, D; Taccini, C; Tackmann, K; Taenzer, J; Taffard, A; Tafirout, R; Taiblum, N; Takai, H; Takashima, R; Takeda, H; Takeshita, T; Takubo, Y; Talby, M; Talyshev, A A; Tam, J Y C; Tan, K G; Tanaka, J; Tanaka, R; Tanaka, S; Tannenwald, B B; Tannoury, N; Tapprogge, S; Tarem, S; Tarrade, F; Tartarelli, G F; Tas, P; Tasevsky, M; Tashiro, T; Tassi, E; Tavares Delgado, A; Tayalati, Y; Taylor, F E; Taylor, G N; Taylor, W; Teischinger, F A; Teixeira Dias Castanheira, M; Teixeira-Dias, P; Temming, K K; Ten Kate, H; Teng, P K; Teoh, J J; Tepel, F; Terada, S; Terashi, K; Terron, J; Terzo, S; Testa, M; Teuscher, R J; Therhaag, J; Theveneaux-Pelzer, T; Thomas, J P; Thomas-Wilsker, J; Thompson, E N; Thompson, P D; Thompson, R J; Thompson, A S; Thomsen, L A; Thomson, E; Thomson, M; Thun, R P; Tibbetts, M J; Ticse Torres, R E; Tikhomirov, V O; Tikhonov, Yu A; Timoshenko, S; Tiouchichine, E; Tipton, P; Tisserant, S; Todorov, T; Todorova-Nova, S; Tojo, J; Tokár, S; Tokushuku, K; Tollefson, K; Tolley, E; Tomlinson, L; Tomoto, M; Tompkins, L; Toms, K; Torrence, E; Torres, H; Torró Pastor, E; Toth, J; Touchard, F; Tovey, D R; Trefzger, T; Tremblet, L; Tricoli, A; Trigger, I M; Trincaz-Duvoid, S; Tripiana, M F; Trischuk, W; Trocmé, B; Troncon, C; Trottier-McDonald, M; Trovatelli, M; True, P; Truong, L; Trzebinski, M; Trzupek, A; Tsarouchas, C; Tseng, J C-L; Tsiareshka, P V; Tsionou, D; Tsipolitis, G; Tsirintanis, N; Tsiskaridze, S; Tsiskaridze, V; Tskhadadze, E G; Tsukerman, I I; Tsulaia, V; Tsuno, S; Tsybychev, D; Tudorache, A; Tudorache, V; Tuna, A N; Tupputi, S A; Turchikhin, S; Turecek, D; Turra, R; Turvey, A J; Tuts, P M; Tykhonov, A; Tylmad, M; Tyndel, M; Ueda, I; Ueno, R; Ughetto, M; Ugland, M; Uhlenbrock, M; Ukegawa, F; Unal, G; Undrus, A; Unel, G; Ungaro, F C; Unno, Y; Unverdorben, C; Urban, J; Urquijo, P; Urrejola, P; Usai, G; Usanova, A; Vacavant, L; Vacek, V; Vachon, B; Valderanis, C; Valencic, N; Valentinetti, S; Valero, A; Valery, L; Valkar, S; Valladolid Gallego, E; Vallecorsa, S; Valls Ferrer, J A; Van Den Wollenberg, W; Van Der Deijl, P C; van der Geer, R; van der Graaf, H; Van Der Leeuw, R; van Eldik, N; van Gemmeren, P; Van Nieuwkoop, J; van Vulpen, I; van Woerden, M C; Vanadia, M; Vandelli, W; Vanguri, R; Vaniachine, A; Vannucci, F; Vardanyan, G; Vari, R; Varnes, E W; Varol, T; Varouchas, D; Vartapetian, A; Varvell, K E; Vazeille, F; Vazquez Schroeder, T; Veatch, J; Veloce, L M; Veloso, F; Velz, T; Veneziano, S; Ventura, A; Ventura, D; Venturi, M; Venturi, N; Venturini, A; Vercesi, V; Verducci, M; Verkerke, W; Vermeulen, J C; Vest, A; Vetterli, M C; Viazlo, O; Vichou, I; Vickey, T; Vickey Boeriu, O E; Viehhauser, G H A; Viel, S; Vigne, R; Villa, M; Villaplana Perez, M; Vilucchi, E; Vincter, M G; Vinogradov, V B; Vivarelli, I; Vives Vaque, F; Vlachos, S; Vladoiu, D; Vlasak, M; Vogel, M; Vokac, P; Volpi, G; Volpi, M; von der Schmitt, H; von Radziewski, H; von Toerne, E; Vorobel, V; Vorobev, K; Vos, M; Voss, R; Vossebeld, J H; Vranjes, N; Vranjes Milosavljevic, M; Vrba, V; Vreeswijk, M; Vuillermet, R; Vukotic, I; Vykydal, Z; Wagner, P; Wagner, W; Wahlberg, H; Wahrmund, S; Wakabayashi, J; Walder, J; Walker, R; Walkowiak, W; Wang, C; Wang, F; Wang, H; Wang, H; Wang, J; Wang, J; Wang, K; Wang, R; Wang, S M; Wang, T; Wang, X; Wanotayaroj, C; Warburton, A; Ward, C P; Wardrope, D R; Warsinsky, M; Washbrook, A; Wasicki, C; Watkins, P M; Watson, A T; Watson, I J; Watson, M F; Watts, G; Watts, S; Waugh, B M; Webb, S; Weber, M S; Weber, S W; Webster, J S; Weidberg, A R; Weinert, B; Weingarten, J; Weiser, C; Weits, H; Wells, P S; Wenaus, T; Wengler, T; Wenig, S; Wermes, N; Werner, M; Werner, P; Wessels, M; Wetter, J; Whalen, K; Wharton, A M; White, A; White, M J; White, R; White, S; Whiteson, D; Wickens, F J; Wiedenmann, W; Wielers, M; Wienemann, P; Wiglesworth, C; Wiik-Fuchs, L A M; Wildauer, A; Wilkens, H G; Williams, H H; Williams, S; Willis, C; Willocq, S; Wilson, A; Wilson, J A; Wingerter-Seez, I; Winklmeier, F; Winter, B T; Wittgen, M; Wittkowski, J; Wollstadt, S J; Wolter, M W; Wolters, H; Wosiek, B K; Wotschack, J; Woudstra, M J; Wozniak, K W; Wu, M; Wu, M; Wu, S L; Wu, X; Wu, Y; Wyatt, T R; Wynne, B M; Xella, S; Xu, D; Xu, L; Yabsley, B; Yacoob, S; Yakabe, R; Yamada, M; Yamaguchi, Y; Yamamoto, A; Yamamoto, S; Yamanaka, T; Yamauchi, K; Yamazaki, Y; Yan, Z; Yang, H; Yang, H; Yang, Y; Yao, L; Yao, W-M; Yasu, Y; Yatsenko, E; Yau Wong, K H; Ye, J; Ye, S; Yeletskikh, I; Yen, A L; Yildirim, E; Yorita, K; Yoshida, R; Yoshihara, K; Young, C; Young, C J S; Youssef, S; Yu, D R; Yu, J; Yu, J M; Yu, J; Yuan, L; Yurkewicz, A; Yusuff, I; Zabinski, B; Zaidan, R; Zaitsev, A M; Zalieckas, J; Zaman, A; Zambito, S; Zanello, L; Zanzi, D; Zeitnitz, C; Zeman, M; Zemla, A; Zengel, K; Zenin, O; Ženiš, T; Zerwas, D; Zhang, D; Zhang, F; Zhang, J; Zhang, L; Zhang, R; Zhang, X; Zhang, Z; Zhao, X; Zhao, Y; Zhao, Z; Zhemchugov, A; Zhong, J; Zhou, B; Zhou, C; Zhou, L; Zhou, L; Zhou, N; Zhu, C G; Zhu, H; Zhu, J; Zhu, Y; Zhuang, X; Zhukov, K; Zibell, A; Zieminska, D; Zimine, N I; Zimmermann, C; Zimmermann, S; Zinonos, Z; Zinser, M; Ziolkowski, M; Živković, L; Zobernig, G; Zoccoli, A; Zur Nedden, M; Zurzolo, G; Zwalinski, L; Khachatryan, V; Sirunyan, A M; Tumasyan, A; Adam, W; Asilar, E; Bergauer, T; Brandstetter, J; Brondolin, E; Dragicevic, M; Erö, J; Flechl, M; Friedl, M; Frühwirth, R; Ghete, V M; Hartl, C; Hörmann, N; Hrubec, J; Jeitler, M; Knünz, V; König, A; Krammer, M; Krätschmer, I; Liko, D; Matsushita, T; Mikulec, I; Rabady, D; Rahbaran, B; Rohringer, H; Schieck, J; Schöfbeck, R; Strauss, J; Treberer-Treberspurg, W; Waltenberger, W; Wulz, C-E; Mossolov, V; Shumeiko, N; Suarez Gonzalez, J; Alderweireldt, S; Cornelis, T; De Wolf, E A; Janssen, X; Knutsson, A; Lauwers, J; Luyckx, S; Ochesanu, S; Rougny, R; Van De Klundert, M; Van Haevermaet, H; Van Mechelen, P; Van Remortel, N; Van Spilbeeck, A; Abu Zeid, S; Blekman, F; D'Hondt, J; Daci, N; De Bruyn, I; Deroover, K; Heracleous, N; Keaveney, J; Lowette, S; Moreels, L; Olbrechts, A; Python, Q; Strom, D; Tavernier, S; Van Doninck, W; Van Mulders, P; Van Onsem, G P; Van Parijs, I; Barria, P; Caillol, C; Clerbaux, B; De Lentdecker, G; Delannoy, H; Dobur, D; Fasanella, G; Favart, L; Gay, A P R; Grebenyuk, A; Lenzi, T; Léonard, A; Maerschalk, T; Mohammadi, A; Perniè, L; Randle-Conde, A; Reis, T; Seva, T; Thomas, L; Vander Velde, C; Vanlaer, P; Wang, J; Yonamine, R; Zenoni, F; Zhang, F; Beernaert, K; Benucci, L; Cimmino, A; Crucy, S; Fagot, A; Garcia, G; Gul, M; Mccartin, J; Ocampo Rios, A A; Poyraz, D; Ryckbosch, D; Salva Diblen, S; Sigamani, M; Strobbe, N; Tytgat, M; Van Driessche, W; Yazgan, E; Zaganidis, N; Basegmez, S; Beluffi, C; Bondu, O; Bruno, G; Castello, R; Caudron, A; Ceard, L; Da Silveira, G G; Delaere, C; Favart, D; Forthomme, L; Giammanco, A; Hollar, J; Jafari, A; Jez, P; Komm, M; Lemaitre, V; Mertens, A; Nuttens, C; Perrini, L; Pin, A; Piotrzkowski, K; Popov, A; Quertenmont, L; Selvaggi, M; Vidal Marono, M; Beliy, N; Caebergs, T; Hammad, G H; Aldá Júnior, W L; Alves, G A; Brito, L; Correa Martins Junior, M; Dos Reis Martins, T; Hensel, C; Mora Herrera, C; Moraes, A; Pol, M E; Rebello Teles, P; Belchior Batista Das Chagas, E; Carvalho, W; Chinellato, J; Custódio, A; Da Costa, E M; De Jesus Damiao, D; De Oliveira Martins, C; Fonseca De Souza, S; Huertas Guativa, L M; Malbouisson, H; Matos Figueiredo, D; Mundim, L; Nogima, H; Prado Da Silva, W L; Santoro, A; Sznajder, A; Tonelli Manganote, E J; Vilela Pereira, A; Ahuja, S; Bernardes, C A; De Souza Santos, A; Dogra, S; Tomei, T R Fernandez Perez; Gregores, E M; Mercadante, P G; Moon, C S; Novaes, S F; Padula, Sandra S; Romero Abad, D; Ruiz Vargas, J C; Aleksandrov, A; Genchev, V; Hadjiiska, R; Iaydjiev, P; Marinov, A; Piperov, S; Rodozov, M; Stoykova, S; Sultanov, G; Vutova, M; Dimitrov, A; Glushkov, I; Litov, L; Pavlov, B; Petkov, P; Ahmad, M; Bian, J G; Chen, G M; Chen, H S; Chen, M; Cheng, T; Du, R; Jiang, C H; Plestina, R; Romeo, F; Shaheen, S M; Tao, J; Wang, C; Wang, Z; Zhang, H; Asawatangtrakuldee, C; Ban, Y; Chen, G; Li, Q; Liu, S; Mao, Y; Qian, S J; Wang, D; Wang, M; Wang, Q; Xu, Z; Yang, D; Zhang, Z; Zou, W; Avila, C; Cabrera, A; Chaparro Sierra, L F; Florez, C; Gomez, J P; Gomez Moreno, B; Sanabria, J C; Godinovic, N; Lelas, D; Polic, D; Puljak, I; Antunovic, Z; Kovac, M; Brigljevic, V; Kadija, K; Luetic, J; Sudic, L; Attikis, A; Mavromanolakis, G; Mousa, J; Nicolaou, C; Ptochos, F; Razis, P A; Rykaczewski, H; Bodlak, M; Finger, M; Finger, M; Ali, A; Aly, R; Aly, S; Assran, Y; Ellithi Kamel, A; Lotfy, A; Mahmoud, M A; Masod, R; Radi, A; Calpas, B; Kadastik, M; Murumaa, M; Raidal, M; Tiko, A; Veelken, C; Eerola, P; Pekkanen, J; Voutilainen, M; Härkönen, J; Karimäki, V; Kinnunen, R; Lampén, T; Lassila-Perini, K; Lehti, S; Lindén, T; Luukka, P; Mäenpää, T; Peltola, T; Tuominen, E; Tuominiemi, J; Tuovinen, E; Wendland, L; Talvitie, J; Tuuva, T; Besancon, M; Couderc, F; Dejardin, M; Denegri, D; Fabbro, B; Faure, J L; Favaro, C; Ferri, F; Ganjour, S; Givernaud, A; Gras, P; Hamel de Monchenault, G; Jarry, P; Locci, E; Machet, M; Malcles, J; Rander, J; Rosowsky, A; Titov, M; Zghiche, A; Baffioni, S; Beaudette, F; Busson, P; Cadamuro, L; Chapon, E; Charlot, C; Dahms, T; Davignon, O; Filipovic, N; Florent, A; Granier de Cassagnac, R; Lisniak, S; Mastrolorenzo, L; Miné, P; Naranjo, I N; Nguyen, M; Ochando, C; Ortona, G; Paganini, P; Regnard, S; Salerno, R; Sauvan, J B; Sirois, Y; Strebler, T; Yilmaz, Y; Zabi, A; Agram, J-L; Andrea, J; Aubin, A; Bloch, D; Brom, J-M; Buttignol, M; Chabert, E C; Chanon, N; Collard, C; Conte, E; Fontaine, J-C; Gelé, D; Goerlach, U; Goetzmann, C; Le Bihan, A-C; Merlin, J A; Skovpen, K; Van Hove, P; Gadrat, S; Beauceron, S; Bernet, C; Boudoul, G; Bouvier, E; Brochet, S; Carrillo Montoya, C A; Chasserat, J; Chierici, R; Contardo, D; Courbon, B; Depasse, P; El Mamouni, H; Fan, J; Fay, J; Gascon, S; Gouzevitch, M; Ille, B; Laktineh, I B; Lethuillier, M; Mirabito, L; Pequegnot, A L; Perries, S; Ruiz Alvarez, J D; Sabes, D; Sgandurra, L; Sordini, V; Vander Donckt, M; Verdier, P; Viret, S; Xiao, H; Tsamalaidze, Z; Autermann, C; Beranek, S; Bontenackels, M; Edelhoff, M; Feld, L; Heister, A; Kiesel, M K; Klein, K; Lipinski, M; Ostapchuk, A; Preuten, M; Raupach, F; Sammet, J; Schael, S; Schulte, J F; Verlage, T; Weber, H; Wittmer, B; Zhukov, V; Ata, M; Brodski, M; Dietz-Laursonn, E; Duchardt, D; Endres, M; Erdmann, M; Erdweg, S; Esch, T; Fischer, R; Güth, A; Hebbeker, T; Heidemann, C; Hoepfner, K; Klingebiel, D; Knutzen, S; Kreuzer, P; Merschmeyer, M; Meyer, A; Millet, P; Olschewski, M; Padeken, K; Papacz, P; Pook, T; Radziej, M; Reithler, H; Rieger, M; Scheuch, F; Sonnenschein, L; Teyssier, D; Thüer, S; Cherepanov, V; Erdogan, Y; Flügge, G; Geenen, H; Geisler, M; Haj Ahmad, W; Hoehle, F; Kargoll, B; Kress, T; Kuessel, Y; Künsken, A; Lingemann, J; Nehrkorn, A; Nowack, A; Nugent, I M; Pistone, C; Pooth, O; Stahl, A; Aldaya Martin, M; Asin, I; Bartosik, N; Behnke, O; Behrens, U; Bell, A J; Borras, K; Burgmeier, A; Cakir, A; Calligaris, L; Campbell, A; Choudhury, S; Costanza, F; Diez Pardos, C; Dolinska, G; Dooling, S; Dorland, T; Eckerlin, G; Eckstein, D; Eichhorn, T; Flucke, G; Gallo, E; Garay Garcia, J; Geiser, A; Gizhko, A; Gunnellini, P; Hauk, J; Hempel, M; Jung, H; Kalogeropoulos, A; Karacheban, O; Kasemann, M; Katsas, P; Kieseler, J; Kleinwort, C; Korol, I; Lange, W; Leonard, J; Lipka, K; Lobanov, A; Lohmann, W; Mankel, R; Marfin, I; Melzer-Pellmann, I-A; Meyer, A B; Mittag, G; Mnich, J; Mussgiller, A; Naumann-Emme, S; Nayak, A; Ntomari, E; Perrey, H; Pitzl, D; Placakyte, R; Raspereza, A; Ribeiro Cipriano, P M; Roland, B; Sahin, M Ö; Salfeld-Nebgen, J; Saxena, P; Schoerner-Sadenius, T; Schröder, M; Seitz, C; Spannagel, S; Trippkewitz, K D; Wissing, C; Blobel, V; Centis Vignali, M; Draeger, A R; Erfle, J; Garutti, E; Goebel, K; Gonzalez, D; Görner, M; Haller, J; Hoffmann, M; Höing, R S; Junkes, A; Klanner, R; Kogler, R; Lapsien, T; Lenz, T; Marchesini, I; Marconi, D; Nowatschin, D; Ott, J; Pantaleo, F; Peiffer, T; Perieanu, A; Pietsch, N; Poehlsen, J; Rathjens, D; Sander, C; Schettler, H; Schleper, P; Schlieckau, E; Schmidt, A; Schwandt, J; Seidel, M; Sola, V; Stadie, H; Steinbrück, G; Tholen, H; Troendle, D; Usai, E; Vanelderen, L; Vanhoefer, A; Akbiyik, M; Amstutz, C; Barth, C; Baus, C; Berger, J; Beskidt, C; Böser, C; Butz, E; Caspart, R; Chwalek, T; Colombo, F; De Boer, W; Descroix, A; Dierlamm, A; Eber, R; Feindt, M; Fink, S; Fischer, M; Frensch, F; Freund, B; Friese, R; Funke, D; Giffels, M; Gilbert, A; Haitz, D; Harbaum, T; Harrendorf, M A; Hartmann, F; Husemann, U; Kassel, F; Katkov, I; Kornmayer, A; Kudella, S; Lobelle Pardo, P; Maier, B; Mildner, H; Mozer, M U; Müller, T; Müller, Th; Plagge, M; Printz, M; Quast, G; Rabbertz, K; Röcker, S; Roscher, F; Shvetsov, I; Sieber, G; Simonis, H J; Stober, F M; Ulrich, R; Wagner-Kuhr, J; Wayand, S; Weiler, T; Williamson, S; Wöhrmann, C; Wolf, R; Anagnostou, G; Daskalakis, G; Geralis, T; Giakoumopoulou, V A; Kyriakis, A; Loukas, D; Markou, A; Psallidas, A; Topsis-Giotis, I; Agapitos, A; Kesisoglou, S; Panagiotou, A; Saoulidou, N; Tziaferi, E; Evangelou, I; Flouris, G; Foudas, C; Kokkas, P; Loukas, N; Manthos, N; Papadopoulos, I; Paradas, E; Strologas, J; Bencze, G; Hajdu, C; Hazi, A; Hidas, P; Horvath, D; Sikler, F; Veszpremi, V; Vesztergombi, G; Zsigmond, A J; Beni, N; Czellar, S; Karancsi, J; Molnar, J; Szillasi, Z; Bartók, M; Makovec, A; Raics, P; Trocsanyi, Z L; Ujvari, B; Mal, P; Mandal, K; Sahoo, N; Swain, S K; Bansal, S; Beri, S B; Bhatnagar, V; Chawla, R; Gupta, R; Bhawandeep, U; Kalsi, A K; Kaur, A; Kaur, M; Kumar, R; Mehta, A; Mittal, M; Nishu, N; Singh, J B; Walia, G; Kumar, Ashok; Kumar, Arun; Bhardwaj, A; Choudhary, B C; Garg, R B; Kumar, A; Malhotra, S; Naimuddin, M; Ranjan, K; Sharma, R; Sharma, V; Banerjee, S; Bhattacharya, S; Chatterjee, K; Dey, S; Dutta, S; Jain, Sa; Jain, Sh; Khurana, R; Majumdar, N; Modak, A; Mondal, K; Mukherjee, S; Mukhopadhyay, S; Roy, A; Roy, D; Roy Chowdhury, S; Sarkar, S; Sharan, M; Abdulsalam, A; Chudasama, R; Dutta, D; Jha, V; Kumar, V; Mohanty, A K; Pant, L M; Shukla, P; Topkar, A; Aziz, T; Banerjee, S; Bhowmik, S; Chatterjee, R M; Dewanjee, R K; Dugad, S; Ganguly, S; Ghosh, S; Guchait, M; Gurtu, A; Kole, G; Kumar, S; Mahakud, B; Maity, M; Majumder, G; Mazumdar, K; Mitra, S; Mohanty, G B; Parida, B; Sarkar, T; Sudhakar, K; Sur, N; Sutar, B; Wickramage, N; Sharma, S; Bakhshiansohi, H; Behnamian, H; Etesami, S M; Fahim, A; Goldouzian, R; Khakzad, M; Mohammadi Najafabadi, M; Naseri, M; Paktinat Mehdiabadi, S; Rezaei Hosseinabadi, F; Safarzadeh, B; Zeinali, M; Felcini, M; Grunewald, M; Abbrescia, M; Calabria, C; Caputo, C; Chhibra, S S; Colaleo, A; Creanza, D; Cristella, L; De Filippis, N; De Palma, M; Fiore, L; Iaselli, G; Maggi, G; Maggi, M; Miniello, G; My, S; Nuzzo, S; Pompili, A; Pugliese, G; Radogna, R; Ranieri, A; Selvaggi, G; Silvestris, L; Venditti, R; Verwilligen, P; Abbiendi, G; Battilana, C; Benvenuti, A C; Bonacorsi, D; Braibant-Giacomelli, S; Brigliadori, L; Campanini, R; Capiluppi, P; Castro, A; Cavallo, F R; Codispoti, G; Cuffiani, M; Dallavalle, G M; Fabbri, F; Fanfani, A; Fasanella, D; Giacomelli, P; Grandi, C; Guiducci, L; Marcellini, S; Masetti, G; Montanari, A; Navarria, F L; Perrotta, A; Rossi, A M; Rovelli, T; Siroli, G P; Tosi, N; Travaglini, R; Cappello, G; Chiorboli, M; Costa, S; Giordano, F; Potenza, R; Tricomi, A; Tuve, C; Barbagli, G; Ciulli, V; Civinini, C; D'Alessandro, R; Focardi, E; Gonzi, S; Gori, V; Lenzi, P; Meschini, M; Paoletti, S; Sguazzoni, G; Tropiano, A; Viliani, L; Benussi, L; Bianco, S; Fabbri, F; Piccolo, D; Calvelli, V; Ferro, F; Lo Vetere, M; Robutti, E; Tosi, S; Dinardo, M E; Fiorendi, S; Gennai, S; Gerosa, R; Ghezzi, A; Govoni, P; Malvezzi, S; Manzoni, R A; Marzocchi, B; Menasce, D; Moroni, L; Paganoni, M; Pedrini, D; Ragazzi, S; Redaelli, N; Tabarelli de Fatis, T; Buontempo, S; Cavallo, N; Di Guida, S; Esposito, M; Fabozzi, F; Iorio, A O M; Lanza, G; Lista, L; Meola, S; Merola, M; Paolucci, P; Sciacca, C; Thyssen, F; Azzi, P; Bacchetta, N; Bisello, D; Branca, A; Carlin, R; Carvalho Antunes De Oliveira, A; Checchia, P; Dall'Osso, M; Dorigo, T; Dosselli, U; Gasparini, F; Gasparini, U; Gozzelino, A; Kanishchev, K; Lacaprara, S; Margoni, M; Meneguzzo, A T; Pazzini, J; Pozzobon, N; Ronchese, P; Simonetto, F; Torassa, E; Tosi, M; Zanetti, M; Zotto, P; Zucchetta, A; Zumerle, G; Braghieri, A; Gabusi, M; Magnani, A; Ratti, S P; Re, V; Riccardi, C; Salvini, P; Vai, I; Vitulo, P; Alunni Solestizi, L; Biasini, M; Bilei, G M; Ciangottini, D; Fanò, L; Lariccia, P; Mantovani, G; Menichelli, M; Saha, A; Santocchia, A; Spiezia, A; Androsov, K; Azzurri, P; Bagliesi, G; Bernardini, J; Boccali, T; Broccolo, G; Castaldi, R; Ciocci, M A; Dell'Orso, R; Donato, S; Fedi, G; Foà, L; Giassi, A; Grippo, M T; Ligabue, F; Lomtadze, T; Martini, L; Messineo, A; Palla, F; Rizzi, A; Savoy-Navarro, A; Serban, A T; Spagnolo, P; Squillacioti, P; Tenchini, R; Tonelli, G; Venturi, A; Verdini, P G; Barone, L; Cavallari, F; D'imperio, G; Del Re, D; Diemoz, M; Gelli, S; Jorda, C; Longo, E; Margaroli, F; Meridiani, P; Micheli, F; Organtini, G; Paramatti, R; Preiato, F; Rahatlou, S; Rovelli, C; Santanastasio, F; Traczyk, P; Amapane, N; Arcidiacono, R; Argiro, S; Arneodo, M; Bellan, R; Biino, C; Cartiglia, N; Costa, M; Covarelli, R; Degano, A; Demaria, N; Finco, L; Kiani, B; Mariotti, C; Maselli, S; Migliore, E; Monaco, V; Monteil, E; Musich, M; Obertino, M M; Pacher, L; Pastrone, N; Pelliccioni, M; Pinna Angioni, G L; Ravera, F; Romero, A; Ruspa, M; Sacchi, R; Solano, A; Staiano, A; Tamponi, U; Belforte, S; Candelise, V; Casarsa, M; Cossutti, F; Della Ricca, G; Gobbo, B; La Licata, C; Marone, M; Schizzi, A; Umer, T; Zanetti, A; Chang, S; Kropivnitskaya, A; Nam, S K; Kim, D H; Kim, G N; Kim, M S; Kong, D J; Lee, S; Oh, Y D; Sakharov, A; Son, D C; Brochero Cifuentes, J A; Kim, H; Kim, T J; Ryu, M S; Song, S; Choi, S; Go, Y; Gyun, D; Hong, B; Jo, M; Kim, H; Kim, Y; Lee, B; Lee, K; Lee, K S; Lee, S; Park, S K; Roh, Y; Yoo, H D; Choi, M; Kim, J H; Lee, J S H; Park, I C; Ryu, G; Choi, Y; Choi, Y K; Goh, J; Kim, D; Kwon, E; Lee, J; Yu, I; Juodagalvis, A; Vaitkus, J; Ibrahim, Z A; Komaragiri, J R; Md Ali, M A B; Mohamad Idris, F; Wan Abdullah, W A T; Casimiro Linares, E; Castilla-Valdez, H; De La Cruz-Burelo, E; Heredia-de La Cruz, I; Hernandez-Almada, A; Lopez-Fernandez, R; Sanchez-Hernandez, A; Carrillo Moreno, S; Vazquez Valencia, F; Carpinteyro, S; Pedraza, I; Salazar Ibarguen, H A; Morelos Pineda, A; Krofcheck, D; Butler, P H; Reucroft, S; Ahmad, A; Ahmad, M; Hassan, Q; Hoorani, H R; Khan, W A; Khurshid, T; Shoaib, M; Bialkowska, H; Bluj, M; Boimska, B; Frueboes, T; Górski, M; Kazana, M; Nawrocki, K; Romanowska-Rybinska, K; Szleper, M; Zalewski, P; Brona, G; Bunkowski, K; Doroba, K; Kalinowski, A; Konecki, M; Krolikowski, J; Misiura, M; Olszewski, M; Walczak, M; Bargassa, P; Beirão Da Cruz E Silva, C; Di Francesco, A; Faccioli, P; Ferreira Parracho, P G; Gallinaro, M; Lloret Iglesias, L; Nguyen, F; Rodrigues Antunes, J; Seixas, J; Toldaiev, O; Vadruccio, D; Varela, J; Vischia, P; Afanasiev, S; Bunin, P; Gavrilenko, M; Golutvin, I; Gorbunov, I; Kamenev, A; Karjavin, V; Konoplyanikov, V; Lanev, A; Malakhov, A; Matveev, V; Moisenz, P; Palichik, V; Perelygin, V; Shmatov, S; Shulha, S; Skatchkov, N; Smirnov, V; Toriashvili, T; Zarubin, A; Golovtsov, V; Ivanov, Y; Kim, V; Kuznetsova, E; Levchenko, P; Murzin, V; Oreshkin, V; Smirnov, I; Sulimov, V; Uvarov, L; Vavilov, S; Vorobyev, A; Andreev, Yu; Dermenev, A; Gninenko, S; Golubev, N; Karneyeu, A; Kirsanov, M; Krasnikov, N; Pashenkov, A; Tlisov, D; Toropin, A; Epshteyn, V; Gavrilov, V; Lychkovskaya, N; Popov, V; Pozdnyakov, I; Safronov, G; Spiridonov, A; Vlasov, E; Zhokin, A; Andreev, V; Azarkin, M; Dremin, I; Kirakosyan, M; Leonidov, A; Mesyats, G; Rusakov, S V; Vinogradov, A; Baskakov, A; Belyaev, A; Boos, E; Bunichev, V; Dubinin, M; Dudko, L; Ershov, A; Gribushin, A; Klyukhin, V; Kodolova, O; Lokhtin, I; Myagkov, I; Obraztsov, S; Petrushanko, S; Savrin, V; Azhgirey, I; Bayshev, I; Bitioukov, S; Kachanov, V; Kalinin, A; Konstantinov, D; Krychkine, V; Petrov, V; Ryutin, R; Sobol, A; Tourtchanovitch, L; Troshin, S; Tyurin, N; Uzunian, A; Volkov, A; Adzic, P; Ekmedzic, M; Milosevic, J; Rekovic, V; Alcaraz Maestre, J; Calvo, E; Cerrada, M; Chamizo Llatas, M; Colino, N; De La Cruz, B; Delgado Peris, A; Domínguez Vázquez, D; Escalante Del Valle, A; Fernandez Bedoya, C; Fernández Ramos, J P; Flix, J; Fouz, M C; Garcia-Abia, P; Gonzalez Lopez, O; Goy Lopez, S; Hernandez, J M; Josa, M I; Navarro De Martino, E; Pérez-Calero Yzquierdo, A; Puerta Pelayo, J; Quintario Olmeda, A; Redondo, I; Romero, L; Soares, M S; Albajar, C; de Trocóniz, J F; Missiroli, M; Moran, D; Brun, H; Cuevas, J; Fernandez Menendez, J; Folgueras, S; Gonzalez Caballero, I; Palencia Cortezon, E; Vizan Garcia, J M; Cabrillo, I J; Calderon, A; Castiñeiras De Saa, J R; Duarte Campderros, J; Fernandez, M; Gomez, G; Graziano, A; Lopez Virto, A; Marco, J; Marco, R; Martinez Rivero, C; Matorras, F; Munoz Sanchez, F J; Piedra Gomez, J; Rodrigo, T; Rodríguez-Marrero, A Y; Ruiz-Jimeno, A; Scodellaro, L; Vila, I; Vilar Cortabitarte, R; Abbaneo, D; Auffray, E; Auzinger, G; Bachtis, M; Baillon, P; Ball, A H; Barney, D; Benaglia, A; Bendavid, J; Benhabib, L; Benitez, J F; Berruti, G M; Bloch, P; Bocci, A; Bonato, A; Botta, C; Breuker, H; Camporesi, T; Cerminara, G; Colafranceschi, S; D'Alfonso, M; d'Enterria, D; Dabrowski, A; Daponte, V; David, A; De Gruttola, M; De Guio, F; De Roeck, A; De Visscher, S; Di Marco, E; Dobson, M; Dordevic, M; du Pree, T; Dupont-Sagorin, N; Elliott-Peisert, A; Franzoni, G; Funk, W; Gigi, D; Gill, K; Giordano, D; Girone, M; Glege, F; Guida, R; Gundacker, S; Guthoff, M; Hammer, J; Hansen, M; Harris, P; Hegeman, J; Innocente, V; Janot, P; Kirschenmann, H; Kortelainen, M J; Kousouris, K; Krajczar, K; Lecoq, P; Lourenço, C; Lucchini, M T; Magini, N; Malgeri, L; Mannelli, M; Marrouche, J; Martelli, A; Masetti, L; Meijers, F; Mersi, S; Meschi, E; Moortgat, F; Morovic, S; Mulders, M; Nemallapudi, M V; Neugebauer, H; Orfanelli, S; Orsini, L; Pape, L; Perez, E; Petrilli, A; Petrucciani, G; Pfeiffer, A; Piparo, D; Racz, A; Rolandi, G; Rovere, M; Ruan, M; Sakulin, H; Schäfer, C; Schwick, C; Sharma, A; Silva, P; Simon, M; Sphicas, P; Spiga, D; Steggemann, J; Stieger, B; Stoye, M; Takahashi, Y; Treille, D; Tsirou, A; Veres, G I; Wardle, N; Wöhri, H K; Zagozdzinska, A; Zeuner, W D; Bertl, W; Deiters, K; Erdmann, W; Horisberger, R; Ingram, Q; Kaestli, H C; Kotlinski, D; Langenegger, U; Rohe, T; Bachmair, F; Bäni, L; Bianchini, L; Buchmann, M A; Casal, B; Dissertori, G; Dittmar, M; Donegà, M; Dünser, M; Eller, P; Grab, C; Heidegger, C; Hits, D; Hoss, J; Kasieczka, G; Lustermann, W; Mangano, B; Marini, A C; Marionneau, M; Martinez Ruiz Del Arbol, P; Masciovecchio, M; Meister, D; Musella, P; Nessi-Tedaldi, F; Pandolfi, F; Pata, J; Pauss, F; Perrozzi, L; Peruzzi, M; Quittnat, M; Rossini, M; Starodumov, A; Takahashi, M; Tavolaro, V R; Theofilatos, K; Wallny, R; Weber, H A; Aarrestad, T K; Amsler, C; Canelli, M F; Chiochia, V; De Cosa, A; Galloni, C; Hinzmann, A; Hreus, T; Kilminster, B; Lange, C; Ngadiuba, J; Pinna, D; Robmann, P; Ronga, F J; Salerno, D; Taroni, S; Yang, Y; Cardaci, M; Chen, K H; Doan, T H; Ferro, C; Konyushikhin, M; Kuo, C M; Lin, W; Lu, Y J; Volpe, R; Yu, S S; Chang, P; Chang, Y H; Chang, Y W; Chao, Y; Chen, K F; Chen, P H; Dietz, C; Fiori, F; Grundler, U; Hou, W-S; Hsiung, Y; Liu, Y F; Lu, R-S; Miñano Moya, M; Petrakou, E; Tsai, J F; Tzeng, Y M; Wilken, R; Asavapibhop, B; Kovitanggoon, K; Singh, G; Srimanobhas, N; Suwonjandee, N; Adiguzel, A; Cerci, S; Dozen, C; Girgis, S; Gokbulut, G; Guler, Y; Gurpinar, E; Hos, I; Kangal, E E; Kayis Topaksu, A; Onengut, G; Ozdemir, K; Ozturk, S; Tali, B; Topakli, H; Vergili, M; Zorbilmez, C; Akin, I V; Bilin, B; Bilmis, S; Isildak, B; Karapinar, G; Surat, U E; Yalvac, M; Zeyrek, M; Albayrak, E A; Gülmez, E; Kaya, M; Kaya, O; Yetkin, T; Cankocak, K; Sen, S; Vardarlı, F I; Grynyov, B; Levchuk, L; Sorokin, P; Aggleton, R; Ball, F; Beck, L; Brooke, J J; Clement, E; Cussans, D; Flacher, H; Goldstein, J; Grimes, M; Heath, G P; Heath, H F; Jacob, J; Kreczko, L; Lucas, C; Meng, Z; Newbold, D M; Paramesvaran, S; Poll, A; Sakuma, T; Seif El Nasr-Storey, S; Senkin, S; Smith, D; Smith, V J; Bell, K W; Belyaev, A; Brew, C; Brown, R M; Cockerill, D J A; Coughlan, J A; Harder, K; Harper, S; Olaiya, E; Petyt, D; Shepherd-Themistocleous, C H; Thea, A; Tomalin, I R; Williams, T; Womersley, W J; Worm, S D; Baber, M; Bainbridge, R; Buchmuller, O; Bundock, A; Burton, D; Casasso, S; Citron, M; Colling, D; Corpe, L; Cripps, N; Dauncey, P; Davies, G; De Wit, A; Della Negra, M; Dunne, P; Elwood, A; Ferguson, W; Fulcher, J; Futyan, D; Hall, G; Iles, G; Karapostoli, G; Kenzie, M; Lane, R; Lucas, R; Lyons, L; Magnan, A-M; Malik, S; Nash, J; Nikitenko, A; Pela, J; Pesaresi, M; Petridis, K; Raymond, D M; Richards, A; Rose, A; Seez, C; Sharp, P; Tapper, A; Uchida, K; Vazquez Acosta, M; Virdee, T; Zenz, S C; Cole, J E; Hobson, P R; Khan, A; Kyberd, P; Leggat, D; Leslie, D; Reid, I D; Symonds, P; Teodorescu, L; Turner, M; Borzou, A; Dittmann, J; Hatakeyama, K; Kasmi, A; Liu, H; Pastika, N; Charaf, O; Cooper, S I; Henderson, C; Rumerio, P; Avetisyan, A; Bose, T; Fantasia, C; Gastler, D; Lawson, P; Rankin, D; Richardson, C; Rohlf, J; St John, J; Sulak, L; Zou, D; Alimena, J; Berry, E; Bhattacharya, S; Cutts, D; Dhingra, N; Ferapontov, A; Garabedian, A; Heintz, U; Laird, E; Landsberg, G; Mao, Z; Narain, M; Sagir, S; Sinthuprasith, T; Breedon, R; Breto, G; Calderon De La Barca Sanchez, M; Chauhan, S; Chertok, M; Conway, J; Conway, R; Cox, P T; Erbacher, R; Gardner, M; Ko, W; Lander, R; Mulhearn, M; Pellett, D; Pilot, J; Ricci-Tam, F; Shalhout, S; Smith, J; Squires, M; Stolp, D; Tripathi, M; Wilbur, S; Yohay, R; Cousins, R; Everaerts, P; Farrell, C; Hauser, J; Ignatenko, M; Rakness, G; Saltzberg, D; Takasugi, E; Valuev, V; Weber, M; Burt, K; Clare, R; Ellison, J; Gary, J W; Hanson, G; Heilman, J; Ivova Rikova, M; Jandir, P; Kennedy, E; Lacroix, F; Long, O R; Luthra, A; Malberti, M; Olmedo Negrete, M; Shrinivas, A; Sumowidagdo, S; Wei, H; Wimpenny, S; Branson, J G; Cerati, G B; Cittolin, S; D'Agnolo, R T; Holzner, A; Kelley, R; Klein, D; Letts, J; Macneill, I; Olivito, D; Padhi, S; Pieri, M; Sani, M; Sharma, V; Simon, S; Tadel, M; Tu, Y; Vartak, A; Wasserbaech, S; Welke, C; Würthwein, F; Yagil, A; Zevi Della Porta, G; Barge, D; Bradmiller-Feld, J; Campagnari, C; Dishaw, A; Dutta, V; Flowers, K; Franco Sevilla, M; Geffert, P; George, C; Golf, F; Gouskos, L; Gran, J; Incandela, J; Justus, C; Mccoll, N; Mullin, S D; Richman, J; Stuart, D; Suarez, I; To, W; West, C; Yoo, J; Anderson, D; Apresyan, A; Bornheim, A; Bunn, J; Chen, Y; Duarte, J; Mott, A; Newman, H B; Pena, C; Pierini, M; Spiropulu, M; Vlimant, J R; Xie, S; Zhu, R Y; Azzolini, V; Calamba, A; Carlson, B; Ferguson, T; Iiyama, Y; Paulini, M; Russ, J; Sun, M; Vogel, H; Vorobiev, I; Cumalat, J P; Ford, W T; Gaz, A; Jensen, F; Johnson, A; Krohn, M; Mulholland, T; Nauenberg, U; Smith, J G; Stenson, K; Wagner, S R; Alexander, J; Chatterjee, A; Chaves, J; Chu, J; Dittmer, S; Eggert, N; Mirman, N; Nicolas Kaufman, G; Patterson, J R; Rinkevicius, A; Ryd, A; Skinnari, L; Soffi, L; Sun, W; Tan, S M; Teo, W D; Thom, J; Thompson, J; Tucker, J; Weng, Y; Wittich, P; Abdullin, S; Albrow, M; Anderson, J; Apollinari, G; Bauerdick, L A T; Beretvas, A; Berryhill, J; Bhat, P C; Bolla, G; Burkett, K; Butler, J N; Cheung, H W K; Chlebana, F; Cihangir, S; Elvira, V D; Fisk, I; Freeman, J; Gottschalk, E; Gray, L; Green, D; Grünendahl, S; Gutsche, O; Hanlon, J; Hare, D; Harris, R M; Hirschauer, J; Hooberman, B; Hu, Z; Jindariani, S; Johnson, M; Joshi, U; Jung, A W; Klima, B; Kreis, B; Kwan, S; Lammel, S; Linacre, J; Lincoln, D; Lipton, R; Liu, T; Lopes De Sá, R; Lykken, J; Maeshima, K; Marraffino, J M; Martinez Outschoorn, V I; Maruyama, S; Mason, D; McBride, P; Merkel, P; Mishra, K; Mrenna, S; Nahn, S; Newman-Holmes, C; O'Dell, V; Prokofyev, O; Sexton-Kennedy, E; Soha, A; Spalding, W J; Spiegel, L; Taylor, L; Tkaczyk, S; Tran, N V; Uplegger, L; Vaandering, E W; Vernieri, C; Verzocchi, M; Vidal, R; Whitbeck, A; Yang, F; Yin, H; Acosta, D; Avery, P; Bortignon, P; Bourilkov, D; Carnes, A; Carver, M; Curry, D; Das, S; Di Giovanni, G P; Field, R D; Fisher, M; Furic, I K; Hugon, J; Konigsberg, J; Korytov, A; Low, J F; Ma, P; Matchev, K; Mei, H; Milenovic, P; Mitselmakher, G; Muniz, L; Rank, D; Shchutska, L; Snowball, M; Sperka, D; Wang, S J; Yelton, J; Hewamanage, S; Linn, S; Markowitz, P; Martinez, G; Rodriguez, J L; Ackert, A; Adams, J R; Adams, T; Askew, A; Bochenek, J; Diamond, B; Haas, J; Hagopian, S; Hagopian, V; Johnson, K F; Khatiwada, A; Prosper, H; Veeraraghavan, V; Weinberg, M; Bhopatkar, V; Hohlmann, M; Kalakhety, H; Mareskas-Palcek, D; Roy, T; Yumiceva, F; Adams, M R; Apanasevich, L; Berry, D; Betts, R R; Bucinskaite, I; Cavanaugh, R; Evdokimov, O; Gauthier, L; Gerber, C E; Hofman, D J; Kurt, P; O'Brien, C; Sandoval Gonzalez, I D; Silkworth, C; Turner, P; Varelas, N; Wu, Z; Zakaria, M; Bilki, B; Clarida, W; Dilsiz, K; Durgut, S; Gandrajula, R P; Haytmyradov, M; Khristenko, V; Merlo, J-P; Mermerkaya, H; Mestvirishvili, A; Moeller, A; Nachtman, J; Ogul, H; Onel, Y; Ozok, F; Penzo, A; Snyder, C; Tan, P; Tiras, E; Wetzel, J; Yi, K; Anderson, I; Barnett, B A; Blumenfeld, B; Fehling, D; Feng, L; Gritsan, A V; Maksimovic, P; Martin, C; Nash, K; Osherson, M; Swartz, M; Xiao, M; Xin, Y; Baringer, P; Bean, A; Benelli, G; Bruner, C; Gray, J; Kenny, R P; Majumder, D; Malek, M; Murray, M; Noonan, D; Sanders, S; Stringer, R; Wang, Q; Wood, J S; Chakaberia, I; Ivanov, A; Kaadze, K; Khalil, S; Makouski, M; Maravin, Y; Saini, L K; Skhirtladze, N; Svintradze, I; Toda, S; Lange, D; Rebassoo, F; Wright, D; Anelli, C; Baden, A; Baron, O; Belloni, A; Calvert, B; Eno, S C; Ferraioli, C; Gomez, J A; Hadley, N J; Jabeen, S; Kellogg, R G; Kolberg, T; Kunkle, J; Lu, Y; Mignerey, A C; Pedro, K; Shin, Y H; Skuja, A; Tonjes, M B; Tonwar, S C; Apyan, A; Barbieri, R; Baty, A; Bierwagen, K; Brandt, S; Busza, W; Cali, I A; Di Matteo, L; Gomez Ceballos, G; Goncharov, M; Gulhan, D; Innocenti, G M; Klute, M; Kovalskyi, D; Lai, Y S; Lee, Y-J; Levin, A; Luckey, P D; Mcginn, C; Niu, X; Paus, C; Ralph, D; Roland, C; Roland, G; Stephans, G S F; Sumorok, K; Varma, M; Velicanu, D; Veverka, J; Wang, J; Wang, T W; Wyslouch, B; Yang, M; Zhukova, V; Dahmes, B; Finkel, A; Gude, A; Hansen, P; Kalafut, S; Kao, S C; Klapoetke, K; Kubota, Y; Lesko, Z; Mans, J; Nourbakhsh, S; Ruckstuhl, N; Rusack, R; Tambe, N; Turkewitz, J; Acosta, J G; Oliveros, S; Avdeeva, E; Bloom, K; Bose, S; Claes, D R; Dominguez, A; Fangmeier, C; Gonzalez Suarez, R; Kamalieddin, R; Keller, J; Knowlton, D; Kravchenko, I; Lazo-Flores, J; Meier, F; Monroy, J; Ratnikov, F; Siado, J E; Snow, G R; Alyari, M; Dolen, J; George, J; Godshalk, A; Iashvili, I; Kaisen, J; Kharchilava, A; Kumar, A; Rappoccio, S; Alverson, G; Barberis, E; Baumgartel, D; Chasco, M; Hortiangtham, A; Massironi, A; Morse, D M; Nash, D; Orimoto, T; Teixeira De Lima, R; Trocino, D; Wang, R-J; Wood, D; Zhang, J; Hahn, K A; Kubik, A; Mucia, N; Odell, N; Pollack, B; Pozdnyakov, A; Schmitt, M; Stoynev, S; Sung, K; Trovato, M; Velasco, M; Won, S; Brinkerhoff, A; Dev, N; Hildreth, M; Jessop, C; Karmgard, D J; Kellams, N; Lannon, K; Lynch, S; Marinelli, N; Meng, F; Mueller, C; Musienko, Y; Pearson, T; Planer, M; Ruchti, R; Smith, G; Valls, N; Wayne, M; Wolf, M; Woodard, A; Antonelli, L; Brinson, J; Bylsma, B; Durkin, L S; Flowers, S; Hart, A; Hill, C; Hughes, R; Kotov, K; Ling, T Y; Liu, B; Luo, W; Puigh, D; Rodenburg, M; Winer, B L; Wulsin, H W; Driga, O; Elmer, P; Hardenbrook, J; Hebda, P; Koay, S A; Lujan, P; Marlow, D; Medvedeva, T; Mooney, M; Olsen, J; Palmer, C; Piroué, P; Quan, X; Saka, H; Stickland, D; Tully, C; Werner, J S; Zuranski, A; Barnes, V E; Benedetti, D; Bortoletto, D; Gutay, L; Jha, M K; Jones, M; Jung, K; Kress, M; Leonardo, N; Miller, D H; Neumeister, N; Primavera, F; Radburn-Smith, B C; Shi, X; Shipsey, I; Silvers, D; Sun, J; Svyatkovskiy, A; Wang, F; Xie, W; Xu, L; Zablocki, J; Parashar, N; Stupak, J; Adair, A; Akgun, B; Chen, Z; Ecklund, K M; Geurts, F J M; Guilbaud, M; Li, W; Michlin, B; Northup, M; Padley, B P; Redjimi, R; Roberts, J; Rorie, J; Tu, Z; Zabel, J; Betchart, B; Bodek, A; de Barbaro, P; Demina, R; Eshaq, Y; Ferbel, T; Galanti, M; Garcia-Bellido, A; Goldenzweig, P; Han, J; Harel, A; Hindrichs, O; Khukhunaishvili, A; Petrillo, G; Verzetti, M; Demortier, L; Arora, S; Barker, A; Chou, J P; Contreras-Campana, C; Contreras-Campana, E; Duggan, D; Ferencek, D; Gershtein, Y; Gray, R; Halkiadakis, E; Hidas, D; Hughes, E; Kaplan, S; Kunnawalkam Elayavalli, R; Lath, A; Panwalkar, S; Park, M; Salur, S; Schnetzer, S; Sheffield, D; Somalwar, S; Stone, R; Thomas, S; Thomassen, P; Walker, M; Foerster, M; Riley, G; Rose, K; Spanier, S; York, A; Bouhali, O; Castaneda Hernandez, A; Dalchenko, M; De Mattia, M; Delgado, A; Dildick, S; Eusebi, R; Flanagan, W; Gilmore, J; Kamon, T; Krutelyov, V; Montalvo, R; Mueller, R; Osipenkov, I; Pakhotin, Y; Patel, R; Perloff, A; Roe, J; Rose, A; Safonov, A; Tatarinov, A; Ulmer, K A; Akchurin, N; Cowden, C; Damgov, J; Dragoiu, C; Dudero, P R; Faulkner, J; Kunori, S; Lamichhane, K; Lee, S W; Libeiro, T; Undleeb, S; Volobouev, I; Appelt, E; Delannoy, A G; Greene, S; Gurrola, A; Janjam, R; Johns, W; Maguire, C; Mao, Y; Melo, A; Sheldon, P; Snook, B; Tuo, S; Velkovska, J; Xu, Q; Arenton, M W; Boutle, S; Cox, B; Francis, B; Goodell, J; Hirosky, R; Ledovskoy, A; Li, H; Lin, C; Neu, C; Wolfe, E; Wood, J; Xia, F; Clarke, C; Harr, R; Karchin, P E; Kottachchi Kankanamge Don, C; Lamichhane, P; Sturdy, J; Belknap, D A; Carlsmith, D; Cepeda, M; Christian, A; Dasu, S; Dodd, L; Duric, S; Friis, E; Gomber, B; Hall-Wilton, R; Herndon, M; Hervé, A; Klabbers, P; Lanaro, A; Levine, A; Long, K; Loveless, R; Mohapatra, A; Ojalvo, I; Perry, T; Pierro, G A; Polese, G; Ross, I; Ruggles, T; Sarangi, T; Savin, A; Sharma, A; Smith, N; Smith, W H; Taylor, D; Woods, N 2015-05-15 A measurement of the Higgs boson mass is presented based on the combined data samples of the ATLAS and CMS experiments at the CERN LHC in the H→γγ and H→ZZ→4ℓ decay channels. The results are obtained from a simultaneous fit to the reconstructed invariant mass peaks in the two channels and for the two experiments. The measured masses from the individual channels and the two experiments are found to be consistent among themselves. The combined measured mass of the Higgs boson is m_{H}=125.09±0.21 (stat)±0.11 (syst) GeV. 13. Penning Trap Experiments with the Most Exotic Nuclei on Earth: Precision Mass Measurements of Halo Nuclei Science.gov (United States) Brodeur, M.; Brunner, T.; Ettenauer, S.; Lapierre, A.; Ringle, R.; Delheij, P.; Dilling, J. 2009-05-01 Exotic nuclei are characterized with an extremely unbalanced protons-neutrons ratio (p/n) where for instance, the halo isotopes of He and Li have up to 3X more n than p (compared to p/n = 1 in ^12C). The properties of these exotic halo nuclei have long been recognized as the most stringent tests of our understanding of the strong force. ^11Li belongs to a special category of halos called Borromean, bound as a three-body family, while the two-body siblings, ^10Li and 2 n, are unbound as separate entities. Last year, a first mass measurement of the radioisotope ^11Li using a Penning trap spectrometer was carried out at the TITAN (Triumf's Ion Trap for Atomic and Nuclear science) facility at TRIUMF-ISAC. Penning traps are proven to be the most precise device to make mass measurements, yet until now they were unable to reach these nuclei. At TRIUMF we managed to measure the mass of ^11Li to an unprecedented precision of dm/m = 60 ppb, which is remarkable since it has a half-life of only 8.8 ms which it the shortest-lived nuclide to be measured with this technique. Furthermore, new and improved masses for the 2 and 4 n halo ^6,8He, as well has the 1 n halo ^11Be have been performed. An overview of the TITAN mass measurement program and its impact in understanding the most exotic nuclei will be given. 14. Constraining Aerosol Optical Models Using Ground-Based, Collocated Particle Size and Mass Measurements in Variable Air Mass Regimes During the 7-SEAS/Dongsha Experiment Science.gov (United States) Bell, Shaun W.; Hansell, Richard A.; Chow, Judith C.; Tsay, Si-Chee; Wang, Sheng-Hsiang; Ji, Qiang; Li, Can; Watson, John G.; Khlystov, Andrey 2012-01-01 During the spring of 2010, NASA Goddard's COMMIT ground-based mobile laboratory was stationed on Dongsha Island off the southwest coast of Taiwan, in preparation for the upcoming 2012 7-SEAS field campaign. The measurement period offered a unique opportunity for conducting detailed investigations of the optical properties of aerosols associated with different air mass regimes including background maritime and those contaminated by anthropogenic air pollution and mineral dust. What appears to be the first time for this region, a shortwave optical closure experiment for both scattering and absorption was attempted over a 12-day period during which aerosols exhibited the most change. Constraints to the optical model included combined SMPS and APS number concentration data for a continuum of fine and coarse-mode particle sizes up to PM2.5. We also take advantage of an IMPROVE chemical sampler to help constrain aerosol composition and mass partitioning of key elemental species including sea-salt, particulate organic matter, soil, non sea-salt sulphate, nitrate, and elemental carbon. Our results demonstrate that the observed aerosol scattering and absorption for these diverse air masses are reasonably captured by the model, where peak aerosol events and transitions between key aerosols types are evident. Signatures of heavy polluted aerosol composed mostly of ammonium and non sea-salt sulphate mixed with some dust with transitions to background sea-salt conditions are apparent in the absorption data, which is particularly reassuring owing to the large variability in the imaginary component of the refractive indices. Extinctive features at significantly smaller time scales than the one-day sample period of IMPROVE are more difficult to reproduce, as this requires further knowledge concerning the source apportionment of major chemical components in the model. Consistency between the measured and modeled optical parameters serves as an important link for advancing remote 15. Measurement of the Hadronic Mass Spectrum in B to Xulnu Decaysand Determination of the b-Quark Mass at the BaBar Experiment Energy Technology Data Exchange (ETDEWEB) Tackmann, Kerstin [Univ. of California, Berkeley, CA (United States) 2008-06-26 I present preliminary results of the measurement of the hadronic mass spectrum and its first three spectral moments in inclusive charmless semileptonic B-meson decays. The truncated hadronic mass moments are used for the first determination of the b-quark mass and the nonperturbative parameters μπ2 and ρD3 in this B-meson decay channel. The study is based on 383 x 106 B$\\bar{B}$decays collected with the BABAR experiment at the PEP-II e+e- storage rings, located at the Stanford Linear Accelerator Center. The first, second central, and third central hadronic mass moment with a cut on the hadronic mass mX2 < 6.4GeV2 and the lepton momentum p* > 1 GeV are measured to be: M1 = (1.96 ± 0.34stat ± 0.53syst) GeV2; U2 = (1.92 ± 0.59stat ± 0.87syst) GeV4; and U3 = (1.79 ± 0.62stat ± 0.78syst) GeV6; with correlation coefficients ρ12 = 0.99, ρ23 = 0.94, and ρ13 = 0.88, respectively. Using Heavy Quark Effective Theory-based predictions in the kinetic scheme we extract: mb = (4.60 ± 0.13stat ± 0.19syst ± 0.10theo GeV); μπ2 = (0.40 ± 0.14stat ± 0.20syst ± 0.04theo) GeV2; ρD3 = (0.10 ± 0.02stat ± 0.02syst ± 0.07theo) GeV3; at μ = 1 GeV, with correlation coefficients ρmbμπ2 = -0.99, ρ μπ2ρD3 = 0.57, and ρmbρD3 = -0.59. The results are in good agreement with earlier determinations in inclusive charmed semileptonic and radiative penguin B-meson decays and have a 16. Double-diffractive processes in high-resolution missing-mass experiments at the Tevatron International Nuclear Information System (INIS) Khoze, V.A.; Martin, A.D. 2001-01-01 We evaluate, in a model-independent way, the signal-to-background ratio for Higgs→b anti b detection in exclusive double-diffractive events at the Tevatron and the LHC. For the missing-mass approach to be able to identify the Higgs boson, it will be necessary to use a central jet detector and to tag b quark jets. The signal is predicted to be very small at the Tevatron, but observable at the LHC. However we note that the background, that is double-diffractive dijet production, may serve as a unique gluon factory. We also give estimates for the double-diffractive production of χ c and χ b mesons at the Tevatron. We emphasize that a high-resolution missing-mass measurement, on its own, is insufficient to identify rare processes. (orig.) 17. Precise measurement of the charged B meson mass at the LHCb experiment CERN Document Server AUTHOR|(CDS)2076439 In the SM framework hadrons are colourless particles composed of quarks and gluons that interact by means the strong interactions described by Quantum chromodynamics (QCD). In this framework the theoretical predictions have to be compared with experimental data in order to verify if QCD is the correct theory of strong interactions.$b$physics is considered a stringent Standard Model (SM) test and recent developments in$b$hadron physics gave significant improvement in experimental determination of SM parameters.$b$mass is a fundamental parameter of SM but, since quarks are confined inside hadrons and are not observed as physical particles, it has to be extrapolated by theoretical calculations: a non-perturbative tool formulated on a discrete Euclidean space time grid for calculating the hadronic spectrum and the matrix elements of any operator called lattice QCD (LQCD) has been developed. One of the purposes of LQCD is to provide predictions on$B$meson mass starting from$bquark value: a fundamental r... 18. Shooting Alone: The Pre-Attack Experiences and Behaviors of U.S. Solo Mass Murderers. Science.gov (United States) Gill, Paul; Silver, James; Horgan, John; Corner, Emily 2017-05-01 This paper outlines the sociodemographic, developmental, antecedent attack, attack preparation, and commission properties of 115 mass murderers between 1990 and 2014. The results indicate that mass murderer attacks are usually the culmination of a complex mix of personal, political, and social drivers that crystalize at the same time to drive the individual down the path of violent action. We specifically focus upon areas related to prior criminal engagement, leakage, and attack location familiarity. Whether the violence comes to fruition is usually a combination of the availability and vulnerability of suitable targets that suit the heady mix of personal and political grievances and the individual's capability to engage in an attack from both a psychological and technical capability standpoint. Many individual cases share a mixture of unfortunate personal life circumstances coupled with an intensification of beliefs/grievances that later developed into the idea to engage in violence. © 2016 American Academy of Forensic Sciences. 19. Mass coral spawning: A natural large-scale nutrien t addition experiment DEFF Research Database (Denmark) Eyre, B.D.; Glud, Ronnie Nøhr; Patten, N. 2008-01-01 A mass coral spawning event on the Heron Island reef flat in 2005 provided a unique opportunity to examine the response of a coral reef ecosystem to a large episodic nutrient addition. A post-major spawning phytoplankton bloom resulted in only a small drawdown of dissolved inorganic phosphorus (DIP......), and dissolved organic phosphorus were used in the production of biomass, and mass balance calculations highlighted the importance of organic forms of N and P for benthic and pelagic production in tropical coral reef environments characterized by low inorganic N and P. The input of N and P via the deposition...... potential N limitation of benthic coral reef communities. For example, there was sufficient bioavailable P stored in the top 10 cm of the sediment column to sustain the prespawning rates of benthic production for over 200 d. Most of the change in benthic N cycling occurred via DON and N-2 pathways, driven... 20. Prospects for experiments on neutrino masses and mixing via neutrino oscillations at future accelerators International Nuclear Information System (INIS) Lanou, R.E. Jr. 1982-01-01 A study is made of the requirements necessary for improvement in our knowledge of limits in mass and mixing parameters for neutrinos via oscillation phenomena at accelerators. It is concluded that increased neutrino event rate (flux x energy) at modest energy machines (e.g., AGS and LAMPF) is the single most important requirement. This will permit smaller E/L ratios and refinement of systematics 1. Quantitative shear wave ultrasound elastography: initial experience in solid breast masses OpenAIRE Evans, Andrew; Whelehan, Patsy; Thomson, Kim; McLean, Denis; Brauer, Katrin; Purdie, Colin; Jordan, Lee; Baker, Lee; Thompson, Alastair 2010-01-01 Introduction Shear wave elastography is a new method of obtaining quantitative tissue elasticity data during breast ultrasound examinations. The aims of this study were (1) to determine the reproducibility of shear wave elastography (2) to correlate the elasticity values of a series of solid breast masses with histological findings and (3) to compare shear wave elastography with greyscale ultrasound for benign/malignant classification. Methods Using the Aixplorer® ultrasound system (SuperSoni... 2. Operational Experience of an Open-Access, Subscription-Based Mass Spectrometry and Proteomics Facility Science.gov (United States) Williamson, Nicholas A. 2018-03-01 This paper discusses the successful adoption of a subscription-based, open-access model of service delivery for a mass spectrometry and proteomics facility. In 2009, the Mass Spectrometry and Proteomics Facility at the University of Melbourne (Australia) moved away from the standard fee for service model of service provision. Instead, the facility adopted a subscription- or membership-based, open-access model of service delivery. For a low fixed yearly cost, users could directly operate the instrumentation but, more importantly, there were no limits on usage other than the necessity to share available instrument time with all other users. All necessary training from platform staff and many of the base reagents were also provided as part of the membership cost. These changes proved to be very successful in terms of financial outcomes for the facility, instrument access and usage, and overall research output. This article describes the systems put in place as well as the overall successes and challenges associated with the operation of a mass spectrometry/proteomics core in this manner. [Figure not available: see fulltext. 3. The Gillette Stadium Experience: A Retrospective Review of Mass Gathering Events From 2010 to 2015. Science.gov (United States) Goldberg, Scott A; Maggin, Jeremy; Molloy, Michael S; Baker, Olesya; Sarin, Ritu; Kelleher, Michael; Mont, Kevin; Fajana, Adedeji; Goralnick, Eric 2018-03-19 Mass gathering events can substantially impact public safety. Analyzing patient presentation and transport rates at various mass gathering events can help inform staffing models and improve preparedness. A retrospective review of all patients seeking medical attention across a variety of event types at a single venue with a capacity of 68,756 from January 2010 through September 2015. We examined 232 events with a total of 8,260,349 attendees generating 8157 medical contacts. Rates were 10 presentations and 1.6 transports per 10,000 attendees with a non-significant trend towards increased rates in postseason National Football League games. Concerts had significantly higher rates of presentation and transport than all other event types. Presenting concern varied significantly by event type and gender, and transport rate increased predictably with age. For cold weather events, transport rates increased at colder temperatures. Overall, on-site physicians did not impact rates. At a single venue hosting a variety of events across a 6-year period, we demonstrated significant variations in presentation and transport rates. Weather, gender, event type, and age all play important roles. Our analysis, while representative only of our specific venue, may be useful in developing response plans and staffing models for similar mass gathering venues. (Disaster Med Public Health Preparedness. 2018;page 1 of 7). 4. 2DB: a Proteomics database for storage, analysis, presentation, and retrieval of information from mass spectrometric experiments. Science.gov (United States) Allmer, Jens; Kuhlgert, Sebastian; Hippler, Michael 2008-07-07 The amount of information stemming from proteomics experiments involving (multi dimensional) separation techniques, mass spectrometric analysis, and computational analysis is ever-increasing. Data from such an experimental workflow needs to be captured, related and analyzed. Biological experiments within this scope produce heterogenic data ranging from pictures of one or two-dimensional protein maps and spectra recorded by tandem mass spectrometry to text-based identifications made by algorithms which analyze these spectra. Additionally, peptide and corresponding protein information needs to be displayed. In order to handle the large amount of data from computational processing of mass spectrometric experiments, automatic import scripts are available and the necessity for manual input to the database has been minimized. Information is in a generic format which abstracts from specific software tools typically used in such an experimental workflow. The software is therefore capable of storing and cross analysing results from many algorithms. A novel feature and a focus of this database is to facilitate protein identification by using peptides identified from mass spectrometry and link this information directly to respective protein maps. Additionally, our application employs spectral counting for quantitative presentation of the data. All information can be linked to hot spots on images to place the results into an experimental context. A summary of identified proteins, containing all relevant information per hot spot, is automatically generated, usually upon either a change in the underlying protein models or due to newly imported identifications. The supporting information for this report can be accessed in multiple ways using the user interface provided by the application. We present a proteomics database which aims to greatly reduce evaluation time of results from mass spectrometric experiments and enhance result quality by allowing consistent data handling 5. 2DB: a Proteomics database for storage, analysis, presentation, and retrieval of information from mass spectrometric experiments Directory of Open Access Journals (Sweden) Hippler Michael 2008-07-01 Full Text Available Abstract Background The amount of information stemming from proteomics experiments involving (multi dimensional separation techniques, mass spectrometric analysis, and computational analysis is ever-increasing. Data from such an experimental workflow needs to be captured, related and analyzed. Biological experiments within this scope produce heterogenic data ranging from pictures of one or two-dimensional protein maps and spectra recorded by tandem mass spectrometry to text-based identifications made by algorithms which analyze these spectra. Additionally, peptide and corresponding protein information needs to be displayed. Results In order to handle the large amount of data from computational processing of mass spectrometric experiments, automatic import scripts are available and the necessity for manual input to the database has been minimized. Information is in a generic format which abstracts from specific software tools typically used in such an experimental workflow. The software is therefore capable of storing and cross analysing results from many algorithms. A novel feature and a focus of this database is to facilitate protein identification by using peptides identified from mass spectrometry and link this information directly to respective protein maps. Additionally, our application employs spectral counting for quantitative presentation of the data. All information can be linked to hot spots on images to place the results into an experimental context. A summary of identified proteins, containing all relevant information per hot spot, is automatically generated, usually upon either a change in the underlying protein models or due to newly imported identifications. The supporting information for this report can be accessed in multiple ways using the user interface provided by the application. Conclusion We present a proteomics database which aims to greatly reduce evaluation time of results from mass spectrometric experiments and enhance 6. Observation of a new boson at a mass of 125 GeV with the CMS experiment at the LHC CERN Document Server Chatrchyan, Serguei; Sirunyan, Albert M; Tumasyan, Armen; Adam, Wolfgang; Aguilo, Ernest; Bergauer, Thomas; Dragicevic, Marko; Erö, Janos; Fabjan, Christian; Friedl, Markus; Fruehwirth, Rudolf; Ghete, Vasile Mihai; Hammer, Josef; Hoch, Michael; Hörmann, Natascha; Hrubec, Josef; Jeitler, Manfred; Kiesenhofer, Wolfgang; Knünz, Valentin; Krammer, Manfred; Krätschmer, Ilse; Liko, Dietrich; Majerotto, Walter; Mikulec, Ivan; Pernicka, Manfred; Rahbaran, Babak; Rohringer, Christine; Rohringer, Herbert; Schöfbeck, Robert; Strauss, Josef; Szoncsó, Fritz; Taurok, Anton; Waltenberger, Wolfgang; Walzel, Gerhard; Widl, Edmund; Wulz, Claudia-Elisabeth; Chekhovsky, Vladimir; Emeliantchik, Igor; Litomin, Aliaksandr; Makarenko, Vladimir; Mossolov, Vladimir; Shumeiko, Nikolai; Solin, Alexander; Stefanovitch, Roman; Suarez Gonzalez, Juan; Fedorov, Andrey; Korzhik, Mikhail; Missevitch, Oleg; Zuyeuski, Raman; Bansal, Monika; Bansal, Sunil; Beaumont, Willem; Cornelis, Tom; De Wolf, Eddi A; Druzhkin, Dmitry; Janssen, Xavier; Luyckx, Sten; Mucibello, Luca; Ochesanu, Silvia; Roland, Benoit; Rougny, Romain; Selvaggi, Michele; Staykova, Zlatka; Van Haevermaet, Hans; Van Mechelen, Pierre; Van Remortel, Nick; Van Spilbeeck, Alex; Blekman, Freya; Blyweert, Stijn; D'Hondt, Jorgen; Devroede, Olivier; Gonzalez Suarez, Rebeca; Goorens, Robert; Kalogeropoulos, Alexis; Maes, Michael; Olbrechts, Annik; Tavernier, Stefaan; Van Doninck, Walter; Van Lancker, Luc; Van Mulders, Petra; Van Onsem, Gerrit Patrick; Villella, Ilaria; Clerbaux, Barbara; De Lentdecker, Gilles; Dero, Vincent; Dewulf, Jean-Paul; Favart, Laurent; Gay, Arnaud; Hreus, Tomas; Léonard, Alexandre; Marage, Pierre Edouard; Mohammadi, Abdollah; Reis, Thomas; Rugovac, Shkelzen; Thomas, Laurent; Vander Velde, Catherine; Vanlaer, Pascal; Wang, Jian; Wickens, John; Adler, Volker; Beernaert, Kelly; Cimmino, Anna; Costantini, Silvia; Garcia, Guillaume; Grunewald, Martin; Klein, Benjamin; Lellouch, Jérémie; Marinov, Andrey; Mccartin, Joseph; Ocampo Rios, Alberto Andres; Ryckbosch, Dirk; Strobbe, Nadja; Thyssen, Filip; Tytgat, Michael; Verwilligen, Piet; Walsh, Sinead; Yazgan, Efe; Zaganidis, Nicolas; Basegmez, Suzan; Bruno, Giacomo; Castello, Roberto; Ceard, Ludivine; De Favereau De Jeneret, Jerome; Delaere, Christophe; Demin, Pavel; Du Pree, Tristan; Favart, Denis; Forthomme, Laurent; Giammanco, Andrea; Grégoire, Ghislain; Hollar, Jonathan; Lemaitre, Vincent; Liao, Junhui; Militaru, Otilia; Nuttens, Claude; Pagano, Davide; Pin, Arnaud; Piotrzkowski, Krzysztof; Schul, Nicolas; Vizan Garcia, Jesus Manuel; Beliy, Nikita; Caebergs, Thierry; Daubie, Evelyne; Hammad, Gregory Habib; Alves, Gilvan; Brito, Lucas; Correa Martins Junior, Marcos; De Jesus Damiao, Dilson; Martins, Thiago; Pol, Maria Elena; Henrique Gomes E Souza, Moacyr; Aldá Júnior, Walter Luiz; Carvalho, Wagner; Custódio, Analu; Melo Da Costa, Eliza; De Oliveira Martins, Carley; Fonseca De Souza, Sandro; Matos Figueiredo, Diego; Mundim, Luiz; Nogima, Helio; Oguri, Vitor; Prado Da Silva, Wanda Lucia; Santoro, Alberto; Sznajder, Andre; Souza Dos Anjos, Tiago; Bernardes, Cesar Augusto; De Almeida Dias, Flavia; Tomei, Thiago; De Moraes Gregores, Eduardo; Iope, Rogerio Luiz; Lagana, Caio; Lietti, Sergio M; Da Cunha Marinho, Franciole; Mercadante, Pedro G; Novaes, Sergio F; Padula, Sandra; Dimitrov, Lubomir; Genchev, Vladimir; Iaydjiev, Plamen; Piperov, Stefan; Rodozov, Mircho; Stoykova, Stefka; Sultanov, Georgi; Tcholakov, Vanio; Trayanov, Rumen; Vankov, Ivan; Vutova, Mariana; Roumenin, Chavdar; Uzunova, Daniela; Zahariev, Roman; Dimitrov, Anton; Hadjiiska, Roumyana; Kozhuharov, Venelin; Litov, Leander; Pavlov, Borislav; Petkov, Peicho; Bian, Jian-Guo; Chen, Guo-Ming; Chen, He-Sheng; He, Kang-Lin; Jiang, Chun-Hua; Li, Wei-Guo; Liang, Dong; Liang, Song; Meng, Xiangwei; Sun, Gongxing; Sun, Han-Sheng; Tao, Junquan; Wang, Jian; Wang, Xianyou; Wang, Zheng; Xiao, Hong; Xu, Ming; Yang, Min; Zang, Jingjing; Zhang, Xiaomei; Zhang, Zhen; Zhang, Zhenxia; Zhao, Wei-Ren; Zhu, Zian; Asawatangtrakuldee, Chayanit; Ban, Yong; Cai, Jianxin; Guo, Shuang; Guo, Yifei; Li, Wenbo; Liu, Hong-Tao; Liu, Shuai; Mao, Yajun; Qian, Si-Jin; Teng, Haiyun; Wang, Dayong; Ye, Yan-Lin; Zhang, Linlin; Zhu, Bo; Zou, Wei; Avila, Carlos; Gomez, Juan Pablo; Gomez Moreno, Bernardo; Osorio Oliveros, Andres Felipe; Sanabria, Juan Carlos; Godinovic, Nikola; Lelas, Damir; Plestina, Roko; Polic, Dunja; Puljak, Ivica; Antunovic, Zeljko; Kovac, Marko; Brigljevic, Vuko; Duric, Senka; Kadija, Kreso; Luetic, Jelena; Morovic, Srecko; Attikis, Alexandros; Galanti, Mario; Mavromanolakis, Georgios; Mousa, Jehad; Nicolaou, Charalambos; Ptochos, Fotios; Razis, Panos A; Finger, Miroslav; Finger Jr, Michael; Aly, Ayman; Assran, Yasser; Awad, Adel; Elgammal, Sherif; Ellithi Kamel, Ali; Khalil, Shaaban; Mahmoud, Mohammed; Mahrous, Ayman; Radi, Amr; Hektor, Andi; Kadastik, Mario; Kannike, Kristjan; Müntel, Mait; Murumaa, Marion; Raidal, Martti; Rebane, Liis; Strumia, Alessandro; Tiko, Andres; Eerola, Paula; Fedi, Giacomo; Voutilainen, Mikko; Anttila, Erkki; Härkönen, Jaakko; Heikkinen, Mika Aatos; Karimäki, Veikko; Katajisto, Harri Mikael; Kinnunen, Ritva; Kortelainen, Matti J; Kotamäki, Miikka; Lampén, Tapio; Lassila-Perini, Kati; Lehti, Sami; Lindén, Tomas; Luukka, Panja-Riina; Mäenpää, Teppo; Peltola, Timo; Tuominen, Eija; Tuominiemi, Jorma; Tuovinen, Esa; Ungaro, Donatella; Vanhala, Tommi Pekka; Wendland, Lauri; Banzuzi, Kukka; Karjalainen, Ahti; Korpela, Arja; Tuuva, Tuure; Anfreville, Marc; Besancon, Marc; Choudhury, Somnath; Dejardin, Marc; Denegri, Daniel; Fabbro, Bernard; Faure, Jean-Louis; Ferri, Federico; Ganjour, Serguei; Gentit, François-Xavier; Givernaud, Alain; Gras, Philippe; Hamel de Monchenault, Gautier; Jarry, Patrick; Kircher, François; Lemaire, Marie-Claude; Locci, Elizabeth; Malcles, Julie; Mandjavidze, Irakli; Nayak, Aruna; Pansart, Jean-Pierre; Rander, John; Reymond, Jean-Marc; Rosowsky, André; Shreyber, Irina; Titov, Maksym; Verrecchia, Patrice; Badier, Jean; Baffioni, Stephanie; Beaudette, Florian; Becheva, Emilia; Benhabib, Lamia; Bianchini, Lorenzo; Bluj, Michal; Broutin, Clementine; Busson, Philippe; Cerutti, Muriel; Chamont, David; Charlot, Claude; Daci, Nadir; Dahms, Torsten; Dalchenko, Mykhailo; Dobrzynski, Ludwik; Florent, Alice; Geerebaert, Yannick; Granier de Cassagnac, Raphael; Haguenauer, Maurice; Hennion, Pascale; Milleret, Gérard; Miné, Philippe; Mironov, Camelia; Naranjo, Ivo Nicolas; Nguyen, Matthew; Ochando, Christophe; Paganini, Pascal; Romanteau, Thierry; Sabes, David; Salerno, Roberto; Sartirana, Andrea; Sirois, Yves; Thiebaux, Christophe; Veelken, Christian; Zabi, Alexandre; Agram, Jean-Laurent; Andrea, Jeremy; Besson, Auguste; Bloch, Daniel; Bodin, David; Brom, Jean-Marie; Cardaci, Marco; Chabert, Eric Christian; Collard, Caroline; Conte, Eric; Drouhin, Frédéric; Ferro, Cristina; Fontaine, Jean-Charles; Gelé, Denis; Goerlach, Ulrich; Goetzmann, Christophe; Gross, Laurent; Huss, Daniel; Juillot, Pierre; Kieffer, Eric; Le Bihan, Anne-Catherine; Pansanel, Jérôme; Patois, Yannick; Van Hove, Pierre; Boutigny, Dominique; Mercier, Damien; Baulieu, Guillaume; Beauceron, Stephanie; Beaupere, Nicolas; Bedjidian, Marc; Bondu, Olivier; Boudoul, Gaelle; Brochet, Sébastien; Chasserat, Julien; Chierici, Roberto; Combaret, Christophe; Contardo, Didier; Depasse, Pierre; El Mamouni, Houmani; Fay, Jean; Gascon, Susan; Giraud, Noël; Gouzevitch, Maxime; Haroutunian, Roger; Ille, Bernard; Kurca, Tibor; Lethuillier, Morgan; Lumb, Nicholas; Mathez, Hervé; Mirabito, Laurent; Perries, Stephane; Sgandurra, Louis; Sordini, Viola; Tschudi, Yohann; Vander Donckt, Muriel; Verdier, Patrice; Viret, Sébastien; Roinishvili, Vladimir; Rurua, Lali; Amaglobeli, Nodar; Bagaturia, Iuri; Chiladze, Badri; Kvatadze, Ramazi; Lomidze, David; Shanidze, Revaz; Tsamalaidze, Zviad; Adolphi, Roman; Anagnostou, Georgios; Autermann, Christian; Beranek, Sarah; Brauer, Richard; Braunschweig, Wolfgang; Calpas, Betty; Edelhoff, Matthias; Feld, Lutz; Heracleous, Natalie; Hindrichs, Otto; Jussen, Ruediger; Karpinski, Waclaw; Klein, Katja; Lübelsmeyer, Klaus; Merz, Jennifer; Ostapchuk, Andrey; Pandoulas, Demetrios; Perieanu, Adrian; Raupach, Frank; Sammet, Jan; Schael, Stefan; Schmitz, Detlef; Schultz von Dratzig, Arndt; Siedling, Rolf; Sprenger, Daniel; Weber, Hendrik; Wittmer, Bruno; Wlochal, Michael; Zhukov, Valery; Ata, Metin; Biallass, Philipp; Caudron, Julien; Dietz-Laursonn, Erik; Duchardt, Deborah; Erdmann, Martin; Fischer, Robert; Güth, Andreas; Hebbeker, Thomas; Heidemann, Carsten; Hilgers, Guenter; Hoepfner, Kerstin; Hof, Carsten; Klimkovich, Tatsiana; Klingebiel, Dennis; Kreuzer, Peter; Magass, Carsten; Merschmeyer, Markus; Meyer, Arnd; Olschewski, Mark; Papacz, Paul; Philipps, Barthel; Pieta, Holger; Reithler, Hans; Schmitz, Stefan Antonius; Sonnenschein, Lars; Sowa, Michael; Steggemann, Jan; Teyssier, Daniel; Weber, Martin; Bontenackels, Michael; Cherepanov, Vladimir; Erdogan, Yusuf; Flügge, Günter; Geenen, Heiko; Geisler, Matthias; Haj Ahmad, Wael; Hoehle, Felix; Kargoll, Bastian; Kress, Thomas; Kuessel, Yvonne; Nowack, Andreas; Perchalla, Lars; Pooth, Oliver; Sauerland, Philip; Stahl, Achim; Zoeller, Marc Henning; Aldaya Martin, Maria; Behr, Joerg; Behrenhoff, Wolf; Behrens, Ulf; Bergholz, Matthias; Bethani, Agni; Borras, Kerstin; Burgmeier, Armin; Cakir, Altan; Calligaris, Luigi; Campbell, Alan; Castro, Elena; Costanza, Francesco; Dammann, Dirk; Diez Pardos, Carmen; Eckerlin, Guenter; Eckstein, Doris; Flossdorf, Alexander; Flucke, Gero; Geiser, Achim; Glushkov, Ivan; Goettlicher, Peter; Grebenyuk, Anastasia; Gunnellini, Paolo; Habib, Shiraz; Hauk, Johannes; Hellwig, Gregor; Jung, Hannes; Kasemann, Matthias; Katsas, Panagiotis; Kleinwort, Claus; Kluge, Hannelies; Knutsson, Albert; Krämer, Mira; Krücker, Dirk; Kuznetsova, Ekaterina; Lange, Wolfgang; Lewendel, Birgit; Lohmann, Wolfgang; Lutz, Benjamin; Mankel, Rainer; Marfin, Ihar; Marienfeld, Markus; Melzer-Pellmann, Isabell-Alissandra; Meyer, Andreas Bernhard; Mnich, Joachim; Muhl, Carsten; Mussgiller, Andreas; Naumann-Emme, Sebastian; Novgorodova, Olga; Olzem, Jan; Parenti, Andrea; Perrey, Hanno; Petrukhin, Alexey; Pitzl, Daniel; Raspereza, Alexei; Ribeiro Cipriano, Pedro M; Riedl, Caroline; Ron, Elias; Rosemann, Christoph; Rosin, Michele; Salfeld-Nebgen, Jakob; Schmidt, Ringo; Schoerner-Sadenius, Thomas; Sen, Niladri; Spiridonov, Alexander; Stein, Matthias; Tomaszewska, Justyna; Volyanskyy, Dmytro; Walsh, Roberval; Wissing, Christoph; Youngman, Christopher; Blobel, Volker; Draeger, Jula; Enderle, Holger; Erfle, Joachim; Gebbert, Ulla; Görner, Martin; Hermanns, Thomas; Höing, Rebekka Sophie; Kaschube, Kolja; Kaussen, Gordon; Kirschenmann, Henning; Klanner, Robert; Lange, Jörn; Mura, Benedikt; Nowak, Friederike; Peiffer, Thomas; Pietsch, Niklas; Rathjens, Denis; Sander, Christian; Schettler, Hannes; Schleper, Peter; Schlieckau, Eike; Schmidt, Alexander; Schröder, Matthias; Schum, Torben; Seidel, Markus; Sola, Valentina; Stadie, Hartmut; Steinbrück, Georg; Thomsen, Jan; Vanelderen, Lukas; Barth, Christian; Bauer, Julia; Berger, Joram; Blüm, Peter; Böser, Christian; Buege, Volker; Chen, Zheng-Yu; Chowdhury, Shantanu; Chwalek, Thorsten; Daeuwel, Daniel; De Boer, Wim; Descroix, Alexis; Dierlamm, Alexander; Dirkes, Guido; Fahrer, Manuel; Feindt, Michael; Felzmann, Ulrich; Frey, Martin; Furgeri, Alexander; Gebauer, Iris; Gessler, Andreas; Gruschke, Jasmin; Guthoff, Moritz; Hackstein, Christoph; Hartmann, Frank; Hauler, Florian; Hauth, Thomas; Heier, Stefan; Heindl, Stefan Michael; Heinrich, Michael; Heiss, Andreas; Held, Hauke; Hoffmann, Karl-Heinz; Honc, Simon; Husemann, Ulrich; Imhof, Markus; Jung, Christopher; Junghans, Sascha; Katkov, Igor; Kerzel, Ulrich; Knoblauch, Dieter; Komaragiri, Jyothsna Rani; Kräber, Michael; Kuhr, Thomas; Liamsuwan, Thiansin; Lobelle Pardo, Patricia; Martschei, Daniel; Menchikov, Alexandre; Mol, Xavier; Mörmann, Dirk; Mueller, Steffen; Müller, Thomas; Neuberger, Dirk; Neuland, Maike Brigitte; Niegel, Martin; Nürnberg, Andreas; Oberst, Oliver; Oehler, Andreas; Ortega Gomez, Tino; Ott, Jochen; Piasecki, Christian; Poschlad, Angela; Quast, Gunter; Rabbertz, Klaus; Ratnikov, Fedor; Ratnikova, Natalia; Renz, Manuel; Röcker, Steffen; Roederer, Frank; Sabellek, Andreas; Saout, Christophe; Scheurer, Armin; Schieferdecker, Dennis; Schieferdecker, Philipp; Schilling, Frank-Peter; Schmanau, Mike; Schott, Gregory; Schwerdtfeger, Wolfgang; Simonis, Hans-Jürgen; Skiba, Alexander; Stober, Fred-Markus Helmut; Theel, Andreas; Thümmel, Wolf Hagen; Troendle, Daniel; Trunov, Artem; Ulrich, Ralf; Wagner-Kuhr, Jeannine; Wayand, Stefan; Weber, Markus; Weiler, Thomas; Zeise, Manuel; Ziebarth, Eva Barbara; Zvada, Marian; Daskalakis, Georgios; Geralis, Theodoros; Kesisoglou, Stilianos; Kyriakis, Aristotelis; Loukas, Demetrios; Manolakos, Ioannis; Markou, Athanasios; Markou, Christos; Mavrommatis, Charalampos; Ntomari, Eleni; Gouskos, Loukas; Panagiotou, Apostolos; Saoulidou, Niki; Evangelou, Ioannis; Foudas, Costas; Kokkas, Panagiotis; Manthos, Nikolaos; Papadopoulos, Ioannis; Patras, Vaios; Bencze, Gyorgy; Hajdu, Csaba; Hidas, Pàl; Horvath, Dezso; Sikler, Ferenc; Veszpremi, Viktor; Vesztergombi, Gyorgy; Zalan, Peter; Beni, Noemi; Czellar, Sandor; Fenyvesi, Andras; Molnar, Jozsef; Palinkas, Jozsef; Szillasi, Zoltan; Karancsi, János; Raics, Peter; Trocsanyi, Zoltan Laszlo; Ujvari, Balazs; Zilizi, Gyula; Beri, Suman Bala; Bhandari, Virender; Bhatnagar, Vipin; Dhingra, Nitish; Gupta, Ruchi; Kaur, Manjit; Kohli, Jatinder Mohan; Mehta, Manuk Zubin; Nishu, Nishu; Saini, Lovedeep Kaur; Sharma, Archana; Singh, Jasbir; Kumar, Ashok; Kumar, Arun; Ahuja, Sudha; Bhardwaj, Ashutosh; Chatterji, Sudeep; Choudhary, Brajesh C; Gupta, Pooja; Kumar, Ajay; Malhotra, Shivali; Naimuddin, Md; Ranjan, Kirti; Sharma, Varun; Shivpuri, Ram Krishen; Banerjee, Sunanda; Bhattacharya, Satyaki; Dutta, Suchandra; Gomber, Bhawna; Jain, Sandhya; Jain, Shilpi; Khurana, Raman; Sarkar, Subir; Sharan, Manoj; Abdulsalam, Abdulla; Choudhury, Rajani Kant; Dutta, Dipanwita; Ghodgaonkar, Manohar; Kailas, Swaminathan; Kataria, Sushil Kumar; Kumar, Vineet; Mehta, Pourus; Mohanty, Ajit Kumar; Pant, Lalit Mohan; Shukla, Prashant; Topkar, Anita; Aziz, Tariq; Chatterjee, Rajdeep Mohan; Chendvankar, Sanjay; Deshpande, Pandurang Vishnu; Ganguli, Som N; Ganguly, Sanmay; Ghosh, Saranya; Guchait, Monoranjan; Gurtu, Atul; Kole, Gouranga; Maity, Manas; Mazumdar, Kajari; Mohanty, Gagan Bihari; Parida, Bibhuti; Patil, Mandakini Ravindra; Raghavan, R; Sudhakar, Katta; Wickramage, Nadeesha; Acharya, Bannaje Sripathi; Banerjee, Sudeshna; Bheesette, Satyanarayana; Dugad, Shashikant; Kalmani, Suresh Devendrappa; Krishnaswamy, Marthi Ramaswamy; Lakkireddi, Venkatesam Reddy; Mondal, Naba Kumar; Narasimham, Vemuri Syamala; Panyam, Nagaraj; Verma, Piyush; Ardalan, Farhad; Arfaei, Hessamaddin; Bakhshiansohi, Hamed; Etesami, Seyed Mohsen; Fahim, Ali; Hashemi, Majid; Jafari, Abideh; Khakzad, Mohsen; Mohammadi Najafabadi, Mojtaba; Paktinat Mehdiabadi, Saeid; Safarzadeh, Batool; Zeinali, Maryam; Abbrescia, Marcello; Barbone, Lucia; Calabria, Cesare; Chhibra, Simranjit Singh; Colaleo, Anna; Creanza, Donato; De Filippis, Nicola; De Palma, Mauro; De Robertis, Giuseppe; Donvito, Giacinto; Fiore, Luigi; Iaselli, Giuseppe; Loddo, Flavio; Lusito, Letizia; Maggi, Giorgio; Maggi, Marcello; Manna, Norman; Marangelli, Bartolomeo; My, Salvatore; Natali, Sergio; Nuzzo, Salvatore; Pacifico, Nicola; Pompili, Alexis; Pugliese, Gabriella; Ranieri, Antonio; Romano, Francesco; Selvaggi, Giovanna; Silvestris, Lucia; Singh, Gurpreet; Spinoso, Vincenzo; Venditti, Rosamaria; Zito, Giuseppe; Abbiendi, Giovanni; Benvenuti, Alberto; Bonacorsi, Daniele; Braibant-Giacomelli, Sylvie; Brigliadori, Luca; Capiluppi, Paolo; Castro, Andrea; Cavallo, Francesca Romana; Cuffiani, Marco; Dallavalle, Gaetano-Marco; Fabbri, Fabrizio; Fanfani, Alessandra; Fasanella, Daniele; Giacomelli, Paolo; Grandi, Claudio; Guiducci, Luigi; Marcellini, Stefano; Masetti, Gianni; Meneghelli, Marco; Montanari, Alessandro; Navarria, Francesco; Odorici, Fabrizio; Perrotta, Andrea; Primavera, Federica; Rossi, Antonio; Rovelli, Tiziano; Siroli, Gianni; Travaglini, Riccardo; Albergo, Sebastiano; Cappello, Gigi; Chiorboli, Massimiliano; Costa, Salvatore; Noto, Francesco; Potenza, Renato; Saizu, Mirela Angela; Tricomi, Alessia; Tuve, Cristina; Barbagli, Giuseppe; Ciulli, Vitaliano; Civinini, Carlo; D'Alessandro, Raffaello; Focardi, Ettore; Frosali, Simone; Gallo, Elisabetta; Genta, Chiara; Gonzi, Sandro; Meschini, Marco; Paoletti, Simone; Parrini, Giuliano; Ranieri, Riccardo; Sguazzoni, Giacomo; Tropiano, Antonio; Benussi, Luigi; Bianco, Stefano; Colafranceschi, Stefano; Fabbri, Franco; Piccolo, Davide; Fabbricatore, Pasquale; Farinon, Stefania; Greco, Michela; Musenich, Riccardo; Tosi, Silvano; Benaglia, Andrea; Carbone, Luca; D'Angelo, Pasqualino; De Guio, Federico; Di Matteo, Leonardo; Dini, Paolo; Farina, Fabio Mario; Fiorendi, Sara; Gennai, Simone; Ghezzi, Alessio; Malvezzi, Sandra; Manzoni, Riccardo Andrea; Martelli, Arabella; Massironi, Andrea; Menasce, Dario; Moroni, Luigi; Negri, Pietro; Paganoni, Marco; Pedrini, Daniele; Pullia, Antonino; Ragazzi, Stefano; Redaelli, Nicola; Sala, Silvano; Tabarelli de Fatis, Tommaso; Buontempo, Salvatore; Carrillo Montoya, Camilo Andres; Cavallo, Nicola; De Cosa, Annapaola; Dogangun, Oktay; Fabozzi, Francesco; Iorio, Alberto Orso Maria; Lista, Luca; Meola, Sabino; Merola, Mario; Paolucci, Pierluigi; Azzi, Patrizia; Bacchetta, Nicola; Bellato, Marco; Benettoni, Massimo; Biasotto, Massimo; Bisello, Dario; Branca, Antonio; Carlin, Roberto; Checchia, Paolo; Dorigo, Tommaso; Dosselli, Umberto; Gasparini, Fabrizio; Gasparini, Ugo; Giubilato, Piero; Gonella, Franco; Gozzelino, Andrea; Gulmini, Michele; Kanishchev, Konstantin; Lacaprara, Stefano; Lazzizzera, Ignazio; Loreti, Maurizio; Margoni, Martino; Maron, Gaetano; Mazzucato, Mirco; Meneguzzo, Anna Teresa; Montecassiano, Fabio; Passaseo, Marina; Pazzini, Jacopo; Pegoraro, Matteo; Pozzobon, Nicola; Ronchese, Paolo; Simonetto, Franco; Torassa, Ezio; Tosi, Mia; Vanini, Sara; Ventura, Sandro; Zotto, Pierluigi; Zumerle, Gianni; Berzano, Umberto; Gabusi, Michele; Ratti, Sergio P; Riccardi, Cristina; Torre, Paola; Vitulo, Paolo; Biasini, Maurizio; Bilei, Gian Mario; Fanò, Livio; Lariccia, Paolo; Lucaroni, Andrea; Mantovani, Giancarlo; Menichelli, Mauro; Nappi, Aniello; Passeri, Daniele; Placidi, Pisana; Romeo, Francesco; Saha, Anirban; Santocchia, Attilio; Servoli, Leonello; Spiezia, Aniello; Taroni, Silvia; Valdata, Marisa; Angelini, Franco; Arezzini, Silvia; Azzurri, Paolo; Bagliesi, Giuseppe; Basti, Andrea; Bellazzini, Ronaldo; Bernardini, Jacopo; Boccali, Tommaso; Bosi, Filippo; Brez, Alessandro; Broccolo, Giuseppe; Calzolari, Federico; Carboni, Andrea; Castaldi, Rino; Cerri, Claudio; Ciampa, Alberto; D'Agnolo, Raffaele Tito; Dell'Orso, Roberto; Fiori, Francesco; Foà, Lorenzo; Giassi, Alessandro; Giusti, Simone; Kraan, Aafke; Latronico, Luca; Ligabue, Franco; Linari, Stefano; Lomtadze, Teimuraz; Martini, Luca; Massa, Maurizio; Massai, Marco Maria; Mazzoni, Enrico; Messineo, Alberto; Moggi, Andrea; Palla, Fabrizio; Raffaelli, Fabrizio; Rizzi, Andrea; Sanguinetti, Giulio; Segneri, Gabriele; Serban, Alin Titus; Spagnolo, Paolo; Spandre, Gloria; Squillacioti, Paola; Tenchini, Roberto; Tonelli, Guido; Venturi, Andrea; Verdini, Piero Giorgio; Baccaro, Stefania; Barone, Luciano; Bartoloni, Alessandro; Cavallari, Francesca; Dafinei, Ioan; Del Re, Daniele; Diemoz, Marcella; Fanelli, Cristiano; Grassi, Marco; Longo, Egidio; Meridiani, Paolo; Micheli, Francesco; Nourbakhsh, Shervin; Organtini, Giovanni; Paramatti, Riccardo; Rahatlou, Shahram; Sigamani, Michael; Soffi, Livia; Talamo, Ivano Giuseppe; Amapane, Nicola; Arcidiacono, Roberta; Argiro, Stefano; Arneodo, Michele; Biino, Cristina; Cartiglia, Nicolo; Costa, Marco; Demaria, Natale; Mariotti, Chiara; Maselli, Silvia; Migliore, Ernesto; Monaco, Vincenzo; Musich, Marco; Obertino, Maria Margherita; Pastrone, Nadia; Pelliccioni, Mario; Peroni, Cristiana; Potenza, Alberto; Romero, Alessandra; Ruspa, Marta; Sacchi, Roberto; Solano, Ada; Staiano, Amedeo; Vilela Pereira, Antonio; Ambroglini, Filippo; Belforte, Stefano; Candelise, Vieri; Casarsa, Massimo; Cossutti, Fabio; Della Ricca, Giuseppe; Gobbo, Benigno; Kavka, Carlos; Marone, Matteo; Montanino, Damiana; Penzo, Aldo; Schizzi, Andrea; Kim, Tae Yeon; Nam, Soon-Kwon; Butanov, Khakimjan; Chang, Sunghyun; Chung, Jin Hyuk; Ham, Seung Woo; Han, Daehee; Kang, Juheon; Kim, Dong Hee; Kim, Gui Nyun; Kim, Ji Eun; Kim, Kyung Sook; Kong, Dae Jung; Lee, Man Woo; Lee, Sangeun; Oh, Young Do; Park, Hyangkyu; Ro, Sang-Ryul; Son, Dohhee; Son, Dong-Chul; Suh, Jun Suhk; Yang, Yu Chul; Yusupov, Hammid; Kim, Jae Yool; Kim, Zero Jaeho; Song, Sanghyeon; Choi, Suyong; Gyun, Dooyeon; Hong, Byung-Sik; Jo, Mihee; Jo, Youngkwon; Kang, Minho; Kim, Hyunchul; Kim, Tae Jeong; Lee, Kyong Sei; Moon, Dong Ho; Park, Sung Keun; Sim, Kwang Souk; Choi, Minkyoo; Hahn, Garam; Kang, Seokon; Kim, Hyunyong; Kim, Ji Hyun; Park, Chawon; Park, Inkyu; Park, Sangnam; Ryu, Geonmo; Choi, Young-Il; Choi, Young Kyu; Goh, Junghwan; Kim, Min Suk; Kwon, Eunhyang; Lee, Byounghoon; Lee, Jongseok; Lee, Sungeun; Seo, Hyunkwan; Yu, Intae; Janulis, Mindaugas; Juodagalvis, Andrius; Naujikas, Rolandas; Castilla-Valdez, Heriberto; De La Cruz-Burelo, Eduard; Heredia-de La Cruz, Ivan; Lopez-Fernandez, Ricardo; Magaña Villalba, Ricardo; Martínez-Ortega, Jorge; Sánchez-Hernández, Alberto; Villasenor-Cendejas, Luis Manuel; Carrillo Moreno, Salvador; Vazquez Valencia, Fabiola; Salazar Ibarguen, Humberto Antonio; Casimiro Linares, Edgar; Morelos Pineda, Antonio; Reyes-Santos, Marco A; Allfrey, Philip; Krofcheck, David; Bell, Alan James; Bernardino Rodrigues, Nuno; Butler, Anthony Philip Howard; Butler, Philip H; Doesburg, Robert; Pfeiffer, Dorothea; Reucroft, Steve; Silverwood, Hamish; Williams, Jennifer C; Ahmad, Muhammad; Ansari, Muhammad Hamid; Asghar, Muhammad Irfan; Hoorani, Hafeez R; Khalid, Shoaib; Khan, Wajid Ali; Khurshid, Taimoor; Qazi, Shamona; Shah, Mehar Ali; Shoaib, Muhammad; Brona, Grzegorz; Bunkowski, Karol; Cwiok, Mikolaj; Czyrkowski, Henryk; Dabrowski, Ryszard; Dominik, Wojciech; Doroba, Krzysztof; Kalinowski, Artur; Konecki, Marcin; Krolikowski, Jan; Oklinski, Wojciech; Pozniak, Krzysztof; Zabolotny, Wojciech; Zych, Pawel; Bialkowska, Helena; Boimska, Bozena; Frueboes, Tomasz; Gokieli, Ryszard; Goscilo, Lukasz; Górski, Maciej; Kazana, Malgorzata; Kudla, Ignacy Maciek; Nawrocki, Krzysztof; Romanowska-Rybinska, Katarzyna; Szleper, Michal; Wrochna, Grzegorz; Zalewski, Piotr; Kasprowicz, Grzegorz; Romaniuk, Ryszard; Alemany-Fernandez, Reyes; Almeida, Nuno; Bargassa, Pedrame; David Tinoco Mendes, Andre; Faccioli, Pietro; Ferreira Parracho, Pedro Guilherme; Gallinaro, Michele; Ribeiro, Pedro Quinaz; Seixas, Joao; Rasteiro Da Silva, Jose; Varela, Joao; Vischia, Pietro; Afanasiev, Serguei; Belotelov, Ivan; Bunin, Pavel; Ershov, Yuri; Gavrilenko, Mikhail; Golunov, Alexander; Golutvin, Igor; Gorbounov, Nikolai; Gorbunov, Ilya; Gramenitski, Igor; Kalagin, Vladimir; Kamenev, Alexey; Karjavin, Vladimir; Konoplyanikov, Viktor; Korenkov, Vladimir; Kozlov, Guennady; Kurenkov, Alexander; Lanev, Alexander; Makankin, Alexander; Malakhov, Alexander; Melnitchenko, Igor; Mitsyn, Valeri Valentinovitch; Moisenz, Petr; Oleynik, Danila; Orlov, Alexandre; Palichik, Vladimir; Perelygin, Victor; Petrosyan, Artem; Savina, Maria; Semenov, Roman; Shmatov, Sergey; Shulha, Siarhei; Skachkova, Anna; Skatchkov, Nikolai; Smetannikov, Vladimir; Smirnov, Vitaly; Smolin, Dmitry; Tikhonenko, Elena; Vasilyev, Sergey; Volodko, Anton; Zarubin, Anatoli; Zhiltsov, Victor; Evstyukhin, Sergey; Golovtsov, Victor; Ivanov, Yury; Kim, Victor; Levchenko, Petr; Murzin, Victor; Oreshkin, Vadim; Smirnov, Igor; Sulimov, Valentin; Uvarov, Lev; Vavilov, Sergey; Vorobyev, Alexey; Vorobyev, Andrey; Andreev, Yuri; Anisimov, Alexander; Dermenev, Alexander; Gninenko, Sergei; Golubev, Nikolai; Gorbunov, Dmitry; Karneyeu, Anton; Kirsanov, Mikhail; Krasnikov, Nikolai; Matveev, Viktor; Pashenkov, Anatoli; Pivovarov, Grigory; Postoev, Vladimir E; Rubakov, Valery; Shirinyants, Valery; Solovey, Anatoly; Tlisov, Danila; Toropin, Alexander; Troitsky, Sergey; Epshteyn, Vladimir; Erofeeva, Maria; Gavrilov, Vladimir; Kaftanov, Vitali; Kiselevich, Ivan; Kolosov, Victor; Konoplyannikov, Anatoly; Kossov, Mikhail; Kozlov, Yury; Krokhotin, Andrey; Litvintsev, Dmitri; Lychkovskaya, Natalia; Oulianov, Alexei; Popov, Vladimir; Safronov, Grigory; Semenov, Sergey; Stepanov, Nikita; Stolin, Viatcheslav; Vlasov, Evgueni; Zaytsev, Valentin; Zhokin, Alexander; Belyaev, Andrey; Boos, Edouard; Bunichev, Viacheslav; Demiyanov, Andrey; Dubinin, Mikhail; Dudko, Lev; Ershov, Alexander; Gribushin, Andrey; Ilyin, Viacheslav; Kaminskiy, Alexandre; Klyukhin, Vyacheslav; Kodolova, Olga; Korotkikh, Vladimir; Kryukov, Alexander; Lokhtin, Igor; Markina, Anastasia; Obraztsov, Stepan; Perfilov, Maxim; Petrushanko, Sergey; Popov, Andrey; Proskuryakov, Alexander; Sarycheva, Ludmila; Savrin, Viktor; Snigirev, Alexander; Vardanyan, Irina; Andreev, Vladimir; Azarkin, Maksim; Dremin, Igor; Kirakosyan, Martin; Leonidov, Andrey; Mesyats, Gennady; Rusakov, Sergey V; Vinogradov, Alexey; Azhgirey, Igor; Bayshev, Igor; Bitioukov, Sergei; Grishin, Viatcheslav; Kachanov, Vassili; Kalinin, Alexey; Konstantinov, Dmitri; Korablev, Andrey; Krychkine, Victor; Levine, Andrey; Petrov, Vladimir; Ryabov, Alexander; Ryutin, Roman; Sobol, Andrei; Talov, Vladimir; Tourtchanovitch, Leonid; Troshin, Sergey; Tyurin, Nikolay; Uzunian, Andrey; Volkov, Alexey; Adzic, Petar; Djordjevic, Milos; Ekmedzic, Marko; Krpic, Dragomir; Milosevic, Jovan; Smiljkovic, Nebojsa; Zupan, Marko; Aguilar-Benitez, Manuel; Alcaraz Maestre, Juan; Arce, Pedro; Battilana, Carlo; Calvo, Enrique; Cerrada, Marcos; Chamizo Llatas, Maria; Colino, Nicanor; De La Cruz, Begona; Delgado Peris, Antonio; Domínguez Vázquez, Daniel; Fernandez Bedoya, Cristina; Fernández Ramos, Juan Pablo; Ferrando, Antonio; Flix, Jose; Fouz, Maria Cruz; Garcia-Abia, Pablo; Gonzalez Lopez, Oscar; Goy Lopez, Silvia; Hernandez, Jose M; Josa, Maria Isabel; Merino, Gonzalo; Puerta Pelayo, Jesus; Quintario Olmeda, Adrián; Redondo, Ignacio; Romero, Luciano; Santaolalla, Javier; Senghi Soares, Mara; Willmott, Carlos; Albajar, Carmen; Codispoti, Giuseppe; de Trocóniz, Jorge F; Brun, Hugues; Cuevas, Javier; Fernandez Menendez, Javier; Folgueras, Santiago; Gonzalez Caballero, Isidro; Lloret Iglesias, Lara; Piedra Gomez, Jonatan; Brochero Cifuentes, Javier Andres; Cabrillo, Iban Jose; Calderon, Alicia; Chuang, Shan-Huei; Duarte Campderros, Jordi; Felcini, Marta; Fernandez, Marcos; Gomez, Gervasio; Gonzalez Sanchez, Javier; Graziano, Alberto; Jorda, Clara; Lopez Virto, Amparo; Marco, Jesus; Marco, Rafael; Martinez Rivero, Celso; Matorras, Francisco; Munoz Sanchez, Francisca Javiela; Rodrigo, Teresa; Rodríguez-Marrero, Ana Yaiza; Ruiz-Jimeno, Alberto; Scodellaro, Luca; Sobron Sanudo, Mar; Vila, Ivan; Vilar Cortabitarte, Rocio; Abbaneo, Duccio; Aspell, Paul; Auffray, Etiennette; Auzinger, Georg; Bachtis, Michail; Baechler, Joachim; Baillon, Paul; Ball, Austin; Barney, David; Benitez, Jose F; Bernet, Colin; Bialas, Wojciech; Bianchi, Giovanni; Bloch, Philippe; Bocci, Andrea; Bonato, Alessio; Botta, Cristina; Breuker, Horst; Campi, Domenico; Camporesi, Tiziano; Cano, Eric; Cerminara, Gianluca; Charkiewicz, Andrzej; Christiansen, Tim; Coarasa Perez, Jose Antonio; Curé, Benoît; D'Enterria, David; Dabrowski, Anne; Daguin, Jerome; De Roeck, Albert; Di Guida, Salvatore; Dobson, Marc; Dupont-Sagorin, Niels; Elliott-Peisert, Anna; Eppard, Michael; Frisch, Benjamin; Funk, Wolfgang; Gaddi, Andrea; Gastal, Martin; Georgiou, Georgios; Gerwig, Hubert; Giffels, Manuel; Gigi, Dominique; Gill, Karl; Giordano, Domenico; Girone, Maria; Giunta, Marina; Glege, Frank; Gomez-Reino Garrido, Robert; Govoni, Pietro; Gowdy, Stephen; Guida, Roberto; Gutleber, Johannes; Hansen, Magnus; Harris, Philip; Hartl, Christian; Harvey, John; Hegner, Benedikt; Hinzmann, Andreas; Honma, Alan; Innocente, Vincenzo; Janot, Patrick; Kaadze, Ketino; Karavakis, Edward; Kloukinas, Kostas; Kousouris, Konstantinos; Lecoq, Paul; Lee, Yen-Jie; Lenzi, Piergiulio; Loos, Robert; Lourenco, Carlos; Magini, Nicolo; Maki, Tuula; Malberti, Martina; Malgeri, Luca; Mannelli, Marcello; Marchioro, Alessandro; Marques Pinho Noite, João; Masetti, Lorenzo; Meijers, Frans; Mersi, Stefano; Meschi, Emilio; Moneta, Lorenzo; Mozer, Matthias Ulrich; Mulders, Martijn; Musella, Pasquale; Onnela, Antti; Orimoto, Toyoko; Orsini, Luciano; Osborne, John Andrew; Palencia Cortezon, Enrique; Perez, Emmanuelle; Perrozzi, Luca; Petagna, Paolo; Petrilli, Achille; Petrucci, Andrea; Pfeiffer, Andreas; Pierini, Maurizio; Pimiä, Martti; Piparo, Danilo; Polese, Giovanni; Postema, Hans; Quertenmont, Loic; Racz, Attila; Reece, William; Ricci, Daniel; Rodrigues Antunes, Joao; Rolandi, Gigi; Rovelli, Chiara; Rovere, Marco; Ryjov, Vladimir; Sakulin, Hannes; Samyn, Dirk; Santanastasio, Francesco; Schäfer, Christoph; Schwick, Christoph; Sciaba, Andrea; Segoni, Ilaria; Sekmen, Sezen; Sharma, Archana; Siegrist, Patrice; Silva, Pedro; Simon, Michal; Sphicas, Paraskevas; Spiga, Daniele; Taylor, Bruce G; Tropea, Paola; Troska, Jan; Tsirou, Andromachi; Vasey, François; Veillet, Lucien; Veres, Gabor Istvan; Vichoudis, Paschalis; Vlimant, Jean-Roch; Wertelaers, Piet; Wöhri, Hermine Katharina; Worm, Steven; Zeuner, Wolfram Dietrich; Bertl, Willi; Deiters, Konrad; Erdmann, Wolfram; Feichtinger, Derek; Gabathuler, Kurt; Horisberger, Roland; Ingram, Quentin; Kaestli, Hans-Christian; König, Stefan; Kotlinski, Danek; Langenegger, Urs; Meier, Beat; Meier, Frank; Renker, Dieter; Rohe, Tilman; Sakhelashvili, Tariel; Sibille, Jennifer; Bäni, Lukas; Behner, Frank; Betev, Botjo; Blau, Bertrand; Bortignon, Pierluigi; Buchmann, Marco-Andrea; Casal, Bruno; Chanon, Nicolas; Chen, Zhiling; Da Silva Di Calafiori, Diogo Raphael; Dambach, Sarah; Davatz, Giovanna; Deisher, Amanda; Dissertori, Günther; Dittmar, Michael; Djambazov, Lubomir; Donegà, Mauro; Dünser, Marc; Eggel, Christina; Eugster, Jürg; Faber, Gérard; Freudenreich, Klaus; Grab, Christoph; Hintz, Wieland; Hits, Dmitry; Hofer, Hans; Holme, Oliver; Horvath, Istvan; Lecomte, Pierre; Lustermann, Werner; Marchica, Carmelo; Marini, Andrea Carlo; Martinez Ruiz del Arbol, Pablo; Mohr, Niklas; Moortgat, Filip; Nägeli, Christoph; Nef, Pascal; Nessi-Tedaldi, Francesca; Pandolfi, Francesco; Pape, Luc; Pauss, Felicitas; Peruzzi, Marco; Punz, Thomas; Ronga, Frederic Jean; Röser, Ulf; Rossini, Marco; Sala, Leonardo; Sanchez, Ann - Karin; Sawley, Marie-Christine; Schinzel, Dietrich; Starodumov, Andrei; Stieger, Benjamin; Suter, Henry; Takahashi, Maiko; Tauscher, Ludwig; Thea, Alessandro; Theofilatos, Konstantinos; Treille, Daniel; Trüb, Peter; Udriot, Stève; Urscheler, Christina; Viertel, Gert; von Gunten, Hans Peter; Wallny, Rainer; Weber, Hannsjoerg Artur; Wehrli, Lukas; Weng, Joanna; Zelepoukine, Serguei; Amsler, Claude; Chiochia, Vincenzo; De Visscher, Simon; Favaro, Carlotta; Ivova Rikova, Mirena; Millan Mejias, Barbara; Otiougova, Polina; Robmann, Peter; Snoek, Hella; Tupputi, Salvatore; Verzetti, Mauro; Chang, Yuan-Hann; Chen, Kuan-Hsin; Chen, Wan-Ting; Go, Apollo; Kuo, Chia-Ming; Li, Syue-Wei; Lin, Willis; Liu, Ming-Hsiung; Liu, Zong-Kai; Lu, Yun-Ju; Mekterovic, Darko; Singh, Anil; Volpe, Roberta; Wu, Jing-Han; Yu, Shin-Shan; Bartalini, Paolo; Chang, Paoti; Chang, You-Hao; Chang, Yu-Wei; Chao, Yuan; Chen, Kai-Feng; Dietz, Charles; Gao, Zhengwei; Grundler, Ulysses; Hou, George Wei-Shu; Hsiung, Yee; Kao, Kai-Yi; Lei, Yeong-Jyi; Liau, Jiun-jie; Lin, Sheng-Wen; Lu, Rong-Shyang; Majumder, Devdatta; Petrakou, Eleni; Shi, Xin; Shiu, Jing-Ge; Tzeng, Yeng-Ming; Ueno, Koji; Velikzhanin, Yury; Wan, Xia; Wang, Chin-chi; Wang, Minzu; Wei, Jui-Te; Yeh, Ping; Asavapibhop, Burin; Simili, Emanuele; Srimanobhas, Norraphat; Adiguzel, Aytul; Bakirci, Mustafa Numan; Cerci, Salim; Dozen, Candan; Dumanoglu, Isa; Eskut, Eda; Girgis, Semiray; Gokbulut, Gul; Gurpinar, Emine; Hos, Ilknur; Kangal, Evrim Ersin; Karaman, Turker; Karapinar, Guler; Kayis Topaksu, Aysel; Onengut, Gulsen; Ozdemir, Kadri; Ozturk, Sertac; Polatoz, Ayse; Sogut, Kenan; Sunar Cerci, Deniz; Tali, Bayram; Topakli, Huseyin; Vergili, Latife Nukhet; Vergili, Mehmet; Akin, Ilina Vasileva; Aliev, Takhmasib; Bilin, Bugra; Deniz, Muhammed; Gamsizkan, Halil; Guler, Ali Murat; Ocalan, Kadir; Ozpineci, Altug; Serin, Meltem; Sever, Ramazan; Surat, Ugur Emrah; Zeyrek, Mehmet; Deliomeroglu, Mehmet; Gülmez, Erhan; Isildak, Bora; Kaya, Mithat; Kaya, Ozlem; Ozkorucuklu, Suat; Sonmez, Nasuf; Cankocak, Kerem; Grynyov, Boris; Levchuk, Leonid; Lukyanenko, Sergiy; Soroka, Dmytro; Sorokin, Pavel; Ahmad, Mian Khawar Hasham; Branson, Andrew; McClatchey, Richard; Odeh, Mohammed; Shamdasani, Jetendr; Soomro, Kamran; Barrass, Timothy; Bostock, Francis; Brooke, James John; Clement, Emyr; Cussans, David; Flacher, Henning; Frazier, Robert; Goldstein, Joel; Grimes, Mark; Heath, Greg P; Heath, Helen F; Kreczko, Lukasz; Lacesso, Winnie; Metson, Simon; Newbold, Dave M; Nirunpong, Kachanon; Poll, Anthony; Senkin, Sergey; Smith, Vincent J; Williams, Thomas; Basso, Lorenzo; Bateman, Eddie; Bell, Ken W; Belyaev, Alexander; Brew, Christopher; Brown, Robert M; Camanzi, Barbara; Cockerill, David JA; Connolly, John F; Coughlan, John A; Denton, Len G; Flower, Paul S; French, Marcus; Greenhalgh, Justin; Halsall, Robert; Harder, Kristian; Harper, Sam; Hill, John; Jackson, James; Kennedy, Bruce W; Lintern, Laurence Albert; Lodge, Anthony B; Olaiya, Emmanuel; Pearson, Matthew; Petyt, David; Radburn-Smith, Benjamin Charles; Shepherd-Themistocleous, Claire; Smith, Brian; Sproston, Martin; Stephenson, Richard; Tomalin, Ian R; Torbet, Martin; Williams, Julian; Womersley, William John; Bainbridge, Robert; Ball, Gordon; Ballin, Jamie; Bauer, Daniela; Beuselinck, Raymond; Buchmuller, Oliver; Colling, David; Cripps, Nicholas; Cutajar, Michael; Dauncey, Paul; Davies, Gavin; Della Negra, Michel; Fayer, Simon; Ferguson, William; Fulcher, Jonathan; Futyan, David; Gilbert, Andrew; Guneratne Bryer, Arlo; Hall, Geoffrey; Hatherell, Zoe; Hays, Jonathan; Iles, Gregory; Jarvis, Martyn; Jones, John; Karapostoli, Georgia; Kenzie, Matthew; Leaver, James; Lyons, Louis; Magnan, Anne-Marie; Marrouche, Jad; Mathias, Bryn; Miller, Derek George; Nandi, Robin; Nash, Jordan; Nikitenko, Alexander; Noy, Matthew; Papageorgiou, Anastasios; Pela, Joao; Pesaresi, Mark; Petridis, Konstantinos; Pioppi, Michele; Rand, Duncan; Raymond, David Mark; Rogerson, Samuel; Rose, Andrew; Ryan, Matthew John; Seez, Christopher; Sharp, Peter; Sparrow, Alex; Stoye, Markus; Tapper, Alexander; Timlin, Claire; Tourneur, Stephane; Vazquez Acosta, Monica; Virdee, Tejinder; Wakefield, Stuart; Wardle, Nicholas; Whyntie, Tom; Wingham, Matthew; Zorba, Osman; Chadwick, Matthew; Cole, Joanne; Hobson, Peter R; Khan, Akram; Kyberd, Paul; Leggat, Duncan; Leslie, Dawn; Martin, William; Reid, Ivan; Symonds, Philip; Teodorescu, Liliana; Turner, Mark; Dittmann, Jay; Hatakeyama, Kenichi; Liu, Hongxuan; Scarborough, Tara; Charaf, Otman; Henderson, Conor; Rumerio, Paolo; Avetisyan, Aram; Bose, Tulika; Carrera Jarrin, Edgar; Fantasia, Cory; Hazen, Eric; Heister, Arno; St John, Jason; Lawson, Philip; Lazic, Dragoslav; Rohlf, James; Sperka, David; Sulak, Lawrence; Varela Rodriguez, Fernando; Wu, Shouxiang; Alimena, Juliette; Bhattacharya, Saptaparna; Cutts, David; Demiragli, Zeynep; Ferapontov, Alexey; Garabedian, Alex; Heintz, Ulrich; Hooper, Ryan; Jabeen, Shabnam; Kukartsev, Gennadiy; Laird, Edward; Landsberg, Greg; Luk, Michael; Narain, Meenakshi; Nguyen, Duong; Segala, Michael; Sinthuprasith, Tutanon; Speer, Thomas; Tsang, Ka Vang; Unalan, Zeynep; Breedon, Richard; Breto, Guillermo; Calderon De La Barca Sanchez, Manuel; Case, Michael; Chauhan, Sushil; Chertok, Maxwell; Conway, John; Conway, Rylan; Cox, Peter Timothy; Dolen, James; Erbacher, Robin; Gardner, Michael; Grim, Gary; Gunion, Jack; Holbrook, Britt; Ko, Winston; Kopecky, Alexandra; Lander, Richard; Lin, Feng-Cheng; Miceli, Tia; Murray, Patrick; Nikolic, Milan; Pellett, Dave; Ricci-tam, Francesca; Rowe, Jeff; Rutherford, Britney; Searle, Matthew; Smith, John; Squires, Michael; Tripathi, Mani; Vasquez Sierra, Ricardo; Yohay, Rachel; Andreev, Valeri; Arisaka, Katsushi; Cline, David; Cousins, Robert; Duris, Joseph; Erhan, Samim; Everaerts, Pieter; Farrell, Chris; Hauser, Jay; Ignatenko, Mikhail; Jarvis, Chad; Kubic, Jonathan; Otwinowski, Stanislaw; Plager, Charles; Rakness, Gregory; Schlein, Peter; Traczyk, Piotr; Valuev, Vyacheslav; Weber, Matthias; Yang, Xiaofeng; Zheng, Yangheng; Babb, John; Clare, Robert; Dinardo, Mauro Emanuele; Ellison, John Anthony; Gary, J William; Giordano, Ferdinando; Hanson, Gail; Jeng, Geng-yuan; Layter, John G; Liu, Hongliang; Long, Owen Rosser; Luthra, Arun; Nguyen, Harold; Paramesvaran, Sudarshan; Shen, Benjamin C; Sturdy, Jared; Sumowidagdo, Suharyo; Wilken, Rachel; Wimpenny, Stephen; Andrews, Warren; Branson, James G; Cerati, Giuseppe Benedetto; Cittolin, Sergio; Evans, David; Golf, Frank; Holzner, André; Kelley, Ryan; Lebourgeois, Matthew; Letts, James; Macneill, Ian; Mangano, Boris; Martin, Terrence; Mrak-Tadel, Alja; Padhi, Sanjay; Palmer, Christopher; Petrucciani, Giovanni; Pieri, Marco; Sani, Matteo; Sfiligoi, Igor; Sharma, Vivek; Simon, Sean; Sudano, Elizabeth; Tadel, Matevz; Tu, Yanjun; Vartak, Adish; Wasserbaech, Steven; Würthwein, Frank; Yagil, Avraham; Yoo, Jaehyeok; Barge, Derek; Bellan, Riccardo; Campagnari, Claudio; D'Alfonso, Mariarosaria; Danielson, Thomas; Flowers, Kristen; Geffert, Paul; Incandela, Joe; Justus, Christopher; Kalavase, Puneeth; Koay, Sue Ann; Kovalskyi, Dmytro; Krutelyov, Vyacheslav; Kyre, Susanne; Lowette, Steven; Magazzu, Guido; Mccoll, Nickolas; Pavlunin, Viktor; Rebassoo, Finn; Ribnik, Jacob; Richman, Jeffrey; Rossin, Roberto; Stuart, David; To, Wing; West, Christopher; White, Dean; Adamczyk, David; Apresyan, Artur; Barczyk, Artur; Bornheim, Adolf; Bunn, Julian; Chen, Yi; Denis, Gregory; Di Marco, Emanuele; Duarte, Javier; Galvez, Philippe; Gataullin, Marat; Kcira, Dorian; Legrand, Iosif; Litvine, Vladimir; Ma, Yousi; Maxa, Zdenek; Mott, Alexander; Mughal, Azher; Nae, Dan; Newman, Harvey B; Ravot, Sylvain; Rogan, Christopher; Rozsa, Sandor Gyula; Shevchenko, Sergey; Shin, Kyoungha; Spiropulu, Maria; Steenberg, Conrad; Thomas, Michael; Timciuc, Vladlen; van Lingen, Frank; Veverka, Jan; Voicu, Bucoveanu-Ramiro; Wilkinson, Richard; Xie, Si; Yang, Yong; Zhang, Liyuan; Zhu, Kejun; Zhu, Ren-Yuan; Akgun, Bora; Azzolini, Virginia; Calamba, Aristotle; Carroll, Ryan; Ferguson, Thomas; Iiyama, Yutaro; Jang, Dong Wook; Jun, Soon Yung; Liu, Yueh-Feng; Paulini, Manfred; Russ, James; Terentyev, Nikolay; Vogel, Helmut; Vorobiev, Igor; Cumalat, John Perry; Drell, Brian Robert; Ford, William T; Gaz, Alessandro; Heyburn, Bernadette; Johnson, Douglas; Luiggi Lopez, Eduardo; Nauenberg, Uriel; Smith, James; Stenson, Kevin; Ulmer, Keith; Wagner, Stephen Robert; Zang, Shi-Lei; Agostino, Lorenzo; Alexander, James; Chatterjee, Avishek; Eggert, Nicholas; Gibbons, Lawrence Kent; Heltsley, Brian; Khukhunaishvili, Aleko; Kreis, Benjamin; Kuznetsov, Valentin; Mirman, Nathan; Nicolas Kaufman, Gala; Patterson, Juliet Ritchie; Riley, Daniel; Ryd, Anders; Salvati, Emmanuele; Stroiney, Steven; Sun, Werner; Teo, Wee Don; Thom, Julia; Thompson, Joshua; Tucker, Jordan; Vaughan, Jennifer; Weng, Yao; Winstrom, Lucas; Wittich, Peter; Winn, Dave; Abdullin, Salavat; Albert, Merina; Albrow, Michael; Anderson, Jacob; Apollinari, Giorgio; Atac, Muzaffer; Badgett, William; Bakken, Jon Alan; Baldin, Boris; Banicz, Karoly; Bauerdick, Lothar AT; Beretvas, Andrew; Berryhill, Jeffrey; Bhat, Pushpalatha C; Binkley, Morris; Borcherding, Frederick; Burkett, Kevin; Butler, Joel Nathan; Chetluru, Vasundhara; Cheung, Harry; Chlebana, Frank; Cihangir, Selcuk; Dagenhart, William; Derylo, Greg; Dumitrescu, Catalin; Dykstra, David; Eartly, David P; Elias, John E; Elvira, Victor Daniel; Eulisse, Giulio; Evans, David; Fagan, David; Fisk, Ian; Foulkes, Stephen; Freeman, Jim; Gaines, Irwin; Gao, Yanyan; Gartung, Patrick; Giacchetti, Lisa; Gottschalk, Erik; Green, Dan; Guo, Yuyi; Gutsche, Oliver; Hahn, Alan; Hanlon, Jim; Harris, Robert M; Hirschauer, James; Holzman, Burt; Hooberman, Benjamin; Howell, Joseph; Huang, Chih-hao; Hufnagel, Dirk; Jindariani, Sergo; Johnson, Marvin; Jones, Christopher Duncan; Joshi, Umesh; Juska, Evaldas; Kilminster, Benjamin; Klima, Boaz; Kunori, Shuichi; Kwan, Simon; Larson, Krista; Leonidopoulos, Christos; Linacre, Jacob; Lincoln, Don; Lipton, Ron; Lopez Perez, Juan Antonio; Los, Serguei; Lykken, Joseph; Maeshima, Kaori; Marraffino, John Michael; Maruyama, Sho; Mason, David; McBride, Patricia; McCauley, Thomas; Mishra, Kalanand; Moccia, Stefano; Mommsen, Remigius K; Mrenna, Stephen; Musienko, Yuri; Muzaffar, Shahzad; Newman-Holmes, Catherine; O'Dell, Vivian; Osborne, Ianna; Pivarski, James; Popescu, Sorina; Pordes, Ruth; Prokofyev, Oleg; Rapsevicius, Valdas; Ronzhin, Anatoly; Rossman, Paul; Ryu, Seangchan; Sexton-Kennedy, Elizabeth; Sharma, Seema; Shaw, Theresa; Smith, Richard P; Soha, Aron; Spalding, William J; Spiegel, Leonard; Tanenbaum, William; Taylor, Lucas; Thompson, Rich; Tiradani, Anthony; Tkaczyk, Slawek; Tran, Nhan Viet; Tuura, Lassi; Uplegger, Lorenzo; Vaandering, Eric Wayne; Vidal, Richard; Whitmore, Juliana; Wu, Weimin; Yang, Fan; Yarba, Julia; Yumiceva, Francisco; Yun, Jae Chul; Zimmerman, Thomas; Acosta, Darin; Avery, Paul; Barashko, Victor; Bourilkov, Dimitri; Chen, Mingshui; Cheng, Tongguang; Das, Souvik; De Gruttola, Michele; Di Giovanni, Gian Piero; Dobur, Didar; Dolinsky, Sergei; Drozdetskiy, Alexey; Field, Richard D; Fisher, Matthew; Fu, Yu; Furic, Ivan-Kresimir; Gartner, Joseph; Gorn, Lisa; Holmes, Daniel; Hugon, Justin; Kim, Bockjoo; Konigsberg, Jacobo; Korytov, Andrey; Kropivnitskaya, Anna; Kypreos, Theodore; Low, Jia Fu; Madorsky, Alexander; Matchev, Konstantin; Milenovic, Predrag; Mitselmakher, Guenakh; Muniz, Lana; Park, Myeonghun; Remington, Ronald; Rinkevicius, Aurelijus; Scurlock, Bobby; Skhirtladze, Nikoloz; Snowball, Matthew; Stasko, John; Yelton, John; Zakaria, Mohammed; Gaultney, Vanessa; Hewamanage, Samantha; Lebolo, Luis Miguel; Linn, Stephan; Markowitz, Pete; Martinez, German; Rodriguez, Jorge Luis; Adams, Todd; Askew, Andrew; Bertoldi, Maurizio; Bochenek, Joseph; Chen, Jie; Dharmaratna, Welathantri GD; Diamond, Brendan; Gleyzer, Sergei V; Haas, Jeff; Hagopian, Sharon; Hagopian, Vasken; Jenkins, Merrill; Johnson, Kurtis F; Prosper, Harrison; Tentindo, Silvia; Veeraraghavan, Venkatesh; Weinberg, Marc; Baarmand, Marc M; Dorney, Brian; Hohlmann, Marcus; Kalakhety, Himali; Ralich, Robert; Vodopiyanov, Igor; Adams, Mark Raymond; Anghel, Ioana Maria; Apanasevich, Leonard; Bai, Yuting; Bazterra, Victor Eduardo; Betts, Russell Richard; Bucinskaite, Inga; Callner, Jeremy; Cavanaugh, Richard; Chung, Man-Ho; Evdokimov, Olga; Garcia-Solis, Edmundo Javier; Gauthier, Lucie; Gerber, Cecilia Elena; Hofman, David Jonathan; Hollis, Richard; Iordanova, Aneta; Khalatyan, Samvel; Kunde, Gerd J; Lacroix, Florent; Malek, Magdalena; O'Brien, Christine; Silkworth, Christopher; Silvestre, Catherine; Smoron, Agata; Strom, Derek; Turner, Paul; Varelas, Nikos; Akgun, Ugur; Albayrak, Elif Asli; Ayan, Ahmet Sedat; Bilki, Burak; Clarida, Warren; Debbins, Paul; Duru, Firdevs; Ingram, F Duane; McCliment, Edward; Merlo, Jean-Pierre; Mermerkaya, Hamit; Mestvirishvili, Alexi; Miller, Michael Jan; Moeller, Anthony; Nachtman, Jane; Newsom, Charles Ray; Norbeck, Edwin; Olson, Jonathan; Onel, Yasar; Ozok, Ferhat; Schmidt, Ianos; Sen, Sercan; Tan, Ping; Tiras, Emrah; Wetzel, James; Yetkin, Taylan; Yi, Kai; Barnett, Bruce Arnold; Blumenfeld, Barry; Bolognesi, Sara; Fehling, David; Giurgiu, Gavril; Gritsan, Andrei; Guo, Zijin; Hu, Guofan; Maksimovic, Petar; Rappoccio, Salvatore; Swartz, Morris; Whitbeck, Andrew; Baringer, Philip; Bean, Alice; Benelli, Gabriele; Coppage, Don; Grachov, Oleg; Kenny Iii, Raymond Patrick; Murray, Michael; Noonan, Daniel; Radicci, Valeria; Sanders, Stephen; Stringer, Robert; Tinti, Gemma; Wood, Jeffrey Scott; Zhukova, Victoria; Barfuss, Anne-Fleur; Bolton, Tim; Chakaberia, Irakli; Ivanov, Andrew; Khalil, Sadia; Makouski, Mikhail; Maravin, Yurii; Shrestha, Shruti; Svintradze, Irakli; Gronberg, Jeffrey; Lange, David; Wright, Douglas; Baden, Drew; Bard, Robert; Boutemeur, Madjid; Calvert, Brian; Eno, Sarah Catherine; Gomez, Jaime; Grassi, Tullio; Hadley, Nicholas John; Kellogg, Richard G; Kirn, Malina; Kolberg, Ted; Lu, Ying; Marionneau, Matthieu; Mignerey, Alice; Pedro, Kevin; Peterman, Alison; Rossato, Kenneth; Skuja, Andris; Temple, Jeffrey; Tonjes, Marguerite; Tonwar, Suresh C; Toole, Terrence; Twedt, Elizabeth; Apyan, Aram; Bauer, Gerry; Bendavid, Joshua; Busza, Wit; Butz, Erik; Cali, Ivan Amos; Chan, Matthew; Dutta, Valentina; Gomez Ceballos, Guillelmo; Goncharov, Maxim; Hahn, Kristan Allan; Kim, Yongsun; Klute, Markus; Krajczar, Krisztian; Levin, Andrew; Luckey, Paul David; Ma, Teng; Nahn, Steve; Paus, Christoph; Ralph, Duncan; Roland, Christof; Roland, Gunther; Rudolph, Matthew; Stephans, George; Stöckli, Fabian; Sumorok, Konstanty; Sung, Kevin; Velicanu, Dragos; Wenger, Edward Allen; Wolf, Roger; Wyslouch, Bolek; Yang, Mingming; Yilmaz, Yetkin; Yoon, Sungho; Zanetti, Marco; Bailleux, David; Cooper, Seth; Cushman, Priscilla; Dahmes, Bryan; De Benedetti, Abraham; Egeland, Ricky; Franzoni, Giovanni; Gude, Alexander; Haupt, Jason; Inyakin, Alexandre; Kao, Shih-Chuan; Klapoetke, Kevin; Kubota, Yuichi; Mans, Jeremy; Pastika, Nathaniel; Rusack, Roger; Singovsky, Alexander; Tambe, Norbert; Turkewitz, Jared; Cremaldi, Lucien Marcus; Kroeger, Rob; Perera, Lalith; Rahmat, Rahmat; Reidy, Jim; Sanders, David A; Summers, Don; Attebury, Garhan; Avdeeva, Ekaterina; Bloom, Kenneth; Bockelman, Brian; Bose, Suvadeep; Butt, Jamila; Claes, Daniel R; Dominguez, Aaron; Eads, Michael; Keller, Jason; Kravchenko, Ilya; Lazo-Flores, Jose; Lundstedt, Carl; Malbouisson, Helena; Malik, Sudhir; Snihur, Robert; Snow, Gregory R; Swanson, David; Baur, Ulrich; Godshalk, Andrew; Iashvili, Ia; Jain, Supriya; Kharchilava, Avto; Kumar, Ashish; Shipkowski, Simon Peter; Smith, Kenneth; Alverson, George; Barberis, Emanuela; Baumgartel, Darin; Chasco, Matthew; Haley, Joseph; Moromisato, Jorge; Nash, David; Swain, John; Trocino, Daniele; Von Goeler, Eberhard; Wood, Darien; Zhang, Jinzhong; Anastassov, Anton; Gobbi, Bruno; Kubik, Andrew; Odell, Nathaniel; Ofierzynski, Radoslaw Adrian; Pollack, Brian; Pozdnyakov, Andrey; Schmitt, Michael Henry; Stoynev, Stoyan; Velasco, Mayda; Won, Steven; Antonelli, Louis; Baumbaugh, Barry; Berry, Douglas; Brinkerhoff, Andrew; Chan, Kwok Ming; Heering, Arjan Hendrix; Hildreth, Michael; Jessop, Colin; Karmgard, Daniel John; Kellams, Nathan; Kolb, Jeff; Lannon, Kevin; Luo, Wuming; Lynch, Sean; Marinelli, Nancy; Morse, David Michael; Pearson, Tessa; Planer, Michael; Ruchti, Randy; Slaunwhite, Jason; Valls, Nil; Wayne, Mitchell; Wolf, Matthias; Woodard, Anna; Bylsma, Ben; Durkin, Lloyd Stanley; Hill, Christopher; Hughes, Richard; Kotov, Khristian; Ling, Ta-Yung; Puigh, Darren; Rodenburg, Marissa; Rush, Chuck J; Sehgal, Veejay; Vuosalo, Carl; Williams, Grayson; Winer, Brian L; Adam, Nadia; Berry, Edmund; Elmer, Peter; Gerbaudo, Davide; Halyo, Valerie; Hebda, Philip; Hegeman, Jeroen; Hunt, Adam; Jindal, Pratima; Lopes Pegna, David; Lujan, Paul; Marlow, Daniel; Medvedeva, Tatiana; Mooney, Michael; Olsen, James; Piroué, Pierre; Quan, Xiaohang; Raval, Amita; Saka, Halil; Stickland, David; Tully, Christopher; Werner, Jeremy Scott; Wildish, Tony; Xie, Zhen; Zenz, Seth Conrad; Zuranski, Andrzej; Acosta, Jhon Gabriel; Bonnett Del Alamo, Miguel; Brownson, Eric; Huang, Xing Tao; Lopez, Angel; Mendez, Hector; Oliveros, Sandra; Ramirez Vargas, Juan Eduardo; Zatserklyaniy, Andriy; Alagoz, Enver; Arndt, Kirk; Barnes, Virgil E; Benedetti, Daniele; Bolla, Gino; Bortoletto, Daniela; Bujak, Adam; De Mattia, Marco; Everett, Adam; Gutay, Laszlo; Hu, Zhen; Jones, Matthew; Koybasi, Ozhan; Kress, Matthew; Laasanen, Alvin T; Lee, Jik; Leonardo, Nuno; Liu, Chang; Maroussov, Vassili; Merkel, Petra; Miller, David Harry; Miyamoto, Jun; Neumeister, Norbert; Rott, Carsten; Roy, Amitava; Shipsey, Ian; Silvers, David; Svyatkovskiy, Alexey; Vidal Marono, Miguel; Yoo, Hwi Dong; Zablocki, Jakub; Zheng, Yu; Guragain, Samir; Parashar, Neeti; Adair, Antony; Boulahouache, Chaouki; Cuplov, Vesna; Ecklund, Karl Matthew; Geurts, Frank JM; Lee, Sang Joon; Li, Wei; Liu, Jinghua H; Matveev, Mikhail; Padley, Brian Paul; Redjimi, Radia; Roberts, Jay; Tumanov, Alexander; Yepes, Pablo; Zabel, James; Betchart, Burton; Bodek, Arie; Budd, Howard; Chung, Yeon Sei; Covarelli, Roberto; de Barbaro, Pawel; Demina, Regina; Eshaq, Yossof; Ferbel, Thomas; Garcia-Bellido, Aran; Ginther, George; Goldenzweig, Pablo; Gotra, Yury; Han, Jiyeon; Harel, Amnon; Korjenevski, Sergey; Miner, Daniel Carl; Orbaker, Douglas; Sakumoto, Willis; Slattery, Paul; Vishnevskiy, Dmitry; Zielinski, Marek; Bhatti, Anwar; Ciesielski, Robert; Demortier, Luc; Goulianos, Konstantin; Lungu, Gheorghe; Malik, Sarah; Mesropian, Christina; Arora, Sanjay; Barker, Anthony; Chou, John Paul; Contreras-Campana, Christian; Contreras-Campana, Emmanuel; Duggan, Daniel; Ferencek, Dinko; Gershtein, Yuri; Gray, Richard; Halkiadakis, Eva; Hidas, Dean; Lath, Amitabh; Panwalkar, Shruti; Park, Michael; Patel, Rishi; Rekovic, Vladimir; Robles, Jorge; Rose, Keith; Salur, Sevil; Schnetzer, Steve; Seitz, Claudia; Somalwar, Sunil; Stone, Robert; Thomas, Scott; Cerizza, Giordano; Hollingsworth, Matthew; Ragghianti, Gerald; Spanier, Stefan; Yang, Zong-Chang; York, Andrew; Bouhali, Othmane; Eusebi, Ricardo; Flanagan, Will; Gilmore, Jason; Kamon, Teruki; Khotilovich, Vadim; Montalvo, Roy; Nguyen, Chi Nhan; Osipenkov, Ilya; Pakhotin, Yuriy; Perloff, Alexx; Roe, Jeffrey; Safonov, Alexei; Sakuma, Tai; Sengupta, Sinjini; Suarez, Indara; Tatarinov, Aysen; Toback, David; Akchurin, Nural; Damgov, Jordan; Dragoiu, Cosmin; Dudero, Phillip Russell; Jeong, Chiyoung; Kovitanggoon, Kittikul; Lee, Sung Won; Libeiro, Terence; Roh, Youn; Sill, Alan; Volobouev, Igor; Wigmans, Richard; Appelt, Eric; Delannoy, Andrés G; Engh, Daniel; Florez, Carlos; Gabella, William; Greene, Senta; Gurrola, Alfredo; Johns, Willard; Kurt, Pelin; Maguire, Charles; Melo, Andrew; Sharma, Monika; Sheldon, Paul; Snook, Benjamin; Tuo, Shengquan; Velkovska, Julia; Andelin, Daniel; Arenton, Michael Wayne; Balazs, Michael; Boutle, Sarah; Conetti, Sergio; Cox, Bradley; Francis, Brian; Goodell, Joseph; Hirosky, Robert; Ledovskoy, Alexander; Lin, Chuanzhe; Neu, Christopher; Phillips II, David; Wood, John; Gollapinni, Sowjanya; Harr, Robert; Karchin, Paul Edmund; Kottachchi Kankanamge Don, Chamath; Lamichhane, Pramod; Mattson, Mark; Milstène, Caroline; Sakharov, Alexandre; Anderson, Michael; Belknap, Donald; Bellinger, James Nugent; Borrello, Laura; Bradley, Daniel; Carlsmith, Duncan; Cepeda, Maria; Crotty, Ian; Dasu, Sridhara; Feyzi, Farshid; Friis, Evan; Gorski, Thomas; Gray, Lindsey; Grogg, Kira Suzanne; Grothe, Monika; Hall-Wilton, Richard; Herndon, Matthew; Hervé, Alain; Klabbers, Pamela; Klukas, Jeffrey; Lackey, Joe; Lanaro, Armando; Lazaridis, Christos; Leonard, Jessica; Loveless, Richard; Lusin, Sergei; Magrans de Abril, Marc; Maier, Will; Mohapatra, Ajit; Ojalvo, Isabel; Palmonari, Francesco; Pierro, Giuseppe Antonio; Reeder, Don; Ross, Ian; Savin, Alexander; Smith, Wesley H; Swanson, Joshua; Wenman, Daniel 2012-08-17 Results are presented from searches for the standard model Higgs boson in proton-proton collisions at\\sqrt{s}$=7 and 8 TeV in the CMS experiment at the LHC, using data samples corresponding to integrated luminosities of up to 5.1 inverse femtobarns at 7 TeV and 5.3 inverse femtobarns at 8 TeV. The search is performed in five decay modes:$\\gamma\\gamma$, ZZ, WW,$\\tau^+ \\tau^-$, and$b\\bar{b}$. An excess of events is observed above the expected background, a local significance of 5.0 standard deviations, at a mass near 125 GeV, signalling the production of a new particle. The expected significance for a standard model Higgs boson of that mass is 5.8 standard deviations. The excess is most significant in the two decay modes with the best mass resolution,$\\gamma\\gamma$and ZZ; a fit to these signals gives a mass of 125.3$\\pm$0.4 (stat.)$\\pm0.5 (syst.) GeV. The decay to two photons indicates that the new particle is a boson with spin different from one. 7. Helical paths, gravitaxis, and separation phenomena for mass-anisotropic self-propelling colloids: Experiment versus theory Science.gov (United States) Campbell, Andrew I.; Wittkowski, Raphael; ten Hagen, Borge; Löwen, Hartmut; Ebbens, Stephen J. 2017-08-01 The self-propulsion mechanism of active colloidal particles often generates not only translational but also rotational motion. For particles with an anisotropic mass density under gravity, the motion is usually influenced by a downwards oriented force and an aligning torque. Here we study the trajectories of self-propelled bottom-heavy Janus particles in three spatial dimensions both in experiments and by theory. For a sufficiently large mass anisotropy, the particles typically move along helical trajectories whose axis is oriented either parallel or antiparallel to the direction of gravity (i.e., they show gravitaxis). In contrast, if the mass anisotropy is small and rotational diffusion is dominant, gravitational alignment of the trajectories is not possible. Furthermore, the trajectories depend on the angular self-propulsion velocity of the particles. If this component of the active motion is strong and rotates the direction of translational self-propulsion of the particles, their trajectories have many loops, whereas elongated swimming paths occur if the angular self-propulsion is weak. We show that the observed gravitational alignment mechanism and the dependence of the trajectory shape on the angular self-propulsion can be used to separate active colloidal particles with respect to their mass anisotropy and angular self-propulsion, respectively. 8. Observation of a new boson at a mass of 125 GeV with the CMS experiment at the LHC Energy Technology Data Exchange (ETDEWEB) Chatrchyan, S.; Khachatryan, V.; Sirunyan, A. M.; Tumasyan, A.; Adam, W.; Aguilo, E.; Bergauer, T.; Dragicevic, M.; Erö, J.; Fabjan, C.; Friedl, M.; Frühwirth, R.; Ghete, V. M.; Hammer, J.; Hoch, M.; Hörmann, N.; Hrubec, J.; Jeitler, M.; Kiesenhofer, W.; Knünz, V.; Krammer, M.; Krätschmer, I.; Liko, D.; Majerotto, W.; Mikulec, I.; Pernicka, M.; Rahbaran, B.; Rohringer, C.; Rohringer, H.; Schöfbeck, R.; Strauss, J.; Szoncsó, F.; Taurok, A.; Waltenberger, W.; Walzel, G.; Widl, E.; Wulz, C. -E.; Chekhovsky, V.; Emeliantchik, I.; Litomin, A.; Makarenko, V.; Mossolov, V.; Shumeiko, N.; Solin, A.; Stefanovitch, R.; Suarez Gonzalez, J.; Fedorov, A.; Korzhik, M.; Missevitch, O.; Zuyeuski, R.; Bansal, M.; Bansal, S.; Beaumont, W.; Cornelis, T.; De Wolf, E. A.; Druzhkin, D.; Janssen, X.; Luyckx, S.; Mucibello, L.; Ochesanu, S.; Roland, B.; Rougny, R.; Selvaggi, M.; Staykova, Z.; Van Haevermaet, H.; Van Mechelen, P.; Van Remortel, N.; Van Spilbeeck, A.; Blekman, F.; Blyweert, S.; DʼHondt, J.; Devroede, O.; Gonzalez Suarez, R.; Goorens, R.; Kalogeropoulos, A.; Maes, M.; Olbrechts, A.; Tavernier, S.; Van Doninck, W.; Van Lancker, L.; Van Mulders, P.; Van Onsem, G. P.; Villella, I.; Clerbaux, B.; De Lentdecker, G.; Dero, V.; Dewulf, J. P.; Gay, A. P. R.; Hreus, T.; Léonard, A.; Marage, P. E.; Mohammadi, A.; Reis, T.; Rugovac, S.; Thomas, L.; Vander Velde, C.; Vanlaer, P.; Wang, J.; Wickens, J.; Adler, V.; Beernaert, K.; Cimmino, A.; Costantini, S.; Garcia, G.; Grunewald, M.; Klein, B.; Lellouch, J.; Marinov, A.; Mccartin, J.; Ocampo Rios, A. A.; Ryckbosch, D.; Strobbe, N.; Thyssen, F.; Tytgat, M.; Walsh, S.; Yazgan, E.; Zaganidis, N.; Basegmez, S.; Bruno, G.; Castello, R.; Ceard, L.; De Favereau De Jeneret, J.; Delaere, C.; Demin, P.; du Pree, T.; Favart, D.; Forthomme, L.; Giammanco, A.; Grégoire, G.; Hollar, J.; Lemaitre, V.; Liao, J.; Militaru, O.; Nuttens, C.; Pagano, D.; Pin, A.; Piotrzkowski, K.; Schul, N.; Vizan Garcia, J. M.; Beliy, N.; Caebergs, T.; Daubie, E.; Hammad, G. H.; Alves, G. A.; Brito, L.; Correa Martin, M.; Martins, T.; Pol, M. E.; Souza, M. H. G.; Aldá Júnior, W. L.; Carvalho, W.; Custódio, A.; Da Costa, E. M.; De Jesus Damiao, D.; De Oliveira Martins, C.; Fonseca De Souza, S.; Matos Figueiredo, D.; Mundim, L.; Nogima, H.; Oguri, V.; Prado Da Silva, W. L.; Santoro, A.; Sznajder, A.; Vilela Pereira, A.; Anjos, T. S.; Bernardes, C. A.; Dias, F. A.; Fernandez Perez Tomei, T. R.; Gregores, E. M.; Iope, R. L.; Lagana, C.; Lietti, S. M.; Marinho, F.; Mercadante, P. G.; Novaes, S. F.; Padula, Sandra S.; Dimitrov, L.; Genchev, V.; Iaydjiev, P.; Piperov, S.; Rodozov, M.; Stoykova, S.; Sultanov, G.; Tcholakov, V.; Trayanov, R.; Vankov, I.; Vutova, M.; Roumenin, C.; Uzunova, D.; Zahariev, R.; Dimitrov, A.; Hadjiiska, R.; Kozhuharov, V.; Litov, L.; Pavlov, B.; Petkov, P.; Bian, J. G.; Chen, G. M.; Chen, H. S.; He, K. L.; Jiang, C. H.; Li, W. G.; Liang, D.; Liang, S.; Meng, X.; Sun, G.; Sun, H. S.; Tao, J.; Wang, J.; Wang, X.; Wang, Z.; Xiao, H.; Xu, M.; Yang, M.; Zang, J.; Zhang, X.; Zhang, Z.; Zhang, Z.; Zhao, W. R.; Zhu, Z.; Asawatangtrakuldee, C.; Ban, Y.; Cai, J.; Guo, S.; Guo, Y.; Li, W.; Liu, H. T.; Liu, S.; Mao, Y.; Qian, S. J.; Teng, H.; Wang, D.; Ye, Y. L.; Zhang, L.; Zhu, B.; Zou, W.; Avila, C.; Gomez, J. P.; Gomez Moreno, B.; Osorio Oliveros, A. F.; Sanabria, J. C.; Godinovic, N.; Lelas, D.; Plestina, R.; Polic, D.; Puljak, I.; Antunovic, Z.; Kovac, M.; Brigljevic, V.; Duric, S.; Kadija, K.; Luetic, J.; Morovic, S.; Attikis, A.; Galanti, M.; Mavromanolakis, G.; Mousa, J.; Nicolaou, C.; Ptochos, F.; Razis, P. A.; Finger, M.; Finger, M.; Aly, A.; Assran, Y.; Awad, A.; Elgammal, S.; Ellithi Kamel, A.; Khalil, S.; Mahmoud, M. A.; Mahrous, A.; Radi, A.; Hektor, A.; Kadastik, M.; Kannike, K.; Müntel, M.; Raidal, M.; Rebane, L.; Strumia, A.; Tiko, A.; Eerola, P.; Fedi, G.; Voutilainen, M.; Anttila, E.; Härkönen, J.; Heikkinen, A.; Karimäki, V.; Katajisto, H. M.; Kinnunen, R.; Kortelainen, M. J.; Kotamäki, M.; Lampén, T.; Lassila-Perini, K.; Lehti, S.; Lindén, T.; Luukka, P.; Mäenpää, T.; Peltola, T.; Tuominen, E.; Tuominiemi, J.; Tuovinen, E.; Ungaro, D.; Vanhala, T. P.; Wendland, L.; Banzuzi, K.; Karjalainen, A.; Korpela, A.; Tuuva, T.; Anfreville, M.; Besancon, M.; Choudhury, S.; Dejardin, M.; Denegri, D.; Fabbro, B.; Faure, J. L.; Ferri, F.; Ganjour, S.; Gentit, F. X.; Givernaud, A.; Gras, P.; Hamel de Monchenault, G.; Jarry, P.; Kircher, F.; Lemaire, M. C.; Locci, E.; Malcles, J.; Mandjavidze, I.; Nayak, A.; Pansart, J. P.; Rander, J.; Reymond, J. M.; Rosowsky, A.; Shreyber, I.; Titov, M.; Verrecchia, P.; Badier, J.; Baffioni, S.; Beaudette, F.; Becheva, E.; Benhabib, L.; Bianchini, L.; Bluj, M.; Broutin, C.; Busson, P.; Cerutti, M.; Chamont, D.; Charlot, C.; Daci, N.; Dahms, T.; Dalchenko, M.; Dobrzynski, L.; Geerebaert, Y.; Granier de Cassagnac, R.; Haguenauer, M.; Hennion, P.; Milleret, G.; Miné, P.; Mironov, C.; Naranjo, I. N.; Nguyen, M.; Ochando, C.; Paganini, P.; Romanteau, T.; Sabes, D.; Salerno, R.; Sartirana, A.; Sirois, Y.; Thiebaux, C.; Veelken, C.; Zabi, A.; Agram, J. -L.; Andrea, J.; Besson, A.; Bloch, D.; Bodin, D.; Brom, J. -M.; Cardaci, M.; Chabert, E. C.; Collard, C.; Conte, E.; Drouhin, F.; Ferro, C.; Fontaine, J. -C.; Gelé, D.; Goerlach, U.; Goetzmann, C.; Gross, L.; Huss, D.; Juillot, P.; Kieffer, E.; Le Bihan, A. -C.; Pansanel, J.; Patois, Y.; Van Hove, P.; Boutigny, D.; Mercier, D.; Baulieu, G.; Beauceron, S.; Beaupere, N.; Bedjidian, M.; Bondu, O.; Boudoul, G.; Brochet, S.; Chasserat, J.; Chierici, R.; Combaret, C.; Contardo, D.; Depasse, P.; El Mamouni, H.; Fay, J.; Gascon, S.; Giraud, N.; Gouzevitch, M.; Haroutunian, R.; Ille, B.; Kurca, T.; Lethuillier, M.; Lumb, N.; Mathez, H.; Mirabito, L.; Perries, S.; Sgandurra, L.; Sordini, V.; Tschudi, Y.; Vander Donckt, M.; Verdier, P.; Viret, S.; Roinishvili, V.; Rurua, L.; Amaglobeli, N.; Bagaturia, I.; Chiladze, B.; Kvatadze, R.; Lomidze, D.; Shanidze, R.; Tsamalaidze, Z.; Adolphi, R.; Anagnostou, G.; Autermann, C.; Beranek, S.; Brauer, R.; Braunschweig, W.; Calpas, B.; Edelhoff, M.; Feld, L.; Heracleous, N.; Hindrichs, O.; Jussen, R.; Karpinski, W.; Klein, K.; Lübelsmeyer, K.; Merz, J.; Ostapchuk, A.; Pandoulas, D.; Perieanu, A.; Raupach, F.; Sammet, J.; Schael, S.; Schmitz, D.; Schultz von Dratzig, A.; Siedling, R.; Sprenger, D.; Weber, H.; Wittmer, B.; Wlochal, M.; Zhukov, V.; Ata, M.; Biallass, P.; Caudron, J.; Dietz-Laursonn, E.; Duchardt, D.; Erdmann, M.; Fischer, R.; Güth, A.; Hebbeker, T.; Heidemann, C.; Hilgers, G.; Hoepfner, K.; Hof, C.; Klimkovich, T.; Klingebiel, D.; Kreuzer, P.; Magass, C.; Merschmeyer, M.; Meyer, A.; Olschewski, M.; Papacz, P.; Philipps, B.; Pieta, H.; Reithler, H.; Schmitz, S. A.; Sonnenschein, L.; Sowa, M.; Steggemann, J.; Teyssier, D.; Weber, M.; Bontenackels, M.; Cherepanov, V.; Erdogan, Y.; Flügge, G.; Geenen, H.; Geisler, M.; Haj Ahmad, W.; Hoehle, F.; Kargoll, B.; Kress, T.; Kuessel, Y.; Lingemann, J.; Nowack, A.; Perchalla, L.; Pooth, O.; Sauerland, P.; Stahl, A.; Zoeller, M. H.; Aldaya Martin, M.; Behr, J.; Behrenhoff, W.; Behrens, U.; Bergholz, M.; Bethani, A.; Borras, K.; Burgmeier, A.; Cakir, A.; Calligaris, L.; Campbell, A.; Castro, E.; Costanza, F.; Dammann, D.; Diez Pardos, C.; Eckerlin, G.; Eckstein, D.; Flossdorf, A.; Flucke, G.; Geiser, A.; Glushkov, I.; Goettlicher, P.; Grebenyuk, A.; Gunnellini, P.; Habib, S.; Hauk, J.; Hellwig, G.; Jung, H.; Kasemann, M.; Katsas, P.; Kleinwort, C.; Kluge, H.; Knutsson, A.; Krämer, M.; Krücker, D.; Kuznetsova, E.; Lange, W.; Lewendel, B.; Lohmann, W.; Lutz, B.; Mankel, R.; Marfin, I.; Marienfeld, M.; Melzer-Pellmann, I. -A.; Meyer, A. B.; Mnich, J.; Muhl, C.; Mussgiller, A.; Naumann-Emme, S.; Novgorodova, O.; Olzem, J.; Parenti, A.; Perrey, H.; Petrukhin, A.; Pitzl, D.; Raspereza, A.; Ribeiro Cipriano, P. M.; Riedl, C.; Ron, E.; Rosemann, C.; Rosin, M.; Salfeld-Nebgen, J.; Schmidt, R.; Schoerner-Sadenius, T.; Sen, N.; Spiridonov, A.; Stein, M.; Tomaszewska, J.; Volyanskyy, D.; Walsh, R.; Wissing, C.; Youngman, C.; Blobel, V.; Draeger, J.; Enderle, H.; Erfle, J.; Gebbert, U.; Görner, M.; Hermanns, T.; Höing, R. S.; Kaschube, K.; Kaussen, G.; Kirschenmann, H.; Klanner, R.; Lange, J.; Mura, B.; Nowak, F.; Peiffer, T.; Pietsch, N.; Rathjens, D.; Sander, C.; Schettler, H.; Schleper, P.; Schlieckau, E.; Schmidt, A.; Schröder, M.; Schum, T.; Seidel, M.; Sibille, J.; Sola, V.; Stadie, H.; Steinbrück, G.; Thomsen, J.; Vanelderen, L.; Barth, C.; Bauer, J.; Berger, J.; Blüm, P.; Böser, C.; Buege, V.; Chen, Z. Y.; Chowdhury, S.; Chwalek, T.; Daeuwel, D.; De Boer, W.; Descroix, A.; Dierlamm, A.; Dirkes, G.; Fahrer, M.; Feindt, M.; Felzmann, U.; Frey, M.; Furgeri, A.; Gebauer, I.; Gessler, A.; Gruschke, J.; Guthoff, M.; Hackstein, C.; Hartmann, F.; Hauler, F.; Hauth, T.; Heier, S.; Heindl, S. M.; Heinrich, M.; Heiss, A.; Held, H.; Hoffmann, K. H.; Honc, S.; Husemann, U.; Imhof, M.; Jung, C.; Junghans, S.; Katkov, I.; Kerzel, U.; Knoblauch, D.; Komaragiri, J. R.; Kräber, M.; Kuhr, T.; Liamsuwan, T.; Lobelle Pardo, P.; Martschei, D.; Menchikov, A.; Mol, X.; Mörmann, D.; Mueller, S.; Müller, Th.; Neuberger, D.; Neuland, M. B.; Niegel, M.; Nürnberg, A.; Oberst, O.; Oehler, A.; Ortega Gomez, T.; Ott, J.; Piasecki, C.; Poschlad, A.; Quast, G.; Rabbertz, K.; Ratnikov, F.; Ratnikova, N.; Renz, M.; Röcker, S.; Roederer, F.; Sabellek, A.; Saout, C.; Scheurer, A.; Schieferdecker, D.; Schieferdecker, P.; Schilling, F. -P.; Schmanau, M.; Schott, G.; Schwerdtfeger, W.; Simonis, H. J.; Skiba, A.; Stober, F. M.; Theel, A.; Thümmel, W. H.; Troendle, D.; Trunov, A.; Ulrich, R.; Wagner-Kuhr, J.; Wayand, S.; Weber, M.; Weiler, T.; Zeise, M.; Ziebarth, E. B.; Zvada, M.; Daskalakis, G.; Geralis, T.; Kesisoglou, S.; Kyriakis, A.; Loukas, D.; Manolakos, I.; Markou, A.; Markou, C.; Mavrommatis, C.; Ntomari, E.; Gouskos, L.; Panagiotou, A.; Saoulidou, N.; Evangelou, I.; Foudas, C.; Kokkas, P.; Manthos, N.; Papadopoulos, I.; Patras, V.; Triantis, F. A.; Bencze, G.; Hajdu, C.; Hidas, P.; Horvath, D.; Sikler, F.; Veszpremi, V.; Vesztergombi, G.; Zalan, P.; Beni, N.; Czellar, S.; Fenyvesi, A.; Molnar, J.; Palinkas, J.; Szillasi, Z.; Karancsi, J.; Raics, P.; Trocsanyi, Z. L.; Ujvari, B.; Zilizi, G.; Beri, S. B.; Bhandari, V.; Bhatnagar, V.; Dhingra, N.; Gupta, R.; Kaur, M.; Kohli, J. M.; Mehta, M. Z.; Nishu, N.; Saini, L. K.; Sharma, A.; Singh, J. B.; Kumar, Ashok; Kumar, Arun; Ahuja, S.; Bhardwaj, A.; Chatterji, S.; Choudhary, B. C.; Gupta, P.; Malhotra, S.; Naimuddin, M.; Ranjan, K.; Sharma, V.; Shivpuri, R. K.; Banerjee, S.; Bhattacharya, S.; Dutta, S.; Gomber, B.; Jain, Sa.; Jain, Sh.; Khurana, R.; Sarkar, S.; Sharan, M.; Abdulsalam, A.; Choudhury, R. K.; Dutta, D.; Ghodgaonkar, M.; Kailas, S.; Kataria, S. K.; Kumar, V.; Mehta, P.; Mohanty, A. K.; Pant, L. M.; Shukla, P.; Topkar, A.; Aziz, T.; Chendvankar, S.; Deshpande, P. V.; Ganguli, S. N.; Ganguly, S.; Guchait, M.; Gurtu, A.; Maity, M.; Mazumdar, K.; Mohanty, G. B.; Parida, B.; Patil, M. R.; Raghavan, R.; Sudhakar, K.; Wickramage, N.; Acharya, B. S.; Banerjee, S.; Bheesette, S.; Dugad, S.; Kalmani, S. D.; Krishnaswamy, M. R.; Lakkireddi, V. R.; Mondal, N. K.; Narasimham, V. S.; Panyam, N.; Verma, P.; Ardalan, F.; Arfaei, H.; Bakhshiansohi, H.; Etesami, S. M.; Fahim, A.; Hashemi, M.; Jafari, A.; Khakzad, M.; Mohammadi Najafabadi, M.; Paktinat Mehdiabadi, S.; Safarzadeh, B.; Zeinali, M.; Abbrescia, M.; Barbone, L.; Calabria, C.; Chhibra, S. S.; Colaleo, A.; Creanza, D.; De Filippis, N.; De Palma, M.; De Robertis, G.; Donvito, G.; Fiore, L.; Iaselli, G.; Loddo, F.; Maggi, G.; Maggi, M.; Manna, N.; Marangelli, B.; My, S.; Natali, S.; Nuzzo, S.; Pacifico, N.; Pompili, A.; Pugliese, G.; Ranieri, A.; Romano, F.; Selvaggi, G.; Silvestris, L.; Singh, G.; Spinoso, V.; Venditti, R.; Verwilligen, P.; Zito, G.; Abbiendi, G.; Benvenuti, A. C.; Bonacorsi, D.; Braibant-Giacomelli, S.; Brigliadori, L.; Capiluppi, P.; Castro, A.; Cavallo, F. R.; Cuffiani, M.; Dallavalle, G. M.; Fabbri, F.; Fanfani, A.; Fasanella, D.; Giacomelli, P.; Grandi, C.; Guiducci, L.; Marcellini, S.; Masetti, G.; Meneghelli, M.; Montanari, A.; Navarria, F. L.; Odorici, F.; Perrotta, A.; Primavera, F.; Rossi, A. M.; Rovelli, T.; Siroli, G. P.; Travaglini, R.; Albergo, S.; Cappello, G.; Chiorboli, M.; Costa, S.; Noto, F.; Potenza, R.; Saizu, M. A.; Tricomi, A.; Tuve, C.; Barbagli, G.; Ciulli, V.; Civinini, C.; DʼAlessandro, R.; Focardi, E.; Frosali, S.; Gallo, E.; Genta, C.; Gonzi, S.; Meschini, M.; Paoletti, S.; Parrini, G.; Ranieri, R.; Sguazzoni, G.; Tropiano, A.; Benussi, L.; Bianco, S.; Colafranceschi, S.; Fabbri, F.; Piccolo, D.; Fabbricatore, P.; Farinon, S.; Greco, M.; Musenich, R.; Tosi, S.; Benaglia, A.; Carbone, L.; DʼAngelo, P.; De Guio, F.; Di Matteo, L.; Dini, P.; Farina, F. M.; Fiorendi, S.; Gennai, S.; Ghezzi, A.; Malvezzi, S.; Manzoni, R. A.; Martelli, A.; Massironi, A.; Menasce, D.; Moroni, L.; Negri, P.; Paganoni, M.; Pedrini, D.; Pullia, A.; Ragazzi, S.; Redaelli, N.; Sala, S.; Tabarelli de Fatis, T.; Buontempo, S.; Carrillo Montoya, C. A.; Cavallo, N.; De Cosa, A.; Dogangun, O.; Fabozzi, F.; Iorio, A. O. M.; Lista, L.; Meola, S.; Merola, M.; Paolucci, P.; Azzi, P.; Bacchetta, N.; Bellato, M.; Benettoni, M.; Biasotto, M.; Bisello, D.; Branca, A.; Carlin, R.; Checchia, P.; Dorigo, T.; Dosselli, U.; Fanzago, F.; Gasparini, F.; Gasparini, U.; Giubilato, P.; Gonella, F.; Gozzelino, A.; Gulmini, M.; Kanishchev, K.; Lacaprara, S.; Lazzizzera, I.; Loreti, M.; Margoni, M.; Maron, G.; Mazzucato, M.; Meneguzzo, A. T.; Montecassiano, F.; Passaseo, M.; Pazzini, J.; Pegoraro, M.; Pozzobon, N.; Ronchese, P.; Simonetto, F.; Torassa, E.; Tosi, M.; Vanini, S.; Ventura, S.; Zotto, P.; Zumerle, G.; Berzano, U.; Gabusi, M.; Ratti, S. P.; Riccardi, C.; Torre, P.; Vitulo, P.; Biasini, M.; Bilei, G. M.; Fanò, L.; Lariccia, P.; Lucaroni, A.; Mantovani, G.; Menichelli, M.; Nappi, A.; Passeri, D.; Placidi, P.; Romeo, F.; Saha, A.; Santocchia, A.; Servoli, L.; Spiezia, A.; Taroni, S.; Valdata, M.; Angelini, F.; Arezzini, S.; Azzurri, P.; Bagliesi, G.; Basti, A.; Bellazzini, R.; Bernardini, J.; Boccali, T.; Bosi, F.; Brez, A.; Broccolo, G.; Calzolari, F.; Carboni, A.; Castaldi, R.; Cerri, C.; Ciampa, A.; DʼAgnolo, R. T.; DellʼOrso, R.; Fiori, F.; Foà, L.; Giassi, A.; Giusti, S.; Kraan, A.; Latronico, L.; Ligabue, F.; Linari, S.; Lomtadze, T.; Martini, L.; Massa, M.; Massai, M. M.; Mazzoni, E.; Messineo, A.; Moggi, A.; Palla, F.; Raffaelli, F.; Rizzi, A.; Sanguinetti, G.; Segneri, G.; Serban, A. T.; Spagnolo, P.; Spandre, G.; Squillacioti, P.; Tenchini, R.; Tonelli, G.; Venturi, A.; Verdini, P. G.; Baccaro, S.; Barone, L.; Bartoloni, A.; Cavallari, F.; Dafinei, I.; Del Re, D.; Diemoz, M.; Fanelli, C.; Grassi, M.; Longo, E.; Meridiani, P.; Micheli, F.; Nourbakhsh, S.; Organtini, G.; Paramatti, R.; Rahatlou, S.; Sigamani, M.; Soffi, L.; Talamo, I. G.; Amapane, N.; Arcidiacono, R.; Argiro, S.; Arneodo, M.; Biino, C.; Cartiglia, N.; Costa, M.; Demaria, N.; Mariotti, C.; Maselli, S.; Migliore, E.; Monaco, V.; Musich, M.; Obertino, M. M.; Pastrone, N.; Pelliccioni, M.; Peroni, C.; Potenza, A.; Romero, A.; Ruspa, M.; Sacchi, R.; Solano, A.; Staiano, A.; Ambroglini, F.; Belforte, S.; Candelise, V.; Casarsa, M.; Cossutti, F.; Della Ricca, G.; Gobbo, B.; Kavka, C.; Marone, M.; Montanino, D.; Penzo, A.; Schizzi, A.; Kim, T. Y.; Nam, S. K.; Chang, S.; Chung, J.; Ham, S. W.; Han, D.; Kang, J.; Kim, D. H.; Kim, G. N.; Kim, J. E.; Kim, K. S.; Kong, D. J.; Lee, M. W.; Oh, Y. D.; Park, H.; Ro, S. R.; Son, D.; Son, D. C.; Suh, J. S.; Kim, J. Y.; Kim, Zero J.; Song, S.; Choi, S.; Gyun, D.; Hong, B.; Jo, M.; Jo, Y.; Kang, M.; Kim, H.; Kim, T. J.; Lee, K. S.; Moon, D. H.; Park, S. K.; Sim, K. S.; Choi, M.; Hahn, G.; Kang, S.; Kim, H.; Kim, J. H.; Park, C.; Park, I. C.; Park, S.; Ryu, G.; Choi, Y.; Choi, Y. K.; Goh, J.; Kim, M. S.; Kwon, E.; Lee, B.; Lee, J.; Lee, S.; Seo, H.; Yu, I.; Janulis, M.; Juodagalvis, A.; Naujikas, R.; Castilla-Valdez, H.; De La Cruz-Burelo, E.; Heredia-de La Cruz, I.; Lopez-Fernandez, R.; Magaña Villalba, R.; Martínez-Ortega, J.; Sánchez-Hernández, A.; Villasenor-Cendejas, L. M.; Carrillo Moreno, S.; Vazquez Valencia, F.; Salazar Ibarguen, H. A.; Casimiro Linares, E.; Morelos Pineda, A.; Reyes-Santos, M. A.; Allfrey, P.; Krofcheck, D.; Bell, A. J.; Bernardino Rodrigues, N.; Butler, A. P. H.; Butler, P. H.; Doesburg, R.; Pfeiffer, D.; Reucroft, S.; Silverwood, H.; Williams, J. C.; Ahmad, M.; Ansari, M. H.; Asghar, M. I.; Butt, J.; Hoorani, H. R.; Khalid, S.; Khan, W. A.; Khurshid, T.; Qazi, S.; Shah, M. A.; Shoaib, M.; Bialkowska, H.; Boimska, B.; Frueboes, T.; Gokieli, R.; Goscilo, L.; Górski, M.; Kazana, M.; Kudla, I. M.; Nawrocki, K.; Romanowska-Rybinska, K.; Szleper, M.; Wrochna, G.; Zalewski, P.; Brona, G.; Bunkowski, K.; Cwiok, M.; Czyrkowski, H.; Dabrowski, R.; Dominik, W.; Doroba, K.; Kalinowski, A.; Konecki, M.; Krolikowski, J.; Oklinski, W.; Pozniak, K.; Zabolotny, W.; Zych, P.; Kasprowicz, G.; Romaniuk, R.; Alemany-Fernandez, R.; Almeida, N.; Bargassa, P.; David, A.; Faccioli, P.; Ferreira Parracho, P. G.; Gallinaro, M.; Ribeiro, P. Q.; Seixas, J.; Silva, J.; Varela, J.; Vischia, P.; Afanasiev, S.; Belotelov, I.; Bunin, P.; Ershov, Y.; Gavrilenko, M.; Golunov, A.; Golutvin, I.; Gorbounov, N.; Gorbunov, I.; Gramenitski, I.; Kalagin, V.; Kamenev, A.; Karjavin, V.; Konoplyanikov, V.; Korenkov, V.; Kozlov, G.; Kurenkov, A.; Lanev, A.; Makankin, A.; Malakhov, A.; Melnitchenko, I.; Mitsyn, V. V.; Moisenz, P.; Oleynik, D.; Orlov, A.; Palichik, V.; Perelygin, V.; Petrosyan, A.; Savina, M.; Semenov, R.; Shmatov, S.; Shulha, S.; Skachkova, A.; Skatchkov, N.; Smetannikov, V.; Smirnov, V.; Smolin, D.; Tikhonenko, E.; Vasilʼev, S.; Volodko, A.; Zarubin, A.; Zhiltsov, V.; Evstyukhin, S.; Golovtsov, V.; Ivanov, Y.; Kim, V.; Levchenko, P.; Murzin, V.; Oreshkin, V.; Smirnov, I.; Sulimov, V.; Uvarov, L.; Vavilov, S.; Vorobyev, A.; Vorobyev, An.; Andreev, Yu.; Anisimov, A.; Dermenev, A.; Gninenko, S.; Golubev, N.; Gorbunov, D.; Karneyeu, A.; Kirsanov, M.; Krasnikov, N.; Matveev, V.; Pashenkov, A.; Pivovarov, G.; Postoev, V. E.; Rubakov, V.; Shirinyants, V.; Solovey, A.; Tlisov, D.; Toropin, A.; Troitsky, S.; Epshteyn, V.; Erofeeva, M.; Gavrilov, V.; Kaftanov, V.; Kiselevich, I.; Kolosov, V.; Konoplyannikov, A.; Kossov, M.; Kozlov, Y.; Krokhotin, A.; Litvintsev, D.; Lychkovskaya, N.; Oulianov, A.; Popov, V.; Safronov, G.; Semenov, S.; Stepanov, N.; Stolin, V.; Vlasov, E.; Zaytsev, V.; Zhokin, A.; Belyaev, A.; Boos, E.; Bunichev, V.; Demiyanov, A.; Dubinin, M.; Dudko, L.; Ershov, A.; Gribushin, A.; Ilyin, V.; Kaminskiy, A.; Klyukhin, V.; Kodolova, O.; Korotkikh, V.; Kryukov, A.; Lokhtin, I.; Markina, A.; Obraztsov, S.; Perfilov, M.; Petrushanko, S.; Popov, A.; Proskuryakov, A.; Sarycheva, L.; Savrin, V.; Snigirev, A.; Vardanyan, I.; Andreev, V.; Azarkin, M.; Dremin, I.; Kirakosyan, M.; Leonidov, A.; Mesyats, G.; Rusakov, S. V.; Vinogradov, A.; Azhgirey, I.; Bayshev, I.; Bitioukov, S.; Grishin, V.; Kachanov, V.; Kalinin, A.; Konstantinov, D.; Korablev, A.; Krychkine, V.; Levine, A.; Petrov, V.; Ryabov, A.; Ryutin, R.; Sobol, A.; Talov, V.; Tourtchanovitch, L.; Troshin, S.; Tyurin, N.; Uzunian, A.; Volkov, A.; Adzic, P.; Djordjevic, M.; Ekmedzic, M.; Krpic, D.; Milosevic, J.; Smiljkovic, N.; Zupan, M.; Aguilar-Benitez, M.; Alcaraz Maestre, J.; Arce, P.; Battilana, C.; Calvo, E.; Cerrada, M.; Chamizo Llatas, M.; Colino, N.; De La Cruz, B.; Delgado Peris, A.; Domínguez Vázquez, D.; Fernandez Bedoya, C.; Fernández Ramos, J. P.; Ferrando, A.; Flix, J.; Fouz, M. C.; Garcia-Abia, P.; Gonzalez Lopez, O.; Goy Lopez, S.; Hernandez, J. M.; Josa, M. I.; Merino, G.; Puerta Pelayo, J.; Quintario Olmeda, A.; Redondo, I.; Romero, L.; Santaolalla, J.; Soares, M. S.; Willmott, C.; Albajar, C.; Codispoti, G.; de Trocóniz, J. F.; Brun, H.; Cuevas, J.; Fernandez Menendez, J.; Folgueras, S.; Gonzalez Caballero, I.; Lloret Iglesias, L.; Piedra Gomez, J.; Brochero Cifuentes, J. A.; Cabrillo, I. J.; Calderon, A.; Chuang, S. H.; Duarte Campderros, J.; Felcini, M.; Fernandez, M.; Gomez, G.; Gonzalez Sanchez, J.; Graziano, A.; Jorda, C.; Lopez Virto, A.; Marco, J.; Marco, R.; Martinez Rivero, C.; Matorras, F.; Munoz Sanchez, F. J.; Rodrigo, T.; Rodríguez-Marrero, A. Y.; Ruiz-Jimeno, A.; Scodellaro, L.; Sobron Sanudo, M.; Vila, I.; Vilar Cortabitarte, R.; Abbaneo, D.; Aspell, P.; Auffray, E.; Auzinger, G.; Bachtis, M.; Baechler, J.; Baillon, P.; Ball, A. H.; Barney, D.; Benitez, J. F.; Bernet, C.; Bialas, W.; Bianchi, G.; Bloch, P.; Bocci, A.; Bonato, A.; Botta, C.; Breuker, H.; Campi, D.; Camporesi, T.; Cano, E.; Cerminara, G.; Charkiewicz, A.; Christiansen, T.; Coarasa Perez, J. A.; Curé, B.; DʼEnterria, D.; Dabrowski, A.; Daguin, J.; De Roeck, A.; Di Guida, S.; Dobson, M.; Dupont-Sagorin, N.; Elliott-Peisert, A.; Eppard, M.; Frisch, B.; Funk, W.; Gaddi, A.; Gastal, M.; Georgiou, G.; Gerwig, H.; Giffels, M.; Gigi, D.; Gill, K.; Giordano, D.; Girone, M.; Giunta, M.; Glege, F.; Gomez-Reino Garrido, R.; Govoni, P.; Gowdy, S.; Guida, R.; Gutleber, J.; Hansen, M.; Harris, P.; Hartl, C.; Harvey, J.; Hegner, B.; Hinzmann, A.; Honma, A.; Innocente, V.; Janot, P.; Kaadze, K.; Karavakis, E.; Kloukinas, K.; Kousouris, K.; Lecoq, P.; Lee, Y. -J.; Lenzi, P.; Loos, R.; Lourenço, C.; Magini, N.; Mäki, T.; Malberti, M.; Malgeri, L.; Mannelli, M.; Marchioro, A.; Marques Pinho Noite, J.; Masetti, L.; Meijers, F.; Mersi, S.; Meschi, E.; Moneta, L.; Mozer, M. U.; Mulders, M.; Musella, P.; Onnela, A.; Orimoto, T.; Orsini, L.; Osborne, J. A.; Palencia Cortezon, E.; Perez, E.; Perrozzi, L.; Petagna, P.; Petrilli, A.; Petrucci, A.; Pfeiffer, A.; Pierini, M.; Pimiä, M.; Piparo, D.; Polese, G.; Postema, H.; Quertenmont, L.; Racz, A.; Reece, W.; Ricci, D.; Rodrigues Antunes, J.; Rolandi, G.; Rovelli, C.; Rovere, M.; Ryjov, V.; Sakulin, H.; Samyn, D.; Santanastasio, F.; Schäfer, C.; Schwick, C.; Sciaba, A.; Segoni, I.; Sekmen, S.; Sharma, A.; Siegrist, P.; Silva, P.; Simon, M.; Sphicas, P.; Spiga, D.; Taylor, B. G.; Tropea, P.; Troska, J.; Tsirou, A.; Vasey, F.; Veillet, L.; Veres, G. I.; Vichoudis, P.; Vlimant, J. R.; Wertelaers, P.; Wöhri, H. K.; Worm, S. D.; Zeuner, W. D.; Bertl, W.; Deiters, K.; Erdmann, W.; Feichtinger, D.; Gabathuler, K.; Horisberger, R.; Ingram, Q.; Kaestli, H. C.; König, S.; Kotlinski, D.; Langenegger, U.; Meier, B.; Meier, F.; Renker, D.; Rohe, T.; Sakhelashvili, T.; Bäni, L.; Behner, F.; Betev, B.; Blau, B.; Bortignon, P.; Buchmann, M. A.; Casal, B.; Chanon, N.; Chen, Z.; Da Silva Di Calafiori, D. R.; Dambach, S.; Davatz, G.; Deisher, A.; Dissertori, G.; Dittmar, M.; Djambazov, L.; Donegà, M.; Dünser, M.; Eggel, C.; Eugster, J.; Faber, G.; Freudenreich, K.; Grab, C.; Hintz, W.; Hits, D.; Hofer, H.; Holme, O.; Horvath, I.; Lecomte, P.; Lustermann, W.; Marchica, C.; Marini, A. C.; Martinez Ruiz del Arbol, P.; Mohr, N.; Moortgat, F.; Nägeli, C.; Nef, P.; Nessi-Tedaldi, F.; Pandolfi, F.; Pape, L.; Pauss, F.; Peruzzi, M.; Punz, T.; Ronga, F. J.; Röser, U.; Rossini, M.; Sala, L.; Sanchez, A. K.; Sawley, M. -C.; Schinzel, D.; Starodumov, A.; Stieger, B.; Suter, H.; Takahashi, M.; Tauscher, L.; Thea, A.; Theofilatos, K.; Treille, D.; Trüb, P.; Udriot, S.; Urscheler, C.; Viertel, G.; von Gunten, H. P.; Wallny, R.; Weber, H. A.; Wehrli, L.; Weng, J.; Zelepoukine, S.; Amsler, C.; Chiochia, V.; De Visscher, S.; Favaro, C.; Ivova Rikova, M.; Millan Mejias, B.; Otiougova, P.; Robmann, P.; Snoek, H.; Tupputi, S.; Verzetti, M.; Chang, Y. H.; Chen, K. H.; Chen, W. T.; Go, A.; Kuo, C. M.; Li, S. W.; Lin, W.; Liu, M. H.; Liu, Z. K.; Lu, Y. J.; Mekterovic, D.; Singh, A. P.; Volpe, R.; Wu, J. H.; Yu, S. S.; Bartalini, P.; Chang, P.; Chang, Y. H.; Chang, Y. W.; Chao, Y.; Chen, K. F.; Dietz, C.; Gao, Z.; Grundler, U.; Hou, W. -S.; Hsiung, Y.; Kao, K. Y.; Lei, Y. J.; Liau, J. j.; Lin, S. W.; Lu, R. -S.; Majumder, D.; Petrakou, E.; Shi, X.; Shiu, J. G.; Tzeng, Y. M.; Ueno, K.; Velikzhanin, Y.; Wan, X.; Wang, C. C.; Wang, M.; Wei, J. T.; Yeh, P.; Asavapibhop, B.; Srimanobhas, N.; Adiguzel, A.; Bakirci, M. N.; Cerci, S.; Dozen, C.; Dumanoglu, I.; Eskut, E.; Girgis, S.; Gokbulut, G.; Gurpinar, E.; Hos, I.; Kangal, E. E.; Karaman, T.; Karapinar, G.; Kayis Topaksu, A.; Onengut, G.; Ozdemir, K.; Ozturk, S.; Polatoz, A.; Sogut, K.; Sunar Cerci, D.; Tali, B.; Topakli, H.; Vergili, L. N.; Vergili, M.; Akin, I. V.; Aliev, T.; Bilin, B.; Deniz, M.; Gamsizkan, H.; Guler, A. M.; Ocalan, K.; Ozpineci, A.; Serin, M.; Sever, R.; Surat, U. E.; Zeyrek, M.; Deliomeroglu, M.; Gülmez, E.; Isildak, B.; Kaya, M.; Kaya, O.; Ozkorucuklu, S.; Sonmez, N.; Cankocak, K.; Grynyov, B.; Levchuk, L.; Lukyanenko, S.; Soroka, D.; Sorokin, P.; Ahmad, M. K. H.; Branson, A.; McClatchey, R.; Odeh, M.; Shamdasani, J.; Soomro, K.; Barrass, T.; Bostock, F.; Brooke, J. J.; Clement, E.; Cussans, D.; Flacher, H.; Frazier, R.; Goldstein, J.; Grimes, M.; Heath, G. P.; Heath, H. F.; Kreczko, L.; Lacesso, W.; Metson, S.; Newbold, D. M.; Nirunpong, K.; Poll, A.; Senkin, S.; Smith, V. J.; Williams, T.; Basso, L.; Bateman, E.; Bell, K. W.; Belyaev, A.; Brew, C.; Brown, R. M.; Camanzi, B.; Cockerill, D. J. A.; Connolly, J. F.; Coughlan, J. A.; Denton, L. G.; Flower, P. S.; French, M. J.; Greenhalgh, R. J. S.; Halsall, R. N. J.; Harder, K.; Harper, S.; Hill, J. A.; Jackson, J.; Kennedy, B. W.; Lintern, A. L.; Lodge, A. B.; Olaiya, E.; Pearson, M. R.; Petyt, D.; Radburn-Smith, B. C.; Shepherd-Themistocleous, C. H.; Smith, B. J.; Sproston, M.; Stephenson, R.; Tomalin, I. R.; Torbet, M. J.; Williams, J. H.; Womersley, W. J.; Bainbridge, R.; Ball, G.; Ballin, J.; Bauer, D.; Beuselinck, R.; Buchmuller, O.; Colling, D.; Cripps, N.; Cutajar, M.; Dauncey, P.; Davies, G.; Della Negra, M.; Ferguson, W.; Fulcher, J.; Futyan, D.; Gilbert, A.; Guneratne Bryer, A.; Hall, G.; Hatherell, Z.; Hays, J.; Iles, G.; Jarvis, M.; Jones, J.; Karapostoli, G.; Kenzie, M.; Leaver, J.; Lyons, L.; Magnan, A. -M.; Marrouche, J.; Mathias, B.; Miller, D. G.; Nandi, R.; Nash, J.; Nikitenko, A.; Noy, M.; Papageorgiou, A.; Pela, J.; Pesaresi, M.; Petridis, K.; Pioppi, M.; Rand, D.; Raymond, D. M.; Rogerson, S.; Rose, A.; Ryan, M. J.; Seez, C.; Sharp, P.; Sparrow, A.; Stoye, M.; Tapper, A.; Timlin, C.; Tourneur, S.; Vazquez Acosta, M.; Virdee, T.; Wakefield, S.; Wardle, N.; Whyntie, T.; Wingham, M.; Zorba, O.; Chadwick, M.; Cole, J. E.; Hobson, P. R.; Khan, A.; Kyberd, P.; Leggat, D.; Leslie, D.; Martin, W.; Reid, I. D.; Symonds, P.; Teodorescu, L.; Turner, M.; Dittmann, J.; Hatakeyama, K.; Liu, H.; Scarborough, T.; Charaf, O.; Henderson, C.; Rumerio, P.; Avetisyan, A.; Bose, T.; Carrera Jarrin, E.; Fantasia, C.; Hazen, E.; Heister, A.; John, J. St.; Lawson, P.; Lazic, D.; Rohlf, J.; Sperka, D.; Sulak, L.; Varela Rodriguez, F.; Wu, S.; Alimena, J.; Bhattacharya, S.; Cutts, D.; Demiragli, Z.; Ferapontov, A.; Garabedian, A.; Heintz, U.; Hooper, R.; Jabeen, S.; Kukartsev, G.; Laird, E.; Landsberg, G.; Luk, M.; Narain, M.; Nguyen, D.; Segala, M.; Sinthuprasith, T.; Speer, T.; Tsang, K. V.; Unalan, Z.; Breedon, R.; Breto, G.; Calderon De La Barca Sanchez, M.; Case, M.; Chauhan, S.; Chertok, M.; Conway, J.; Conway, R.; Cox, P. T.; Dolen, J.; Erbacher, R.; Gardner, M.; Grim, G.; Gunion, J.; Holbrook, B.; Ko, W.; Kopecky, A.; Lander, R.; Lin, F. C.; Miceli, T.; Murray, P.; Nikolic, M.; Pellett, D.; Ricci-tam, F.; Rowe, J.; Rutherford, B.; Searle, M.; Smith, J.; Squires, M.; Tripathi, M.; Vasquez Sierra, R.; Yohay, R.; Andreev, V.; Arisaka, K.; Cline, D.; Cousins, R.; Duris, J.; Erhan, S.; Everaerts, P.; Farrell, C.; Hauser, J.; Ignatenko, M.; Jarvis, C.; Kubic, J.; Otwinowski, S.; Plager, C.; Rakness, G.; Schlein, P.; Traczyk, P.; Valuev, V.; Weber, M.; Yang, X.; Zheng, Y.; Babb, J.; Clare, R.; Dinardo, M. E.; Ellison, J.; Gary, J. W.; Giordano, F.; Hanson, G.; Jeng, G. Y.; Layter, J. G.; Liu, H.; Long, O. R.; Luthra, A.; Nguyen, H.; Paramesvaran, S.; Shen, B. C.; Sturdy, J.; Sumowidagdo, S.; Wilken, R.; Wimpenny, S.; Andrews, W.; Branson, J. G.; Cerati, G. B.; Cinquilli, M.; Cittolin, S.; Evans, D.; Golf, F.; Holzner, A.; Kelley, R.; Lebourgeois, M.; Letts, J.; Macneill, I.; Mangano, B.; Martin, T.; Mrak-Tadel, A.; Padhi, S.; Palmer, C.; Petrucciani, G.; Pieri, M.; Sani, M.; Sfiligoi, I.; Sharma, V.; Simon, S.; Sudano, E.; Tadel, M.; Tu, Y.; Vartak, A.; Wasserbaech, S.; Würthwein, F.; Yagil, A.; Yoo, J.; Barge, D.; Bellan, R.; Campagnari, C.; DʼAlfonso, M.; Danielson, T.; Flowers, K.; Geffert, P.; Incandela, J.; Justus, C.; Kalavase, P.; Koay, S. A.; Kovalskyi, D.; Krutelyov, V.; Kyre, S.; Lowette, S.; Magazzu, G.; Mccoll, N.; Pavlunin, V.; Rebassoo, F.; Ribnik, J.; Richman, J.; Rossin, R.; Stuart, D.; To, W.; West, C.; White, D.; Adamczyk, D.; Apresyan, A.; Barczyk, A.; Bornheim, A.; Bunn, J.; Chen, Y.; Denis, G.; Di Marco, E.; Duarte, J.; Galvez, P.; Gataullin, M.; Kcira, D.; Legrand, I.; Litvine, V.; Ma, Y.; Maxa, Z.; Mott, A.; Mughal, A.; Nae, D.; Newman, H. B.; Ravot, S.; Rogan, C.; Rozsa, S. G.; Shevchenko, S.; Shin, K.; Spiropulu, M.; Steenberg, C.; Thomas, M.; Timciuc, V.; van Lingen, F.; Veverka, J.; Voicu, B. R.; Wilkinson, R.; Xie, S.; Yang, Y.; Zhang, L.; Zhu, K.; Zhu, R. Y.; Akgun, B.; Azzolini, V.; Calamba, A.; Carroll, R.; Ferguson, T.; Iiyama, Y.; Jang, D. W.; Jun, S. Y.; Liu, Y. F.; Paulini, M.; Russ, J.; Terentyev, N.; Vogel, H.; Vorobiev, I.; Cumalat, J. P.; Drell, B. R.; Ford, W. T.; Gaz, A.; Heyburn, B.; Johnson, D.; Luiggi Lopez, E.; Nauenberg, U.; Smith, J. G.; Stenson, K.; Ulmer, K. A.; Wagner, S. R.; Zang, S. L.; Agostino, L.; Alexander, J.; Chatterjee, A.; Eggert, N.; Gibbons, L. K.; Heltsley, B.; Khukhunaishvili, A.; Kreis, B.; Kuznetsov, V.; Mirman, N.; Nicolas Kaufman, G.; Patterson, J. R.; Riley, D.; Ryd, A.; Salvati, E.; Stroiney, S.; Sun, W.; Teo, W. D.; Thom, J.; Thompson, J.; Tucker, J.; Vaughan, J.; Weng, Y.; Winstrom, L.; Wittich, P.; Winn, D.; Abdullin, S.; Albert, M.; Albrow, M.; Anderson, J.; Apollinari, G.; Atac, M.; Badgett, W.; Bakken, J. A.; Baldin, B.; Banicz, K.; Bauerdick, L. A. T.; Beretvas, A.; Berryhill, J.; Bhat, P. C.; Binkley, M.; Borcherding, F.; Burkett, K.; Butler, J. N.; Chetluru, V.; Cheung, H. W. K.; Chlebana, F.; Cihangir, S.; Dagenhart, W.; Derylo, G.; Dumitrescu, C.; Dykstra, D.; Eartly, D. P.; Elias, J. E.; Elvira, V. D.; Eulisse, G.; Evans, D.; Fagan, D.; Fisk, I.; Foulkes, S.; Freeman, J.; Gaines, I.; Gao, Y.; Gartung, P.; Giacchetti, L.; Gottschalk, E.; Green, D.; Guo, Y.; Gutsche, O.; Hahn, A.; Hanlon, J.; Harris, R. M.; Hirschauer, J.; Holzman, B.; Hooberman, B.; Howell, J.; Huang, C. h.; Hufnagel, D.; Jindariani, S.; Johnson, M.; Jones, C. D.; Joshi, U.; Juska, E.; Kilminster, B.; Klima, B.; Kunori, S.; Kwan, S.; Larson, K.; Leonidopoulos, C.; Linacre, J.; Lincoln, D.; Lipton, R.; Lopez Perez, J. A.; Los, S.; Lykken, J.; Maeshima, K.; Marraffino, J. M.; Maruyama, S.; Mason, D.; McBride, P.; McCauley, T.; Mishra, K.; Moccia, S.; Mommsen, R. K.; Mrenna, S.; Musienko, Y.; Muzaffar, S.; Newman-Holmes, C.; OʼDell, V.; Osborne, I.; Pivarski, J.; Popescu, S.; Pordes, R.; Prokofyev, O.; Rapsevicius, V.; Ronzhin, A.; Rossman, P.; Ryu, S.; Sexton-Kennedy, E.; Sharma, S.; Shaw, T. M.; Smith, R. P.; Soha, A.; Spalding, W. J.; Spiegel, L.; Tanenbaum, W.; Taylor, L.; Thompson, R.; Tiradani, A.; Tkaczyk, S.; Tran, N. V.; Tuura, L.; Uplegger, L.; Vaandering, E. W.; Vidal, R.; Whitmore, J.; Wu, W.; Yang, F.; Yarba, J.; Yun, J. C.; Zimmerman, T.; Acosta, D.; Avery, P.; Barashko, V.; Bourilkov, D.; Chen, M.; Cheng, T.; Das, S.; De Gruttola, M.; Di Giovanni, G. P.; Dobur, D.; Dolinsky, S.; Drozdetskiy, A.; Field, R. D.; Fisher, M.; Fu, Y.; Furic, I. K.; Gartner, J.; Gorn, L.; Holmes, D.; Hugon, J.; Kim, B.; Konigsberg, J.; Korytov, A.; Kropivnitskaya, A.; Kypreos, T.; Low, J. F.; Madorsky, A.; Matchev, K.; Milenovic, P.; Mitselmakher, G.; Muniz, L.; Park, M.; Remington, R.; Rinkevicius, A.; Scurlock, B.; Skhirtladze, N.; Snowball, M.; Stasko, J.; Yelton, J.; Zakaria, M.; Gaultney, V.; Hewamanage, S.; Lebolo, L. M.; Linn, S.; Markowitz, P.; Martinez, G.; Rodriguez, J. L.; Adams, T.; Askew, A.; Bertoldi, M.; Bochenek, J.; Chen, J.; Dharmaratna, W. G. D.; Diamond, B.; Gleyzer, S. V.; Haas, J.; Hagopian, S.; Hagopian, V.; Jenkins, M.; Johnson, K. F.; Prosper, H.; Tentindo, S.; Veeraraghavan, V.; Weinberg, M.; Baarmand, M. M.; Dorney, B.; Hohlmann, M.; Kalakhety, H.; Ralich, R.; Vodopiyanov, I.; Yumiceva, F.; Adams, M. R.; Anghel, I. M.; Apanasevich, L.; Bai, Y.; Bazterra, V. E.; Betts, R. R.; Bucinskaite, I.; Callner, J.; Cavanaugh, R.; Chung, M. H.; Evdokimov, O.; Garcia-Solis, E. J.; Gauthier, L.; Gerber, C. E.; Hofman, D. J.; Hollis, R.; Iordanova, A.; Khalatyan, S.; Kunde, G. J.; Lacroix, F.; Malek, M.; OʼBrien, C.; Silkworth, C.; Silvestre, C.; Smoron, A.; Strom, D.; Turner, P.; Varelas, N.; Akgun, U.; Albayrak, E. A.; Ayan, A. S.; Bilki, B.; Clarida, W.; Debbins, P.; Duru, F.; Ingram, F. D.; McCliment, E.; Merlo, J. -P.; Mermerkaya, H.; Mestvirishvili, A.; Miller, M. J.; Moeller, A.; Nachtman, J.; Newsom, C. R.; Norbeck, E.; Olson, J.; Onel, Y.; Ozok, F.; Schmidt, I.; Sen, S.; Tan, P.; Tiras, E.; Wetzel, J.; Yetkin, T.; Yi, K.; Barnett, B. A.; Blumenfeld, B.; Bolognesi, S.; Fehling, D.; Giurgiu, G.; Gritsan, A. V.; Guo, Z. J.; Hu, G.; Maksimovic, P.; Rappoccio, S.; Swartz, M.; Whitbeck, A.; Baringer, P.; Bean, A.; Benelli, G.; Coppage, D.; Grachov, O.; Kenny Iii, R. P.; Murray, M.; Noonan, D.; Radicci, V.; Sanders, S.; Stringer, R.; Tinti, G.; Wood, J. S.; Zhukova, V.; Barfuss, A. F.; Bolton, T.; Chakaberia, I.; Ivanov, A.; Khalil, S.; Makouski, M.; Maravin, Y.; Shrestha, S.; Svintradze, I.; Gronberg, J.; Lange, D.; Wright, D.; Baden, A.; Bard, R.; Boutemeur, M.; Calvert, B.; Eno, S. C.; Gomez, J. A.; Grassi, T.; Hadley, N. J.; Kellogg, R. G.; Kirn, M.; Kolberg, T.; Lu, Y.; Marionneau, M.; Mignerey, A. C.; Pedro, K.; Peterman, A.; Rossato, K.; Skuja, A.; Temple, J.; Tonjes, M. B.; Tonwar, S. C.; Toole, T.; Twedt, E.; Apyan, A.; Bauer, G.; Bendavid, J.; Busza, W.; Butz, E.; Cali, I. A.; Chan, M.; Dutta, V.; Gomez Ceballos, G.; Goncharov, M.; Hahn, K. A.; Kim, Y.; Klute, M.; Krajczar, K.; Levin, A.; Luckey, P. D.; Ma, T.; Nahn, S.; Paus, C.; Ralph, D.; Roland, C.; Roland, G.; Rudolph, M.; Stephans, G. S. F.; Stöckli, F.; Sumorok, K.; Sung, K.; Velicanu, D.; Wenger, E. A.; Wolf, R.; Wyslouch, B.; Yang, M.; Yilmaz, Y.; Yoon, A. S.; Zanetti, M.; Bailleux, D.; Cooper, S. I.; Cushman, P.; Dahmes, B.; De Benedetti, A.; Egeland, R.; Franzoni, G.; Gude, A.; Haupt, J.; Inyakin, A.; Kao, S. C.; Klapoetke, K.; Kubota, Y.; Mans, J.; Pastika, N.; Rusack, R.; Singovsky, A.; Tambe, N.; Turkewitz, J.; Cremaldi, L. M.; Kroeger, R.; Perera, L.; Rahmat, R.; Reidy, J.; Sanders, D. A.; Summers, D.; Attebury, G.; Avdeeva, E.; Bloom, K.; Bockelman, B.; Bose, S.; Claes, D. R.; Dominguez, A.; Eads, M.; Keller, J.; Kravchenko, I.; Lazo-Flores, J.; Lundstedt, C.; Malbouisson, H.; Malik, S.; Snihur, R.; Snow, G. R.; Swanson, D.; Baur, U.; Godshalk, A.; Iashvili, I.; Jain, S.; Kharchilava, A.; Kumar, A.; Shipkowski, S. P.; Smith, K.; Alverson, G.; Barberis, E.; Baumgartel, D.; Chasco, M.; Haley, J.; Moromisato, J.; Nash, D.; Swain, J.; Trocino, D.; Von Goeler, E.; Wood, D.; Zhang, J.; Anastassov, A.; Gobbi, B.; Kubik, A.; Lusito, L.; Odell, N.; Ofierzynski, R. A.; Pollack, B.; Pozdnyakov, A.; Schmitt, M.; Stoynev, S.; Velasco, M.; Won, S.; Antonelli, L.; Baumbaugh, B.; Berry, D.; Brinkerhoff, A.; Chan, K. M.; Heering, A. H.; Hildreth, M.; Jessop, C.; Karmgard, D. J.; Kellams, N.; Kolb, J.; Lannon, K.; Luo, W.; Lynch, S.; Marinelli, N.; Morse, D. M.; Pearson, T.; Planer, M.; Ruchti, R.; Slaunwhite, J.; Valls, N.; Wayne, M.; Wolf, M.; Woodard, A.; Bylsma, B.; Durkin, L. S.; Hill, C.; Hughes, R.; Kotov, K.; Ling, T. Y.; Puigh, D.; Rodenburg, M.; Rush, C. J.; Sehgal, V.; Vuosalo, C.; Williams, G.; Winer, B. L.; Adam, N.; Berry, E.; Elmer, P.; Gerbaudo, D.; Halyo, V.; Hebda, P.; Hegeman, J.; Hunt, A.; Jindal, P.; Lopes Pegna, D.; Lujan, P.; Marlow, D.; Medvedeva, T.; Mooney, M.; Olsen, J.; Piroué, P.; Quan, X.; Raval, A.; Saka, H.; Stickland, D.; Tully, C.; Werner, J. S.; Wildish, T.; Xie, Z.; Zenz, S. C.; Zuranski, A.; Acosta, J. G.; Bonnett Del Alamo, M.; Brownson, E.; Huang, X. T.; Lopez, A.; Mendez, H.; Oliveros, S.; Ramirez Vargas, J. E.; Zatserklyaniy, A.; Alagoz, E.; Arndt, K.; Barnes, V. E.; Benedetti, D.; Bolla, G.; Bortoletto, D.; Bujak, A.; De Mattia, M.; Everett, A.; Gutay, L.; Hu, Z.; Jones, M.; Koybasi, O.; Kress, M.; Laasanen, A. T.; Lee, J.; Leonardo, N.; Liu, C.; Maroussov, V.; Merkel, P.; Miller, D. H.; Miyamoto, J.; Neumeister, N.; Rott, C.; Roy, A.; Shipsey, I.; Silvers, D.; Svyatkovskiy, A.; Vidal Marono, M.; Yoo, H. D.; Zablocki, J.; Zheng, Y.; Guragain, S.; Parashar, N.; Adair, A.; Boulahouache, C.; Cuplov, V.; Ecklund, K. M.; Geurts, F. J. M.; Lee, S. J.; Li, W.; Liu, J. H.; Matveev, M.; Padley, B. P.; Redjimi, R.; Roberts, J.; Tumanov, A.; Yepes, P.; Zabel, J.; Betchart, B.; Bodek, A.; Budd, H.; Chung, Y. S.; Covarelli, R.; de Barbaro, P.; Demina, R.; Eshaq, Y.; Ferbel, T.; Garcia-Bellido, A.; Ginther, G.; Goldenzweig, P.; Gotra, Y.; Han, J.; Harel, A.; Korjenevski, S.; Miner, D. C.; Orbaker, D.; Sakumoto, W.; Slattery, P.; Vishnevskiy, D.; Zielinski, M.; Bhatti, A.; Ciesielski, R.; Demortier, L.; Goulianos, K.; Lungu, G.; Malik, S.; Mesropian, C.; Arora, S.; Barker, A.; Chou, J. P.; Contreras-Campana, C.; Contreras-Campana, E.; Duggan, D.; Ferencek, D.; Gershtein, Y.; Gray, R.; Halkiadakis, E.; Hidas, D.; Lath, A.; Panwalkar, S.; Park, M.; Patel, R.; Rekovic, V.; Robles, J.; Rose, K.; Salur, S.; Schnetzer, S.; Seitz, C.; Somalwar, S.; Stone, R.; Thomas, S.; Cerizza, G.; Hollingsworth, M.; Ragghianti, G.; Spanier, S.; Yang, Z. C.; York, A.; Bouhali, O.; Eusebi, R.; Flanagan, W.; Gilmore, J.; Kamon, T.; Khotilovich, V.; Montalvo, R.; Nguyen, C. N.; Osipenkov, I.; Pakhotin, Y.; Perloff, A.; Roe, J.; Safonov, A.; Sakuma, T.; Sengupta, S.; Suarez, I.; Tatarinov, A.; Toback, D.; Akchurin, N.; Damgov, J.; Dragoiu, C.; Dudero, P. R.; Jeong, C.; Kovitanggoon, K.; Lee, S. W.; Libeiro, T.; Roh, Y.; Sill, A.; Volobouev, I.; Wigmans, R.; Appelt, E.; Delannoy, A. G.; Engh, D.; Florez, C.; Gabella, W.; Greene, S.; Gurrola, A.; Johns, W.; Kurt, P.; Maguire, C.; Melo, A.; Sharma, M.; Sheldon, P.; Snook, B.; Tuo, S.; Velkovska, J.; Andelin, D.; Arenton, M. W.; Balazs, M.; Boutle, S.; Conetti, S.; Cox, B.; Francis, B.; Goodell, J.; Hirosky, R.; Ledovskoy, A.; Lin, C.; Neu, C.; Phillips II, D.; Wood, J.; Gollapinni, S.; Harr, R.; Karchin, P. E.; Kottachchi Kankanamge Don, C.; Lamichhane, P.; Mattson, M.; Milstène, C.; Sakharov, A.; Anderson, M.; Belknap, D.; Bellinger, J. N.; Borrello, L.; Bradley, D.; Carlsmith, D.; Cepeda, M.; Crotty, I.; Dasu, S.; Feyzi, F.; Friis, E.; Gorski, T.; Gray, L.; Grogg, K. S.; Grothe, M.; Hall-Wilton, R.; Herndon, M.; Hervé, A.; Klabbers, P.; Klukas, J.; Lackey, J.; Lanaro, A.; Lazaridis, C.; Leonard, J.; Loveless, R.; Lusin, S.; Magrans de Abril, M.; Maier, W.; Mohapatra, A.; Ojalvo, I.; Palmonari, F.; Pierro, G. A.; Reeder, D.; Ross, I.; Savin, A.; Smith, W. H.; Swanson, J.; Wenman, D. 2012-09-01 Results are presented from searches for the standard model Higgs boson in proton-proton collisions at \\sqrt{s}=7 $and 8 TeV in the Compact Muon Solenoid experiment at the LHC, using data samples corresponding to integrated luminosities of up to 5.1 inverse femtobarns at 7 TeV and 5.3 inverse femtobarns at 8 TeV. The search is performed in five decay modes: gamma gamma, ZZ, WW, tau tau, and b b-bar. An excess of events is observed above the expected background, with a local significance of 5.0 standard deviations, at a mass near 125 GeV, signalling the production of a new particle. The expected significance for a standard model Higgs boson of that mass is 5.8 standard deviations. The excess is most significant in the two decay modes with the best mass resolution, gamma gamma and ZZ; a fit to these signals gives a mass of 125.3 +/- 0.4 (stat.) +/- 0.5 (syst.) GeV. The decay to two photons indicates that the new particle is a boson with spin different from one. 9. New results from the Mainz neutrino mass experiment and perspective of a new large tritium-β-spectrometer International Nuclear Information System (INIS) Bonn, J.; Bornschein, B.; Bornschein, L.; Fickinger, L.; Kraus, Ch.; Otten, E.W.; Ulrich, H.; Weinheimer, Ch.; Kazachenko, O.; Kovalik, A. 2001-01-01 The Mainz neutrino mass experiment investigates the endpoint region of the tritium β decay spectrum to determine the mass of the electron antineutrino. By the recent upgrade the former problem of de-wetting T 2 films has been solved and the signal-to-background-ratio was improved by a factor of 10. The latest measurement leads to m ν 2 = -1.1 ± 2.6 stat ± 1.8 sys eV 2 /c 4 (preliminary), which corresponds to an upper limit of m ν 2 (95 % C.L.) (preliminary). Some indication for the anomaly, reported by the Troitsk group, was found, but its postulated half year period is contradicted by our data. The perspectives of a new Large Tritium-β-Spectrometer to reach sub eV sensitivity will be presented. (authors) 10. Nike Experiment to Observe Strong Areal Mass Oscillations in a Rippled Target Hit by a Short Laser Pulse Science.gov (United States) Aglitskiy, Y.; Karasik, M.; Velikovich, A. L.; Serlin, V.; Weaver, J. L.; Kessler, T. J.; Schmitt, A. J.; Obenschain, S. P.; Metzler, N.; Oh, J. 2010-11-01 When a short (sub-ns) laser pulse deposits finite energy in a target, the shock wave launched into it is immediately followed by a rarefaction wave. If the irradiated surface is rippled, theory and simulations predict strong oscillations of the areal mass perturbation amplitude in the target [A. L. Velikovich et al., Phys. Plasmas 10, 3270 (2003).] The first experiment designed to observe this effect has become possible by adding short-driving-pulse capability to the Nike laser, and has been scheduled for the fall of 2010. Simulations show that while the driving pulse of 0.3 ns is on, the areal mass perturbation amplitude grows by a factor ˜2 due to ablative Richtmyer-Meshkov instability. It then decreases, reverses phase, and reaches another maximum, also about twice its initial value, shortly after the shock breakout at the rear target surface. This signature behavior is observable with the monochromatic x-ray imaging diagnostics fielded on Nike. 11. Integration of Family Planning Counselling to Mass Screening Campaign for Cervical Cancer: Experience from Guinea Directory of Open Access Journals (Sweden) D. W. A. Leno 2018-01-01 Full Text Available Aim. To assess feasibility of integrating family planning counselling into mass screening for cervical cancer in Guinea. Methodology. This was a descriptive cross-sectional study conducted over a month in Guinea regional capital cities. The targeted population comprised women aged 15 to 49 years. Nearly 4000 women were expected for the screening campaigns that utilized VIA and VIL methods with confirmation of positive tests through biopsy. A local treatment was immediately performed when the patient was eligible. Results. Overall 5673 women aged 15 to 60 years were received, a surplus of 42% of the expected population. 92.3% of women were aged 15–49 years and 90.1% were 25–49 years. Long-acting methods were the most utilized (89.2% of family planning users. 154 precancerous and cancerous lesions were screened, a global positivity rate of 2.7%. Conclusion. Integration of counselling and family planning services provision during cervical cancer mass screening is a feasible strategy. A cost-effective analysis of this approach would help a better planning of future campaigns and its replication in other contexts. 12. Sensitivity of the Top quark mass measurement with the CMS experiment at LHC using t-tbar multijet simulated events CERN Document Server Codispoti, Giuseppe This thesis comes after a strong contribution on the realization of the CMS computing system, which can be seen as a relevant part of the experiment itself. A physics analysis completes the road from Monte Carlo production and analysis tools realization to the final physics study which is the actual goal of the experiment. The topic of physics work of this thesis is the study of tt events fully hadronic decay in the CMS experiment. A multi-jet trigger has been provided to fix a reasonable starting point, reducing the multi-jet sample to the nominal trigger rate. An offline selection has been provided to reduce the S/B ratio. The b-tag is applied to provide a further S/B improvement. The selection is applied to the background sample and to the samples generated at different top quark masses. The top quark mass candidate is reconstructed for all those samples using a kinematic fitter. The resulting distributions are used to build p.d.f.’s, interpolating them with a continuous arbitrary curve. These curves are... 13. Mass transfer in corrugated-plate membrane modules. I. Hyperfiltration experiments NARCIS (Netherlands) van der Waal, M.J.; Racz, I.G. 1989-01-01 The application of corrugations as turbulence promoters in membrane filtration was studied. This study showed that it is possible to deform an originally flat membrane to a corrugated shape without damaging it. In hyperfiltration experiments using corrugated cellulose acetate membranes it was found 14. Mass transfer in corrugated-plate membrane modules. II. Ultrafiltration experiments NARCIS (Netherlands) van der Waal, M.J.; Stevanovic, S.; Racz, I.G. 1989-01-01 The application of corrugations as turbulence promoters in membrane filtration was studied. In ultrafiltration experiments with polysulfone membranes using Dextran T70 as solute, it was found that the corrugations result in reduced energy consumption or pressure drop compared with flat membranes at 15. Wave-packet treatment of reactor neutrino oscillation experiments and its implications on determining the neutrino mass hierarchy Energy Technology Data Exchange (ETDEWEB) Chan, Yat-Long; Chu, M.C.; Xu, Jianyi [The Chinese University of Hong Kong, Department of Physics, Shatin (China); Tsui, Ka Ming [University of Tokyo, RCCN, ICRR, Kashiwa, Chiba (Japan); Wong, Chan Fai [Sun Yat-Sen University, Guangzhou (China) 2016-06-15 We derive the neutrino flavor transition probabilities with the neutrino treated as a wave packet. The decoherence and dispersion effects from the wave-packet treatment show up as damping and phase-shifting of the plane-wave neutrino oscillation patterns. If the energy uncertainty in the initial neutrino wave packet is larger than around 0.01 of the neutrino energy, the decoherence and dispersion effects would degrade the sensitivity of reactor neutrino experiments to mass hierarchy measurement to lower than 3 σ confidence level. (orig.) 16. Bestimmung der Mas'se des neutralen Bs-Mesons mit dem ALEPH-Experiment CERN Document Server Stehle, M 2001-01-01 Gegenstand der vorliegenden Arbeit ist die Bestimmung der Masse des neutralen Bs-Mesons. Dazu wurden B~-Mesonen in den beiden Zerfallskanalen B~ -+ J/W P und B~ -+ W(2S) P rekonstruiert, wobei die Subresonanzen in den Zerfallsmoden J /w -+ l+ l-, W(2S) -+ l+ l- und P -+ K+ K- untersucht wurden. Diese beiden Kanale werden auf Grund ihrer eindeutigen Signatur auch als "goldene Kanale" bezeichnet und eignen sich deshalb sehr gut fur eine exklusive Rekonstruktion, wie sie hier angewendet wurde. Grundlage der Analyse waren ca. 4 Millionen hadronische ZO-Zerfalle, die in den Jahren 1991 1995 mit dem ALEPH-Detektor am e+e--Speicherring LEP am CERN aufgezeichnet wurden. Die zwischenzeitliche Reprozessierung der Daten ermoglichte eine prazisere und effizientere Rekonstruktion als das in fruheren Messungen der Fall war. Wegen der niedrigen Verzweigungsverhaltnisse der untersuchten Zerfallsmoden wurden nur wenige einzelne Ereignisse in den Daten erwartet. Die selektierten Kandidaten wurden durch Schnitte in mehreren Ere... 17. Atmospheric proton and deuterium energy spectra determination with the MASS2 experiment Energy Technology Data Exchange (ETDEWEB) Grimani, C.; Brunetti, M.T.; Codino, A.; Finetti, N. [Perugia Univ. (Italy)]|[INFN, Perugia (Italy); Papini, P.; Massimo Brancaccio, F. [Florence Univ. (Italy)]|[INFN, Florence (Italy); Basini, G.; Bongiorno, F. [INFN, Laboratori Nazionali di Frascati, Rome (Italy); Golden, R.L. [New Mexico State Univ., Las Cruces, NM (United States). Particle Astrophysics Lab.; Hof, M. [Siegen Univ. (Germany). Fachbereich Physik 1995-09-01 The energy spectra of atmospheric-secondary protons and deuterium nuclei have been measured during the September 23, 1991, balloon flight of the NMSU/Wizard - MASS2 instrument. The apparatus was launched from Fort Sumner, New Mexico. The geomagnetic cutoff at the launch site is about 4.5 GV/c. The instrument was flown for 9.8 hours at an altitude of over 100,000 feet. Particles detected below the geomagnetic cutoff have been produced mainly by the interactions of the primary cosmic rays with the atmosphere. The measurement of cosmic ray energy spectra below the geomagnetic cutoff provide direct insights into the particle production mechanism and allows comparison to atmospheric cascade calculations. 18. Low-Mass Dark Matter Search with the DarkSide-50 Experiment Energy Technology Data Exchange (ETDEWEB) Agnes, P.; et al. 2018-02-20 We present the results of a search for dark matter WIMPs in the mass range below 20 GeV/c^2 using a target of low-radioactivity argon. The data were obtained using the DarkSide-50 apparatus at Laboratori Nazionali del Gran Sasso (LNGS). The analysis is based on the ionization signal, for which the DarkSide-50 time projection chamber is fully efficient at 0.1 keVee. The observed rate in the detector at 0.5 keVee is about 1.5 events/keVee/kg/day and is almost entirely accounted for by known background sources. We obtain a 90% C.L. exclusion limit above 1.8 GeV/c^2 for the spin-independent cross section of dark matter WIMPs on nucleons, extending the exclusion region for dark matter below previous limits in the range 1.8-6 GeV/c^2. 19. Simultaneous measurement of top quark mass and jet energy scale using template fits at the CMS experiment Energy Technology Data Exchange (ETDEWEB) Naumann-Emme, Sebastian 2011-07-15 In this thesis, pairs of top quarks produced in proton-proton collisions at a center-of-mass energy of 7 TeV and decaying in the muon+jets channel t anti t {yields} (b{mu}{nu})(bqq{sup '}) are analyzed using data that were recorded by the CMS detector in the year 2010 and correspond to an integrated luminosity of 35.9 pb{sup -1}. A sample of 78 events is selected by requiring exactly one isolated muon and at least four jets, two of them being identified as jets from the decay of b quarks. Given these selection criteria, the expected fraction of t anti t events is 94%. The trijet mass, M3, and the dijet mass, M2, are reconstructed, taking into account the b-tagging information. M3 and M2 are estimators of the masses of hadronically decaying top quarks and the corresponding W bosons, respectively. Templates for M2 and for the event-wise mass difference {delta}M{sub 32}=M3-M2 are parametrized as linear functions of the top quark mass, m{sub t}, and the jet energy scale (JES). Based on the precise knowledge of the W boson mass, M2 provides a strong handle on the energy scale of jets from light quarks. The reconstructed M2 and {delta}M{sub 32} in data are compared to the template functions from simulation in a combined likelihood fit. The overall JES in the selected sample is found to be 1.048{+-}0.040(stat){+-}0.015(syst) relative to the simulated JES and the measured m{sub t} is 167.8{+-}7.1(stat+JES){+-}3.1(syst) GeV. This is one of the first measurements of m{sub t} at the Large Hadron Collider. Furthermore, the JES measurement is an important input for the commissioning of the CMS experiment for the upcoming measurements with more data in the near future. (orig.) 20. Simultaneous measurement of top quark mass and jet energy scale using template fits at the CMS experiment International Nuclear Information System (INIS) Naumann-Emme, Sebastian 2011-07-01 In this thesis, pairs of top quarks produced in proton-proton collisions at a center-of-mass energy of 7 TeV and decaying in the muon+jets channel t anti t → (bμν)(bqq ' ) are analyzed using data that were recorded by the CMS detector in the year 2010 and correspond to an integrated luminosity of 35.9 pb -1 . A sample of 78 events is selected by requiring exactly one isolated muon and at least four jets, two of them being identified as jets from the decay of b quarks. Given these selection criteria, the expected fraction of t anti t events is 94%. The trijet mass, M3, and the dijet mass, M2, are reconstructed, taking into account the b-tagging information. M3 and M2 are estimators of the masses of hadronically decaying top quarks and the corresponding W bosons, respectively. Templates for M2 and for the event-wise mass difference ΔM 32 =M3-M2 are parametrized as linear functions of the top quark mass, m t , and the jet energy scale (JES). Based on the precise knowledge of the W boson mass, M2 provides a strong handle on the energy scale of jets from light quarks. The reconstructed M2 and ΔM 32 in data are compared to the template functions from simulation in a combined likelihood fit. The overall JES in the selected sample is found to be 1.048±0.040(stat)±0.015(syst) relative to the simulated JES and the measured m t is 167.8±7.1(stat+JES)±3.1(syst) GeV. This is one of the first measurements of m t at the Large Hadron Collider. Furthermore, the JES measurement is an important input for the commissioning of the CMS experiment for the upcoming measurements with more data in the near future. (orig.) 1. Determination of Total Arsenic and Speciation in Apple Juice by Liquid Chromatography-Inductively Coupled Plasma Mass Spectrometry: An Experiment for the Analytical Chemistry Laboratory Science.gov (United States) He, Ping; Colon, Luis A.; Aga, Diana S. 2016-01-01 A two-part laboratory experiment was designed for upper-level analytical chemistry students to provide hands-on experience in the use of high performance liquid chromatography (HPLC) for separation and inductively coupled plasma mass spectrometry (ICP-MS) for detection. In the first part of the experiment, the students analyze total arsenic in… 2. Experiments and CFD Modelling of Turbulent Mass Transfer in a Mixing Channel DEFF Research Database (Denmark) Hjertager Osenbroch, Lene Kristin; Hjertager, Bjørn H.; Solberg, Tron 2006-01-01 . Three different flow cases are studied. The 2D numerical predictions of the mixing channel show that none of the k-ε turbulence models tested is suitable for the flow cases studied here. The turbulent Schmidt number is reduced to obtain a better agreement between measured and predicted mean......Experiments are carried out for passive mixing in order to obtain local mean and turbulent velocities and concentrations. The mixing takes place in a square channel with two inlets separated by a block. A combined PIV/PLIF technique is used to obtain instantaneous velocity and concentration fields...... and fluctuating concentrations. The multi-peak presumed PDF mixing model is tested.... 3. Radiation safety in the Moscow region: Experience of cooperation SUE SIA Radon with mass media and public Energy Technology Data Exchange (ETDEWEB) Grishin, O.; Rakov, S. [SUE SIA RADON, Moscow (Russian Federation) 2012-07-01 Radiation safety in the present period acquires complex character and closely corresponds with other elements of social and political process, with various spheres of public life. After earthquake in Japan in March 11, 2011 and emergency on the nuclear power plant Fukushima-1 the theme of radiating safety of megapolises has today become urgent, as never. Provision of radiation safety of the Moscow region, taking into account a number of factors, is an important problem in a context of modernization state. Today in sphere of radio ecological safety there are certain achievements: new monitoring systems are developed, technological processes are improved, new information-communicative channels of interaction with mass media and public are formed. Information policy of enterprises the functioning of which is connected with provision of safe ecological conditions and its monitoring is focused on constant and duly informing of public through mass-media. Experience and technologies of interaction with mass-media and public of Moscow State Unitary-Enterprise- united ecological, scientific and research centre of decontamination of radioactive waste and environmental protection (State Unitarian Enterprise, Scientific and Industrial Association Radon, SUE SIA RADON) is submitted in the article. (author) 4. Hadronic recoil in the W boson production at LHC for a W mass measurement with the CMS experiment CERN Document Server AUTHOR|(CDS)2099463 In the first chapter of this work, an overall picture of the theoretical basis is pre- sented. Starting from the foundations of the Standard Model, Higgs mechanism and electroweak symmetry breaking are introduced, focusing on their role of providing SM gauge boson with masses. The important facts of electroweak precision test are also introduced in the last part of the first chapter. After an overview of the Large Hadron Collider (LHC), which is currently operating at CERN, the second part of this work describes the Compact Muon Solenoid (CMS) experiment, aimed to explore in depth particle physics up to the TeV scale: the main features of the subdetectors are briefly described, together with the reconstruction al- gorithms; focus has been put mostly on those features of interest for W mass physics. The third chapter is devoted to discuss the past and the on going efforts for the W boson mass measurement. The original work developed during the thesis is fully discussed in chapters four, five and six. Two are t... 5. Precision measurement of the mass difference between light nuclei and anti-nuclei with the ALICE experiment at the LHC CERN Document Server 2015-01-01 We report on a measurement of the difference$\\Delta \\mu = \\Delta (m / |z|)$between the mass-over-charge ratio of deuteron (d) and anti-deuteron ($\\overline{\\rm d}$), and$^3{\\rm He}$and$^3{\\overline{\\rm He}}$nuclei, carried out with ALICE (A Large Ion Collider Experiment) in Pb-Pb collisions at a centre-of-mass energy per nucleon pair$\\sqrt{s_{\\rm NN}}=~2.76~\\rm{TeV}$. Our measurement yields${\\Delta \\mu}_{\\rm{d\\bar{d}}}/{\\mu}_{\\rm{d}} = [0.9 \\pm 0.5 (\\rm{stat.}) \\pm 1.4 (\\rm{syst.})] \\times 10^{-4}$and${\\Delta \\mu}_{\\rm{^{3}He ^{3}\\overline{He}}}/{\\mu}_{\\rm{^{3}He}} = [-1.2 \\pm 0.9 (\\rm{stat.}) \\pm 1.0 (\\rm{syst.})] \\times 10^{-3}$. Combining these results with existing measurements of the masses of the (anti-)nucleons, the relative binding energy differences are extracted,${\\Delta \\varepsilon}_{\\rm{d\\bar{d}}}/{\\varepsilon}_{\\rm{d}} = -0.04 \\pm 0.05(\\rm{stat.}) \\pm 0.12(\\rm{syst.})$and${\\Delta \\varepsilon}_{\\rm{^{3}He ^{3}\\overline{He}}}/{\\varepsilon}_{\\rm{^{3}He}} = 0.24 \\pm 0.16(\\rm{stat.}) \\pm...
6. Sensitivity analysis of efficiency thermal energy storage on selected rock mass and grout parameters using design of experiment method
International Nuclear Information System (INIS)
Wołoszyn, Jerzy; Gołaś, Andrzej
2014-01-01
Highlights: • Paper propose a new methodology to sensitivity study of underground thermal storage. • Using MDF model and DOE technique significantly shorter of calculations time. • Calculation of one time step was equal to approximately 57 s. • Sensitivity study cover five thermo-physical parameters. • Conductivity of rock mass and grout material have a significant impact on efficiency. - Abstract: The aim of this study was to investigate the influence of selected parameters on the efficiency of underground thermal energy storage. In this paper, besides thermal conductivity, the effect of such parameters as specific heat, density of the rock mass, thermal conductivity and specific heat of grout material was investigated. Implementation of this objective requires the use of an efficient computational method. The aim of the research was achieved by using a new numerical model, Multi Degree of Freedom (MDF), as developed by the authors and Design of Experiment (DoE) techniques with a response surface. The presented methodology can significantly reduce the time that is needed for research and to determine the effect of various parameters on the efficiency of underground thermal energy storage. Preliminary results of the research confirmed that thermal conductivity of the rock mass has the greatest impact on the efficiency of underground thermal energy storage, and that other parameters also play quite significant role
7. Radiation safety in the Moscow region: Experience of cooperation SUE SIA Radon with mass media and public
International Nuclear Information System (INIS)
Grishin, O.; Rakov, S.
2012-01-01
Radiation safety in the present period acquires complex character and closely corresponds with other elements of social and political process, with various spheres of public life. After earthquake in Japan in March 11, 2011 and emergency on the nuclear power plant Fukushima-1 the theme of radiating safety of megapolises has today become urgent, as never. Provision of radiation safety of the Moscow region, taking into account a number of factors, is an important problem in a context of modernization state. Today in sphere of radio ecological safety there are certain achievements: new monitoring systems are developed, technological processes are improved, new information-communicative channels of interaction with mass media and public are formed. Information policy of enterprises the functioning of which is connected with provision of safe ecological conditions and its monitoring is focused on constant and duly informing of public through mass-media. Experience and technologies of interaction with mass-media and public of Moscow State Unitary-Enterprise- united ecological, scientific and research centre of decontamination of radioactive waste and environmental protection (State Unitarian Enterprise, Scientific and Industrial Association Radon, SUE SIA RADON) is submitted in the article. (author)
8. Comparison of Pu isotopic composition between gamma and mass spectrometry: Experience from IAEA-SAL
International Nuclear Information System (INIS)
Parus, J.L.; Raab, W.
1998-01-01
About 2000 Pu containing samples have been analysed during the last 8 years at SAL using gamma spectrometry (GS) in parallel with mass spectrometry (MS). Four different detectors have been used for the measurement of gamma-ray spectra and several versions of the MGA program have been used for spectra evaluation. The results of Pu isotopic composition obtained by both methods have neem systematically compared. Attempts to improve the agreement between GS and MS are described. This was done by adjustment of the emission probabilities for some gamma energies and the development of a new correlation equation for 242 Pu. These improvements have been applied for evaluation of two sets containing 320 and 404 samples, respectively analysed in 1991 and in 1992-93. The mean differences and their standard deviations between MS and GS were calculated, showing mean relative differences for 238-241 Pu isotopes in the range from 0.1 to 0.5% with standard deviations within ± 0.4 to ±1%. For 242 Pu these values are about 0.5% and ± 5%, respectively. (author)
9. qcML: an exchange format for quality control metrics from mass spectrometry experiments.
Science.gov (United States)
Walzer, Mathias; Pernas, Lucia Espona; Nasso, Sara; Bittremieux, Wout; Nahnsen, Sven; Kelchtermans, Pieter; Pichler, Peter; van den Toorn, Henk W P; Staes, An; Vandenbussche, Jonathan; Mazanek, Michael; Taus, Thomas; Scheltema, Richard A; Kelstrup, Christian D; Gatto, Laurent; van Breukelen, Bas; Aiche, Stephan; Valkenborg, Dirk; Laukens, Kris; Lilley, Kathryn S; Olsen, Jesper V; Heck, Albert J R; Mechtler, Karl; Aebersold, Ruedi; Gevaert, Kris; Vizcaíno, Juan Antonio; Hermjakob, Henning; Kohlbacher, Oliver; Martens, Lennart
2014-08-01
Quality control is increasingly recognized as a crucial aspect of mass spectrometry based proteomics. Several recent papers discuss relevant parameters for quality control and present applications to extract these from the instrumental raw data. What has been missing, however, is a standard data exchange format for reporting these performance metrics. We therefore developed the qcML format, an XML-based standard that follows the design principles of the related mzML, mzIdentML, mzQuantML, and TraML standards from the HUPO-PSI (Proteomics Standards Initiative). In addition to the XML format, we also provide tools for the calculation of a wide range of quality metrics as well as a database format and interconversion tools, so that existing LIMS systems can easily add relational storage of the quality control data to their existing schema. We here describe the qcML specification, along with possible use cases and an illustrative example of the subsequent analysis possibilities. All information about qcML is available at http://code.google.com/p/qcml. © 2014 by The American Society for Biochemistry and Molecular Biology, Inc.
10. qcML: An Exchange Format for Quality Control Metrics from Mass Spectrometry Experiments*
Science.gov (United States)
Walzer, Mathias; Pernas, Lucia Espona; Nasso, Sara; Bittremieux, Wout; Nahnsen, Sven; Kelchtermans, Pieter; Pichler, Peter; van den Toorn, Henk W. P.; Staes, An; Vandenbussche, Jonathan; Mazanek, Michael; Taus, Thomas; Scheltema, Richard A.; Kelstrup, Christian D.; Gatto, Laurent; van Breukelen, Bas; Aiche, Stephan; Valkenborg, Dirk; Laukens, Kris; Lilley, Kathryn S.; Olsen, Jesper V.; Heck, Albert J. R.; Mechtler, Karl; Aebersold, Ruedi; Gevaert, Kris; Vizcaíno, Juan Antonio; Hermjakob, Henning; Kohlbacher, Oliver; Martens, Lennart
2014-01-01
Quality control is increasingly recognized as a crucial aspect of mass spectrometry based proteomics. Several recent papers discuss relevant parameters for quality control and present applications to extract these from the instrumental raw data. What has been missing, however, is a standard data exchange format for reporting these performance metrics. We therefore developed the qcML format, an XML-based standard that follows the design principles of the related mzML, mzIdentML, mzQuantML, and TraML standards from the HUPO-PSI (Proteomics Standards Initiative). In addition to the XML format, we also provide tools for the calculation of a wide range of quality metrics as well as a database format and interconversion tools, so that existing LIMS systems can easily add relational storage of the quality control data to their existing schema. We here describe the qcML specification, along with possible use cases and an illustrative example of the subsequent analysis possibilities. All information about qcML is available at http://code.google.com/p/qcml. PMID:24760958
11. Purification and Quantification of an Isomeric Compound in a Mixture by Collisional Excitation in Multistage Mass Spectrometry Experiments.
Science.gov (United States)
Jeanne Dit Fouque, Dany; Maroto, Alicia; Memboeuf, Antony
2016-11-15
The differentiation, characterization, and quantification of isomers and/or isobars in mixtures is a recurrent problem in mass spectrometry and more generally in analytical chemistry. Here we present a new strategy to assess the purity of a compound that is susceptible to be contaminated with another isomeric side-product in trace levels. Providing one of the isomers is available as pure sample, this new strategy allows the detection of isomeric contamination. This is done thanks to a "gas-phase collisional purification" inside an ion trap mass spectrometer paving the way for an improved analysis of at least similar samples. This strategy consists in using collision induced dissociation (CID) multistage mass spectrometry (MS 2 and MS 3 ) experiments and the survival yield (SY) technique. It has been successfully applied to mixtures of cyclic poly( L -lactide) (PLA) with increasing amounts of its linear topological isomer. Purification in gas phase of PLA mixtures was established based on SY curves obtained in MS 3 mode: all samples gave rise to the same SY curve corresponding then to the pure cyclic component. This new strategy was sensitive enough to detect traces of linear PLA (<3%) in a sample of cyclic PLA that was supposedly pure according to other characterization techniques ( 1 H NMR, MALDI-HRMS, and size-exclusion chromatography). Moreover, in this case, the presence of linear isomer was undetectable according to MS/MS or MS/MS/MS analysis only as fragment ions are also of the same m/z values. This type of approach could easily be implemented in hyphenated mass spectrometric techniques to improve the structural and quantitative analysis of complex samples.
12. Outcomes of robot-assisted simple enucleation of renal masses: A single European center experience.
Science.gov (United States)
Matei, Deliu Victor; Vartolomei, Mihai Dorin; Musi, Gennaro; Renne, Giuseppe; Tringali, Valeria Maria Lucia; Mistretta, Francesco Alessandro; Delor, Maurizio; Russo, Andrea; Cioffi, Antonio; Bianchi, Roberto; Cozzi, Gabriele; Di Trapani, Ettore; Bottero, Danilo; Cordima, Giovanni; Lucarelli, Giuseppe; Ferro, Matteo; de Cobelli, Ottavio
2017-05-01
The aim of this study was to assess the ability of pre-and intraoperative parameters, to predict the risk of perioperative complications after robot-assisted laparoscopic simple enucleation (RASE) of renal masses, and to evaluate the rate of trifecta achievement of this approach stratifying the cohort according to the use of ischemia during the enucleation.From April 2009 to June 2016, 129 patients underwent RASE at our Institution. We stratified the procedures in 2 groups: clamping and clamp-less RASE. After RASE, all specimens were retrospectively reviewed to assess the surface-intermediate-base (SIB) scoring system. Patients were followed-up according to the European Association of Urology guidelines recommendations. All pre-, intra-, and postoperative outcomes were prospectively collected in a customized database and retrospectively analyzed.A total of 112 (86.8%) patients underwent a pure RASE and 17 (13.2%) had a hybrid according to SIB classification system. The mean age was 61.17 years. In 21 patients (16.3%), complications occurred, 13 (61.9%) were Clavien 1 and 2, while 8 were Clavien 3a and b complications. Statistical significant association with complications was found in patients with American Society of Anestesiology (ASA) score 3 (44.5%, P = .04), longer mean operative time (OT) 195 versus 161.36 minutes (P =.03), mean postoperative hemoglobin (Hb) 10.1 versus 11.8 (P 3, longer OT, and ΔHb. RASE is suitable for the clamp-less approach, which allows to perform easier the pure enucleation (SIB 0) and to obtain higher rates of trifecta outcomes.
13. Percutaneous CT-guided radiofrequency ablation of solitary small renal masses. A single center experience
Energy Technology Data Exchange (ETDEWEB)
Pieper, C.C.; Fischer, S.; Strunk, H.; Meyer, C.; Thomas, D.; Willinek, W.A.; Schild, H. [Univ. Bonn (Germany). Dept. of Radiology; Hauser, S. [Univ. Bonn (Germany). Dept. of Urology; Nadal, J. [Univ. Bonn (Germany). Inst. for Medical Biometry; Wilhelm, K. [Johanniter Hospital Bonn (Germany). Dept. of Radiology
2015-07-15
To analyze the outcome of patients undergoing percutaneous CT-guided radiofrequency ablation (RFA) of small renal masses (SRM) at a single center during a ten-year time period. Patient records of renal RFAs (07/2003 - 11/2013) were reviewed. Indications were SRM suspicious of malignancy on imaging and one of the following: severe comorbidity; old age; solitary kidney; impaired renal function; patient wish. Biopsy was performed at the time of RFA. Patients were excluded if no follow-up was available. Patient and procedural characteristics were recorded. Survival rates were calculated using the Kaplan-Meier's method and compared with log-rank or cox tests. 38 patients (16 females, mean age 70.0 years [range 52 - 87]) presenting with a solitary SRM were included in the study. Biopsy showed malignancy in 29 patients; 9 had benign tumors. 26 patients suffered from cardiovascular, respiratory or hepatic comorbidities. Technical success (complete ablation on first follow-up) was achieved in 95 % of cases. Two major complications (bowel perforation; hematothorax) occurred. The 3- and 7-year overall survival (OS) [any cause] rates were 73.4 ± 0.8 % and 50.3 ± 1.0 %, respectively (mean follow-up 54.6 months, range 1 - 127). 4 recurrences and 2 metastases were observed. The presence of comorbidities was the only independent predictor of OS. There was no difference in survival between patients with benign and malignant tumors. RFA of SRM is successful in a large percentage of cases with a low complication rate and durable local control. As RFA is typically performed in multimorbid patients, overall survival seems to depend primarily on comorbidities rather than cancer progression.
14. Is mass drug administration against lymphatic filariasis required in urban settings? The experience in Kano, Nigeria.
Science.gov (United States)
Pam, Dung D; de Souza, Dziedzom K; D'Souza, Susan; Opoku, Millicent; Sanda, Safiya; Nazaradden, Ibrahim; Anagbogu, Ifeoma N; Okoronkwo, Chukwu; Davies, Emmanuel; Elhassan, Elisabeth; Molyneux, David H; Bockarie, Moses J; Koudou, Benjamin G
2017-10-01
The Global Programme to Eliminate Lymphatic Filariasis (GPELF), launched in 2000, has the target of eliminating the disease as a public health problem by the year 2020. The strategy adopted is mass drug administration (MDA) to all eligible individuals in endemic communities and the implementation of measures to reduce the morbidity of those suffering from chronic disease. Success has been recorded in many rural endemic communities in which elimination efforts have centered. However, implementation has been challenging in several urban African cities. The large cities of West Africa, exemplified in Nigeria in Kano are challenging for LF elimination program because reaching 65% therapeutic coverage during MDA is difficult. There is therefore a need to define a strategy which could complement MDA. Thus, in Kano State, Nigeria, while LF MDA had reached 33 of the 44 Local Government Areas (LGAs) there remained eleven 'urban' LGAs which had not been covered by MDA. Given the challenges of achieving at least 65% coverage during MDA implementation over several years in order to achieve elimination, it may be challenging to eliminate LF in such settings. In order to plan the LF control activities, this study was undertaken to confirm the LF infection prevalence in the human and mosquito populations in three urban LGAs. The prevalence of circulating filarial antigen (CFA) of Wuchereria bancrofti was assessed by an immuno-chromatography test (ICT) in 981 people in three urban LGAs of Kano state, Nigeria. Mosquitoes were collected over a period of 4 months from May to August 2015 using exit traps, gravid traps and pyrethrum knock-down spray sheet collections (PSC) in different households. A proportion of mosquitoes were analyzed for W. bancrofti, using dissection, loop-mediated isothermal amplification (LAMP) assay and conventional polymerase chain reaction (PCR). The results showed that none of the 981 subjects (constituted of West Africa may not be of significant importance
15. Distributed mass data acquisition system based on PCs and windows NT for LHD fusion plasma experiment
International Nuclear Information System (INIS)
Nakanishi, H.; Kojima, M.; Ohsuna, M.; Komada, S.; Emoto, M.; Sugisaki, H.; Sudo, S.
2000-12-01
A new data acquisition and management system has been developed for the LHD experiment. It has the capability to process 100 MB - 1 GB raw data within a few tens seconds after every plasma discharge. It employs wholly distributed and loosely-tied parallel tasking structure through a fast network, and the cluster of the distributed database severs seems to be a virtual macro-machine as a whole. A PC/Windows NT computer is installed for each diagnostics data acquisition of about 30 kinds, and it controls CAMAC digitizers through the optical SCSI extenders. The diagnostic timing system consists of some kinds of VME modules that are installed to remotely control the diagnostic devices in real-time. They can, as a whole system, distribute the synchronous sampling clocks and programmable triggers for measurement digitizers. The data retrieving terminals can access database as application service clients, and are functionally separated from the data acquisition severs by way of the switching Ethernet. (author)
16. Is mass drug administration against lymphatic filariasis required in urban settings? The experience in Kano, Nigeria.
Directory of Open Access Journals (Sweden)
Dung D Pam
2017-10-01
Full Text Available The Global Programme to Eliminate Lymphatic Filariasis (GPELF, launched in 2000, has the target of eliminating the disease as a public health problem by the year 2020. The strategy adopted is mass drug administration (MDA to all eligible individuals in endemic communities and the implementation of measures to reduce the morbidity of those suffering from chronic disease. Success has been recorded in many rural endemic communities in which elimination efforts have centered. However, implementation has been challenging in several urban African cities. The large cities of West Africa, exemplified in Nigeria in Kano are challenging for LF elimination program because reaching 65% therapeutic coverage during MDA is difficult. There is therefore a need to define a strategy which could complement MDA. Thus, in Kano State, Nigeria, while LF MDA had reached 33 of the 44 Local Government Areas (LGAs there remained eleven 'urban' LGAs which had not been covered by MDA. Given the challenges of achieving at least 65% coverage during MDA implementation over several years in order to achieve elimination, it may be challenging to eliminate LF in such settings. In order to plan the LF control activities, this study was undertaken to confirm the LF infection prevalence in the human and mosquito populations in three urban LGAs.The prevalence of circulating filarial antigen (CFA of Wuchereria bancrofti was assessed by an immuno-chromatography test (ICT in 981 people in three urban LGAs of Kano state, Nigeria. Mosquitoes were collected over a period of 4 months from May to August 2015 using exit traps, gravid traps and pyrethrum knock-down spray sheet collections (PSC in different households. A proportion of mosquitoes were analyzed for W. bancrofti, using dissection, loop-mediated isothermal amplification (LAMP assay and conventional polymerase chain reaction (PCR.The results showed that none of the 981 subjects (constituted of <21% of children 5-10 years old tested
17. Monitoring system for the GRID Monte Carlo mass production in the H1 experiment at DESY
International Nuclear Information System (INIS)
Bystritskaya, Elena; Fomenko, Alexander; Gogitidze, Nelly; Lobodzinski, Bogdan
2014-01-01
The H1 Virtual Organization (VO), as one of the small VOs, employs most components of the EMI or gLite Middleware. In this framework, a monitoring system is designed for the H1 Experiment to identify and recognize within the GRID the best suitable resources for execution of CPU-time consuming Monte Carlo (MC) simulation tasks (jobs). Monitored resources are Computer Elements (CEs), Storage Elements (SEs), WMS-servers (WMSs), CernVM File System (CVMFS) available to the VO HONE and local GRID User Interfaces (UIs). The general principle of monitoring GRID elements is based on the execution of short test jobs on different CE queues using submission through various WMSs and directly to the CREAM-CEs as well. Real H1 MC Production jobs with a small number of events are used to perform the tests. Test jobs are periodically submitted into GRID queues, the status of these jobs is checked, output files of completed jobs are retrieved, the result of each job is analyzed and the waiting time and run time are derived. Using this information, the status of the GRID elements is estimated and the most suitable ones are included in the automatically generated configuration files for use in the H1 MC production. The monitoring system allows for identification of problems in the GRID sites and promptly reacts on it (for example by sending GGUS (Global Grid User Support) trouble tickets). The system can easily be adapted to identify the optimal resources for tasks other than MC production, simply by changing to the relevant test jobs. The monitoring system is written mostly in Python and Perl with insertion of a few shell scripts. In addition to the test monitoring system we use information from real production jobs to monitor the availability and quality of the GRID resources. The monitoring tools register the number of job resubmissions, the percentage of failed and finished jobs relative to all jobs on the CEs and determine the average values of waiting and running time for the
18. Developing antitobacco mass media campaign messages in a low-resource setting: experience from the Kingdom of Tonga.
Science.gov (United States)
Sugden, C; Phongsavan, P; Gloede, S; Filiai, S; Tongamana, V O
2017-05-01
Tobacco use has become the leading cause of preventable death in Tonga, a small island nation in the South Pacific. One pragmatic and economical strategy to address this worrying trend is to adapt effective antitobacco mass media materials developed in high-income countries for local audiences. Using Tonga as an example, this paper shares the practical steps involved in adapting antitobacco campaign materials for local audiences with minimal resources, a limited budget and without the need for an external production team. The Tongan experience underscores the importance of an adaptation process that draws from evidence-based best-practice models and engages local and regional stakeholders to ensure that campaign materials are tailored to the local context and are embedded within a mix of antitobacco strategies. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
19. Developing antitobacco mass media campaign messages in a low-resource setting: experience from the Kingdom of Tonga
Science.gov (United States)
Sugden, C; Phongsavan, P; Gloede, S; Filiai, S; Tongamana, V O
2017-01-01
Tobacco use has become the leading cause of preventable death in Tonga, a small island nation in the South Pacific. One pragmatic and economical strategy to address this worrying trend is to adapt effective antitobacco mass media materials developed in high-income countries for local audiences. Using Tonga as an example, this paper shares the practical steps involved in adapting antitobacco campaign materials for local audiences with minimal resources, a limited budget and without the need for an external production team. The Tongan experience underscores the importance of an adaptation process that draws from evidence-based best-practice models and engages local and regional stakeholders to ensure that campaign materials are tailored to the local context and are embedded within a mix of antitobacco strategies. PMID:26969171
20. Determining ν mass hierarchy by precise measurements of two Δm2 in νe and νμ disappearance experiments
International Nuclear Information System (INIS)
Minakata, H; Nunokawa, H; Parke, S; Funchal, R Zukanovich
2006-01-01
In this paper, we discuss the possibility of determining the neutrino mass hierarchy by comparing the two effective atmospheric neutrino mass squared differences measured, respectively, in electron and in muon neutrino disappearance oscillation experiments. If the former is larger (smaller) than the latter, the mass hierarchy is of normal (inverted) type. We consider two very high precision (a few per mil) measurements of such mass squared differences by phase II of the T2K (Tokai-to-Kamioka) experiment and by the novel Moessbauer enhanced resonant ν-bar e absorption technique. Under optimistic assumptions for the systematic errors of both measurements, we determine the region of sensitivities where the mass hierarchy can be distinguished. Due to the tight space limitation, we present only the general idea and show a few most important plots
1. Validation of the muon momentum resolution in view of the W mass measurement with the CMS experiment
CERN Document Server
Manca, Elisabetta
2016-01-01
In the framework of the Standard Model, the electroweak theory predictsrelations among observables which can be measured. After the discovery ofthe Higgs boson, all the parameters of the Standard Model are known andit is thus possible to predict those observables with increasing precision inorder to test the consistency of the model. A small deviation of measuredvalues from those predictions would be an indirect hint of physics beyondthe Standard Model.In particular, the mass of the W boson has a far smaller uncertainty inthe theoretical prediction than in the measured value. For this reason, anaccurate measurement of the W mass would provide such a test of validityof the Standard Model.In order to achieve the precision required to have a fair comparison with thetheory, it is necessary to control the distributions of the variables enteringthe measurement at the permil level or even better.The CMS experiment is planning to deliver a precise measurement of MWwithin the next years, analysing events of W decaying...
2. Search for heavy resonance in the top-antitop invariant mass spectrum at the ATLAS experiment in the LHC
International Nuclear Information System (INIS)
Dechenaux, B.
2013-01-01
This report presents the analysis conducted with the ATLAS experiment at the LHC and searching for resonant production of new particles decaying into a pair of top quarks. Top quark reconstruction is mainly build upon the notion of hadronic jets, whose identification and reconstruction is a crucial issue for any measure trying to sign top quark decays from proton-proton collisions processes. After a general description of the theoretical and experimental features of jet reconstruction in the ATLAS detector, we present a first attempt to validate the local hadronic calibration method, which aim at correcting the measurement of these objects from inaccuracies caused by detector effects. In the second part, we present the analysis conducted on 14 fb -1 of proton-proton collision data at √(s)=8 TeV collected during the year 2012 and searching for resonant creation of new heavy particles in top-anti-top invariant mass spectrum. For heavy particles, the quarks produced in the decay of the latter have a high impulsion with respect to their mass and those top quark decays often results in a so called 'boosted topology', where the hadronic-decaying top quark is often reconstructed as a single jet of large radius parameter. In this context, we present a preliminary study to reconstruct and identify as precisely as possible this type of boosted topologies, based on the study of jet substructure. (author)
3. Strategies for precision measurements of the charge asymmetry of the W boson mass at the LHC within the ATLAS experiment
CERN Document Server
Fayette, Florent
This thesis dissertation presents a prospect for a measurement of the charge asymmetry of the W boson mass (MW+ - MW-) at the LHC within the ATLAS experiment. This measurement is of primordial importance for the LHC experimental program, both as a direct test of the charge sign independent coupling of the W bosons to the fermions and as a mandatory preliminary step towards the precision measurement of the charge averaged W boson mass. This last pragmatic point can be understood since the LHC specific collisions will provide unprecedented kinematics for the positive and negative channels while the SPS and Tevatron collider produced W+ and W- on the same footing. For that reason, the study of the asymmetries between W+ and W- in Drell--Yan like processes (production of single W decaying into leptons), studied to extract the properties of the W boson, is described thoroughly in this document. Then, the prospect for a measurement of MW+ - MW- at the LHC is addressed in a perspective intending to decrease as much ...
4. Investigating the presence of omeprazole in waters by liquid chromatography coupled to low and high resolution mass spectrometry: degradation experiments.
Science.gov (United States)
Boix, C; Ibáñez, M; Sancho, J V; Niessen, W M A; Hernández, F
2013-10-01
Omeprazole is one of the most consumed pharmaceuticals around the world. However, this compound is scarcely detected in urban wastewater and surface water. The absence of this pharmaceutical in the aquatic ecosystem might be due to its degradation in wastewater treatment plants, as well as in receiving water. In this work, different laboratory-controlled degradation experiments have been carried out on surface water in order to elucidate generated omeprazole transformation products (TPs). Surface water spiked with omeprazole was subjected to hydrolysis, photo-degradation under both sunlight and ultraviolet radiation and chlorination. Analyses by liquid chromatography coupled to quadrupole time-of-flight mass spectrometry (LC-QTOF MS) permitted identification of up to 17 omeprazole TPs. In a subsequent step, the TPs identified were sought in surface water and urban wastewater by LC-QTOF MS and by LC coupled to tandem mass spectrometry with triple quadrupole. The parent omeprazole was not detected in any of the samples, but four TPs were found in several water samples. The most frequently detected compound was OTP 5 (omeprazole sulfide), which might be a reasonable candidate to be included in monitoring programs rather than the parent omeprazole. Copyright © 2013 John Wiley & Sons, Ltd.
5. Search for high mass resonances in the dimuon channel using the muon spectrometer of the atlas experiment at CERN
International Nuclear Information System (INIS)
Helsens, C.
2009-06-01
This thesis covers the search of new neutral gauge bosons decaying into a pair of muons in the ATLAS detector. The Large Hadron Collider (LHC) at CERN will produce parton collisions with very high center of mass energy and may produce Z' predicted by many theories beyond the standard model. Such a resonance should be detected by the ATLAS experiment. For the direct search of Z' decaying into two muons, a small number of events is enough for its discovery, which is possible with the first data. We shall study in particular the effects of the muon spectrometer alignment on high p T tracks and on the Z' discovery potential in the ATLAS experiment. The discovery potentials computed with this method have been officially approved by the ATLAS collaboration and published. At the start of the LHC operation, the muon spectrometer alignment will not have reached the nominal performances. This analysis aims at optimizing the discovery potential of ATLAS for a Z' boson in this degraded initial conditions. The impact on track reconstruction of a degraded alignment is estimated with simulated high p T tracks. Results are given in terms of reconstruction efficiency, momentum and invariant mass resolutions, charge identification and sensitivity to discovery or exclusion. With the first data, an analysis using only the muon spectrometer in stand alone mode will be very useful. Finally, a study on how to determine the initial geometry of the spectrometer (needed for its absolute alignment) is performed. This study uses straight tracks without a magnetic field and also calculates the beam time necessary for reaching a given accuracy of the alignment system. (author)
6. Combined Measurement of the Higgs Boson Mass in $pp$ Collisions at $\\sqrt{s}$ = 7 and 8 TeV with the ATLAS and CMS Experiments
CERN Document Server
2015-05-14
A measurement of the Higgs boson mass is presented based on the combined data samples of the ATLAS and CMS experiments at the CERN LHC in the $H \\rightarrow \\gamma\\gamma$ and $H \\rightarrow ZZ\\rightarrow 4\\ell$ decay channels. The results are obtained from a simultaneous fit to the reconstructed invariant mass peaks in the two channels and for the two experiments. The measured masses from the individual channels and the two experiments are found to be consistent among themselves. The combined measured mass of the Higgs boson is $m_{H} = 125.09\\pm0.21\\,\\mathrm{(stat.)}\\pm0.11\\,\\mathrm{(syst.)}~\\mathrm{GeV}$.
7. Combining Experiments and Simulation of Gas Absorption for Teaching Mass Transfer Fundamentals: Removing CO2 from Air Using Water and NaOH
Science.gov (United States)
Clark, William M.; Jackson, Yaminah Z.; Morin, Michael T.; Ferraro, Giacomo P.
2011-01-01
Laboratory experiments and computer models for studying the mass transfer process of removing CO2 from air using water or dilute NaOH solution as absorbent are presented. Models tie experiment to theory and give a visual representation of concentration profiles and also illustrate the two-film theory and the relative importance of various…
8. Batch experiments of Cs, Co and Eu sorption onto cement with dissolved fibre mass UP2 in the liquid phase
International Nuclear Information System (INIS)
Holgersson, Stellan; Dubois, Isabelle; Boerstell, Lisa
2011-05-01
The potential effects of alkaline degradation products of the fibre mass UP2 on metal sorption onto fresh and degraded cement have been investigated. For this purpose, crushed cement have been leached to support material for the subsequent batch sorption experiments. Alkaline leaching of UP2 were also made, which gave leaching solutions of 30 ppm DOC after 300 days. These solutions were used in the batch experiments. Continued leaching shows that even higher concentrations can be expected: 45 ppm DOC in the leaching with a low-alkaline (pH 12.5) artificial cement pore-water and 150 ppm DOC with a high alkaline (pH 13.3) artificial pore-water. Batch sorption experiments with 134 Cs and 60 Co show no effects on metal sorption onto leached or fresh cement when the 30 ppm DOC leaching solutions of UP2 were used as liquid phase. The measured R d values were 10 -3 m 3 /kg for Cs and in the range of 0.03-0.1 m 3 /kg for Co. Separate experiments with other organics ligands were also made: EDTA, ISA and citric acid with maximum concentrations of 500, 300 and 300 ppm DOC, respectively. Also here no effects on Cs and Co sorption onto leached and fresh cement were established. Batch experiments of 152 Eu were not successful since results were above detection level of about 2-3 m 3 /kg. The addition of the aforementioned 30 ppm UP2 leaching solution or the other organic ligands did not affect the detection level. Measurements of background concentrations of total Eu show a peculiar result of Eu apparently increasing in batch experiments with cement to final values of about 5-10 nM. The underlying reason for this effect, whether real or artificial, could not be established. Background concentrations of Th were about 1 nM in both cement and blank samples. Background concentrations of Zr were about 100-700 nM in both cement and blank samples, the high values were measured acidic blanks, which suggests either acidic leaching from tube walls or contamination from the acid itself. No
9. Batch experiments of Cs, Co and Eu sorption onto cement with dissolved fibre mass UP2 in the liquid phase
Energy Technology Data Exchange (ETDEWEB)
Holgersson, Stellan; Dubois, Isabelle; Boerstell, Lisa (Department of Chemical and Biochemical Engineering, Nuclear Chemistry, Chalmers University of Technology (Sweden))
2011-05-15
The potential effects of alkaline degradation products of the fibre mass UP2 on metal sorption onto fresh and degraded cement have been investigated. For this purpose, crushed cement have been leached to support material for the subsequent batch sorption experiments. Alkaline leaching of UP2 were also made, which gave leaching solutions of 30 ppm DOC after 300 days. These solutions were used in the batch experiments. Continued leaching shows that even higher concentrations can be expected: 45 ppm DOC in the leaching with a low-alkaline (pH 12.5) artificial cement pore-water and 150 ppm DOC with a high alkaline (pH 13.3) artificial pore-water. Batch sorption experiments with 134Cs and 60Co show no effects on metal sorption onto leached or fresh cement when the 30 ppm DOC leaching solutions of UP2 were used as liquid phase. The measured R{sub d} values were 10-3 m3/kg for Cs and in the range of 0.03-0.1 m3/kg for Co. Separate experiments with other organics ligands were also made: EDTA, ISA and citric acid with maximum concentrations of 500, 300 and 300 ppm DOC, respectively. Also here no effects on Cs and Co sorption onto leached and fresh cement were established. Batch experiments of 152Eu were not successful since results were above detection level of about 2-3 m3/kg. The addition of the aforementioned 30 ppm UP2 leaching solution or the other organic ligands did not affect the detection level. Measurements of background concentrations of total Eu show a peculiar result of Eu apparently increasing in batch experiments with cement to final values of about 5-10 nM. The underlying reason for this effect, whether real or artificial, could not be established. Background concentrations of Th were about 1 nM in both cement and blank samples. Background concentrations of Zr were about 100-700 nM in both cement and blank samples, the high values were measured acidic blanks, which suggests either acidic leaching from tube walls or contamination from the acid itself. No
10. What to expect from immersive virtual environment exposure: influences of gender, body mass index, and past experience.
Science.gov (United States)
Stanney, Kay M; Hale, Kelly S; Nahmens, Isabelina; Kennedy, Robert S
2003-01-01
For those interested in using head-coupled PC-based immersive virtual environment (VE) technology to train, entertain, or inform, it is essential to understand the effects this technology has on its users. This study investigated potential adverse effects, including the sickness associated with exposure and extreme responses (emesis, flashbacks). Participants were exposed to a VE for 15 to 60 min, with either complete or streamlined navigational control and simple or complex scenes, after which time measures of sickness were obtained. More than 80% of participants experienced nausea, oculomotor disturbances, and/or disorientation, with disorientation potentially lasting > 24 hr. Of the participants, 12.9% prematurely ended their exposure because of adverse effects; of these, 9.2% experienced an emetic response, whereas only 1.2% of all participants experienced emesis. The results indicate that designers may be able to reduce these rates by limiting exposure duration and reducing the degrees of freedom of the user's navigational control. Results from gender, body mass, and past experience comparisons indicated it may be possible to identify those who will experience adverse effects attributable to exposure and warn such individuals. Applications for this research include military, entertainment, and any other interactive systems for which designers seek to avoid adverse effects associated with exposure.
11. Solution-mass transfer and grain boundary sliding in mafic shear zones - comparison between experiments and nature
Science.gov (United States)
Marti, Sina; Heilbronner, Renée; Stünitz, Holger; Plümper, Oliver; Drury, Martyn
2017-04-01
Grain size sensitive creep (GSSC) mechanisms are widely recognized to be the most efficient deformation mechanisms in shear zones. With or without initial fracturing and fluid infiltration, the onset of heterogeneous nucleation leading to strong grain size reduction is a frequently described process for the initiation of GSSC. Phase mixing due to reaction and heterogeneous nucleation during GSSC impedes grain growth, sustaining small grain sizes as a prerequisite for GSSC. Here we present rock deformation experiments on 'wet' plagioclase - pyroxene mixtures at T=800°C, P=1.0 and 1.5GPa and strain rates of 2e-5 - 2e-6 1/s, performed with a Griggs-type solid medium deformation apparatus. Microstructural criteria are used to show that both, grain boundary sliding (GBS) and solution-mass transfer processes are active and are interpreted to be the dominant strain accommodating processes. Displacement is localized within shear bands formed by fine-grained ( 300 - 500nm) plagioclase (Pl) and the syn-kinematic reaction products amphibole (Amph), quartz (Qz) and zoisite (Zo). We compare our experiments with a natural case - a sheared mafic pegmatite (P-T during deformation 0.7 - 0.9 GPa, 610 - 710 °C; Getsinger et al., 2013) from Northern Norway. Except for the difference in grain size of the experimental and natural samples, microstructures are strikingly alike. The experimental and natural P- and especially T-conditions are very similar. Consequently, extrapolation from experiments to nature must be made without a significant 'temperature-time' trade-off, which is normally taken advantage of when relating experimental to natural strain rates. We will discuss under which assumptions extrapolation to nature in our case is likely feasible. Syn-kinematic reactions during GBS and solution-mass transport are commonly interpreted to result in an ordered (anticlustered) phase mixture. However, phase mixing in our case is restricted: Mixing is extensive between Pl + Zo + Qz and
12. Quantitative correlations between collision induced dissociation mass spectrometry coupled with electrospray ionization or atmospheric pressure chemical ionization mass spectrometry - Experiment and theory
Science.gov (United States)
Ivanova, Bojidarka; Spiteller, Michael
2018-04-01
The problematic that we consider in this paper treats the quantitative correlation model equations between experimental kinetic and thermodynamic parameters of coupled electrospray ionization (ESI) mass spectrometry (MS) or atmospheric pressure chemical ionization (APCI) mass spectrometry with collision induced dissociation mass spectrometry, accounting for the fact that the physical phenomena and mechanisms of ESI- and APCI-ion formation are completely different. There are described forty two fragment reactions of three analytes under independent ESI- and APCI-measurements. The developed new quantitative models allow us to study correlatively the reaction kinetics and thermodynamics using the methods of mass spectrometry, which complementary application with the methods of the quantum chemistry provide 3D structural information of the analytes. Both static and dynamic quantum chemical computations are carried out. The object of analyses are [2,3-dimethyl-4-(4-methyl-benzoyl)-2,3-di-p-tolyl-cyclobutyl]-p-tolyl-methanone (1) and the polycyclic aromatic hydrocarbons derivatives of dibenzoperylen (2) and tetrabenzo [a,c,fg,op]naphthacene (3), respectively. As far as (1) is known to be a product of [2π+2π] cycloaddition reactions of chalcone (1,3-di-p-tolyl-propenone), however producing cyclic derivatives with different stereo selectivity, so that the study provide crucial data about the capability of mass spectrometry to provide determine the stereo selectivity of the analytes. This work also first provides quantitative treatment of the relations '3D molecular/electronic structures'-'quantum chemical diffusion coefficient'-'mass spectrometric diffusion coefficient', thus extending the capability of the mass spectrometry for determination of the exact 3D structure of the analytes using independent measurements and computations of the diffusion coefficients. The determination of the experimental diffusion parameters is carried out within the 'current monitoring method
13. A method for separating Antarctic postglacial rebound and ice mass balance using future ICESat Geoscience Laser Altimeter System, Gravity Recovery and Climate Experiment, and GPS satellite data
OpenAIRE
Velicogna, Isabella; Wahr, John
2002-01-01
Measurements of ice elevation from the Geoscience Laser Altimeter System (GLAS) aboard the Ice, Cloud, and Land Elevation Satellite can be combined with time-variable geoid measurements from the Gravity Recovery and Climate Experiment (GRACE) satellite mission to learn about ongoing changes in polar ice mass and viscoelastic rebound of the lithosphere under the ice sheet. We estimate the accuracy in recovering the spatially varying ice mass trend and postglacial rebound signals for Antarctica...
14. Search for high mass dilepton resonances in pp collisions at $\\sqrt{s}$=7 TeV with the ATLAS experiment
CERN Document Server
2013-07-16
This article presents a search for high mass e e or mu mu resonances in pp collisions at $\\sqrt{s}$ = 7 TeV at the LHC. The data were recorded by the ATLAS experiment during 2010 and correspond to a total integrated luminosity of ~40/pb. No statistically significant excess above the Standard Model expectation is observed in the search region of dilepton invariant mass above 110 GeV. Upper limits at the 95% confidence level are set on the cross section times branching ratio of Z' resonances decaying to dielectrons and dimuons as a function of the resonance mass. A lower mass limit of 1.048 TeV on the Sequential Standard Model Z' boson is derived, as well as mass limits on Z* and E(6)-motivated Z' models.
15. Search for high mass dilepton resonances in pp collisions at √(s)=7 TeV with the ATLAS experiment
International Nuclear Information System (INIS)
2011-01-01
This Letter presents a search for high mass e + e - or μ + μ - resonances in pp collisions at √(s)=7 TeV at the LHC. The data were recorded by the ATLAS experiment during 2010 and correspond to a total integrated luminosity of ∼40 pb -1 . No statistically significant excess above the Standard Model expectation is observed in the search region of dilepton invariant mass above 110 GeV. Upper limits at the 95% confidence level are set on the cross section times branching ratio of Z ' resonances decaying to dielectrons and dimuons as a function of the resonance mass. A lower mass limit of 1.048 TeV on the Sequential Standard Model Z ' boson is derived, as well as mass limits on Z * and E 6 -motivated Z ' models.
16. Search for high mass dilepton resonances in pp collisions at {radical}(s)=7 TeV with the ATLAS experiment
Energy Technology Data Exchange (ETDEWEB)
Aad, G [Fakultaet fuer Mathematik und Physik, Albert-Ludwigs-Universitaet, Freiburg i.Br. (Germany); Abbott, B [Homer L. Dodge Department of Physics and Astronomy, University of Oklahoma, Norman, OK (United States); Abdallah, J [Institut de Fisica d' Altes Energies and Universitat Autonoma de Barcelona and ICREA, Barcelona (Spain); Abdelalim, A A [Section de Physique, Universite de Geneve, Geneva (Switzerland); Abdesselam, A [Department of Physics, Oxford University, Oxford (United Kingdom); Abdinov, O [Institute of Physics, Azerbaijan Academy of Sciences, Baku (Azerbaijan); Abi, B [Department of Physics, Oklahoma State University, Stillwater, OK (United States); Abolins, M [Department of Physics and Astronomy, Michigan State University, East Lansing, MI (United States); Abramowicz, H [Raymond and Beverly Sackler School of Physics and Astronomy, Tel Aviv University, Tel Aviv (Israel); Abreu, H [LAL, Univ. Paris-Sud and CNRS/IN2P3, Orsay (France); Acerbi, E [INFN Sezione di Milano (Italy); Dipartimento di Fisica, Universita di Milano, Milano (Italy); Acharya, B S [INFN Gruppo Collegato di Udine (Italy); ICTP, Trieste [Italy; Adams, D L [Physics Department, Brookhaven National Laboratory, Upton, NY (United States); Addy, T N [Department of Physics, Hampton University, Hampton, VA (United States); Adelman, J [Department of Physics, Yale University, New Haven, CT (United States); Aderholz, M [Max-Planck-Institut fuer Physik (Werner-Heisenberg-Institut), Muenchen (Germany); Adomeit, S [Fakultaet fuer Physik, Ludwig-Maximilians-Universitaet Muenchen, Muenchen (Germany); Adragna, P [Department of Physics, Queen Mary University of London, London (United Kingdom); Adye, T [Particle Physics Department, Rutherford Appleton Laboratory, Didcot (United Kingdom); Aefsky, S [Department of Physics, Brandeis University, Waltham, MA (United States)
2011-06-13
This Letter presents a search for high mass e{sup +}e{sup -} or {mu}{sup +}{mu}{sup -} resonances in pp collisions at {radical}(s)=7 TeV at the LHC. The data were recorded by the ATLAS experiment during 2010 and correspond to a total integrated luminosity of {approx}40 pb{sup -1}. No statistically significant excess above the Standard Model expectation is observed in the search region of dilepton invariant mass above 110 GeV. Upper limits at the 95% confidence level are set on the cross section times branching ratio of Z{sup '} resonances decaying to dielectrons and dimuons as a function of the resonance mass. A lower mass limit of 1.048 TeV on the Sequential Standard Model Z{sup '} boson is derived, as well as mass limits on Z{sup *} and E{sub 6}-motivated Z{sup '} models.
17. Measurement of the top quark mass using dilepton events and a neutrino weighting algorithm with the DOe experiment at the Tevatron (Run II)
Energy Technology Data Exchange (ETDEWEB)
Meyer, J.
2007-07-01
Several measurements of the top quark mass in the dilepton final states with the DOe experiment are presented. The theoretical and experimental properties of the top quark are described together with a brief introduction of the Standard Model of particle physics and the physics of hadron collisions. An overview over the experimental setup is given. The Tevatron at Fermilab is presently the highest-energy hadron collider in the world with a center-of-mass energy of 1.96 TeV. There are two main experiments called CDF and DOe, A description of the components of the multipurpose DOe detector is given. The reconstruction of simulated events and data events is explained and the criteria for the identification of electrons, muons, jets, and missing transverse energy is given. The kinematics in the dilepton final state is underconstraint. Therefore, the top quark mass is extracted by the so-called Neutrino Weighting method. This method is introduced and several different approaches are described, compared, and enhanced. Results for the international summer conferences 2006 and winter 2007 are presented. The top quark mass measurement for the combination of all three dilepton channels with a dataset of 1.05 1/fb yields: mtop=172.5{+-}5.5 (stat.) {+-} 5.8 (syst.) GeV. This result is presently the most precise top quark mass measurement of the DOe experiment in the dilepton chann el. It entered the top quark mass wold average from March 2007. (orig.)
18. Experience gained in the process of the variable mass heat flow control implemented in the district heat supply system of the city of Gyor
Energy Technology Data Exchange (ETDEWEB)
Nagy, F.; Milanovich, L.; Lelek, J.; Kekk, I. [District Heating LTD. of Gyor (Hungary)
1996-11-01
The district heating system of the city of Gyor is fed from a hot water boiler plant. The total heat demand for 23,000 residential homes and several public facilities is 260 MW. The variable mass flow control was implemented in 1991 through 1992. Design, preparatory job and the majority of implementation was carried out without external involvement. The paper presents historical background and brief project presentation which is followed by comparative presentation of the variable mass flow control and constant mass flow control. This comparative survey has been conducted on the basis of operating data for 1993 and those for 1988. In the conclusion the gained experience is summarized.
19. The mass-hierarchy and CP-violation discovery reach of the LBNO long-baseline neutrino experiment
CERN Document Server
Agarwalla, S.K.; Aittola, M.; Alekou, A.; Andrieu, B.; Angus, D.; Antoniou, F.; Ariga, A.; Ariga, T.; Asfandiyarov, R.; Autiero, D.; Ballett, P.; Bandac, I.; Banerjee, D.; Barker, G.J.; Barr, G.; Bartmann, W.; Bay, F.; Berardi, V.; Bertram, I.; Bésida, O.; Blebea-Apostu, A.M.; Blondel, A.; Bogomilov, M.; Borriello, E.; Boyd, S.; Brancus, I.; Bravar, A.; Buizza-Avanzini, M.; Cafagna, F.; Calin, M.; Calviani, M.; Campanelli, M.; Cantini, C.; Caretta, O.; Cata-Danil, G.; Catanesi, M.G.; Cervera, A.; Chakraborty, S.; Chaussard, L.; Chesneanu, D.; Chipesiu, F.; Christodoulou, G.; Coleman, J.; Crivelli, P.; Davenne, T.; Dawson, J.; De Bonis, I.; De Jong, J.; Déclais, Y.; del Amo Sanchez, P.; Delbart, A.; Densham, C.; Di Lodovico, F.; Di Luise, S.; Duchesneau, D.; Dumarchez, J.; Efthymiopoulos, I.; Eliseev, A.; Emery, S.; Enqvist, K.; Enqvist, T.; Epprecht, L.; Ereditato, A.; Erykalov, A.N.; Esanu, T.; Finch, A.J.; Fitton, M.D.; Franco, D.; Galymov, V.; Gavrilov, G.; Gendotti, A.; Giganti, C.; Goddard, B.; Gomez, J.J.; Gomoiu, C.M.; Gornushkin, Y.A.; Gorodetzky, P.; Grant, N.; Haesler, A.; Haigh, M.D.; Hasegawa, T.; Haug, S.; Hierholzer, M.; Hissa, J.; Horikawa, S.; Huitu, K.; Ilic, J.; Ioannisian, A.N.; Izmaylov, A.; Jipa, A.; Kainulainen, K.; Kalliokoski, T.; Karadzhov, Y.; Kawada, J.; Khabibullin, M.; Khotjantsev, A.; Kokko, E.; Kopylov, A.N.; Kormos, L.L.; Korzenev, A.; Kosyanenko, S.; Kreslo, I.; Kryn, D.; Kudryavtsev, V.A.; Kudenko, Y.; Kumpulainen, J.; Kuusiniemi, P.; Lagoda, J.; Lazanu, I.; Levy, J. -M.; Litchfield, R.P.; Loo, K.; Loveridge, P.; Maalampi, J.; Magaletti, L.; Margineanu, R.M.; Marteau, J.; Martin-Mari, C.; Matveev, V.; Mavrokoridis, K.; Mazzucato, E.; McCauley, N.; Mercadante, A.; Mineev, O.; Mirizzi, A.; Mitrica, B.; Morgan, B.; Murdoch, M.; Murphy, S.; Narita, S.; Nesterenko, D.A.; Nguyen, K.; Nikolics, K.; Noah, E.; Novikov, Yu.; O'Keeffe, H.; Odell, J.; Oprima, A.; Palladino, V.; Pascoli, S.; Patzak, T.; Payne, D.; Pectu, M.; Pennacchio, E.; Papaphilippou, Y.; Periale, L.; Pessard, H.; Pistillo, C.; Popov, B.; Przewlocki, P.; Quinto, M.; Radicioni, E.; Ramachers, Y.; Ratoff, P.N.; Ravonel, M.; Rayner, M.; Resnati, F.; Ristea, O.; Robert, A.; Rondio, E.; Rubbia, A.; Rummukainen, K.; Sacco, R.; Saftoiu, A.; Sakashita, K.; Sarkamo, J.; Sato, F.; Saviano, N.; Scantamburlo, E.; Sergiampietri, F.; Sgalaberna, D.; Shaposhnikova, E.; Slupecki, M.; Sorel, M.; Spooner, N.J.C.; Stahl, A.; Stanca, D.; Steerenberg, R.; Sterian, A.R.; Sterian, P.; Still, B.; Stoica, S.; Strauss, T.; Suhonen, J.; Suvorov, V.; Szeptycka, M.; Terri, R.; Thompson, L.F.; Toma, G.; Tonazzo, A.; Touramanis, C.; Trzaska, W.H.; Tsenov, R.; Tuominen, K.; Vacheret, A.; Valram, M.; Vankova-Kirilova, G.; Vanucci, F.; Vasseur, G.; Velotti, F.; Velten, P.; Viant, T.; Vincke, H.; Virtanen, A.; Vorobyev, A.; Wark, D.; Weber, A.; Weber, M.; Wiebusch, C.; Wilson, J.R.; Wu, S.; Yershov, N.; Zalipska, J.; Zito, M.
2014-01-01
The next generation neutrino observatory proposed by the LBNO collaboration will address fundamental questions in particle and astroparticle physics. The experiment consists of a far detector, in its first stage a 20 kt LAr double phase TPC and a magnetised iron calorimeter, situated at 2300 km from CERN and a near detector based on a high-pressure argon gas TPC. The long baseline provides a unique opportunity to study neutrino flavour oscillations over their 1st and 2nd oscillation maxima exploring the $L/E$ behaviour, and distinguishing effects arising from $\\delta_{CP}$ and matter. In this paper we have reevaluated the physics potential of this setup for determining the mass hierarchy (MH) and discovering CP-violation (CPV), using a conventional neutrino beam from the CERN SPS with a power of 750 kW. We use conservative assumptions on the knowledge of oscillation parameter priors and systematic uncertainties. The impact of each systematic error and the precision of oscillation prior is shown. We demonstrat...
20. The Colorado Ultraviolet Transit Experiment (CUTE): a dedicated cubesat mission for the study of exoplanetary mass loss and magnetic fields
Science.gov (United States)
Fleming, Brian T.; France, Kevin; Nell, Nicholas; Kohnert, Richard; Pool, Kelsey; Egan, Arika; Fossati, Luca; Koskinen, Tommi; Vidotto, Aline A.; Hoadley, Keri; Desert, Jean-Michel; Beasley, Matthew; Petit, Pascal
2017-08-01
The Colorado Ultraviolet Transit Experiment (CUTE) is a near-UV (2550 - 3300 Å) 6U cubesat mission designed to monitor transiting hot Jupiters to quantify their atmospheric mass loss and magnetic fields. CUTE will probe both atomic (Mg and Fe) and molecular (OH) lines for evidence of enhanced transit absorption, and to search for evidence of early ingress due to bow shocks ahead of the planet's orbital motion. As a dedicated mission, CUTE will observe > 60 spectroscopic transits of hot Jupiters over a nominal seven month mission. This represents the equivalent of > 700 orbits of the only other instrument capable of these measurements, the Hubble Space Telescope. CUTE efficiently utilizes the available cubesat volume by means of an innovative optical design to achieve a projected effective area of ˜ 22 cm2 , low instrumental background, and a spectral resolving power of R ˜ 3000 over the entire science bandpass. These performance characteristics enable CUTE to discern a transit depth of motivation and expected results, and an overview of the projected fabrication, calibration and launch timeline.
1. The Gigatracker: An ultra-fast and low-mass silicon pixel detector for the NA62 experiment
International Nuclear Information System (INIS)
Fiorini, M.; Carassiti, V.; Ceccucci, A.; Cortina, E.; Cotta Ramusino, A.; Dellacasa, G.; Garbolino, S.; Jarron, P.; Kaplon, J.; Kluge, A.; Mapelli, A.; Marchetto, F.; Martin, E.; Martoiu, S.; Mazza, G.; Morel, M.; Noy, M.; Nuessle, G.; Petrucci, F.; Riedler, P.
2011-01-01
The Gigatracker is a hybrid silicon pixel detector developed to track the highly intense NA62 hadron beam with a time resolution of 150 ps (rms). The beam spectrometer of the experiment is composed of three Gigatracker stations installed in vacuum in order to precisely measure momentum, time and direction of every traversing particle. Precise tracking demands a very low mass of the detector assembly ( 0 per station) in order to limit multiple scattering and beam hadronic interactions. The high rate and especially the high timing precision requirements are very demanding: two R and D options are ongoing and the corresponding prototype read-out chips have been recently designed and produced in 0.13μm CMOS technology. One solution makes use of a constant fraction discriminator and on-pixel analogue-based time-to-digital-converter (TDC); the other comprises a delay-locked loop based TDC placed at the end of each pixel column and a time-over-threshold discriminator with time-walk correction technique. The current status of the R and D program is overviewed and results from the prototype read-out chips test are presented.
2. The Gigatracker: An ultra-fast and low-mass silicon pixel detector for the NA62 experiment
Science.gov (United States)
Fiorini, M.; Carassiti, V.; Ceccucci, A.; Cortina, E.; Cotta Ramusino, A.; Dellacasa, G.; Garbolino, S.; Jarron, P.; Kaplon, J.; Kluge, A.; Mapelli, A.; Marchetto, F.; Martin, E.; Martoiu, S.; Mazza, G.; Morel, M.; Noy, M.; Nuessle, G.; Petrucci, F.; Riedler, P.; Aglieri Rinella, G.; Rivetti, A.; Tiuraniemi, S.
2011-02-01
The Gigatracker is a hybrid silicon pixel detector developed to track the highly intense NA62 hadron beam with a time resolution of 150 ps (rms). The beam spectrometer of the experiment is composed of three Gigatracker stations installed in vacuum in order to precisely measure momentum, time and direction of every traversing particle. Precise tracking demands a very low mass of the detector assembly ( beam hadronic interactions. The high rate and especially the high timing precision requirements are very demanding: two R&D options are ongoing and the corresponding prototype read-out chips have been recently designed and produced in 0.13 μm CMOS technology. One solution makes use of a constant fraction discriminator and on-pixel analogue-based time-to-digital-converter (TDC); the other comprises a delay-locked loop based TDC placed at the end of each pixel column and a time-over-threshold discriminator with time-walk correction technique. The current status of the R&D program is overviewed and results from the prototype read-out chips test are presented.
3. The Gigatracker: An ultra-fast and low-mass silicon pixel detector for the NA62 experiment
CERN Document Server
Fiorini, M; Morel, M; Petrucci, F; Marchetto, F; Garbolino, S; Cortina, E; Tiuraniemi, S; Ceccucci, A; Martin, E; Riedler, P; Martoiu, S; Ramusino, A C; Rinella, G A; Mapelli, A; Mazza, G; Noy, M; Jarron, P; Nuessle, G; Dellacasa, G; Kluge, A; Rivetti, A; Kaplon, J
2011-01-01
The Gigatracker is a hybrid silicon pixel detector developed to track the highly intense NA62 hadron beam with a time resolution of 150 ps (rms). The beam spectrometer of the experiment is composed of three Gigatracker stations installed in vacuum in order to precisely measure momentum, time and direction of every traversing particle. Precise tracking demands a very low mass of the detector assembly (<0.5\\% X(O) per station) in order to limit multiple scattering and beam hadronic interactions. The high rate and especially the high timing precision requirements are very demanding: two R\\&D options are ongoing and the corresponding prototype read-out chips have been recently designed and produced in 0.13 mu m CMOS technology. One solution makes use of a constant fraction discriminator and on-pixel analogue-based time-to-digital-converter (TDC); the other comprises a delay-locked loop based TDC placed at the end of each pixel column and a time-over-threshold discriminator with time-walk correction techniq...
4. Colorado Ultraviolet Transit Experiment: a dedicated CubeSat mission to study exoplanetary mass loss and magnetic fields
Science.gov (United States)
Fleming, Brian T.; France, Kevin; Nell, Nicholas; Kohnert, Richard; Pool, Kelsey; Egan, Arika; Fossati, Luca; Koskinen, Tommi; Vidotto, Aline A.; Hoadley, Keri; Desert, Jean-Michel; Beasley, Matthew; Petit, Pascal M.
2018-01-01
The Colorado Ultraviolet Transit Experiment (CUTE) is a near-UV (2550 to 3300 Å) 6U CubeSat mission designed to monitor transiting hot Jupiters to quantify their atmospheric mass loss and magnetic fields. CUTE will probe both atomic (Mg and Fe) and molecular (OH) lines for evidence of enhanced transit absorption, and to search for evidence of early ingress due to bow shocks ahead of the planet's orbital motion. As a dedicated mission, CUTE will observe ≳100 spectroscopic transits of hot Jupiters over a nominal 7-month mission. This represents the equivalent of >700 orbits of the only other instrument capable of these measurements, the Hubble Space Telescope. CUTE efficiently utilizes the available CubeSat volume by means of an innovative optical design to achieve a projected effective area of ˜28 cm2, low instrumental background, and a spectral resolving power of R˜3000 over the primary science bandpass. These performance characteristics enable CUTE to discern transit depths between 0.1% and 1% in individual spectral absorption lines. We present the CUTE optical and mechanical design, a summary of the science motivation and expected results, and an overview of the projected fabrication, calibration, and launch timeline.
5. [The experience of involvement of volunteers into maintenance of infection safety during period of implementation of mass activities].
Science.gov (United States)
Imamov, A A; Balabanova, L A; Zamalieva, M A
2016-01-01
The article presents experience of Rospotrebnadzor in the Republic of Tatarstan in the field of preventive medicine concerning training of volunteers on issues of infection safety with purpose of prevention of ictuses of infection diseases during mass activities with international participation in the period of XXVII World Summer Students Games. The model of hygienic training for volunteers provides two directions: training for volunteers ’ leaders on issues of infection safety and remote course for involved volunteers. During period of preparation for the Students Games-2013 hygienic training was organized for volunteers-leaders in the field of infection safety with following attestation. The modern training technologies were applied. The volunteers-leaders familiarized with groups of infection diseases including the most dangerous ones, investigated with expert algorithm of actions to be applied in case of suspicion on infection disease in gest or participant of the Games-2013 to secure one's health and health of immediate population. The active volunteers-leaders became trainers and coaches in the field of infection safety. The second stage of infection safety training organized by youth trainers' pool in number of 30 individuals the training technology "Equal trains equal" was applied for hygienic training of volunteers involved at epidemiologically significant objects (food objects, hotels, accompaniment of guests and sportsmen). The volunteers-leaders trained to infection safety 1400 volunteers. The format of electronic personal cabinet and remote course were selected as tools of post-training monitoring.
6. Glacier mass variations from recent ITSG-Grace solutions: Experiences with the point-mass modeling technique in the framework of project SPICE.
Science.gov (United States)
Reimond, S.; Klinger, B.; Krauss, S.; Mayer-Gürr, T.; Eicker, A.; Zemp, M.
2017-12-01
In recent years, remotely sensed observations have become one of the most ubiquitous and valuable sources of information for glacier monitoring. In addition to altimetry and interferometry data (as observed, e.g., by the CryoSat-2 and TanDEM-X satellites), time-variable gravity field data from the GRACE satellite mission has been used by several authors to assess mass changes in glacier systems. The main challenges in this context are i) the limited spatial resolution of GRACE, ii) the gravity signal attenuation in space and iii) the problem of isolating the glaciological signal from the gravitational signatures as detected by GRACE.In order to tackle the challenges i) and ii), we thoroughly investigate the point-mass modeling technique to represent the local gravity field. Instead of simply evaluating global spherical harmonics, we operate on the normal equation level and make use of GRACE K-band ranging data (available since April 2002) processed at the Graz University of Technology. Assessing such small-scale mass changes from space-borne gravimetric data is an ill-posed problem, which we aim to stabilize by utilizing a Genetic Algorithm based Tikhonov regularization. Concerning issue iii), we evaluate three different hydrology models (i.e. GLDAS, LSDM and WGHM) for validation purposes and the derivation of error bounds. The non-glaciological signal is calculated for each region of interest and reduced from the GRACE results.We present mass variations of several alpine glacier systems (e.g. the European Alps, Svalbard or Iceland) and compare our results to glaciological observations provided by the World Glacier Monitoring Service (WGMS) and alternative inversion methods (surface density modeling).
7. Measurement of cross section of quark pair production top with the D0 experiment at the Tevatron and determination the top quark mass using this measure
Energy Technology Data Exchange (ETDEWEB)
Chevalier-Thery, Solene [Univ. Pierre et Marie Curie, Paris (France)
2010-06-01
The top quark has been discovered by CDF and D0 experiments in 1995 at the proton-antiproton collider Tevatron. The amount of data recorded by both experiments makes it possible to accurately study the properties of this quark: its mass is now known to better than 1% accuracy. This thesis describes the measurement of the top pair cross section in the electron muon channel with 4, 3 fb -1 recorded data between 2006 and 2009 by the D0 experiment. Since the final state included a muon, improvements of some aspects of its identification have been performed : a study of the contamination of the cosmic muons and a study of the quality of the muon tracks. The cross section measurement is in good agreement with the theoretical calculations and the other experimental measurements. This measurement has been used to extract a value for the top quark mass. This method allows for the extraction of a better defined top mass than direct measurements as it depends less on Monte Carlo simulations. The uncertainty on this extracted mass, dominated by the experimental one, is however larger than for direct measurements. In order to decrease this uncertainty, the ratio of the Z boson and the top pair production cross sections has been studied to look for some possible theoretical correlations. At the Tevatron, the two cross sections are not theoretically correlated: no decrease of the uncertainty on the extracted top mass is therefore possible.
8. Measurement of cross section of quark pair production top with the D0 experiment at the Tevatron and determination the top quark mass using this measure
International Nuclear Information System (INIS)
Chevalier-Thery, Solene
2010-01-01
The top quark has been discovered by CDF and D0 experiments in 1995 at the proton-antiproton collider Tevatron. The amount of data recorded by both experiments makes it possible to accurately study the properties of this quark: its mass is now known to better than 1% accuracy. This thesis describes the measurement of the top pair cross section in the electron muon channel with 4, 3 fb -1 recorded data between 2006 and 2009 by the D0 experiment. Since the final state included a muon, improvements of some aspects of its identification have been performed : a study of the contamination of the cosmic muons and a study of the quality of the muon tracks. The cross section measurement is in good agreement with the theoretical calculations and the other experimental measurements. This measurement has been used to extract a value for the top quark mass. This method allows for the extraction of a better defined top mass than direct measurements as it depends less on Monte Carlo simulations. The uncertainty on this extracted mass, dominated by the experimental one, is however larger than for direct measurements. In order to decrease this uncertainty, the ratio of the Z boson and the top pair production cross sections has been studied to look for some possible theoretical correlations. At the Tevatron, the two cross sections are not theoretically correlated: no decrease of the uncertainty on the extracted top mass is therefore possible.
9. The role of personal opinions and experiences in compliance with mass drug administration for lymphatic filariasis elimination in Kenya.
Science.gov (United States)
Njomo, Doris W; Amuyunzu-Nyamongo, Mary; Magambo, Japheth K; Njenga, Sammy M
2012-01-01
The main strategy adopted for Lymphatic Filariasis (LF) elimination globally is annual mass drug administration (MDA) for 4 to 6 rounds. At least 65% of the population at risk should be treated in each round for LF elimination to occur. In Kenya, MDA using diethylcarbamazine citrate (DEC) and albendazole data shows declining compliance (proportion of eligible populations who receive and swallow the drugs) levels (85%-62.8%). The present study's aim was to determine the role of personal opinions and experiences in compliance with MDA. This was a retrospective cross-sectional study conducted between January and September 2009 in two districts based on December 2008 MDA round. In each district, one location with high and one with low compliance was selected. Through systematic sampling, nine villages were selected and interviewer-based questionnaires administered to 965 household heads or adult representatives also systematically sampled. The qualitative data were generated from opinion leaders, LF patients with clinical signs and community drug distributors (CDDs) all purposively selected and interviewed. Sixteen focus group discussions (FGDs) were also conducted with single-sex adult and youth male and female groups. Chi square test was used to assess the statistical significance of differences in compliance with treatment based on the records reviewed. The house-to-house method of drug distribution influenced compliance. Over one-quarter (27%) in low compared to 15% in high compliance villages disliked this method. Problems related to size, number and taste of the drugs were more common in low (16.4%) than in high (14.4%) compliance villages. Reasons for failure to take the drugs were associated with compliance (pcompliance villages. Experience of side effects influenced compliance (Pcompliance in both types of villages (p>0.05). Community sensitization on treatment, drugs used, their regimen and distribution method involving all leaders should be strengthened by
10. Air Mass Factor Formulation for Spectroscopic Measurements from Satellites: Application to Formaldehyde Retrievals from the Global Ozone Monitoring Experiment
Science.gov (United States)
Palmer, Paul I.; Jacob, Daniel J.; Chance, Kelly; Martin, Randall V.; Spurr, Robert J. D.; Kurosu, Thomas P.; Bey, Isabelle; Yantosca, Robert; Fiore, Arlene; Li, Qinbin
2004-01-01
We present a new formulation for the air mass factor (AMF) to convert slant column measurements of optically thin atmospheric species from space into total vertical columns. Because of atmospheric scattering, the AMF depends on the vertical distribution of the species. We formulate the AMF as the integral of the relative vertical distribution (shape factor) of the species over the depth of the atmosphere, weighted by altitude-dependent coefficients (scattering weights) computed independently from a radiative transfer model. The scattering weights are readily tabulated, and one can then obtain the AMF for any observation scene by using shape factors from a three dimensional (3-D) atmospheric chemistry model for the period of observation. This approach subsequently allows objective evaluation of the 3-D model with the observed vertical columns, since the shape factor and the vertical column in the model represent two independent pieces of information. We demonstrate the AMF method by using slant column measurements of formaldehyde at 346 nm from the Global Ozone Monitoring Experiment satellite instrument over North America during July 1996. Shape factors are cumputed with the Global Earth Observing System CHEMistry (GEOS-CHEM) global 3-D model and are checked for consistency with the few available aircraft measurements. Scattering weights increase by an order of magnitude from the surface to the upper troposphere. The AMFs are typically 20-40% less over continents than over the oceans and are approximately half the values calculated in the absence of scattering. Model-induced errors in the AMF are estimated to be approximately 10%. The GEOS-CHEM model captures 50% and 60% of the variances in the observed slant and vertical columns, respectively. Comparison of the simulated and observed vertical columns allows assessment of model bias.
11. Search for narrow baryon resonances (of masses through 3.4 and 5 GeV) through a π-p large angle elastic scattering formation experiment
International Nuclear Information System (INIS)
Chauveau, J.
1981-01-01
This work describes a search for narrow baryon resonances (of masses between 3.4 and 5 GeV) through a π - p large angle elastic scattering formation experiment. An optimization of the sensitivity of the experiment to detect resonances is obtained by the measurement of the central part of the angular distribution (/cos theta*/ -4 . The apparatus and data analysis are described in details. No narrow resonance has been found, the sensitivity of the experiment being characterized by a width GAMMA approximately equal to 1 MeV and an elasticity x approximately equal to 0.01. Finally, the differential cross section measurement is compared to some parton models [fr
12. Determination of W boson mass from the four quark decay in the DELPHI experiment at LEP; Mesure de la masse du boson W dans la desintegration a quatre quarks dans l'experience DELPHI au LEP
Energy Technology Data Exchange (ETDEWEB)
Duperrin, Arnaud [Universite Claude Bernard, 69 - Lyon (France)
1999-04-20
The accurate determination of W boson mass allows testing the coherence of standard model and imposing new constrains upon some of its parameters as for instance the Higgs boson mass. The W mass obtained by direct reconstruction of WW{yields}qq-bar qq-bar, was measured by means of the data recorded in 1997 and 1998 by DELPHI experimentat the center-of-mass energies 183 GeV and 189 GeV, for a total luminosity of 212 pb{sup -1} at the LEP collider of CERN. A neural network was used to label the signal what led to an efficiency and a selection accuracy of 86% and 80%, respectively. The number of the selected events among the data was 1710. By mass reconstruction, the jet multiplicity was let to vary free between four and eight jets. A new fast algorithm of kinematic fitting of the jets was developed to improve the mass resolution of the multi-jet events, by imposing the energy and momentum conservation. The association of jets has been carried out also by means of a neural network. The W boson mass was extracted starting from a probability fitting of the two-dimensional distribution provided by complete simulation and formed by the mean and the difference of two masses of W reconstructs. A technique of Monte Carlo re-weighting has been developed to obtain the simulation spectra for arbitrary M{sub W} values. The leading systematic uncertainties are studied by gathering a large number of e{sup +}e{sup -}{yields}Z{yields}qq-bar events and by merging them for obtaining events similar to the W pairs. The value of the W mass obtained from the data based on this probability is M{sub W} = 80.350{+-}0.099 (stat.) {+-} 0.038 (exp.) {+-} 0.056 (th.) {+-} 0.018 (LEP) GeV/c{sup 2}. This result is competitive and in good agreement with other determinations. The current world average mass of W boson is M{sub W} = 80.394 {+-} 0.042 GeV/c{sup 2}. This result, included in a global fitting of the electroweak data, constraints the standard Higgs boson to a mass lower than 262 Ge
13. High-precision mass measurements of neutron-deficient Tl isotopes at ISOLTRAP and the development of an ultra-stable voltage source for the PENTATRAP experiment
Energy Technology Data Exchange (ETDEWEB)
Boehm, Christine
2015-01-14
Atomic masses and hence binding energies of nuclides are of great importance for studies of nuclear structure since they reflect all effective interactions in a nucleus. Within this thesis the masses of seven nuclides, namely {sup 194}Au, {sup 194}Hg, {sup 190,193,198}Tl and {sup 202,208}Pb, were determined at the Penning-trap mass spectrometer ISOLTRAP at ISOLDE/CERN. The thallium region in the chart of isotopes is of special interest due to the occurrence of nuclear structure effects like low-lying isomers, level inversion, shape coexistence and deformations. These effects are investigated by applying finite-difference mass formulas, such as the two-neutron separation energies or the so-called empirical pairing gaps. The second topic addressed within the present thesis is an ultra-stable voltage source, called StaReP (Stable Reference for Penning Trap Experiments), which was developed at the Max-Planck-Institut fuer Kernphysik. It is one of the key components of the high-precision mass spectrometer PENTATRAP, containing a tower of five Penning traps. A 25-channel voltage source with a relative stability of few 10{sup -8} over a period of 10 minutes in the range of 0 to -100V is mandatory for PENTATRAP aiming for mass measurements with relative mass uncertainties of ≤ 10{sup -11}. Mass values with such a high precision allow for stringent tests of quantum electrodynamics in strong electric fields, testing Einstein's mass-energy relation E = mc{sup 2} as well as measurements of decay energies (Q-values) with applications in neutrino physics.
14. Jet calibration and top quark mass measurement in the semi-leptonic channel in the ATLAS experiment
International Nuclear Information System (INIS)
Balli, Fabrice
2014-01-01
The main goal of this thesis is to provide a measurement as accurate as possible of the top quark mass in the semi-leptonic decay channel. This experimental measurement is made thanks to the ATLAS detector near LHC, a proton-proton collider. The main interests for this precision measurement are the physics constraints to the theoretical models of fundamental constituents. Besides, the top quark mass is a parameter allowing to have more information on the vacuum stability at the Planck scale within the Standard Model. Jet energy calibration is crucial to this measurement. The impact of real data taking conditions on this calibration and on jet performance is detailed. The top quark mass measurement using 2011 data collected at an energy in the center-of-mass of 7 TeV is presented. It is using a tri dimensional template analysis method. The measured top quark mass is: m(top) = 172.01 ± 0.92 (stat) ± 1.17 (syst) GeV. The 2012 data collected at an energy in the center-of-mass of 8 TeV are also analysed, and a preliminary result for the top quark mass is provided: m(top) = 172.82 ± 0.39 (stat) ± 1.12 (syst) GeV, the combination of both measurements being the most accurate result of this thesis: m(top) = 172.64 ± 0.37 (stat) ± 1.10 (syst) GeV. (author) [fr
15. The role of personal opinions and experiences in compliance with mass drug administration for lymphatic filariasis elimination in Kenya.
Directory of Open Access Journals (Sweden)
Doris W Njomo
Full Text Available The main strategy adopted for Lymphatic Filariasis (LF elimination globally is annual mass drug administration (MDA for 4 to 6 rounds. At least 65% of the population at risk should be treated in each round for LF elimination to occur. In Kenya, MDA using diethylcarbamazine citrate (DEC and albendazole data shows declining compliance (proportion of eligible populations who receive and swallow the drugs levels (85%-62.8%. The present study's aim was to determine the role of personal opinions and experiences in compliance with MDA.This was a retrospective cross-sectional study conducted between January and September 2009 in two districts based on December 2008 MDA round. In each district, one location with high and one with low compliance was selected. Through systematic sampling, nine villages were selected and interviewer-based questionnaires administered to 965 household heads or adult representatives also systematically sampled. The qualitative data were generated from opinion leaders, LF patients with clinical signs and community drug distributors (CDDs all purposively selected and interviewed. Sixteen focus group discussions (FGDs were also conducted with single-sex adult and youth male and female groups. Chi square test was used to assess the statistical significance of differences in compliance with treatment based on the records reviewed. The house-to-house method of drug distribution influenced compliance. Over one-quarter (27% in low compared to 15% in high compliance villages disliked this method. Problems related to size, number and taste of the drugs were more common in low (16.4% than in high (14.4% compliance villages. Reasons for failure to take the drugs were associated with compliance (p0.05.Community sensitization on treatment, drugs used, their regimen and distribution method involving all leaders should be strengthened by the Programme Implementers. The communities need to be made aware of the potential side effects of the
16. Scale effect experiment in a fractured rock mass. Pilot study in the certified Fanay-Augeres mine (F)
International Nuclear Information System (INIS)
Durand, E.; Peaudecerf, P.; Ledoux, E.; De Marsily, G.
1985-01-01
This report (in two volumes) presents the results of a first phase of research about ''scale effect'' on permeability and solute transport in a fractured rock mass, to assess its suitability for future disposal of radioactive wastes. The gallery which was ''certified'' is located in the Fanay-Augeres mine(F), at a depth of about 175 m, in a granite mass. The portion selected for the subsequent experimental work is about 100 m long
17. Arterial spin labeling MR imaging for characterisation of renal masses in patients with impaired renal function: initial experience
International Nuclear Information System (INIS)
Pedrosa, Ivan; Rafatzand, Khashayar; Robson, Philip; Alsop, David C.; Wagner, Andrew A.; Atkins, Michael B.; Rofsky, Neil M.
2012-01-01
To retrospectively evaluate the feasibility of arterial spin labeling (ASL) magnetic resonance imaging (MRI) for the assessment of vascularity of renal masses in patients with impaired renal function. Between May 2007 and November 2008, 11/67 consecutive patients referred for MRI evaluation of a renal mass underwent unenhanced ASL-MRI due to moderate-to-severe chronic or acute renal failure. Mean blood flow in vascularised and non-vascularised lesions and the relation between blood flow and final diagnosis of malignancy were correlated with a 2-sided homogeneous variance t-test and the Fisher Exact Test, respectively. A p value 2 (range 7-39). The average blood flow of 11 renal masses interpreted as ASL-positive (134 +/- 85.7 mL/100 g/min) was higher than that of 6 renal masses interpreted as ASL-negative (20.5 +/- 8.1 mL/100 g/min)(p = 0.015). ASL-positivity correlated with malignancy (n = 3) or epithelial atypia (n = 1) at histopathology or progression at follow up (n = 7). ASL detection of vascularity in renal masses in patients with impaired renal function is feasible and seems to indicate neoplasia although the technique requires further evaluation. (orig.)
18. Arterial spin labeling MR imaging for characterisation of renal masses in patients with impaired renal function: initial experience
Energy Technology Data Exchange (ETDEWEB)
Pedrosa, Ivan [Beth Israel Deaconess Medical Center and Harvard Medical School, Department of Radiology, Boston, MA (United States); UT Southwestern Medical Center, Department of Radiology, Dallas, TX (United States); Rafatzand, Khashayar; Robson, Philip; Alsop, David C. [Beth Israel Deaconess Medical Center and Harvard Medical School, Department of Radiology, Boston, MA (United States); Wagner, Andrew A. [Beth Israel Deaconess Medical Center and Harvard Medical School, Surgery, Division of Urology, Boston, MA (United States); Atkins, Michael B. [Beth Israel Deaconess Medical Center and Harvard Medical School, Hematology/Oncology, Boston, MA (United States); Rofsky, Neil M. [University of Texas Southwestern Medical Center, Departments of Radiology, Dallas, TX (United States)
2012-02-15
To retrospectively evaluate the feasibility of arterial spin labeling (ASL) magnetic resonance imaging (MRI) for the assessment of vascularity of renal masses in patients with impaired renal function. Between May 2007 and November 2008, 11/67 consecutive patients referred for MRI evaluation of a renal mass underwent unenhanced ASL-MRI due to moderate-to-severe chronic or acute renal failure. Mean blood flow in vascularised and non-vascularised lesions and the relation between blood flow and final diagnosis of malignancy were correlated with a 2-sided homogeneous variance t-test and the Fisher Exact Test, respectively. A p value <0.05 was considered statistically significant. Seventeen renal lesions were evaluated in 11 patients (8 male; mean age = 70 years) (range 57-86). The median eGFR was 24 mL/min/1.73 m{sup 2} (range 7-39). The average blood flow of 11 renal masses interpreted as ASL-positive (134 +/- 85.7 mL/100 g/min) was higher than that of 6 renal masses interpreted as ASL-negative (20.5 +/- 8.1 mL/100 g/min)(p = 0.015). ASL-positivity correlated with malignancy (n = 3) or epithelial atypia (n = 1) at histopathology or progression at follow up (n = 7). ASL detection of vascularity in renal masses in patients with impaired renal function is feasible and seems to indicate neoplasia although the technique requires further evaluation. (orig.)
19. PDV-based estimation of ejecta particles' mass-velocity function from shock-loaded tin experiment
Science.gov (United States)
Franzkowiak, J.-E.; Prudhomme, G.; Mercier, P.; Lauriot, S.; Dubreuil, E.; Berthe, L.
2018-03-01
A metallic tin plate with a given surface finish of wavelength λ ≃ 60 μm and amplitude h ≃ 8 μm is explosively driven by an electro-detonator with a shock-induced breakout pressure PSB = 28 GPa (unsupported). The resulting dynamic fragmentation process, the so-called "micro-jetting," is the creation of high-speed jets of matter moving faster than the bulk metallic surface. Hydrodynamic instabilities result in the fragmentation of these jets into micron-sized metallic particles constituting a self-expanding cloud of droplets, whose areal mass, velocity, and particle size distributions are unknown. Lithium-niobate-piezoelectric sensor measured areal mass and Photonic Doppler Velocimetry (PDV) was used to get a time-velocity spectrogram of the cloud. In this article, we present both experimental mass and velocity results and we relate the integrated areal mass of the cloud to the PDV power spectral density with the assumption of a power law particle size distribution. Two models of PDV spectrograms are described. The first one accounts for the speckle statistics of the spectrum and the second one describes an average spectrum for which speckle fluctuations are removed. Finally, the second model is used for a maximum likelihood estimation of the cloud's parameters from PDV data. The estimated integrated areal mass from PDV data is found to agree well with piezoelectric results. We highlight the relevance of analyzing PDV data and correlating different diagnostics to retrieve the physical properties of ejecta particles.
20. Thermal convection loop experiments and analysis of mass transport process in Lithium/Fe-12Cr-1MoVW systems
International Nuclear Information System (INIS)
Bell, G.E.C.
1988-01-01
Lithium is an attractive coolant and breeder material for first- generation fusion reactor blankets. The compatibility of lithium with structural alloys, in the form of mass transport and deposition, may impose restrictions on blanket operating parameters such as temperature and lithium purity. A ferritic steel, such as Fe-12CrlMoVW, is a candidate for use as a structural alloy in a self-cooled lithium blanket design. Experimental data on mass transport in lithium/Fe-12CrlMoVW were obtained from two thermal convection loops which spanned the fusion relevant temperature range; one operated from 360 to 505/degree/C for 3040 hours and the other from 525 to 655/degree/C for 2510 hours. The experimental effort was supported by analysis of the mechanisms and processes of mass transport and deposition. It was found that mass transport and deposition, as measured by specimen weight change, were not simple functions of temperature for the entire temperature range investigated. The mass transfer behavior and surface morphology at low temperatures were dominated by impurity reactions of nitrogen and carbon in the lithium with the steel. In the experiment between 360 and 505/degree/C, nitrogen levels were sufficient below 450/degree/C to allow the formation of the adherent, protective corrosion product Li 9 CrN 5 . Weight losses in the 360 to 505/degree/C experiment were insensitive to temperature below 450/degree/C. Between 450 and 505/degree/C, the precipitation of carbon in the form of chromium-rich M 23 C 6 (M = Fe or Cr) carbides, due to the formation of Li 9 CrN 5 and corresponding release of carbon, resulted in weight gains for the highest temperature specimens in the experiment. 98 refs., 83 figs., 9 tabs
1. 26Al tracer experiment by accelerator mass spectrometry and its application to the studies for amyotrophic lateral sclerosis and Alzheimer's disease, 1
International Nuclear Information System (INIS)
Kobayashi, Koichi; Yumoto, Sakae; Nagai, Hisao; Hosoyama, Yoshiyuki; Imamura, Mineo; Masuzawa, Shin-ichirou; Koizumi, Yoshinobu; Yamashita, Hiroshi.
1990-01-01
Accelerator mass spectrometry (AMS) was applied for 26 Al tracer experiment to study the aluminum toxicity and metabolism in rats. To investigate the cause of amyotrophic lateral sclerosis (ALS) and Alzheimer's disease, the aluminum incorporation into the brain (cerebrum) was studied by AMS using 26 Al as a tracer. When 26 Al was intraperitoneally injected into rats, a considerable amount of 26 Al was incorporated into the cerebrum after 5-35 days of the injection. (author)
2. Teaching Mass Transfer and Filtration Using Crossflow Reverse Osmosis and Nanofiltration: An Experiment for the Undergraduate Unit Operations Lab
Science.gov (United States)
Anastasio, Daniel; McCutcheon, Jeffrey
2012-01-01
A crossflow reverse osmosis (RO) system was built for a senior-level chemical engineering unit operations laboratory course. Intended to teach students mass transfer fundamentals related to membrane separations, students tested several commercial desalination membranes, measuring water flux and salt rejections at various pressures, flow rates, and…
3. The effect of diffusivity on gas-liquid mass transfer in stirred vessels. Experiments at atmospheric and elevated pressures
NARCIS (Netherlands)
Versteeg, G.F.; Blauwhoff, P.M.M.; Swaaij, W.P.M. van
1987-01-01
Mass transfer has been studied in gas-liquid stirred vessels with horizontal interfaces which appeared to the eye to be completely smooth. Special attention has been paid to the influence of the coefficient of molecular diffusion. The results are compared with those published before. The simplifying
4. INITIAL EXPERIENCE WITH ENDOSCOPIC ULTRASOUND-GUIDED FINE NEEDLE ASPIRATION OF RENAL MASSES: indications, applications and limitations
Directory of Open Access Journals (Sweden)
Renata Nobre MOURA
2014-12-01
Full Text Available Context Tissue sampling of renal masses is traditionally performed via the percutaneous approach or laparoscopicaly. The utility of endoscopic ultrasound to biopsy renal lesions it remains unclear and few cases have been reported. Objectives To evaluate the feasibility and outcome of endoscopic ultrasound fine needle aspiration of renal tumors. Methods Consecutive subjects undergoing attempted endoscopic ultrasound fine needle aspiration of a kidney mass after evaluation with computerized tomography or magnetic resonance. Results Ten procedures were performed in nine male patients (median age 54.7 years on the right (n = 4 and left kidney (n = 4 and bilaterally (n = 1. Kidney masses (median diameter 55 mm ; range 13-160 mm were located in the upper pole (n = 3, the lower pole (n = 2 and the mesorenal region (n = 3. In two cases, the mass involved more than one kidney region. Surgical resection confirmed renal cell carcinoma in six patients in whom pre-operative endoscopic ultrasound fine needle aspiration demonstrated renal cell carcinoma. No complications were reported. Conclusions Endoscopic ultrasound fine needle aspiration appears as a safe and feasible procedure with good results and minimal morbidity.
5. Narratives of Agency: The Experiences of Braille Literacy Practitioners in the "Kha Ri Gude" South African Mass Literacy Campaign
Science.gov (United States)
McKay, Veronica I.; Romm, Norma R. A.
2015-01-01
In this article, we locate the "Kha Ri Gude" South African Mass Literacy Campaign within the context of the problem of illiteracy and exclusion in South Africa, while concentrating on various post-apartheid initiatives designed to give visually challenged adults the opportunity to become literate. We shall provide a detailed account of…
6. Measuring Gas-Phase Basicities of Amino Acids Using an Ion Trap Mass Spectrometer: A Physical Chemistry Laboratory Experiment
Science.gov (United States)
Sunderlin, Lee S.; Ryzhov, Victor; Keller, Lanea M. M.; Gaillard, Elizabeth R.
2005-01-01
An experiment is performed to measure the relative gas-phase basicities of a series of five amino acids to compare the results to literature values. The experiments use the kinetic method for deriving ion thermochemistry and allow students to perform accurate measurements of thermodynamics in a relatively short time.
7. Mass measurement of halo nuclides and beam cooling with the mass spectrometer Mistral; Mesure de masse de noyaux a halo et refroidissement de faisceaux avec l'experience MISTRAL
Energy Technology Data Exchange (ETDEWEB)
Bachelet, C
2004-12-01
Halo nuclides are a spectacular drip-line phenomenon and their description pushes nuclear theories to their limits. The most critical input parameter is the nuclear binding energy; a quantity that requires excellent measurement precision, since the two-neutron separation energy is small at the drip-line by definition. Moreover halo nuclides are typically very short-lived. Thus, a high accuracy instrument using a quick method of measurement is necessary. MISTRAL is such an instrument; it is a radiofrequency transmission mass spectrometer located at ISOLDE/CERN. In July 2003 we measured the mass of the Li{sup 11}, a two-neutron halo nuclide. Our measurement improves the precision by a factor 6, with an error of 5 keV. Moreover the measurement gives a two-neutron separation energy 20% higher than the previous value. This measurement has an impact on the radius of the nucleus, and on the state of the two valence neutrons. At the same time, a measurement of the Be{sup 11} was performed with an uncertainty of 4 keV, in excellent agreement with previous measurements. In order to measure the mass of the two-neutron halo nuclide Be{sup 14}, an ion beam cooling system is presently under development which will increase the sensitivity of the spectrometer. The second part of this work presents the development of this beam cooler using a gas-filled Paul trap. (author)
8. Design of pulsed perforated-plate columns for industrial scale mass transfer applications - present experience and the need for a model based approach
International Nuclear Information System (INIS)
Roy, Amitava
2010-01-01
Mass transfer is a vital unit operation in the processing of spent nuclear fuel in the backend of closed fuel cycle and Pulsed perforated plate extraction columns are used as mass transfer device for more than five decades. The pulsed perforated plate column is an agitated differential contactor, which has wide applicability due to its simplicity, high mass transfer efficiency, high through put, suitability for maintenance free remote operation, ease of cleaning/decontamination and cost effectiveness. Design of pulsed columns are based on a model proposed to describe the hydrodynamics and mass transfer. In equilibrium stage model, the HETS values are obtained from pilot plant experiments and then scaled empirically to design columns for industrial application. The dispersion model accounts for mass transfer kinetics and back-mixing. The drop population balance model can describe complex hydrodynamics of dispersed phase, that is, drop formation, break-up and drop-to-drop interactions. In recent years, significant progress has been made to model pulsed columns using CFD, which provides complete mathematical description of hydrodynamics in terms of spatial distribution of flow fields and 3D visualization. Under the condition of pulsation, the poly-dispersed nature of turbulent droplet swarm renders modeling difficult. In the absence of industry acceptance of proposed models, the conventional chemical engineering practice is to use HETS-NTS concept or HTU-NTU approach to design extraction columns. The practicability of HTU-NTU approach has some limitations due to the lack of experimental data on individual film mass transfer coefficients. Presently, the HETS-NTS concept has been used for designing the columns, which has given satisfactory performance. The design objective is mainly to arrive at the diameter and height of the mass transfer section for a specific plate geometry, fluid properties and pulsing condition to meet the intended throughput (capacity) and mass
9. An analysis of nuclear fuel burnup in the AGR-1 TRISO fuel experiment using gamma spectrometry, mass spectrometry, and computational simulation techniques
International Nuclear Information System (INIS)
Harp, Jason M.; Demkowicz, Paul A.; Winston, Philip L.; Sterbentz, James W.
2014-01-01
Highlights: • The burnup of irradiated AGR-1 TRISO fuel was analyzed using gamma spectrometry. • The burnup of irradiated AGR-1 TRISO fuel was also analyzed using mass spectrometry. • Agreement between experimental results and neutron physics simulations was excellent. - Abstract: AGR-1 was the first in a series of experiments designed to test US TRISO fuel under high temperature gas-cooled reactor irradiation conditions. This experiment was irradiated in the Advanced Test Reactor (ATR) at Idaho National Laboratory (INL) and is currently undergoing post-irradiation examination (PIE) at INL and Oak Ridge National Laboratory. One component of the AGR-1 PIE is the experimental evaluation of the burnup of the fuel by two separate techniques. Gamma spectrometry was used to non-destructively evaluate the burnup of all 72 of the TRISO fuel compacts that comprised the AGR-1 experiment. Two methods for evaluating burnup by gamma spectrometry were developed, one based on the Cs-137 activity and the other based on the ratio of Cs-134 and Cs-137 activities. Burnup values determined from both methods compared well with the values predicted from simulations. The highest measured burnup was 20.1% FIMA (fissions per initial heavy metal atom) for the direct method and 20.0% FIMA for the ratio method (compared to 19.56% FIMA from simulations). An advantage of the ratio method is that the burnup of the cylindrical fuel compacts can be determined in small (2.5 mm) axial increments and an axial burnup profile can be produced. Destructive chemical analysis by inductively coupled mass spectrometry (ICP-MS) was then performed on selected compacts that were representative of the expected range of fuel burnups in the experiment to compare with the burnup values determined by gamma spectrometry. The compacts analyzed by mass spectrometry had a burnup range of 19.3% FIMA to 10.7% FIMA. The mass spectrometry evaluation of burnup for the four compacts agreed well with the gamma
10. Advanced Measurement and Simulation Procedure for the Identification of Heat and Mass Transfer Parameters in Dynamic Adsorption Experiments
Directory of Open Access Journals (Sweden)
Andreas Velte
2017-08-01
Full Text Available Thermally-driven heat pumps can help to mitigate CO2 emissions by enhancing the efficiency of heating systems or by driving cooling systems with waste or solar heat. In order to make the thermally-driven systems more attractive for the end consumer, these systems need a higher power density. A higher power density can be achieved by intensifying the heat and mass transfer processes within the adsorption heat exchanger. For the optimization of this key component, a numerical model of the non-isothermal adsorption dynamics can be applied. The calibration of such a model can be difficult, since heat and mass transfer processes are strongly coupled. We present a measurement and simulation procedure that makes it possible to calibrate the heat transfer part of the numerical model separately from the mass transfer part. Furthermore, it is possible to identify the parts of the model that need to be improved. For this purpose, a modification of the well-known large temperature jump method is developed. The newly-introduced measurements are conducted under an inert N2 atmosphere, and the surface temperature of the sample is measured with an infrared sensor. We show that the procedure is applicable for two completely different types of samples: a loose grains configuration and a fibrous structure that is directly crystallized.
11. Mass Enhancement of Nearly Trivalent Compound EuCo2Si2: Studied by the de Haas-van Alphen Experiments and Energy Band Calculations
International Nuclear Information System (INIS)
Ōnuki, Yoshichika; Hedo, Masato; Nakama, Takao; Nakamura, Ai; Aoki, Dai; Boukahil, Mounir; Haga, Yoshinori; Takeuchi, Tetsuya; Harima, Hisatomo
2015-01-01
We succeeded in growing single crystals of EuCo 2 Si 2 by the Bridgman method, and carried out the de Haas-van Alphen (dHvA) experiments. EuCo 2 Si 2 was previously studied from a viewpoint of the trivalent electronic state on the basis of the magnetic susceptibility and X-ray absorption experiments, whereas most of the other Eu compounds order magnetically, with the divalent electronic state. The detected dHvA branches in the present experiments are found to be explained by the results of the full potential linearized augmented plane wave energy band calculations on the basis of a local density approximation (LDA) for YCo 2 Si 2 (LDA) and EuCo 2 Si 2 (LDA + U), revealing the trivalent electronic state. The detected cyclotron effective masses are moderately large, ranging from 1.2 to 2.9 m 0
12. Properties of the Higgs boson in the 4 lepton final state at the LHC with the ATLAS experiment: mass, limits on the high mass contribution and on the total width
CERN Document Server
AUTHOR|(SzGeCERN)719049
The theme of the analyses presented in this Thesis is the measurement of the Higgs boson properties in the H$\\rightarrow$ZZ$\\rightarrow$4l decay channel with the ATLAS experiment at the LHC. A detailed overview on the electron calibration process is first presented. In this regard, the track-cluster combination algorithm is found to improve the energy resolution of low $E_{T}$ electrons by exploiting both track and cluster information into a maximum likelihood fit. The improvement in resolution is approximately 18-20% for J/$\\Psi$ dielectron decays, and of the order of 3% for Z$\\rightarrow$ee events. In addition, the E-p combination algorithm has also been applied to the H$\\rightarrow$ZZ$\\rightarrow$4l channel with electrons in the final state resulting in a non-negligible gain on the invariant mass distribution (4-5%). Secondly, the Higgs mass and its total width are evaluated in the H$\\rightarrow$ZZ$\\rightarrow$4l channel. The Higgs mass is measured in the 4l decay channel with particular interest on the be...
13. Single-experiment simultaneous-measurement of elemental mass-attenuation coefficients of hydrogen, carbon and oxygen for 0.123-1.33 MeV gamma rays
International Nuclear Information System (INIS)
Teli, M.T.; Nathuram, R.; Mahajan, C.S.
2000-01-01
As it is inconvenient to use elements like hydrogen, carbon and oxygen in pure forms for measurement of their gamma mass-attenuation coefficients, the measurements are to be done indirectly, by using compounds of the elements or a mixture of them. We give here a simple method of measuring the total mass-attenuation coefficients μ/ρ of the elements in a compound simultaneously and in a single experiment through the measurements of the μ/ρ values of the concerned compounds and using the mixture rule. The method is applied for the measurement of μ/ρ of hydrogen, carbon and oxygen by using acetone, ethanol and 1-propanol. Our results (for E γ =0.123-1.33 MeV) are seen to be in better agreement with the theoretical values of Hubbell and Seltzer (1995) [Hubbell J.H. and Seltzer S.M. (1995). Tables of X-ray mass attenuation coefficients and mass energy-absorption coefficients 1 keV to 20 MeV for elements Z=1 to 92 and 48 additional substances of dosimetric interest. NISTIR 5632] as compared to the results of El-Kateb and Abdul-Hamid (1991) [El-Kateb, A.H., Abdul-Hamid, A.S., 1991. Photon attenuation coefficient study of some materials containing hydrogen, carbon, and oxygen. Appl. Rad. Isot. 42, 303-307
14. Proton-gamma coincidence experiment on medium mass nuclei at 400MeV and study of reaction mechanisms
International Nuclear Information System (INIS)
Baldit, Alain.
1981-01-01
Previous γ ray production experiments produced by proton on nuclei show important cross sections for residual nuclei corresponding to a four nucleon (2p + 2n) removal. With our (p - γ) coincidence experiment the forward emitted proton reflects the primary interaction and the γ spectra characterizes the final state of the reaction. Protons are detected with a magnetic spectrometer and γ rays are selected with a Ge(Li) diode. Angular and momentum analysis of scattered protons demonstrate a primary quasi free process on nucleons. No indication of knock out reactions on clusters has been seen. The residual nuclei are mainly produced by evaporation processes. A theoretical calculation involving intranuclear cascades and evaporation processes has been performed. The nucleus model is based upon a Fermi gas and nuclear density agrees with diffusion electron experiments. Residual nuclei far from target are well described with a such model. Residual nuclei near the target are sensitive to the nuclear structure [fr
15. Patients' perceptions of cosmetic surgery at a time of globalization, medical consumerism, and mass media culture: a French experience.
Science.gov (United States)
Lazar, Câlin Constantin; Deneuve, Sophie
2013-08-01
The global popularity of cosmetic surgery, combined with mass media attention on medical consumerism, has resulted in misinformation that may have negatively affected the "collective image" of aesthetic practitioners. The authors assess patients' perceptions of cosmetic surgery and analyze their decision-making processes. During a 2-year period, 250 consecutive patients presenting to either of 2 public hospitals for cosmetic surgery treatment were asked to complete a 7-item questionnaire evaluating their knowledge of opinions about, and referring practices for, aesthetic procedures. Patients undergoing oncologic, postbariatric, or reconstructive procedures were not included in the study. After exclusion of 71 cases for refusal or incompletion, 179 questionnaires were retained and analyzed (from 162 women and 17 men). Overall, repair (70.4%), comfort (45.3%), and health (40.8%) were the words most frequently associated with cosmetic surgery. Quality of preoperative information (69.3%), patient-physician relationship (65.4%), and results seen in relatives/friends (46.3%) were the most important criteria for selecting a cosmetic surgeon. Moreover, 82.7% of patients knew the difference between cosmetic surgery and cosmetic medicine. Although potential patients appear to be more educated about cosmetic surgery than they were several years ago, misinformation still persists. As physicians, we must be responsible for disseminating accurate education and strengthening our collaboration with general practitioners to improve not only our results but also the accuracy of information in the mass media.
16. Critical Heat Flux Phenomena at HighPressure & Low Mass Fluxes: NEUP Final Report Part I: Experiments
Energy Technology Data Exchange (ETDEWEB)
Corradini, Michael [Univ. of Wisconsin, Madison, WI (United States); Wu, Qiao [Oregon State Univ., Corvallis, OR (United States)
2015-04-30
This report is a preliminary document presenting an overview of the Critical Heat Flux (CHF) phenomenon, the High Pressure Critical Heat Flux facility (HPCHF), preliminary CHF data acquired, and the future direction of the research. The HPCHF facility has been designed and built to study CHF at high pressure and low mass flux ranges in a rod bundle prototypical of conceptual Small Modular Reactor (SMR) designs. The rod bundle is comprised of four electrically heated rods in a 2x2 square rod bundle with a prototypic chopped-cosine axial power profile and equipped with thermocouples at various axial and circumferential positions embedded in each rod for CHF detection. Experimental test parameters for CHF detection range from pressures of ~80 – 160 bar, mass fluxes of ~400 – 1500 kg/m2s, and inlet water subcooling from ~30 – 70°C. The preliminary data base established will be further extended in the future along with comparisons to existing CHF correlations, models, etc. whose application ranges may be applicable to the conditions of SMRs.
17. Effective Responder Communication Improves Efficiency and Psychological Outcomes in a Mass Decontamination Field Experiment: Implications for Public Behaviour in the Event of a Chemical Incident
Science.gov (United States)
Carter, Holly; Drury, John; Amlôt, Richard; Rubin, G. James; Williams, Richard
2014-01-01
The risk of incidents involving mass decontamination in response to a chemical, biological, radiological, or nuclear release has increased in recent years, due to technological advances, and the willingness of terrorists to use unconventional weapons. Planning for such incidents has focused on the technical issues involved, rather than on psychosocial concerns. This paper presents a novel experimental study, examining the effect of three different responder communication strategies on public experiences and behaviour during a mass decontamination field experiment. Specifically, the research examined the impact of social identity processes on the relationship between effective responder communication, and relevant outcome variables (e.g. public compliance, public anxiety, and co-operative public behaviour). All participants (n = 111) were asked to visualise that they had been involved in an incident involving mass decontamination, before undergoing the decontamination process, and receiving one of three different communication strategies: 1) ‘Theory-based communication’: Health-focused explanations about decontamination, and sufficient practical information; 2) ‘Standard practice communication’: No health-focused explanations about decontamination, sufficient practical information; 3) ‘Brief communication’: No health-focused explanations about decontamination, insufficient practical information. Four types of data were collected: timings of the decontamination process; observational data; and quantitative and qualitative self-report data. The communication strategy which resulted in the most efficient progression of participants through the decontamination process, as well as the fewest observations of non-compliance and confusion, was that which included both health-focused explanations about decontamination and sufficient practical information. Further, this strategy resulted in increased perceptions of responder legitimacy and increased
18. Effective responder communication improves efficiency and psychological outcomes in a mass decontamination field experiment: implications for public behaviour in the event of a chemical incident.
Science.gov (United States)
Carter, Holly; Drury, John; Amlôt, Richard; Rubin, G James; Williams, Richard
2014-01-01
The risk of incidents involving mass decontamination in response to a chemical, biological, radiological, or nuclear release has increased in recent years, due to technological advances, and the willingness of terrorists to use unconventional weapons. Planning for such incidents has focused on the technical issues involved, rather than on psychosocial concerns. This paper presents a novel experimental study, examining the effect of three different responder communication strategies on public experiences and behaviour during a mass decontamination field experiment. Specifically, the research examined the impact of social identity processes on the relationship between effective responder communication, and relevant outcome variables (e.g. public compliance, public anxiety, and co-operative public behaviour). All participants (n = 111) were asked to visualise that they had been involved in an incident involving mass decontamination, before undergoing the decontamination process, and receiving one of three different communication strategies: 1) 'Theory-based communication': Health-focused explanations about decontamination, and sufficient practical information; 2) 'Standard practice communication': No health-focused explanations about decontamination, sufficient practical information; 3) 'Brief communication': No health-focused explanations about decontamination, insufficient practical information. Four types of data were collected: timings of the decontamination process; observational data; and quantitative and qualitative self-report data. The communication strategy which resulted in the most efficient progression of participants through the decontamination process, as well as the fewest observations of non-compliance and confusion, was that which included both health-focused explanations about decontamination and sufficient practical information. Further, this strategy resulted in increased perceptions of responder legitimacy and increased identification with
19. Effective responder communication improves efficiency and psychological outcomes in a mass decontamination field experiment: implications for public behaviour in the event of a chemical incident.
Directory of Open Access Journals (Sweden)
Holly Carter
Full Text Available The risk of incidents involving mass decontamination in response to a chemical, biological, radiological, or nuclear release has increased in recent years, due to technological advances, and the willingness of terrorists to use unconventional weapons. Planning for such incidents has focused on the technical issues involved, rather than on psychosocial concerns. This paper presents a novel experimental study, examining the effect of three different responder communication strategies on public experiences and behaviour during a mass decontamination field experiment. Specifically, the research examined the impact of social identity processes on the relationship between effective responder communication, and relevant outcome variables (e.g. public compliance, public anxiety, and co-operative public behaviour. All participants (n = 111 were asked to visualise that they had been involved in an incident involving mass decontamination, before undergoing the decontamination process, and receiving one of three different communication strategies: 1 'Theory-based communication': Health-focused explanations about decontamination, and sufficient practical information; 2 'Standard practice communication': No health-focused explanations about decontamination, sufficient practical information; 3 'Brief communication': No health-focused explanations about decontamination, insufficient practical information. Four types of data were collected: timings of the decontamination process; observational data; and quantitative and qualitative self-report data. The communication strategy which resulted in the most efficient progression of participants through the decontamination process, as well as the fewest observations of non-compliance and confusion, was that which included both health-focused explanations about decontamination and sufficient practical information. Further, this strategy resulted in increased perceptions of responder legitimacy and increased
20. The giant superconducting magnet system of 10,000 tons mass for fusion experiment at Cadarache, France
International Nuclear Information System (INIS)
Sahu, A.K.
2013-01-01
1. sup 26 Al tracer experiment by accelerator mass spectrometry and its application to the studies for amyotrophic lateral sclerosis and Alzheimer's disease, 1
Energy Technology Data Exchange (ETDEWEB)
Kobayashi, Koichi (Tokyo Univ. (Japan). Research Center for Nuclear Science and Technology); Yumoto, Sakae; Nagai, Hisao; Hosoyama, Yoshiyuki; Imamura, Mineo; Masuzawa, Shin-ichirou; Koizumi, Yoshinobu; Yamashita, Hiroshi
1990-12-01
Accelerator mass spectrometry (AMS) was applied for {sup 26}Al tracer experiment to study the aluminum toxicity and metabolism in rats. To investigate the cause of amyotrophic lateral sclerosis (ALS) and Alzheimer's disease, the aluminum incorporation into the brain (cerebrum) was studied by AMS using {sup 26}Al as a tracer. When {sup 26}Al was intraperitoneally injected into rats, a considerable amount of {sup 26}Al was incorporated into the cerebrum after 5-35 days of the injection. (author).
2. A Self-Sensing Piezoelectric MicroCantilever Biosensor for Detection of Ultrasmall Adsorbed Masses: Theory and Experiments
Directory of Open Access Journals (Sweden)
Srinivas Sridhar
2013-05-01
Full Text Available Detection of ultrasmall masses such as proteins and pathogens has been made possible as a result of advancements in nanotechnology. Development of label-free and highly sensitive biosensors has enabled the transduction of molecular recognition into detectable physical quantities. Microcantilever (MC-based systems have played a widespread role in developing such biosensors. One of the most important drawbacks of all of the available biosensors is that they all come at a very high cost. Moreover, there are certain limitations in the measurement equipments attached to the biosensors which are mostly optical measurement systems. A unique self-sensing detection technique is proposed in this paper in order to address most of the limitations of the current measurement systems. A self-sensing bridge is used to excite piezoelectric MC-based sensor functioning in dynamic mode, which simultaneously measures the system’s response through the self-induced voltage generated in the piezoelectric material. As a result, the need for bulky, expensive read-out equipment is eliminated. A comprehensive mathematical model is presented for the proposed self-sensing detection platform using distributed-parameters system modeling. An adaptation strategy is then implemented in the second part in order to compensate for the time-variation of piezoelectric properties which dynamically improves the behavior of the system. Finally, results are reported from an extensive experimental investigation carried out to prove the capability of the proposed platform. Experimental results verified the proposed mathematical modeling presented in the first part of the study with accuracy of 97.48%. Implementing the adaptation strategy increased the accuracy to 99.82%. These results proved the measurement capability of the proposed self-sensing strategy. It enables development of a cost-effective, sensitive and miniaturized mass sensing platform.
3. Automated Assessment of Left Ventricular Function and Mass Using Heart Deformation Analysis: Initial Experience in 160 Older Adults.
Science.gov (United States)
Lin, Kai; Collins, Jeremy D; Lloyd-Jones, Donald M; Jolly, Marie-Pierre; Li, Debiao; Markl, Michael; Carr, James C
2016-03-01
To assess the performance of automated quantification of left ventricular function and mass based on heart deformation analysis (HDA) in asymptomatic older adults. This study complied with Health Insurance Portability and Accountability Act regulations. Following the approval of the institutional review board, 160 asymptomatic older participants were recruited for cardiac magnetic resonance imaging including two-dimensional cine images covering the entire left ventricle in short-axis view. Data analysis included the calculation of left ventricular ejection fraction (LVEF), left ventricular mass (LVM), and cardiac output (CO) using HDA and standard global cardiac function analysis (delineation of end-systolic and end-diastolic left ventricle epi- and endocardial borders). The agreement between methods was evaluated using intraclass correlation coefficient (ICC) and coefficient of variation (CoV). HDA had a shorter processing time than the standard method (1.5 ± 0.3 min/case vs. 5.8 ± 1.4 min/case, P HDA. There was a systemic bias toward lower LVEF (62.8% ± 8.3% vs. 69.3% ± 6.7%, P HDA compared to the standard technique. Conversely, HDA overestimated LVM (114.8 ± 30.1 g vs. 100.2 ± 29.0 g, P HDA has the potential to measure LVEF, CO, and LVM without the need for user interaction based on standard cardiac two-dimensional cine images. Copyright © 2015 The Association of University Radiologists. Published by Elsevier Inc. All rights reserved.
4. Search for a Structure in the B_{s}^{0}π^{±} Invariant Mass Spectrum with the ATLAS Experiment.
Science.gov (United States)
2018-05-18
A search for the narrow structure, X(5568), reported by the D0 Collaboration in the decay sequence X→B_{s}^{0}π^{±}, B_{s}^{0}→J/ψϕ, is presented. The analysis is based on a data sample recorded with the ATLAS detector at the LHC corresponding to 4.9 fb^{-1} of pp collisions at 7 TeV and 19.5 fb^{-1} at 8 TeV. No significant signal was found. Upper limits on the number of signal events, with properties corresponding to those reported by D0, and on the X production rate relative to B_{s}^{0} mesons, ρ_{X}, were determined at 95% confidence level. The results are N(X)<382 and ρ_{X}<0.015 for B_{s}^{0} mesons with transverse momenta above 10 GeV, and N(X)<356 and ρ_{X}<0.016 for transverse momenta above 15 GeV. Limits are also set for potential B_{s}^{0}π^{±} resonances in the mass range 5550 to 5700 MeV.
5. [Hepatic fine needle aspiration biopsy. Experience in the study of hepatic masses at the Salvador Zubiran National Institute of Nutrition].
Science.gov (United States)
Angeles-Angeles, A; Gamboa-Domínguez, A; Velázquez Fernández, D; Muñoz-Fernández, L
1994-01-01
The results of 114 fine-needle aspiration biopsies (FNAB) of the liver performed during six years (1987-1992) at the Departament of Pathology of the Instituto Nacional de la Nutrición Salvador Zubirán are presented. All were done by radiologists under ultrasonographic (three cases) or computerized tomographic guidance (111 cases). In order to determine the diagnostic accuracy, diagnoses made by FNAB were compared with those made by histological examination (coarse biopsies or surgical specimens) and/or by other diagnostic procedures including the clinical follow-up. Six cases were excluded because clinical information was not available. In 92 cases (85.2%) a correct diagnosis was made, in six (5.5%) the sample was inadequate and in 10 (9.3%) the diagnosis made by FNAB was incorrect. The diagnoses made were as follows: hepatocarcinoma 44, metastatic carcinoma 27, inflammatory lesions 12, regeneration 10, normal eight, unclassified carcinoma five, and lymphoma two. The sensitivity was 96.2, specificity 93.1, positive predictive value 97.4, negative predictive value 90.0, accuracy 95.3 and prevalence 73.1. There were three false negative and two false positive for carcinoma. These figures are similar to those found by other authors. No relevant complications were observed. It is concluded that FNAB of the liver is a safe, inexpensive and reliable method in the diagnoses of liver masses.
6. Search for a Structure in the Bs0π± Invariant Mass Spectrum with the ATLAS Experiment
Science.gov (United States)
2018-05-01
A search for the narrow structure, X (5568 ), reported by the D0 Collaboration in the decay sequence X →Bs0π±, Bs0→J /ψ ϕ , is presented. The analysis is based on a data sample recorded with the ATLAS detector at the LHC corresponding to 4.9 fb-1 of p p collisions at 7 TeV and 19.5 fb-1 at 8 TeV. No significant signal was found. Upper limits on the number of signal events, with properties corresponding to those reported by D0, and on the X production rate relative to Bs0 mesons, ρX, were determined at 95% confidence level. The results are N (X )<382 and ρX<0.015 for Bs0 mesons with transverse momenta above 10 GeV, and N (X )<356 and ρX<0.016 for transverse momenta above 15 GeV. Limits are also set for potential Bs0π± resonances in the mass range 5550 to 5700 MeV.
7. Properties of the Higgs boson in the 4 leptons final state with the ATLAS experiment at the LHC: mass, limit on the high mass contribution and on the Higgs width
International Nuclear Information System (INIS)
Calandri, Alessandro
2015-01-01
The theme of the analyses presented in this Thesis is the measurement of the Higgs boson properties in the H→ZZ→4l decay channel with the ATLAS experiment at the LHC. A detailed overview on the electron calibration process is first presented. In this regard, the track-cluster combination algorithm is found to improve the energy resolution of low ET electrons by exploiting both track and cluster information into a maximum likelihood fit. The improvement in resolution is approximately 18-20% for J/ψ dielectron decays, and of the order of 3% for Z→ee events. In addition, the E-p combination algorithm has also been applied to the H→ZZ→4l channel with electrons in the final state resulting in a non-negligible gain on the invariant mass distribution (4-5%). Secondly, the Higgs mass and its total width are evaluated in the H→ZZ→4l channel. The Higgs mass is measured in the 4l decay channel with particular interest on the beneficial effects brought by the improved electron calibration and the track-cluster combination. The mass on the full 2011 and 2012 datasets is worked out with a 2-dimensional fit on the invariant mass of the 4 lepton final state, m4l, and on a boosted decision tree (BDT)-based output conceived against the main ZZ irreducible background and constructed on variables that are sensitive to the Higgs boson spin-parity state. Regarding the Higgs width, results are based on a relatively recent approach aimed at indirectly constraining the Higgs boson width by exploiting the m4l high-mass region where the Higgs boson acts as a propagator. The Higgs production cross section in the on-shell m4l region, where the Higgs boson is a resonance, depends on the total Higgs width, whereas this is not the case for the high mass m4l (off-shell). Limits on the Higgs width can be therefore set when merging the off-shell results with the on-shell ones. A limit of 6.7 times ΓSMH is obtained in the four lepton channel. Secondly, by combining with the on
8. A Foreground Masking Strategy for [C II] Intensity Mapping Experiments Using Galaxies Selected by Stellar Mass and Redshift
Science.gov (United States)
Sun, G.; Moncelsi, L.; Viero, M. P.; Silva, M. B.; Bock, J.; Bradford, C. M.; Chang, T.-C.; Cheng, Y.-T.; Cooray, A. R.; Crites, A.; Hailey-Dunsheath, S.; Uzgil, B.; Hunacek, J. R.; Zemcov, M.
2018-04-01
Intensity mapping provides a unique means to probe the epoch of reionization (EoR), when the neutral intergalactic medium was ionized by energetic photons emitted from the first galaxies. The [C II] 158 μm fine-structure line is typically one of the brightest emission lines of star-forming galaxies and thus a promising tracer of the global EoR star formation activity. However, [C II] intensity maps at 6 ≲ z ≲ 8 are contaminated by interloping CO rotational line emission (3 ≤ J upp ≤ 6) from lower-redshift galaxies. Here we present a strategy to remove the foreground contamination in upcoming [C II] intensity mapping experiments, guided by a model of CO emission from foreground galaxies. The model is based on empirical measurements of the mean and scatter of the total infrared luminosities of galaxies at z {10}8 {M}ȯ selected in the K-band from the COSMOS/UltraVISTA survey, which can be converted to CO line strengths. For a mock field of the Tomographic Ionized-carbon Mapping Experiment, we find that masking out the “voxels” (spectral–spatial elements) containing foreground galaxies identified using an optimized CO flux threshold results in a z-dependent criterion {m}{{K}}AB}≲ 22 (or {M}* ≳ {10}9 {M}ȯ ) at z cost of a moderate ≲8% loss of total survey volume.
9. Church music inculturation by way of an experiment of arrangement of Dolo-Dolo mass ordinarium accompaniment- composed by Mateus Weruin for woodwind quintet
Directory of Open Access Journals (Sweden)
Yohanes Ruswanto
2017-08-01
Full Text Available Inculturation of Church music in an experiment of creating this arrangement aims to bring a different form of musical ordinarium accompaniment form of Dolo-Dolo Mass from Flores, with a different media that uses the woodwind quintet (flute, oboe, clarinet, French horn, and Basson. The experiment took one of the ordinary songs from Madah Bakti “Tuhan Kasihanilah Kami”. The harmonization fine-tunes to the chorus arrangement composed by Mateus Weruin. The literature study was conducted through collecting references on the art of Dolo-Dolo and woodwind quintet so it can be used to create an idea for this arrangement. The result shows that a rhythmic character that characterizes the traditional Flores music lies in a dotted sixteenth pattern. The richness of sounds and agile characters coming from each instrument creates a percussive atmosphere of Flores folk music. The result of the arrangement experiment can be used to enrich the reference of accompaniment music to the general public and specifically, the Catholic Church.
10. Pilot experiment to control Medfly, Ceratitis capitata (Wied.) (Diptera: Tephritidae) using mass trapping technique in a custard apple (Annona cherimolia Mill.) orchard
International Nuclear Information System (INIS)
Ros, J.P.; Escobar, I.; Garcia Tapia, F.J.; Aranda, G.
2000-01-01
Recently, as a result of assays coordinated by the International Atomic Energy Agency (and participated by Spain), it was decided that ammonium acetate, putrescine and trimethylamine be included in low release polyetilene bag dispensers (Biolure, Consep, Co) as the Mediterranean fruit fly (Medfly) females were greatly attracted by them. These synthetic substances are placed in traps at the frequency of one and a half months to two months. If Tephry traps are used, one DDVP wafer (containing Vapona or a similar substance) is enough to kill the flies that enter them. These attractants make it unnecessary to replenish the liquids in the Mcphail traps and remain effective throughout the entire fruit season. The Caja Rural de Granada (a bank of farmers) encourages all techniques that increase crop profits for farmers. It is even more desirable if such crop profitability can be achieved without the application of insecticides which may result in the likely presence of toxic residues. In the light of the results achieved by the attractants with regard to female Medflies, the Caja Rural de Granada, together with the National Institute of Agricultural Research Counselling (Agricultural Department), performed an experiment on mass trapping to confirm whether it is possible to protect a fruit plantation with the application of this biotechnical method. Due to the great risk of this initial experiment, the farmers were free to use insecticides as often as they deemed necessary so that no damages due to any plague could be blamed on the experiment
11. Definition and calibration of the hadronic recoil in view of the measurement of the $W$ boson mass with the CMS experiment
CERN Document Server
AUTHOR|(CDS)2099460; Rolandi, Gigi
Although the Standard Model of particle physics might appear as a complete theory, there are several theoretical problems and experimental observations which suggest that it is not the complete and definitive theory. For this reason, new physics is searched for both directly, and through precise measurements of standard model observables. A measurement of the mass of the $W$ boson with a relative precision of $10^{-4}$ is of primary interest because of the small uncertainty in its theoretical prediction, and because the LHC experiments have already collected several hundreds of millions of $W$ boson decays, which allow precise measurements of its properties. Such a measurement may represent a turning point in the world of particle physics, showing that the SM is inconsistent at this energy scale. In order to achieve such precision, it is requested to master the detector, the analysis, and the theoretical predictions at an unprecedented level. A study of the hadronic recoil is described in this thesis, resulti...
12. A novel membrane inlet mass spectrometer method to measure ¹⁵NH4₄⁺ for isotope-enrichment experiments in aquatic ecosystems.
Science.gov (United States)
Yin, Guoyu; Hou, Lijun; Liu, Min; Liu, Zhanfei; Gardner, Wayne S
2014-08-19
Nitrogen (N) pollution in aquatic ecosystems has attracted much attention over the past decades, but the dynamics of this bioreactive element are difficult to measure in aquatic oxygen-transition environments. Nitrogen-transformation experiments often require measurement of (15)N-ammonium ((15)NH4(+)) ratios in small-volume (15)N-enriched samples. Published methods to determine N isotope ratios of dissolved ammonium require large samples and/or costly equipment and effort. We present a novel ("OX/MIMS") method to determine N isotope ratios for (15)NH4(+) in experimental waters previously enriched with (15)N compounds. Dissolved reduced (15)N (dominated by (15)NH4(+)) is oxidized with hypobromite iodine to nitrogen gas ((29)N2 and/or (30)N2) and analyzed by membrane inlet mass spectrometry (MIMS) to quantify (15)NH4(+) concentrations. The N isotope ratios, obtained by comparing the (15)NH4(+) to total ammonium (via autoanalyzer) concentrations, are compared to the ratios of prepared standards. The OX/MIMS method requires only small sample volumes of water (ca. 12 mL) or sediment slurries and is rapid, convenient, accurate, and precise (R(2) = 0.9994, p < 0.0001) over a range of salinities and (15)N/(14)N ratios. It can provide data needed to quantify rates of ammonium regeneration, potential ammonium uptake, and dissimilatory nitrate reduction to ammonium (DNRA). Isotope ratio results agreed closely (R = 0.998, P = 0.001) with those determined independently by isotope ratio mass spectrometry for DNRA measurements or by ammonium isotope retention time shift liquid chromatography for water-column N-cycling experiments. Application of OX/MIMS should simplify experimental approaches and improve understanding of N-cycling rates and fate in a variety of freshwater and marine environments.
13. Measurement of the Top Quark Mass using Dilepton Events and a Neutrino Weighting Algorithm with the D0 Experiment at the Tevatron (Run II)
International Nuclear Information System (INIS)
Meyer, Joerg; Bonn U
2007-01-01
measurement of the top quark mass by the D0 experiment at Fermilab in the dilepton final states. The comparison of the measured top quark masses in different final states allows an important consistency check of the Standard Model. Inconsistent results would be a clear hint of a misinterpretation of the analyzed data set. With the exception of the Higgs boson, all particles predicted by the Standard Model have been found. The search for the Higgs boson is one of the main focuses in high energy physics. The theory section will discuss the close relationship between the physics of the Higgs boson and the top quark
14. Evaporation of liquid droplets of nano- and micro-meter size as a function of molecular mass and intermolecular interactions: experiments and molecular dynamics simulations.
Science.gov (United States)
Hołyst, Robert; Litniewski, Marek; Jakubczyk, Daniel
2017-09-13
Transport of heat to the surface of a liquid is a limiting step in the evaporation of liquids into an inert gas. Molecular dynamics (MD) simulations of a two component Lennard-Jones (LJ) fluid revealed two modes of energy transport from a vapour to an interface of an evaporating droplet of liquid. Heat is transported according to the equation of temperature diffusion, far from the droplet of radius R. The heat flux, in this region, is proportional to temperature gradient and heat conductivity in the vapour. However at some distance from the interface, Aλ, (where λ is the mean free path in the gas), the temperature has a discontinuity and heat is transported ballistically i.e. by direct individual collisions of gas molecules with the interface. This ballistic transport reduces the heat flux (and consequently the mass flux) by the factor R/(R + Aλ) in comparison to the flux obtained from temperature diffusion. Thus it slows down the evaporation of droplets of sizes R ∼ Aλ and smaller (practically for sizes from 10 3 nm down to 1 nm). We analyzed parameter A as a function of interactions between molecules and their masses. The rescaled parameter, A(k B T b /ε 11 ) 1/2 , is a linear function of the ratio of the molecular mass of the liquid molecules to the molecular mass of the gas molecules, m 1 /m 2 (for a series of chemically similar compounds). Here ε 11 is the interaction parameter between molecules in the liquid (proportional to the enthalpy of evaporation) and T b is the temperature of the gas in the bulk. We tested the predictions of MD simulations in experiments performed on droplets of ethylene glycol, diethylene glycol, triethylene glycol and tetraethylene glycol. They were suspended in an electrodynamic trap and evaporated into dry nitrogen gas. A changes from ∼1 (for ethylene glycol) to approximately 10 (for tetraethylene glycol) and has the same dependence on molecular parameters as obtained for the LJ fluid in MD simulations. The value of x = A
15. Measurement of the Top Quark Mass using Dilepton Events and a Neutrino Weighting Algorithm with the D0 Experiment at the Tevatron (Run II)
Energy Technology Data Exchange (ETDEWEB)
Meyer, Joerg [Univ. of Bonn (Germany)
2007-01-01
measurement of the top quark mass by the D0 experiment at Fermilab in the dilepton final states. The comparison of the measured top quark masses in different final states allows an important consistency check of the Standard Model. Inconsistent results would be a clear hint of a misinterpretation of the analyzed data set. With the exception of the Higgs boson, all particles predicted by the Standard Model have been found. The search for the Higgs boson is one of the main focuses in high energy physics. The theory section will discuss the close relationship between the physics of the Higgs boson and the top quark.
16. Mass of the photon
International Nuclear Information System (INIS)
Goldhaber, A.S.; Nieto, M.M.
1976-01-01
Several experiments are discussed for which results are equivalent to the catching and weighing of a photon. It is noted that none of the experiments has proved the rest mass to be zero and that such a proof may be impossible. It is shown by such experiments that the rest mass is less than the limit of accuracy of the experiment. These limits have approached ever closer to zero, and the most recent values are exceedingly small
17. An intercomparison experiment on isotope dilution thermal ionisation mass spectrometry using plutonium-239 spike for the determination of plutonium concentration in dissolver solution of irradiated fuel
International Nuclear Information System (INIS)
Aggarwal, S.K.; Shah, P.M.; Saxena, M.K.; Jain, H.C.; Gurba, P.B.; Babbar, R.K.; Udagatti, S.V.; Moorthy, A.D.; Singh, R.K.; Bajpai, D.D.
1996-01-01
Determination of plutonium concentration in the dissolver solution of irradiated fuel is one of the key measurements in the nuclear fuel cycle. This report presents the results of an intercomparison experiment performed between Fuel Chemistry Division (FCD) at BARC and PREFRE, Tarapur for determining plutonium concentration in dissolver solution of irradiated fuel using 239 Pu spike in isotope dilution thermal ionisation mass spectrometry (ID-TIMS). The 239 Pu spike method was previously established at FCD as viable alternative to the imported enriched 242 Pu or 244 Pu; the spike used internationally for plutonium concentration determination by IDMS in dissolver solution of irradiated fuel. Precision and accuracy achievable for determining plutonium concentration are compared under the laboratory and the plant conditions using 239 Pu spike in IDMS. For this purpose, two different dissolver solutions with 240 Pu/ 239 Pu atom ratios of about 0.3 and 0.07 corresponding, respectively, to high and low burn-up fuels, were used. The results of the intercomparison experiment demonstrate that there is no difference in the precision values obtained under the laboratory and the plant conditions; with mean precision values of better than 0.2%. Further, the plutonium concentration values determined by the two laboratories agreed within 0.3%. This exercise, therefore, demonstrates that ID-TIMS method using 239 Pu spike can be used for determining plutonium concentration in dissolver solution of irradiated fuel, under the plant conditions. 7 refs., 8 tabs
18. Hydrodynamic characteristics in the Levantine Basin in autumn 2016 - The CINEL experiment (CIrculation and water mass properties in the North-Eastern Levantine)
Science.gov (United States)
Mauri, Elena; Poulain, Pierre-Marie; Gerin, Riccardo; Hayes, Dan; Gildor, Hezi; Kokkini, Zoi
2017-04-01
During the CINEL experiment, currents and thermohaline properties of the water masses in the eastern areas of the Levantine Basin (Mediterranean Sea) were monitored with mobile autonomous systems in October-December 2016. Two gliders were operated together with satellite-tracked drifters and Argo floats to study the complex circulation features governing the dynamics near the coast and in the open sea. Strong mesoscale and sub-basin scale eddies were detected and were crossed several times by the gliders during the experiment. The physical and biogeochemical parameters were sampled, showing peculiar characteristics in some of the mesoscale features and a probable interaction with a persistent coastal current off Israel. The in-situ observations were interpreted in concert with the distribution of tracers (sea surface temperature, chlorophyll) and altimetry data obtained from satellites. Numerical simulations with a high resolution model in which deep profiles of temperature and salinity from gliders were assimilated, were used in near-real time to fine tune the observational array and to help with the interpretation of the local dynamics.
19. ABSOLUTE NEUTRINO MASSES
DEFF Research Database (Denmark)
Schechter, J.; Shahid, M. N.
2012-01-01
We discuss the possibility of using experiments timing the propagation of neutrino beams over large distances to help determine the absolute masses of the three neutrinos.......We discuss the possibility of using experiments timing the propagation of neutrino beams over large distances to help determine the absolute masses of the three neutrinos....
20. Qupe--a Rich Internet Application to take a step forward in the analysis of mass spectrometry-based quantitative proteomics experiments.
Science.gov (United States)
Albaum, Stefan P; Neuweger, Heiko; Fränzel, Benjamin; Lange, Sita; Mertens, Dominik; Trötschel, Christian; Wolters, Dirk; Kalinowski, Jörn; Nattkemper, Tim W; Goesmann, Alexander
2009-12-01
The goal of present -omics sciences is to understand biological systems as a whole in terms of interactions of the individual cellular components. One of the main building blocks in this field of study is proteomics where tandem mass spectrometry (LC-MS/MS) in combination with isotopic labelling techniques provides a common way to obtain a direct insight into regulation at the protein level. Methods to identify and quantify the peptides contained in a sample are well established, and their output usually results in lists of identified proteins and calculated relative abundance values. The next step is to move ahead from these abstract lists and apply statistical inference methods to compare measurements, to identify genes that are significantly up- or down-regulated, or to detect clusters of proteins with similar expression profiles. We introduce the Rich Internet Application (RIA) Qupe providing comprehensive data management and analysis functions for LC-MS/MS experiments. Starting with the import of mass spectra data the system guides the experimenter through the process of protein identification by database search, the calculation of protein abundance ratios, and in particular, the statistical evaluation of the quantification results including multivariate analysis methods such as analysis of variance or hierarchical cluster analysis. While a data model to store these results has been developed, a well-defined programming interface facilitates the integration of novel approaches. A compute cluster is utilized to distribute computationally intensive calculations, and a web service allows to interchange information with other -omics software applications. To demonstrate that Qupe represents a step forward in quantitative proteomics analysis an application study on Corynebacterium glutamicum has been carried out. Qupe is implemented in Java utilizing Hibernate, Echo2, R and the Spring framework. We encourage the usage of the RIA in the sense of the 'software as a
1. The silicon tracking system of the CBM experiment at FAIR. Development of microstrip sensors and signal transmission lines for a low-mass, low-noise system
International Nuclear Information System (INIS)
Singla, Minni
2014-01-01
2. Effect of heating rate and grain size on the melting behavior of the alloy Nb-47 mass % Ti in pulse-heating experiments
International Nuclear Information System (INIS)
Basak, D.; Boettinger, W.J.; Josell, D.; Coriell, S.R.; McClure, J.L.; Cezairliyan, A.
1999-01-01
The effect of heating rate and grain size on the melting behavior of Nb-47 mass% Ti is measured and modeled. The experimental method uses rapid resistive self-heating of wire specimens at rates between ∼10 2 and ∼10 4 K/s and simultaneous measurement of radiance temperature and normal spectral emissivity as functions of time until specimen collapse, typically between 0.4 and 0.9 fraction melted. During heating, a sharp drop in emissivity is observed at a temperature that is independent of heating rate and grain size. This drop is due to surface and grain boundary melting at the alloy solidus temperature even though there is very little deflection (limited melting) of the temperature-time curve from the imposed heating rate. Above the solidus temperature, the emissivity remains nearly constant with increasing temperature and the temperature vs time curve gradually reaches a sloped plateau over which the major fraction of the specimen melts. As the heating rate and/or grain size is increased, the onset temperature of the sloped plateau approaches the alloy liquidus temperature and the slope of the plateau approaches zero. This interpretation of the shapes of the temperature-time-curves is supported by a model that includes diffusion in the solid coupled with a heat balance during the melting process. There is no evidence of loss of local equilibrium at the melt front during melting in these experiments
3. Tide Gauge Records Reveal Improved Processing of Gravity Recovery and Climate Experiment Time-Variable Mass Solutions over the Coastal Ocean
Science.gov (United States)
Piecuch, Christopher G.; Landerer, Felix W.; Ponte, Rui M.
2018-05-01
Monthly ocean bottom pressure solutions from the Gravity Recovery and Climate Experiment (GRACE), derived using surface spherical cap mass concentration (MC) blocks and spherical harmonics (SH) basis functions, are compared to tide gauge (TG) monthly averaged sea level data over 2003-2015 to evaluate improved gravimetric data processing methods near the coast. MC solutions can explain ≳ 42% of the monthly variance in TG time series over broad shelf regions and in semi-enclosed marginal seas. MC solutions also generally explain ˜5-32 % more TG data variance than SH estimates. Applying a coastline resolution improvement algorithm in the GRACE data processing leads to ˜ 31% more variance in TG records explained by the MC solution on average compared to not using this algorithm. Synthetic observations sampled from an ocean general circulation model exhibit similar patterns of correspondence between modeled TG and MC time series and differences between MC and SH time series in terms of their relationship with TG time series, suggesting that observational results here are generally consistent with expectations from ocean dynamics. This work demonstrates the improved quality of recent MC solutions compared to earlier SH estimates over the coastal ocean, and suggests that the MC solutions could be a useful tool for understanding contemporary coastal sea level variability and change.
4. Baseline Optimization for the Measurement of CP Violation, Mass Hierarchy, and $\\theta_{23}$ Octant in a Long-Baseline Neutrino Oscillation Experiment
Energy Technology Data Exchange (ETDEWEB)
Bass, M. [Colorado State U.; Bishai, M. [Brookhaven; Cherdack, D. [Colorado State U.; Diwan, M. [Brookhaven; Djurcic, Z. [Argonne; Hernandez, J. [Houston U.; Lundberg, B. [Fermilab; Paolone, V. [Pittsburgh U.; Qian, X. [Brookhaven; Rameika, R. [Fermilab; Whitehead, L. [Houston U.; Wilson, R. J. [Colorado State U.; Worcester, E. [Brookhaven; Zeller, G. [Fermilab
2015-03-19
Next-generation long-baseline electron neutrino appearance experiments will seek to discover CP violation, determine the mass hierarchy and resolve the θ23 octant. In light of the recent precision measurements of θ13, we consider the sensitivity of these measurements in a study to determine the optimal baseline, including practical considerations regarding beam and detector performance. We conclude that a detector at a baseline of at least 1000 km in a wide-band muon neutrino beam is the optimal configuration.
5. Direct detection of dark matter with the EDELWEISS-III experiment: signals induced by charge trapping, data analysis and characterization of cryogenic detector sensitivity to low-mass WIMPs
International Nuclear Information System (INIS)
Arnaud, Quentin
2015-01-01
The EDELWEISS-III experiment is dedicated to direct dark matter searches aiming at detecting WIMPS. These massive particles should account for more than 80% of the mass of the Universe and be detectable through their elastic scattering on nuclei constituting the absorber of a detector. As the expected WIMP event rate is extremely low ( 20 GeV). Finally, a study dedicated to the optimization of solid cryogenic detectors to low mass WIMP searches is presented. This study is performed on simulated data using a statistical test based on a profiled likelihood ratio that allows for statistical background subtraction and spectral shape discrimination. This study combined with results from Run308, has lead the EDELWEISS experiment to favor low mass WIMP searches ( [fr
6. Dual-domain mass-transfer parameters from electrical hysteresis: theory and analytical approach applied to laboratory, synthetic streambed, and groundwater experiments
Science.gov (United States)
Briggs, Martin A.; Day-Lewis, Frederick D.; Ong, John B.; Harvey, Judson W.; Lane, John W.
2014-01-01
Models of dual-domain mass transfer (DDMT) are used to explain anomalous aquifer transport behavior such as the slow release of contamination and solute tracer tailing. Traditional tracer experiments to characterize DDMT are performed at the flow path scale (meters), which inherently incorporates heterogeneous exchange processes; hence, estimated “effective” parameters are sensitive to experimental design (i.e., duration and injection velocity). Recently, electrical geophysical methods have been used to aid in the inference of DDMT parameters because, unlike traditional fluid sampling, electrical methods can directly sense less-mobile solute dynamics and can target specific points along subsurface flow paths. Here we propose an analytical framework for graphical parameter inference based on a simple petrophysical model explaining the hysteretic relation between measurements of bulk and fluid conductivity arising in the presence of DDMT at the local scale. Analysis is graphical and involves visual inspection of hysteresis patterns to (1) determine the size of paired mobile and less-mobile porosities and (2) identify the exchange rate coefficient through simple curve fitting. We demonstrate the approach using laboratory column experimental data, synthetic streambed experimental data, and field tracer-test data. Results from the analytical approach compare favorably with results from calibration of numerical models and also independent measurements of mobile and less-mobile porosity. We show that localized electrical hysteresis patterns resulting from diffusive exchange are independent of injection velocity, indicating that repeatable parameters can be extracted under varied experimental designs, and these parameters represent the true intrinsic properties of specific volumes of porous media of aquifers and hyporheic zones.
7. Prevention and surveillance of public health risks during extended mass gatherings in rural areas: the experience of the Tamworth Country Music Festival, Australia.
Science.gov (United States)
Polkinghorne, B G; Massey, P D; Durrheim, D N; Byrnes, T; MacIntyre, C R
2013-01-01
To describe and evaluate the public health response to the Tamworth Country Music Festival, an annual extended mass gathering in rural New South Wales, Australia; and to propose a framework for responding to similar rural mass gatherings. Process evaluation by direct observation, archival analysis and focus group discussion. The various components of the public health response to the 2011 Tamworth Country Music Festival were actively recorded. An archival review of documentation from 2007 to 2010 was performed to provide context. A focus group was also conducted to discuss the evolution of the public health response and the consequences of public health involvement. Public health risks increased with increasing duration of the rural mass gathering. Major events held within the rural mass gathering further strained resources. The prevention, preparedness, response and recovery principles provided a useful framework for public health actions. Particular risks included inadequately trained food preparation volunteers functioning in poorly equipped temporary facilities, heat-related ailments and arboviral disease. Extended mass gatherings in rural areas pose particular public health challenges; surge capacity is limited and local infrastructure may be overwhelmed in the event of an acute incident or outbreak. There is value in proactive public health surveillance and monitoring. Annual mass gatherings provide opportunities for continual systems improvement. Early multi-agency planning can identify key risks and identify opportunities for partnership. Special consideration is required for major events within mass gatherings. Copyright © 2012 The Royal Society for Public Health. Published by Elsevier Ltd. All rights reserved.
8. Cocrystal solubility-pH and drug solubilization capacity of sodium dodecyl sulfate – mass action model for data analysis and simulation to improve design of experiments
Directory of Open Access Journals (Sweden)
Alex Avdeef
2018-06-01
Full Text Available This review discusses the disposition of the anionic surfactant, sodium dodecyl sulfate (SDS; i.e., sodium lauryl sulfate, to solubilize sparingly-soluble drugs above the surfactant critical micelle concentration (CMC, as quantitated by the solubilization capacity (k. A compilation of 101 published SDS k values of mostly poorly-soluble drug molecules was used to develop a prediction model as a function of the drug’s intrinsic solubility, S0, and its calculated H-bond acceptor/donor potential. In almost all cases, the surfactant was found to solubilize the neutral form of the drug. Using the mass action model, the k values were converted to drug-micelle stoichiometric binding constants, Kn, corresponding to drug-micelle equilibria in drug-saturated solutions. An in-depth case study (data from published sources considered the micellization reactions as a function of pH of a weak base, B, (pKa 3.58, S0 52 μg/mL, where at pH 1 the BH.SDS salt was predicted to precipitate both below and above the CMC. At low SDS concentrations, two drug salts were predicted to co-precipitate: BH.Cl and BH.SDS. Solubility products of both were determined from the analysis of the reported solubility-surfactant data. Above the CMC, in a rare example, the charged form of the drug (BH+ appeared to be strongly solubilized by the surfactant. The constant for that reaction was also determined. At pH 7, the reactions were simpler, as only the neutral form of the drug was solubilized, to a significantly lesser extent than at pH 1. Case studies also featured examples of solubilization of solids in the form of cocrystals. For many cocrystal systems studied in aqueous solution, the anticipated supersaturated state is not long-lasting, as the drug component precipitates to a thermodynamically stable form, thus lowering the amount of the active ingredient available for intestinal absorption. Use of surfactant can prevent this. A recently-described method for predicting the
9. Robotic-assisted transperitoneal nephron-sparing surgery for small renal masses with associated surgical procedures: surgical technique and preliminary experience.
Science.gov (United States)
Ceccarelli, Graziano; Codacci-Pisanelli, Massimo; Patriti, Alberto; Ceribelli, Cecilia; Biancafarina, Alessia; Casciola, Luciano
2013-09-01
Small renal masses (T1a) are commonly diagnosed incidentally and can be treated with nephron-sparing surgery, preserving renal function and obtaining the same oncological results as radical surgery. Bigger lesions (T1b) may be treated in particular situations with a conservative approach too. We present our surgical technique based on robotic assistance for nephron-sparing surgery. We retrospectively analysed our series of 32 consecutive patients (two with 2 tumours and one with 4 bilateral tumours), for a total of 37 robotic nephron-sparing surgery (RNSS) performed between June 2008 and July 2012 by a single surgeon (G.C.). The technique differs depending on tumour site and size. The mean tumour size was 3.6 cm; according to the R.E.N.A.L. Nephrometry Score 9 procedures were considered of low, 14 of moderate and 9 of hight complexity with no conversion in open surgery. Vascular clamping was performed in 22 cases with a mean warm ischemia time of 21.5 min and the mean total procedure time was 149.2 min. Mean estimated blood loss was 187.1 ml. Mean hospital stay was 4.4 days. Histopathological evaluation confirmed 19 cases of clear cell carcinoma (all the multiple tumours were of this nature), 3 chromophobe tumours, 1 collecting duct carcinoma, 5 oncocytomas, 1 leiomyoma, 1 cavernous haemangioma and 2 benign cysts. Associated surgical procedures were performed in 10 cases (4 cholecystectomies, 3 important lyses of peritoneal adhesions, 1 adnexectomy, 1 right hemicolectomy, 1 hepatic resection). The mean follow-up time was 28.1 months ± 12.3 (range 6-54). Intraoperative complications were 3 cases of important bleeding not requiring conversion to open or transfusions. Regarding post-operative complications, there were a bowel occlusion, 1 pleural effusion, 2 pararenal hematoma, 3 asymptomatic DVT (deep vein thrombosis) and 1 transient increase in creatinine level. There was no evidence of tumour recurrence in the follow-up. RNSS is a safe and feasible technique
10. Determination of the top-quark pole mass using $t \\bar{t}+1$-jet events collected with the ATLAS experiment in 7 TeV $pp$ collisions
CERN Document Server
2015-10-19
The normalized differential cross section for top-quark pair production in association with at least one jet is studied as a function of the inverse of the invariant mass of the $t \\bar{t}+1$-jet system. This distribution can be used for a precise determination of the top-quark mass since gluon radiation depends on the mass of the quarks. The experimental analysis is based on proton--proton collision data collected by the ATLAS detector at the LHC with a centre-of-mass energy of 7 TeV corresponding to an integrated luminosity of 4.6 fb$^{-1}$. The selected events were identified using the lepton+jets top-quark-pair decay channel, where lepton refers to either an electron or a muon. The observed distribution is compared to a theoretical prediction at next-to-leading-order accuracy in quantum chromodynamics using the pole-mass scheme. With this method, the measured value of the top-quark pole mass, $m_t^{\\rm pole}$, is: $m_t^{\\rm pole}$ =173.7 $\\pm$ 1.5 (stat.) $\\pm$ 1.4 (syst.) $^{+ 1.0}_{-0.5}$ (theory) GeV. ...
11. The origin of mass
International Nuclear Information System (INIS)
Cashmore, R.; Sutton, C.
1992-01-01
The existence of mass in the Universe remains unexplained but recent high-energy experiments, described in this article, are close to testing the most plausible explanation for the masses of fundamental particles, which may, in turn, lead to a clearer understanding of mass on the macro-scale. The Standard Model includes the concept of the Higgs mechanism which endows particles with mass. Actual evidence for the existence of the postulated particle known as the Higgs boson would lead to confirmation of the theory and efforts to detect it at CERN are complex and determined. (UK)
12. Top quark mass measurement in the dilepton electron-muon channel with the matrix element method in the ATLAS experiment at LHC
International Nuclear Information System (INIS)
Demilly, A.
2014-01-01
The LHC produced proton-proton collisions data at 7 TeV of center of mass energy in 2011 and 8 TeV in 2012, corresponding to an integrated luminosity of, respectively, 5 fb"-"1 and 23 fb"-"1. Data acquired by ATLAS have led to a better understanding of the detector and its performance, to many measurements of physical quantities and the discovery of the Higgs boson. Top quark is involved in many processes beyond the Standard Model. Its mass is an important parameter for the Standard Model and any New Physics theory, thus measuring its mass accurately is necessary. After a description of the Standard Model of Particle Physics, and the role of the top quark in it, the first half of this thesis describes the ATLAS detector and its electromagnetic calorimeter, for which a study of the calibration constant patching is detailed. The second half details top quark physics events detected in ATLAS and their selection. Theoretical aspects of the matrix element method and its implementation for the top quark mass measurement in the dilepton electron-muon channel in the experimental framework of ATLAS are discussed. The measurement calibration and optimisation studies for the analysis are presented. Finally, systematic uncertainties are described and estimated. This measurement yields a top quark mass of (173.65 ± 0.70 ± 2.36) GeV ; showing no discrepancy with current worldwide measurements. (author)
13. Increase in transmission loss of a double panel system by addition of mass inclusions to a poro-elastic layer: A comparison between theory and experiment
Science.gov (United States)
Idrisi, Kamal; Johnson, Marty E.; Toso, Alessandro; Carneal, James P.
2009-06-01
This paper is concerned with the modeling and optimization of heterogeneous (HG) blankets, which are used in this investigation to reduce the sound transmission through double panel systems. HG blankets consist of poro-elastic media with small embedded masses, which act similarly to a distributed mass-spring-damper-system. HG blankets have shown significant potential to reduce low frequency radiated sound from structures, where traditional poro-elastic materials have little effect. A mathematical model of a double panel system with an acoustic cavity and HG blanket was developed using impedance and mobility methods. The predicted responses of the source and the receiving panel due to a point force are validated with experimental measurements. The presented results indicate that proper tuning of the HG blankets can result in broadband noise reduction below 500 Hz with less than 10% added mass.
14. Main sequence mass loss
International Nuclear Information System (INIS)
Brunish, W.M.; Guzik, J.A.; Willson, L.A.; Bowen, G.
1987-01-01
It has been hypothesized that variable stars may experience mass loss, driven, at least in part, by oscillations. The class of stars we are discussing here are the δ Scuti variables. These are variable stars with masses between about 1.2 and 2.25 M/sub θ/, lying on or very near the main sequence. According to this theory, high rotation rates enhance the rate of mass loss, so main sequence stars born in this mass range would have a range of mass loss rates, depending on their initial rotation velocity and the amplitude of the oscillations. The stars would evolve rapidly down the main sequence until (at about 1.25 M/sub θ/) a surface convection zone began to form. The presence of this convective region would slow the rotation, perhaps allowing magnetic braking to occur, and thus sharply reduce the mass loss rate. 7 refs
15. Measurement of the Top Quark Mass in the All-Hadronic Top-Antitop Decay Channel Using Proton-Proton Collision Data from the ATLAS Experiment at a Centre-of-Mass Energy of 8 TeV
CERN Document Server
AUTHOR|(INSPIRE)INSPIRE-00220136
A measurement of the top quark mass ($m_{top}$) is presented using top quark candidates reconstructed from the jets in all-hadronic $t\\bar{t}$ decays. The analysis makes use of the full ATLAS dataset collected over the 2012 period at a centre-of-mass energy of $\\sqrt{s}=8$ TeV and with a total integrated luminosity of $\\int Ldt=20.6$ fb$^{-1}$. A five-jet trigger, together with an offline cut requiring five central jets with a transverse momentum of at least 60 GeV, was used to pre-select candidate signal events. A series of selection cuts are subsequently employed to increase the signal fraction in data events, motivated by both measured data and simulation. These cuts, as well the analytic $\\chi^2$ reconstruction algorithm to designate which jets are to be used to reconstruct the top quarks, aim to select those events and candidate top quarks deemed most consistent with an all-hadronic $t\\bar{t}$ decay topology. The most discriminating cut involves the output of a b-tagging algorithm employed to ident...
16. Mass discrimination
Energy Technology Data Exchange (ETDEWEB)
Broeckman, A. [Rijksuniversiteit Utrecht (Netherlands)
1978-12-15
In thermal ionization mass spectrometry the phenomenon of mass discrimination has led to the use of a correction factor for isotope ratio-measurements. The correction factor is defined as the measured ratio divided by the true or accepted value of this ratio. In fact this factor corrects for systematic errors of the whole procedure; however mass discrimination is often associated just with the mass spectrometer.
17. Negative mass
International Nuclear Information System (INIS)
Hammond, Richard T
2015-01-01
Some physical aspects of negative mass are examined. Several unusual properties, such as the ability of negative mass to penetrate any armor, are analysed. Other surprising effects include the bizarre system of negative mass chasing positive mass, naked singularities and the violation of cosmic censorship, wormholes, and quantum mechanical results as well. In addition, a brief look into the implications for strings is given. (paper)
18. Measurements of the Higgs boson mass and width in the four-lepton final state and electron reconstruction in the CMS experiment at the LHC
CERN Document Server
Dalchenko, Mykhailo; Charlot, Claude
This thesis document reports measurements of the mass and width of the new boson re- cently discovered at the Large Hadron Collider (LHC), candidating to be the Standard Model Higgs boson. The analysis uses proton-proton collision data recorded by the Compact Muon Solenoid (CMS) detector at the LHC, corresponding to integrated luminosities of $5.1~fb^{−1}$ at $7~$TeV center of mass energy and $19.7~fb^{−1}$ at $8~$TeV center of mass energy. Set of events selecting Higgs boson via the $H\\to ZZ$ decay channel, where both $Z$ bosons decay to electron or muon lepton pairs, is used for the Higgs boson properties measurements. A precise measurement of its mass has been performed and gives $125.6\\pm0.4\\mbox{(stat)}\\pm0.2\\mbox{(syst)}~$GeV. Constraints on the Higgs boson width were established using its off-shell production and decay to a pair of $Z$ bosons, where one $Z$ boson decays to an electron or muon pair, and the other to an electron, muon, or neutrino pair. The obtained result is an upper limit on the Hi...
19. Analysis of Phospholipid Mixtures from Biological Tissues by Matrix-Assisted Laser Desorption and Ionization Time-of-Flight Mass Spectrometry (MALDI-TOF MS): A Laboratory Experiment
Science.gov (United States)
Eibisch, Mandy; Fuchs, Beate; Schiller, Jurgen; Sub, Rosmarie; Teuber, Kristin
2011-01-01
Matrix-assisted laser desorption and ionization time-of-flight mass spectrometry (MALDI-TOF MS) is increasingly used to investigate the phospholipid (PL) compositions of tissues and body fluids, often without previous separation of the total mixture into the individual PL classes. Therefore, the questions of whether all PL classes are detectable…
20. Electronic states and nature of bonding in the molecule YC by all electron ab initio multiconfiguration self-consistent-field calculations and mass spectrometric equilibrium experiments
DEFF Research Database (Denmark)
Shim, Irene; Pelino, Mario; Gingerich, Karl A.
1992-01-01
, and they hardly contribute to the bonding. The chemical bond in the YC molecule is polar with charge transfer from Y to C giving rise to a dipole moment of 3.90 D at 3.9 a.u. in the 4PI ground state. Mass spectrometric equilibrium investigations in the temperature range 2365-2792 K have resulted...
1. Nominal Mass?
Science.gov (United States)
Attygalle, Athula B; Pavlov, Julius
2017-08-01
The current IUPAC-recommended definition of the term "nominal mass," based on the most abundant naturally occurring stable isotope of an element, is flawed. We propose that Nominal mass should be defined as the sum of integer masses of protons and neutrons in any chemical species. In this way, all isotopes and isotopologues can be assigned a definitive identifier. Graphical Abstract ᅟ.
2. Resonance searches with the t$\\bar{t}$ invariant mass distribution measured with the DØ experiment at √s=1.96 TeV
Energy Technology Data Exchange (ETDEWEB)
Schliephake, Thorsten Dirk [Univ. of Wuppertal (Germany)
2010-06-01
Understanding the universe, its birth and its future is one of the biggest motivations in physics. In order to understand the cosmos, the fundamental particles forming the universe, the components our matter is built of need to be known and understood. Over time physicists have built a theory which describes the physics of the known fundamental particles very well: the Standard Model (SM) of particle physics. The SM describes the particles, their interactions and phenomena with high precision. So far no proven deviations from the SM have been found, though recently evidence for possible physics beyond the SM has been observed. The SM is not describing the mass of the elementary particles however and even with the addition of the Higgs mechanism giving mass to the particles, we have no full theory for all four fundamental forces. We know the model needs to be extended or replaced by another one, as gravitation is not included in the SM. Having a theory which describes all fundamental particles found so far and all but one fundamental interaction is a great success. However, all this describes about 4% of the universe we live in. 23% is dark matter and 73% is dark energy. Dark matter is believed to interact only through gravity and maybe the weak force, which makes it hardly observable. Dark energy is even more elusive. Among other theories the cosmologic constant and scalar fields are discussed to describe it. One should also note that other models exist which for example modify the Newtonian law of gravity. The Higgs mechanism has become the most popular model for mass generation. Alternative theories like Super Symmetry (SUSY), large Extra Dimensions, Technicolor, String Theory, to name just a few, have spread to describe the necessary mass generation or new particles. As proof for new physics beyond the SM has not been found yet, one assumes that new physics will manifest itself at a larger energy scale and therefore a higher particle mass. Particles with high
3. Neutrino mass sum-rule
Science.gov (United States)
Damanik, Asan
2018-03-01
Neutrino mass sum-rele is a very important research subject from theoretical side because neutrino oscillation experiment only gave us two squared-mass differences and three mixing angles. We review neutrino mass sum-rule in literature that have been reported by many authors and discuss its phenomenological implications.
4. Limit on the production of a low-mass vector boson in e+e−→Uγ, U→e+e− with the KLOE experiment
Directory of Open Access Journals (Sweden)
A. Anastasi
2015-11-01
Full Text Available The existence of a new force beyond the Standard Model is compelling because it could explain several striking astrophysical observations which fail standard interpretations. We searched for the light vector mediator of this dark force, the U boson, with the KLOE detector at the DAΦNE e+e− collider. Using an integrated luminosity of 1.54 fb−1, we studied the process e+e−→Uγ, with U→e+e−, using radiative return to search for a resonant peak in the dielectron invariant-mass distribution. We did not find evidence for a signal, and set a 90% CL upper limit on the mixing strength between the Standard Model photon and the dark photon, ε2, at 10−6–10−4 in the 5–520 MeV/c2 mass range.
5. 26kDa endochitinase from barley seeds: real-time monitoring of the enzymatic reaction and substrate binding experiments using electrospray ionization mass spectrometry
DEFF Research Database (Denmark)
Dennhart, Nicole; Weigang, Linda M M; Fujiwara, Maho
2009-01-01
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8844480514526367, "perplexity": 20106.207135251494}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676592636.68/warc/CC-MAIN-20180721145209-20180721165209-00639.warc.gz"}
|
http://mathhelpforum.com/business-math/57880-challenge-finance-gurus.html
|
## A challenge for the finance gurus
Hi guys
I'm new to this site (first post) and I haven't done high school or university maths for 20 years, so what I am attempting is beyond me.
I have the following formula for a home loan:
A = [P(1 + R/f)(Yf)R/f]/[1 + R/f)(Yf) - 1] where (Yf) should be super-scripted, ie "to the power of".
When:
A = periodical repayments
P = initial loan amount (principal)
R = interest rate per annum
f = frequency of repayments per annum (i.e. monthly = 12 etc)
Y = term of loan in years
My data is as follows:
Fortnightly repayments = $1145.44 (a little more than I need to.) Current loan amount =$308,000
Interest rate = 8.0%
Repayment frequency = fortnightly (26 times p.a.)
28 years remaining on a 30 year loan
What I want to know is, on a standard home loan, how do I calculate the following:
1. If I pay a one-off amount into the mortgage on top of my regular repayments, how much does this reduce the term Y by (assuming interest rates do not change)?
2. How much interest do I save over the term of the loan by doing this (assuming interest rates do not change)?
3. How do I calculate the reduction on the term of my loan by paying small extra amounts? That is, as interest rates fall I continue to pay the same amount as I did at the peak of the interest rate cycle
4. How do I calculate the interest I save by paying the extra small amounts?
I'm guessing the above formula is inadequate for what I need, and that I need an amortization table in an Excel spreadsheet so that I can plug the numbers in. I'd also like to graph the data in Excel to get a visual indication of the effects of what I propose.
Any ideas?
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.61408531665802, "perplexity": 1160.7790225495853}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701156520.89/warc/CC-MAIN-20160205193916-00126-ip-10-236-182-209.ec2.internal.warc.gz"}
|
http://mechanicaldesign.asmedigitalcollection.asme.org/article.aspx?articleid=1449295
|
0
RESEARCH PAPERS
# On Calculating the Degrees of Freedom or Mobility of Overconstrained Linkages: Single-Loop Exceptional Linkages
[+] Author and Article Information
J. M. Rico
F.I.M.E.E. Universidad de Guanajuato Salamanca, Gto. 38040, México and Departamento de Ingeniería Mecánica, Instituto Tecnológico de Celaya, Celaya, Gto. 38010, Mé[email protected]; [email protected]
B. Ravani
Department of Mechanical and Aeronautical Engineering, University of California, Davis, CA [email protected]
For a review on the basic definitions and results on vector spaces, algebras and matrices, the reader is referred to (15)
The mobility of trivial linkages can be easily determined using Kutzbach-Grübler criterion and does not need the approach of this paper. It is included here only for completeness.
The process indicated here is that of obtaining the closure algebra of a vector space in the Lie algebra, $se(3)$, of the Euclidean group, $SE(3)$, see (13), pp. 483–487.
This paths must be understood in a graph theoretical sense, where the links are regarded as vertex and the kinematic pairs are regarded as edges of a graph. In some cases, see Example 2, the actual position of the links appears to contradict this definition.
The column spaces of these Jacobian matrices are the clockwise and counterclockwise infinitesimal mechanical liaisons of link $j$ with respect to link $i$, denoted in (13) as $Vc(i,j)$ and $Vcc(i,j)$, respectively.
The column space of these closure matrices are the clockwise and counterclockwise closure algebras of link $j$ with respect to link $i$, denoted in (13) as $Ac(i,j)$ and $Acc(i,j)$, respectively.
The column space of this absolute closure matrix is the absolute closure algebra, $Aa(i,j)$ employed in (13).
It is important to note a subtle difference in linear algebra notation, while $[1,2,3]$ indicates the vector space generated by the screws $1,2,3$, $[123]$ indicates the matrix whose columns are given by $1,2,3$.
In fact, any three screws of the sets of screws ${21,32,43,54}$ satisfy the conditions.
In fact, any three screws of the sets of screws ${65,76,87,18}$ satisfy the conditions.
Every cylindrical pair will be represented by a pair of screws, one associated with the rotation and one associated with the traslation. Although other combinations can be used, this seems to be the simplest one. The same approach can be used for other types of kinematic pairs, for example, spherical or planar pairs.
It should be noted that in terms of the original screws of the Jacobian matrix, this Lie product is given by $[21(21,2*2)]$.
It should be noted that in terms of the original screws of the Jacobian matrix, this Lie product is given by $[2*2(212*2)]$.
It should be noted that in terms of the original screws of the Jacobian matrix, this Lie product is given by $[2*2(212*2)]$.
As usual in linear algebra, [ ] represents the vector space generated by the columns of the matrix (see (15)).
Correspondingly, if $[BAa]≠{0⃗}$, Hervè called the constraints dependent.
As indicated in Sec. 4, it is more proper to talk about the mobility characteristics of a pair of links in a kinematic chain. However, by abusing the language, these characteristics are extended to the kinematic chain.
In Table 1, the comparison of ranks of matrices takes the place of the comparison of subalgebras and subspaces employed in 13. The reason for the success of this replacement is that the subalgebras and subspaces involved satisfy the following equations: $Vc(i,j)⩽Ac(i,j)$, $Vcc(i,j)⩽Acc(i,j)$, $Aa(i,j)⩽Ac(i,j)$, $Aa(i,j)⩽Acc(i,j)$, ${0}⩽Aa(i,j)$.
This chain is particularly interesting, if one considers the mobility of link 2 with respect to link 1, the chain is a trivial chain, the same result is obtained by considering the mobility of links 8 with respect to link 1. Moreover, if one considers the mobility of link 6 with respect to link 1, the chain is an exceptional chain. All analyses yield the same number of degrees of freedom.
In this case, the resulting matrix is just the usual Jacobian matrix.
In fact, the relative motion of links 2 or 3, with respect to links 5 or 6 does include a rotation around the $Z$-axis.
Moreover, the vectors $ω⃗$ and $v⃗O$ can be identified as the angular velocity of a rigid body and the velocity of a point $O$, fixed in the rigid body, respectively.
Here, the notation introduced by von Mises is employed. It should be noted that in many differential geometry oriented contributions, a comma after the first operand is included.
J. Mech. Des 129(3), 301-311 (Mar 10, 2006) (11 pages) doi:10.1115/1.2406101 History: Received May 10, 2005; Revised March 10, 2006
## Abstract
This paper reformulates and extends the new, group theoretic, mobility criterion recently developed by the authors, (Rico, J. M., and Ravani, B., 2003, ASME J. Mech. Des., 125, pp. 70–80). In contrast to the Kutzbach-Grübler criterion, the new mobility criterion and the approach presented apply to a large class of single-loop overconstrained linkages. The criterion is reformulated, in terms of the well-known Jacobian matrices, for exceptional linkages and extended to linkages with partitioned mobility and trivial linkages. Several examples are included.
<>
## Figures
Figure 1
Open-loop kinematic chain
Figure 2
Three basic types of kinematic pairs
Figure 3
Single-loop kinematic chain and the two possible paths between rigid bodies i and j
Figure 4
Figure 5
Figure 6
Kinematic chain with partitioned mobility
Figure 7
Figure 8
Sarrus’s linkage in a singular position
Figure 9
## Discussions
Some tools below are only available to our subscribers or users with an online account.
### Related Content
Customize your page view by dragging and repositioning the boxes below.
Topic Collections
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 30, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7596984505653381, "perplexity": 700.9110556041715}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257648198.55/warc/CC-MAIN-20180323063710-20180323083710-00074.warc.gz"}
|
https://tex.stackexchange.com/questions/472303/required-files-for-writing-physics-related-thesis?noredirect=1
|
# Required Files for Writing Physics-related Thesis
I am trying to write my dissertation for PhD in physics using overleaf website. I am using the template given by our school in this discipline that is called ua-thesis.
I also have a .bib file including all my citations. Besides the .tex file in the project, I also managed to find a bibliography style called aip.bst that seems to be common in my field. However, it seems that while compiling the project, I am receiving lots of error messages related to the in-text citations of the .tex file.
Am I missing some other file in the project required for the compilation (e.g. a .cls file that I am not aware of)?
Here is a MWE producing the same exact error messages:
\documentclass[draft]{ua-thesis}
\usepackage[figuresright]{rotating}
\usepackage{amsmath}
\usepackage{afterpage}
\usepackage{subfloat}
\usepackage{comment}
\usepackage{xspace}
\usepackage{graphicx}
\author{Author's Name}
\title{Title}
\date{2019}
\begin{document}
\maketitle
\tableofcontents
\listoffigures
\listoftables
\begin{abstract}
\end{abstract}
\chapter{Introduction}
Some statement \citep{2016MNRAS.460.1399V}. Another statement \citep{2016MNRAS.463.1666C}. This is known as something. \citep{1993MNRAS.264..201K,1999ApJ...522...82K,1999ApJ...524L..19M}. Still another statement \citep{2017arXiv171106267K}. Final statement \citep{Tolman511,1934rtc..book.....T,1935ApJ....82..302H,1996ApJ...456L..79P}.
\bibliographystyle{aip}
\bibliography{References}
\end{document}
This produces the error
Package natbib Error: Bibliography not compatible with author-year citations.
(natbib) Press to continue in numerical citation style. See the natbib package documentation for explanation. Type H for immediate help. ... l.31 ...mand\NAT@force@numbers{}\NAT@force@numbers Check the bibliography entries for non-compliant syntax, or select author-year BibTeX style, e.g. plainnat
My .bib file contains
@Article{2016MNRAS.460.1399V,
author = {{Vogelsberger}, M. and {Zavala}, J. and {Cyr-Racine}, F.-Y. and {Pfrommer}, C. and {Bringmann}, T. and {Sigurdson}, K.},
title = {{ETHOS - an effective theory of structure formation: dark matter physics as a possible explanation of the small-scale CDM problems}},
journal = {\mnras},
year = {2016},
volume = {460},
pages = {1399-1416},
month = aug,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
doi = {10.1093/mnras/stw1076},
eprint = {1512.05349},
keywords = {methods: numerical, galaxies: haloes, dark matter},
}
@Article{2016MNRAS.463.1666C,
author = {{Castro}, T. and {Marra}, V. and {Quartin}, M.},
title = {{Constraining the halo mass function with observations}},
journal = {\mnras},
year = {2016},
volume = {463},
pages = {1666-1677},
month = dec,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
doi = {10.1093/mnras/stw2072},
eprint = {1605.07548},
keywords = {large-scale structure of Universe, cosmology: observations, cosmological parameters, gravitational lensing: weak, stars: supernovae: general, supernovae: general},
}
@Article{1993MNRAS.264..201K,
author = {{Kauffmann}, G. and {White}, S.~D.~M. and {Guiderdoni}, B.},
title = {{The Formation and Evolution of Galaxies Within Merging Dark Matter Haloes}},
journal = {\mnras},
year = {1993},
volume = {264},
pages = {201},
month = sep,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1093/mnras/264.1.201},
}
@Article{1999ApJ...522...82K,
author = {{Klypin}, A. and {Kravtsov}, A.~V. and {Valenzuela}, O. and {Prada}, F.},
title = {{Where Are the Missing Galactic Satellites?}},
journal = {\apj},
year = {1999},
volume = {522},
pages = {82-92},
month = sep,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1086/307643},
eprint = {astro-ph/9901240},
keywords = {COSMOLOGY: THEORY, GALAXIES: CLUSTERS: GENERAL, GALAXIES: INTERACTIONS, GALAXY: FORMATION, GALAXIES: LOCAL GROUP, METHODS: NUMERICAL, Cosmology: Theory, Galaxies: Clusters: General, Galaxies: Interactions, Galaxy: Formation, Galaxies: Local Group, Methods: Numerical},
}
@Article{1999ApJ...524L..19M,
author = {{Moore}, B. and {Ghigna}, S. and {Governato}, F. and {Lake}, G. and {Quinn}, T. and {Stadel}, J. and {Tozzi}, P.},
title = {{Dark Matter Substructure within Galactic Halos}},
journal = {\apjl},
year = {1999},
volume = {524},
pages = {L19-L22},
month = oct,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1086/312287},
eprint = {astro-ph/9907411},
keywords = {COSMOLOGY: OBSERVATIONS, COSMOLOGY: THEORY, COSMOLOGY: DARK MATTER, GALAXIES: CLUSTERS: GENERAL, GALAXIES: FORMATION, Cosmology: Observations, Cosmology: Theory, Cosmology: Dark Matter, Galaxies: Clusters: General, Galaxies: Formation},
}
@Article{2017arXiv171106267K,
author = {{Kim}, S.~Y. and {Peter}, A.~H.~G. and {Hargis}, J.~R.},
title = {{There is No Missing Satellites Problem}},
journal = {ArXiv e-prints},
year = {2017},
month = nov,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
eprint = {1711.06267},
keywords = {Astrophysics - Cosmology and Nongalactic Astrophysics, Astrophysics - Astrophysics of Galaxies, High Energy Physics - Phenomenology},
}
@Article{Tolman511,
author = {Tolman, Richard C.},
title = {ON THE ESTIMATION OF DISTANCES IN A CURVED UNIVERSE WITH A NON-STATIC LINE ELEMENT},
journal = {Proceedings of the National Academy of Sciences},
year = {1930},
volume = {16},
number = {7},
pages = {511--520},
issn = {0027-8424},
doi = {10.1073/pnas.16.7.511},
eprint = {http://www.pnas.org/content/16/7/511.full.pdf},
publisher = {National Academy of Sciences},
url = {http://www.pnas.org/content/16/7/511},
}
• Can you make a minimal working example with bibliography (MWEB) that reproduces the errors you get? – samcarter_is_at_topanswers.xyz Jan 28 '19 at 20:50
• Note that the message you quote is not the original error message that LaTeX would issue on an undefined command. It is an explanatory text added by your compiler (Overleaf?). In particular the real TeX error would tell you which command is undefined. That could give you a hint for a first online search. – moewe Jan 28 '19 at 21:54
• A classic issue with ADS-exported .bib entries is that they use macros for journal names. Unless you load a file that defines those names or define the names you need manually you will get an error. See tex.stackexchange.com/q/366618/35864 and linked questions. More generally see tex.stackexchange.com/q/386053/35864 (the ADS .bib export has quite some unfortunate idiosyncrasies). – moewe Jan 28 '19 at 21:56
• @moewe Can you please tell me where you bought your crystal ball? It seems to be of very high quality! – samcarter_is_at_topanswers.xyz Jan 28 '19 at 21:58
• But you also need to load natbib differently. The code does not show how you load natbib (so strictly speaking it is not an MWE that reproduces the issue: tex.meta.stackexchange.com/q/228/35864). Since aip.bst does not do author-year citations, you need to load natbib with the option numbers, for example as \usepackage[numbers]{natbib}. – moewe Jan 28 '19 at 22:17
With the documentclass ua-thesis you are using an older version ua-thesis.cls Document Class: ua-thesis 1997/03/08 UA Thesis Class. Using this class I get the error message that commands like \mras are undefined.
In the new version of this class (file uathesis-master.zip, you gave the link in your next question) you can find file aastex_hack.sys, containing the missing definitions. Copy this file into the same directory/project you have your tex code and vbib file.
In the following I'm using one file mwe.tex containing both parts (TeX code and bib file concatenated with package filecontents to one compilable MWE. In real you still have two files, your for example thesis-tex and references.bib or simular).
In your given code are two calls of packages missing:
\usepackage[numbers]{natbib} % <========================================
\usepackage{aastex_hack} % <============================================
Package natbib is needed to allow the usage of command \citep, package aastex_hack adds the commands for \mnras etc.
With this zip file you get bibliography style uabibnat (file uabibnat.bst) you can use. You did not mentions where your used file aip.bst comes from, I found one in the internet (see comment in the following MWE). With both styles you can compile the following code without errors. Depending on the used style there are some warnings I let to you to get rid of them ...
So please copy the following complete MWE to your computer and compile it, following the ususal chain:
\RequirePackage{filecontents}
\begin{filecontents*}{\jobname.bib}
@Article{2016MNRAS.460.1399V,
author = {{Vogelsberger}, M. and {Zavala}, J. and {Cyr-Racine}, F.-Y. and {Pfrommer}, C. and {Bringmann}, T. and {Sigurdson}, K.},
title = {{ETHOS - an effective theory of structure formation: dark matter physics as a possible explanation of the small-scale CDM problems}},
journal = {\mnras},
year = {2016},
volume = {460},
pages = {1399-1416},
month = aug,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
doi = {10.1093/mnras/stw1076},
eprint = {1512.05349},
keywords = {methods: numerical, galaxies: haloes, dark matter},
}
@Article{2016MNRAS.463.1666C,
author = {{Castro}, T. and {Marra}, V. and {Quartin}, M.},
title = {{Constraining the halo mass function with observations}},
journal = {\mnras},
year = {2016},
volume = {463},
pages = {1666-1677},
month = dec,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
doi = {10.1093/mnras/stw2072},
eprint = {1605.07548},
keywords = {large-scale structure of Universe, cosmology: observations, cosmological parameters, gravitational lensing: weak, stars: supernovae: general, supernovae: general},
}
@Article{1993MNRAS.264..201K,
author = {{Kauffmann}, G. and {White}, S.~D.~M. and {Guiderdoni}, B.},
title = {{The Formation and Evolution of Galaxies Within Merging Dark Matter Haloes}},
journal = {\mnras},
year = {1993},
volume = {264},
pages = {201},
month = sep,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1093/mnras/264.1.201},
}
@Article{1999ApJ...522...82K,
author = {{Klypin}, A. and {Kravtsov}, A.~V. and {Valenzuela}, O. and {Prada}, F.},
title = {{Where Are the Missing Galactic Satellites?}},
journal = {\apj},
year = {1999},
volume = {522},
pages = {82-92},
month = sep,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1086/307643},
eprint = {astro-ph/9901240},
keywords = {COSMOLOGY: THEORY, GALAXIES: CLUSTERS: GENERAL, GALAXIES: INTERACTIONS, GALAXY: FORMATION, GALAXIES: LOCAL GROUP, METHODS: NUMERICAL, Cosmology: Theory, Galaxies: Clusters: General, Galaxies: Interactions, Galaxy: Formation, Galaxies: Local Group, Methods: Numerical},
}
@Article{1999ApJ...524L..19M,
author = {{Moore}, B. and {Ghigna}, S. and {Governato}, F. and {Lake}, G. and {Quinn}, T. and {Stadel}, J. and {Tozzi}, P.},
title = {{Dark Matter Substructure within Galactic Halos}},
journal = {\apjl},
year = {1999},
volume = {524},
pages = {L19-L22},
month = oct,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
doi = {10.1086/312287},
eprint = {astro-ph/9907411},
keywords = {COSMOLOGY: OBSERVATIONS, COSMOLOGY: THEORY, COSMOLOGY: DARK MATTER, GALAXIES: CLUSTERS: GENERAL, GALAXIES: FORMATION, Cosmology: Observations, Cosmology: Theory, Cosmology: Dark Matter, Galaxies: Clusters: General, Galaxies: Formation},
}
@Article{2017arXiv171106267K,
author = {{Kim}, S.~Y. and {Peter}, A.~H.~G. and {Hargis}, J.~R.},
title = {{There is No Missing Satellites Problem}},
journal = {ArXiv e-prints},
year = {2017},
month = nov,
adsnote = {Provided by the SAO/NASA Astrophysics Data System},
archiveprefix = {arXiv},
eprint = {1711.06267},
keywords = {Astrophysics - Cosmology and Nongalactic Astrophysics, Astrophysics - Astrophysics of Galaxies, High Energy Physics - Phenomenology},
}
@Article{Tolman511,
author = {Tolman, Richard C.},
title = {ON THE ESTIMATION OF DISTANCES IN A CURVED UNIVERSE WITH A NON-STATIC LINE ELEMENT},
journal = {Proceedings of the National Academy of Sciences},
year = {1930},
volume = {16},
number = {7},
pages = {511--520},
issn = {0027-8424},
doi = {10.1073/pnas.16.7.511},
eprint = {http://www.pnas.org/content/16/7/511.full.pdf},
publisher = {National Academy of Sciences},
url = {http://www.pnas.org/content/16/7/511},
}
\end{filecontents*}
\documentclass[draft]{ua-thesis}
\usepackage[figuresright]{rotating}
\usepackage{amsmath}
\usepackage{afterpage}
\usepackage{subfloat}
\usepackage{comment}
\usepackage{xspace}
\usepackage{graphicx}
\usepackage[numbers]{natbib} % <========================================
\usepackage{aastex_hack} % <============================================
\author{Author's Name}
\title{Title}
\date{2019}
\begin{document}
\maketitle
\tableofcontents
\listoffigures
\listoftables
\begin{abstract}
\end{abstract}
\chapter{Introduction}
Some statement \citep{2016MNRAS.460.1399V}.
Another statement \citep{2016MNRAS.463.1666C}.
This is known as something. \citep{1993MNRAS.264..201K,1999ApJ...522...82K,1999ApJ...524L..19M}.
Still another statement \citep{2017arXiv171106267K}.
Final statement \citep{Tolman511,1934rtc..book.....T,1935ApJ....82..302H,1996ApJ...456L..79P}.
% https://apps.carleton.edu/curricular/physics/assets/aip.bst
\bibliographystyle{uabibnat} % aip uabibnat
\bibliography{\jobname}
\end{document}
Then you will get the following bibliography, based on style uabibnat:
• Yes, I followed all these and I still have an issue with the code as pointed in the other post: tex.stackexchange.com/questions/472678/… – Ash Jan 31 '19 at 18:09
• I have lots of issues with the old class. I wish there was a solution for the new class. – Ash Jan 31 '19 at 19:40
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.712901771068573, "perplexity": 13721.424924083922}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347432237.67/warc/CC-MAIN-20200603050448-20200603080448-00177.warc.gz"}
|
http://www.numdam.org/item/SPS_1999__33__267_0/
|
Brownian filtrations are not stable under equivalent time-changes
Séminaire de probabilités de Strasbourg, Volume 33 (1999), pp. 267-276.
@article{SPS_1999__33__267_0,
author = {\'Emery, Michel and Schachermayer, Walter},
title = {Brownian filtrations are not stable under equivalent time-changes},
journal = {S\'eminaire de probabilit\'es de Strasbourg},
pages = {267--276},
publisher = {Springer - Lecture Notes in Mathematics},
volume = {33},
year = {1999},
zbl = {0949.60087},
mrnumber = {1768000},
language = {en},
url = {http://www.numdam.org/item/SPS_1999__33__267_0/}
}
TY - JOUR
AU - Émery, Michel
AU - Schachermayer, Walter
TI - Brownian filtrations are not stable under equivalent time-changes
JO - Séminaire de probabilités de Strasbourg
PY - 1999
DA - 1999///
SP - 267
EP - 276
VL - 33
PB - Springer - Lecture Notes in Mathematics
UR - http://www.numdam.org/item/SPS_1999__33__267_0/
UR - https://zbmath.org/?q=an%3A0949.60087
UR - https://www.ams.org/mathscinet-getitem?mr=1768000
LA - en
ID - SPS_1999__33__267_0
ER -
%0 Journal Article
%A Émery, Michel
%A Schachermayer, Walter
%T Brownian filtrations are not stable under equivalent time-changes
%J Séminaire de probabilités de Strasbourg
%D 1999
%P 267-276
%V 33
%I Springer - Lecture Notes in Mathematics
%G en
%F SPS_1999__33__267_0
Émery, Michel; Schachermayer, Walter. Brownian filtrations are not stable under equivalent time-changes. Séminaire de probabilités de Strasbourg, Volume 33 (1999), pp. 267-276. http://www.numdam.org/item/SPS_1999__33__267_0/
[BE 99] S. Beghdadi-Sakrani & M. Émery. On certain probabilities equivalent to coin-tossing, d'après Schachermayer. In this volume. | Numdam | Zbl
[DFST 96] L. Dubins, J. Feldman. M. Smorodinsky & B. Tsirelson. Decreasing sequences of σ-fields and a measure change for Brownian motion. Ann. Prob.24, 882-904, 1996. | Zbl
[RY 91] D. Revuz & M. Yor. Continuous Martingales and Brownian Motion. Springer, 1991. | Zbl
[T 97] B. Tsirelson. Triple points: From non-Brownian filtrations to harmonic measures. GAFA, Geom. funct. anal.7, 1096-1142, 1997. | MR | Zbl
[V 73] A.M. Vershik. Approximation in measure theory. Doctor Thesis, Leningrad 1973. Expanded and updated english version: The theory of decreasing sequences of measurable partitions. St. Petersburg Math. J. 6, 705-761, 1995. | Zbl
[vW 83] H. Von Weizsäcker. Exchanging the order of taking suprema and countable intersections of σ-algebras. Ann. Inst. Henri Poincaré 19, 91-100, 1983. | Numdam | MR | Zbl
[W 91] D. Williams. Probability with Martingales. Cambridge University Press, 1991. | MR | Zbl
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7370225787162781, "perplexity": 17285.739267390643}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764500758.20/warc/CC-MAIN-20230208092053-20230208122053-00161.warc.gz"}
|
http://math.stackexchange.com/users/36924/mike-flynn?tab=activity&sort=comments
|
# Mike Flynn
less info
reputation
29
bio website location age member for 2 years seen Aug 15 at 15:20 profile views 40
Mar3 comment How to properly take derivatives in calculus of variations (Euler-Lagrange formula) @SanathDevalapurkar It is not obvious to me why $df'/df$ is $0$. Aug14 comment P(tomorrow is the end of the world) =? "they both", do you mean "they all"? Aug13 comment Mean Distance on a 3-sphere? Can you show an attempt to solve the problem? Aug12 comment Linear programming simplex - can I have a constraint with a multiplication? so if your solution is $x = {x_1, x_2, \dots, x_n}$, you are saying all of $n_1, n_2, t_1, t_2$ are components of $x$? Aug9 comment Understanding detailed balance (crossposted from stats) @al-Hwarizmi Markov Chain Monte Carlo, a method of approximating a probability distribution with a random walk, often used in statistics. The second sentence should read "proportional to their ratio of probability densities". I've now edited it. Jul19 comment A method to test for uniform distribution over a convex polytope hit and run. Pick an initial point, choose a random direction, the next point is a random point between initial point and wall in that direction. Jul19 comment A method to test for uniform distribution over a convex polytope I have no way of knowing $|P|$, in fact, this method is how I would calculate $|P|$, provided the sampling is uniform. Jun24 comment What is the typical method for sampling uniformly in a convex polytope Not in practice actually, if you make the jump length long enough. And are you saying that there is no known way to do it quickly for 1000 dimensions? Because I was considering using CUDA to do it... Jun18 comment Changing a basis and satisfying inequalities Ah, I now realize you must be very careful with each step... Dec11 comment What is a covector and what is it used for? yes, so $dx([1,2])$ would be the function $dx$ acting on the vector $[1,2] \in R^2$. Dec4 comment How would you define a geodesic in $\operatorname{SO}(2n)$ and $\operatorname{SO}(2n+1)$? Thank you for the answer. In $R^{n^2}$, do we compute the Euclidean distance as if the matrix was a long vector, adding up the squares of the components and taking a square root? Dec3 comment What is the diameter of a manifold? @QiaochuYuan, it was a question unattached to a textbook, just "What are the intrinsic diameters of SO(2n) and SO(2n+1)?" The textbook we have been using is Frank Morgan's "Riemannian Geometry". My professor probably assumed that the choice of metric would be immediately obvious, but it is not, at least to me. Dec3 comment What is the diameter of a manifold? @QiaochuYuan does changing it to "intrinsic diameter" change anything? Nov19 comment What is a covector and what is it used for? Yes, I think most of my misunderstanding arose from the fact that I was interpreting $dx$ as a "length" as opposed to a function.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9236931800842285, "perplexity": 521.4167242220439}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500830746.39/warc/CC-MAIN-20140820021350-00273-ip-10-180-136-8.ec2.internal.warc.gz"}
|
http://mathhelpforum.com/math-puzzles/209067-scrambled-equations-problem.html
|
# Math Help - Scrambled Equations problem
1. ## Scrambled Equations problem
OK.. I downloaded this math game for my phone and it says all problems are solvable but I have 2 that have me stumped.
by using +,-,*,/ and ( )
get the solution 46 using the numbers 1,2,6,9
each number can only be used once and can not be put next to each other to form a different number
the other equation that has me stumped is:
solution 35 using 3,3,7,8
good luck!
2. ## Re: Scrambled Equations problem
Hello, marcnsam!
Here's the second one . . .
By using $+,\,-,\,\times,\,\div,\, (\,)$, get 35 using 3, 3, 7, 8.
$8(7-3) + 3 \:=\:35$
3. ## Re: Scrambled Equations problem
Originally Posted by marcnsam
by using +,-,*,/ and ( )
get the solution 46 using the numbers 1,2,6,9
(9 - 1)*6 - 2 = 46
Thanks Soroban. You gave me the idea.
-Dan
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 2, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7014043927192688, "perplexity": 3477.753002429586}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115863825.66/warc/CC-MAIN-20150124161103-00239-ip-10-180-212-252.ec2.internal.warc.gz"}
|
http://link.springer.com/chapter/10.1007%2F978-3-642-22321-1_7
|
Chapter
Developments in Language Theory
Volume 6795 of the series Lecture Notes in Computer Science pp 70-81
# Avoiding Abelian Powers in Partial Words
• Francine Blanchet-SadriAffiliated withDepartment of Computer Science, University of North Carolina
• , Sean SimmonsAffiliated withDepartment of Mathematics, The University of Texas at Austin
* Final gross prices may vary according to local VAT.
## Abstract
We study abelian repetitions in partial words, or sequences that may contain some unknown positions or holes. First, we look at the avoidance of abelian pth powers in infinite partial words, where p > 2, extending recent results regarding the case where p = 2. We investigate, for a given p, the smallest alphabet size needed to construct an infinite partial word with finitely or infinitely many holes that avoids abelian pth powers. We construct in particular an infinite binary partial word with infinitely many holes that avoids 6th powers. Then we show, in a number of cases, that the number of abelian p-free partial words of length n with h holes over a given alphabet grows exponentially as n increases. Finally, we prove that we cannot avoid abelian pth powers under arbitrary insertion of holes in an infinite word.
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8809349536895752, "perplexity": 968.7997746928725}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988718426.35/warc/CC-MAIN-20161020183838-00562-ip-10-171-6-4.ec2.internal.warc.gz"}
|
https://samjshah.com/2008/11/17/keplers-laws/?replytocom=602
|
# Kepler’s Laws
In my multivariable calculus class, we spent last Friday reading the textbook as a group, trying to understand the section on Kepler’s Laws. We got done showing that if there is a sun-like object and another object with a particular initial position and velocity, it will either fall into the sun, be an circle, be an ellipse, be a parabola, or be a hyperbola.
Today we were going to move onto using this result to derive the three Keplerian laws of planetary motion.
But then I decided to scrap that. Because even though we read the book and followed the text, line by line and equation by equation, we lost sense of what we were doing. We lost sense of the conceptual underpinnings for each equation. We didn’t know what motivated the book to make the moves it made. It’s largely the book’s fault, which is really unclear — if you’re a high school student and not used to having your book say “we leave this as an exercise to the reader.” (Seriously, it did that.)
One of the things you’ve heard me say is that I want to foster the skill of students learning to communicate math well.
So, I decided to scrap the plan of moving forward, and we’re devoting two or three days to
WRITING OUR OWN TEXT EXPLAINING THE DERIVATION OF KEPLER’S LAWS.
We started out the class outlining a basic structure to it (Part I: What we want to show; Part II: Initial Conditions; Part III: Gravitational Pull; etc.). Then the four students started talking about what they wanted to say. (One agreed to draw the diagrams we’re going to include in our text.) I just sat up front, and when they decided, I typed it up in my LaTeX editor — projected so that students could tell me to fix or reorder something. Sometimes I prompted them (“you told me to write $\vec{v}$ but you never told the reader what that is” or “does it matter if the initial velocity weren’t orthogonal to the position vector?”). And it took us 50 minutes to get about a third of the way done.
But you know what? It is working. They’re talking, they’re thinking, they’re arguing with each other, they’re asking questions. And they’re learning to work through things, and explain them to someone else.
I was so pleased. Hopefully the next few days go as well.
## 2 comments
1. Matt E says:
That is pretty awesome.
2. samjshah says:
I can’t even tell you how great it is. It’s so relaxed and students are really engaging with the ideas and the text and presentation.
It helps that the students were the ones that wanted to do the Kepler section (I gave them the option to skip it), so they had this inherent desire to do it.
But still. I can’t believe how well this is working. The only downside is that so far we’re only about halfway done, and we’ve taken 2 days to do this so far. I was wondering if it was worth 4 days writing this, on top of the day we spent going through the book.
However, after watching them work — after saying “we should submit this for publication somewhere” and “we should send this to [our teacher last year, who retired],” I know it is worth it.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 1, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5909084677696228, "perplexity": 818.8332410540704}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027316194.18/warc/CC-MAIN-20190821194752-20190821220752-00147.warc.gz"}
|
https://r.prevos.net/euler-problem-28/
|
# The Ulam Spiral: Euler Problem 28
Euler Problem 28 takes us to the world of the Ulam Spiral. This is a spiral that contains sequential positive integers in a square spiral, marking the prime numbers. Stanislaw Ulam discovered that a lot of primes are located along the diagonals. These diagonals can be described as polynomials. The Ulam Spiral is thus a way of generating quadratic primes (Euler Problem 27).
Ulam Spiral (WikiMedia).
## Euler Problem 28 Definition
Starting with the number 1 and moving to the right in a clockwise direction a 5 by 5 spiral is formed as follows:
21 22 23 24 25
20 07 08 09 10
19 06 01 02 11
18 05 04 03 12
17 16 15 14 13
It can be verified that the sum of the numbers on the diagonals is 101. What is the sum of the numbers on the diagonals in a 1001 by 1001 spiral formed in the same way?
## Proposed Solution
To solve this problem we do not need to create a matrix. This code calculates the values of the corners of a matrix with size $n$. The lowest number in the matrix with size $n$ is $n(n-3)+4$. The numbers increase by $n-1$.
The code steps through all matrices from size 3 to 1001. The solution uses only the uneven sized matrices because these have a centre. The answer to the problem is the sum of all numbers.
```size <- 1001 # Size of matrix
answer <- 1 # Starting number
# Define corners of subsequent matrices
for (n in seq(from = 3, to = size, by = 2)) {
corners <- seq(from = n * (n - 3) + 3, by = n - 1, length.out = 4)
}
```
## Plotting the Ulam Spiral
We can go beyond Euler Problem 28 and play with the mathematics. This code snippet plots all the prime numbers in the Ulam Spiral. Watch the video for an explanation of the patterns that appear along the diagonals.
Ulam Spiral prime numbers.
The code creates a matrix of the required size and fills it with the Ulam Spiral. The code then identifies all primes using the is.prime function from Euler Problem 7. A heat map visualises the results.
```# Ulam Spiral
size <- 201 # Size of matrix
ulam <- matrix(ncol = size, nrow = size)
mid <- floor(size / 2 + 1)
ulam[mid, mid] <- 1
for (n in seq(from = 3, to = size, by = 2)) {
numbers <- (n * (n - 4) + 5) : ((n + 2) * ((n + 2) - 4) + 4)
d <- mid - floor(n / 2)
l <- length(numbers)
ulam[d, d:(d + n - 1)] <- numbers[(l - n + 1):l]
ulam[d + n - 1, (d + n - 1):d] <- numbers[(n - 1):(n - 2 + n)]
ulam[(d + 1):(d + n - 2), d] <- numbers[(l - n):(l - 2 * n + 3)]
ulam[(d + 1):(d + n - 2), d + n - 1] <- numbers[1:(n - 2)]
}
ulam.primes <- apply(ulam, c(1, 2), is.prime)
# Visualise
library(ggplot2)
library(reshape2)
ulam.primes <- melt(ulam.primes)
ggplot(ulam.primes, aes(x=Var1, y=Var2, fill=value)) + geom_tile() +
scale_fill_manual(values = c("white", "black")) +
guides(fill=FALSE) +
theme(panel.grid.major = element_blank(),
panel.grid.minor = element_blank(),
panel.background = element_blank()) +
theme(axis.title.x=element_blank(),
axis.text.x=element_blank(),
axis.ticks.x=element_blank(),
axis.title.y=element_blank(),
axis.text.y=element_blank(),
axis.ticks.y=element_blank()
)
```
## 8 thoughts on “The Ulam Spiral: Euler Problem 28”
1. Actually, you should replace:
ulam <- numbers[(l – n + 1):l]
ulam <- numbers[(n – 1):(n – 2 + n)]
with:
ulam[c, c:(c + n -1)] <- numbers[(l – n + 1):l]
ulam[c + n -1, (c + n -1):c] <- numbers[(n – 1):(n – 2 + n)]
• Indeed, see my response to your previous comment. Thanks again for helping out.
• Great!
Thanks for interesting article.
2. Something wrong in the code – first you create a matrix:
ulam <- matrix(ncol = size, nrow = size)
next you correctly set the middle number:
ulam[mid, mid] <- 1
but later, you reference it as a variable:
ulam <- numbers[(l – n + 1):l]
It can't work this way!
• Hi World Observer,
Thanks for pointing this out. The SyntaxHighlighter plugin removed some of the code because it was confused by “[c”. Now that I have changed the variable c to d, the code displays correctly.
I have updated the code in the post.
Peter
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 4, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.613677442073822, "perplexity": 892.7979770706205}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257650188.31/warc/CC-MAIN-20180324093251-20180324113251-00794.warc.gz"}
|
https://brilliant.org/problems/lets-do-some-calculus-28/
|
Let's do some calculus! (28)
Calculus Level 4
$\large {\lim_{x \to 0}} \ \dfrac{\cos^2 x - \cos x - e^x \cos x + e^x - \dfrac{x^3}{2}}{x^n}$
Find the value of $$n$$ for which the above limit is finite and non-zero.
Notations: $$e \approx 2.71828$$ is the Euler's number.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9868950843811035, "perplexity": 574.6111969690453}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463608107.28/warc/CC-MAIN-20170525155936-20170525175936-00032.warc.gz"}
|
http://nrich.maths.org/5433/index?nomenu=1
|
by Henry Kwok
Rules of Pole Star Sudoku
This is a variation of Sudoku on a standard 9x9 grid which contains a set of special clue-numbers.
These are small numbers placed always on the border lines between selected pairs of neighbouring cells of the grid.
Each clue-number is the difference between the two numbers that should be in the neighbouring cells just to the left and to the right from that clue-number. For example, a clue-number 7 on the border line means that one of the possible answers in the cell on the left can be 9, 2, 8 or 1. If we choose 9 for the answer in the cell, it means that the answer in the cell next to it on the right is 2. If we choose the answer 2, it means that the answer on the right is 9.
Not much information is there? However, fortunately for the solver, you can use a starting digit (digit 9 in the last column) as the Pole Star to guide you out of the "wilderness" of the puzzle.
Hence this Sudoku variant is named Pole Star Sudoku.
The remaining rules are as in a standard Sudoku: the object of the puzzle is to fill in the whole 9x9 grid with numbers 1 through 9 (one number per cell) so that each horizontal line, each vertical line, and each of the nine 3x3 squares (outlined with the bold lines) must contain all the nine different numbers 1 through 9.
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8112038969993591, "perplexity": 360.85287596766324}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440645298065.68/warc/CC-MAIN-20150827031458-00326-ip-10-171-96-226.ec2.internal.warc.gz"}
|
https://space.meta.stackexchange.com/questions/371/adding-a-hyphen-to-tagremotesensing/372
|
# Adding a hyphen to [tag:remotesensing]
It seems that "remote sensing" is two separate words, so the tag should have a hyphen (i.e., be ). (I do not think this justifies having a synonym.)
I am tagging this even though I do not think there would be much debate about whether "remote sensing" should be treated as two words.
• On the other hand, there is earth-observation which is a subset of remote sensing. – gerrit Sep 24 '13 at 13:05
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8949571251869202, "perplexity": 976.2930626861611}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141182794.28/warc/CC-MAIN-20201125125427-20201125155427-00595.warc.gz"}
|
http://rogmonthcornfar-blog.logdown.com/posts/5855768-direct-methods-for-sparse-matrices-numerical-math-taglio-assassin
|
3 months ago
Direct Methods For Sparse Matrices (Numerical Math scanner trouble jone
Direct Methods For Sparse Matrices (Numerical Mathematics And Scientific Computation) 12
NUMERICAL MATHEMATICS AND SCIENTIFIC COMPUTATION A.M .
NUMERICAL MATHEMATICS AND SCIENTIFIC COMPUTATION .. Direct methods for sparse matrices * M.J.. Baines: Moving nite .. 8.6 Numerical methods in the parabolic .
0198534213 - Direct Methods for Sparse Matrices Numerical .
Direct Methods for Sparse Matrices (Numerical Mathematics and Scientific Computation) by Duff, I.. S.; Erisman, A.. M.; Reid, J.. K.. and a great selection of similar .
Direct Methods for Sparse Matrices (Numerical Mathematics .
Buy Direct Methods for Sparse Matrices (Numerical Mathematics and Scientific Computation) 2 by I.. S.. Duff, A.. M.. Erisman, J.. K.. Reid (ISBN: 9780198508380) from Amazon .
Direct Methods For Sparse Matrices Numerical Mathematics .
Direct methods for sparse matrices (numerical mathematics , buy direct methods for .
Direct Methods For Sparse Matrices Numerical Mathematics .
Browse and Read Direct Methods For Sparse Matrices Numerical Mathematics And Scientific Computation Direct Methods For Sparse Matrices Numerical
Direct Methods For Sparse Matrices Numerical Mathematics .
Direct Methods For Sparse Matrices Numerical Mathematics And Scientific Computation.pdf DIRECT METHODS FOR SPARSE MATRICES NUMERICAL MATHEMATICS AND SCIENTIFIC
Inverse Eigenvalue Problems (Numerical Mathematics and .
Fortran 95 2003 Explained Numerical Mathematics and Scientific Computation numerical mathematics and scientific computation .. Du, A.. Erisman, and J.. Reid: Direct .
Direct Methods for Sparse Matrices - I.. S.. Duff; A.. M .
Numerical Mathematics and Scientific Computation.. .. Comprehensive view of direct methods for sparse matrices.. .. Science & Mathematics > Mathematics > Numerical .
Direct methods for sparse linear systems
Studies in Applied and Numerical Mathematics; .. textbook on modern sparse direct methods.. .. Direct Methods for Sparse Linear Systems equips readers with the .
Direct methods for sparse matrix solution
Direct methods for sparse matrix solution.. .. Direct Methods for Sparse Matrices, Oxford University Press, London, . 3b9d4819c4
3d page flip full version free 58
pretty little liars book 10 pdf free 34
golka marian socjologia kultury pdf 11
the dip seth godin free ebook 84
amma magana sex stories kannada. 20
corel draw x4 serial number keygen 11
Met Art Feeona A 13
radicals and visionaries ebook 11
touchstone 3 teacher's book pdf 165
maximum ride forever epub free 23
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8223960995674133, "perplexity": 15801.799253071815}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794864405.39/warc/CC-MAIN-20180521142238-20180521162238-00163.warc.gz"}
|
https://intl.siyavula.com/read/science/grade-11/newtons-laws/02-newtons-laws-06
|
We think you are located in South Africa. Is this correct?
# Don't get left behind
Join thousands of learners improving their science marks online with Siyavula Practice.
## Forces and Newton's Laws
Exercise 2.10
A force acts on an object. Name three effects that the force can have on the object.
A force can change the shape of the object. A force can change the direction in which the object is moving. A force can accelerate or stop a body.
Identify each of the following forces as contact or non-contact forces.
The force between the north pole of a magnet and a paper clip.
non-contact
The force required to open the door of a taxi.
contact
The force required to stop a soccer ball.
contact
The force causing a ball, dropped from a height of $$\text{2}$$ $$\text{m}$$, to fall to the floor.
non-contact
A book of mass $$\text{2}$$ $$\text{kg}$$ is lying on a table. Draw a labelled force diagram indicating all the forces on the book.
Where $$\vec{F}_{g}$$ is the force due to gravity and $$\vec{N}$$ is the normal force.
A constant, resultant force acts on a body which can move freely in a straight line. Which physical quantity will remain constant?
1. acceleration
2. velocity
3. momentum
4. kinetic energy
[SC 2003/11]
acceleration
Two forces, $$\text{10}$$ $$\text{N}$$ and $$\text{15}$$ $$\text{N}$$, act at an angle at the same point.
Which of the following cannot be the resultant of these two forces?
1. $$\text{2}$$ $$\text{N}$$
2. $$\text{5}$$ $$\text{N}$$
3. $$\text{8}$$ $$\text{N}$$
4. $$\text{20}$$ $$\text{N}$$
[SC 2005/11 SG1]
$$\text{2}$$ $$\text{N}$$
A concrete block weighing $$\text{250}$$ $$\text{N}$$ is at rest on an inclined surface at an angle of $$\text{20}$$$$\text{°}$$. The magnitude of the normal force, in newtons, is
1. $$\text{250}$$
2. $$250\cos\text{20}\text{°}$$
3. $$\text{250}\sin\text{20}\text{°}$$
4. $$\text{2 500}\cos\text{20}\text{°}$$
$$\text{250}$$
A $$\text{30}$$ $$\text{kg}$$ box sits on a flat frictionless surface. Two forces of $$\text{200}$$ $$\text{N}$$ each are applied to the box as shown in the diagram. Which statement best describes the motion of the box?
1. The box is lifted off the surface.
2. The box moves to the right.
3. The box does not move.
4. The box moves to the left.
The box moves to the left.
A concrete block weighing $$\text{200}$$ $$\text{N}$$ is at rest on an inclined surface at an angle of $$\text{20}$$$$\text{°}$$. The normal force, in newtons, is
1. $$\text{200}$$
2. $$\text{200}\cos\text{20}\text{°}$$
3. $$\text{200}\sin\text{20}\text{°}$$
4. $$\text{2 000}\cos\text{20}\text{°}$$
$$\text{200}$$
A box, mass $$m$$, is at rest on a rough horizontal surface. A force of constant magnitude $$F$$ is then applied on the box at an angle of $$\text{60}$$$$\text{°}$$ to the horizontal, as shown.
If the box has a uniform horizontal acceleration of magnitude, $$a$$, the frictional force acting on the box is...
1. $$F\cos\text{60}\text{°}-ma$$ in the direction of A
2. $$F\cos\text{60}\text{°}-ma$$ in the direction of B
3. $$F\sin\text{60}\text{°}-ma$$ in the direction of A
4. $$F\sin\text{60}\text{°}-ma$$ in the direction of B
[SC 2003/11]
$$F\cos\text{60}\text{°}-ma$$ in the direction of A
Thabo stands in a train carriage which is moving eastwards. The train suddenly brakes. Thabo continues to move eastwards due to the effect of:
1. his inertia.
2. the inertia of the train.
3. the braking force on him.
4. a resultant force acting on him.
[SC 2002/11 SG]
his inertia
A $$\text{100}$$ $$\text{kg}$$ crate is placed on a slope that makes an angle of $$\text{45}$$$$\text{°}$$ with the horizontal. The gravitational force on the box is $$\text{98}$$ $$\text{N}$$. The box does not slide down the slope. Calculate the magnitude and direction of the frictional force and the normal force present in this situation.
Since the box does not move the frictional force acting on the box is equal to the component of the gravitational force parallel to the plane.
\begin{align*} F_{f} & = F_{gx} \\ & = F \cos \theta \\ & = (98) \cos(45) \\ & = \text{69,3}\text{ N} \end{align*}
This force acts at an an angle of $$\text{45}$$$$\text{°}$$ with the horizontal.
The normal force acts in the opposite direction to the gravitational force and has a magnitude of $$\text{98}$$ $$\text{N}$$ at an angle of $$\text{45}$$$$\text{°}$$ with the horizontal.
A body moving at a CONSTANT VELOCITY on a horizontal plane, has a number of unequal forces acting on it. Which one of the following statements is TRUE?
1. At least two of the forces must be acting in the same direction.
2. The resultant of the forces is zero.
3. Friction between the body and the plane causes a resultant force.
4. The vector sum of the forces causes a resultant force which acts in the direction of motion.
[SC 2002/11 HG1]
The resultant of the forces is zero.
Two masses of $$m$$ and $$2m$$ respectively are connected by an elastic band on a frictionless surface. The masses are pulled in opposite directions by two forces each of magnitude $$F$$, stretching the elastic band and holding the masses stationary.
Which of the following gives the magnitude of the tension in the elastic band?
1. zero
2. $$\frac{\text{1}}{\text{2}}F$$
3. $$F$$
4. $$2F$$
[IEB 2005/11 HG]
$$F$$
A rocket takes off from its launching pad, accelerating up into the air.
The rocket accelerates because the magnitude of the upward force, F is greater than the magnitude of the rocket's weight, W. Which of the following statements best describes how force F arises?
1. F is the force of the air acting on the base of the rocket.
2. F is the force of the rocket's gas jet pushing down on the air.
3. F is the force of the rocket's gas jet pushing down on the ground.
4. F is the reaction to the force that the rocket exerts on the gases which escape out through the tail nozzle.
[IEB 2005/11 HG]
F is the reaction to the force that the rocket exerts on the gases which escape out through the tail nozzle.
A box of mass $$\text{20}$$ $$\text{kg}$$ rests on a smooth horizontal surface. What will happen to the box if two forces each of magnitude $$\text{200}$$ $$\text{N}$$ are applied simultaneously to the box as shown in the diagram.
The box will:
1. be lifted off the surface.
2. move to the left.
3. move to the right.
4. remain at rest.
[SC 2001/11 HG1]
move to the left
A $$\text{2}$$ $$\text{kg}$$ mass is suspended from spring balance X, while a $$\text{3}$$ $$\text{kg}$$ mass is suspended from spring balance Y. Balance X is in turn suspended from the $$\text{3}$$ $$\text{kg}$$ mass. Ignore the weights of the two spring balances.
The readings (in N) on balances X and Y are as follows:
X Y a) $$\text{19,6}$$ $$\text{29,4}$$ b) $$\text{19,6}$$ 49 c) $$\text{24,5}$$ $$\text{24,5}$$ d) 49 49
[SC 2001/11 HG1]
b) $$\text{19,6}$$, 49
P and Q are two forces of equal magnitude applied simultaneously to a body at X.
As the angle θ between the forces is decreased from $$\text{180}$$$$\text{°}$$ to $$\text{0}$$$$\text{°}$$, the magnitude of the resultant of the two forces will
1. initially increase and then decrease.
2. initially decrease and then increase.
3. increase only.
4. decrease only.
[SC 2002/03 HG1]
increase only
The graph below shows the velocity-time graph for a moving object:
Which of the following graphs could best represent the relationship between the resultant force applied to the object and time?
(a) (b) (c) (d)
[SC 2002/03 HG1]
graph (b)
Two blocks each of mass $$\text{8}$$ $$\text{kg}$$ are in contact with each other and are accelerated along a frictionless surface by a force of $$\text{80}$$ $$\text{N}$$ as shown in the diagram. The force which block Q will exert on block P is equal to ...
1. $$\text{0}$$ $$\text{N}$$
2. $$\text{40}$$ $$\text{N}$$
3. $$\text{60}$$ $$\text{N}$$
4. $$\text{80}$$ $$\text{N}$$
[SC 2002/03 HG1]
$$\text{40}$$ $$\text{N}$$
A $$\text{12}$$ $$\text{kg}$$ box is placed on a rough surface. A force of $$\text{20}$$ $$\text{N}$$ applied at an angle of $$\text{30}$$$$\text{°}$$ to the horizontal cannot move the box. Calculate the magnitude and direction of the normal and friction forces.
The normal force is:
\begin{align*} N & = mg \\ & = (\text{12})(\text{9,8}) \\ & = \text{117,6}\text{ N} \end{align*}
This force points straight up from the surface.
The friction force is:
\begin{align*} F_{f} & = F_{\text{app}} \cos 30° \\ & = (\text{20})\cos(\text{30}°) \\ & = \text{17,3}\text{ N} \end{align*}
The frictional force points in the opposite direction to the applied force.
Three $$\text{1}$$ $$\text{kg}$$ mass pieces are placed on top of a $$\text{2}$$ $$\text{kg}$$ trolley. When a force of magnitude F is applied to the trolley, it experiences an acceleration a.
If one of the $$\text{1}$$ $$\text{kg}$$ mass pieces falls off while F is still being applied, the trolley will accelerate at ...
1. $$\frac{\text{1}}{\text{5}}a$$
2. $$\frac{\text{4}}{\text{5}}a$$
3. $$\frac{\text{5}}{\text{4}}a$$
4. $$5a$$
[SC 2002/03 HG1]
$$\frac{\text{5}}{\text{4}}a$$
A car moves along a horizontal road at constant velocity. Which of the following statements is true?
1. The car is not in equilibrium.
2. There are no forces acting on the car.
3. There is zero resultant force.
4. There is no frictional force.
[IEB 2004/11 HG1]
There is zero resultant force.
A crane lifts a load vertically upwards at constant speed. The upward force exerted on the load is F. Which of the following statements is correct?
1. The acceleration of the load is $$\text{9,8}$$ $$\text{m·s^{-1}}$$ downwards.
2. The resultant force on the load is F.
3. The load has a weight equal in magnitude to F.
4. The forces of the crane on the load, and the weight of the load, are an example of a Newton's third law 'action-reaction' pair.
[IEB 2004/11 HG1]
The load has a weight equal in magnitude to F.
A body of mass $$M$$ is at rest on a smooth horizontal surface with two forces applied to it as in the diagram below. Force $${F}_{1}$$ is equal to $$Mg$$. The force $${F}_{1}$$ is applied to the right at an angle θ to the horizontal, and a force of $${F}_{2}$$ is applied horizontally to the left.
How is the body affected when the angle θ is increased?
1. It remains at rest.
2. It lifts up off the surface, and accelerates towards the right.
3. It lifts up off the surface, and accelerates towards the left.
4. It accelerates to the left, moving along the smooth horizontal surface.
[IEB 2004/11 HG1]
It accelerates to the left, moving along the smooth horizontal surface.
Which of the following statements correctly explains why a passenger in a car, who is not restrained by the seat belt, continues to move forward when the brakes are applied suddenly?
1. The braking force applied to the car exerts an equal and opposite force on the passenger.
2. A forward force (called inertia) acts on the passenger.
3. A resultant forward force acts on the passenger.
4. A zero resultant force acts on the passenger.
[IEB 2003/11 HG1]
A zero resultant force acts on the passenger.
A rocket (mass $$\text{20 000}$$ $$\text{kg}$$) accelerates from rest to $$\text{40}$$ $$\text{m·s^{-1}}$$ in the first $$\text{1,6}$$ seconds of its journey upwards into space.
The rocket's propulsion system consists of exhaust gases, which are pushed out of an outlet at its base.
Explain, with reference to the appropriate law of Newton, how the escaping exhaust gases exert an upwards force (thrust) on the rocket.
Newton's third law states that if body A exerts a force on body B, then body B exerts a force of equal magnitude on body A, but in the opposite direction.
The exhaust gases exert a force on the rocket. This means that the rocket also exerts a force on the exhaust gases. Since the direction of the force from the exhaust is downwards the direction of the force from the rocket must be upwards. This is what propels the rocket upwards.
What is the magnitude of the total thrust exerted on the rocket during the first $$\text{1,6}$$ $$\text{s}$$?
We first need to find the acceleration. We can do this by using the equations of motion:
\begin{align*} v_{f} & = v_{i} + at \\ 40 & = 0 + (\text{1,6})a \\ a & = \text{25}\text{ m·s$^{-2}$} \end{align*}
Now we can find the magnitude of the total thrust, T:
\begin{align*} T & = W + ma \\ & = (\text{20 000})(\text{9,8}) + (\text{20 000})(\text{25}) \\ & = \text{696 000}\text{ N} \end{align*}
An astronaut of mass $$\text{80}$$ $$\text{kg}$$ is carried in the space capsule. Determine the resultant force acting on him during the first $$\text{1,6}$$ $$\text{s}$$.
\begin{align*} T & = W + ma \\ & = (\text{80})(\text{9,8}) + (\text{80})(\text{25}) \\ & = \text{2 784}\text{ N} \end{align*}
Explain why the astronaut, seated in his chair, feels “heavier” while the rocket is launched.
Weight is measured through normal forces. When the rocket accelerates upwards he feels a greater normal force acting on him as the force required to accelerate him upwards in addition to balancing out the gravitational force.
State Newton's second law of Motion.
If a resultant force acts on a body, it will cause the body to accelerate in the direction of the resultant force. The acceleration of the body will be directly proportional to the resultant force and inversely proportional to the mass of the body. The mathematical representation is: $\vec{F}_{net} = m\vec{a}$
A sports car (mass $$\text{1 000}$$ $$\text{kg}$$) is able to accelerate uniformly from rest to $$\text{30}$$ $$\text{m·s^{-1}}$$ in a minimum time of $$\text{6}$$ $$\text{s}$$.
Calculate the magnitude of the acceleration of the car.
\begin{align*} a & = \frac{\Delta v}{\Delta t} \\ & = \frac{30}{6} \\ & = \text{5}\text{ m·s$^{-2}$} \end{align*}
What is the magnitude of the resultant force acting on the car during these $$\text{6}$$ $$\text{s}$$?
\begin{align*} F & = ma \\ & = (\text{1 000})(5) \\ & = \text{5 000}\text{ N} \end{align*}
The magnitude of the force that the wheels of the vehicle exert on the road surface as it accelerates is $$\text{7 500}$$ $$\text{N}$$. What is the magnitude of the retarding forces acting on this car?
The retarding force is the force that the cars wheels exert minus the resultant force:
\begin{align*} F & = F_{\text{app}} - F_{\text{res}} \\ & = \text{7 500} - \text{5 000} \\ & = \text{2 500}\text{ N} \end{align*}
By reference to a suitable Law of Motion, explain why a headrest is important in a car with such a rapid acceleration.
Newton's first law states that an object continues in a state of rest or uniform motion (motion with a constant velocity) unless it is acted on by an unbalanced (net or resultant) force.
There is a resultant force on the person's head and as the car accelerates their head is pulled back. The headrest stops the person's head from being pulled to far back and prevents whiplash.
A child (mass $$\text{18}$$ $$\text{kg}$$) is strapped in his car seat as the car moves to the right at constant velocity along a straight level road. A tool box rests on the seat beside him.
The driver brakes suddenly, bringing the car rapidly to a halt. There is negligible friction between the car seat and the box.
Draw a labelled free-body diagram of the forces acting on the child during the time that the car is being braked.
Draw a labelled free-body diagram of the forces acting on the box during the time that the car is being braked.
Modern cars are designed with safety features (besides seat belts) to protect drivers and passengers during collisions e.g. the crumple zones on the car's body. Rather than remaining rigid during a collision, the crumple zones allow the car's body to collapse steadily.
State Newton's second law of motion.
If a resultant force acts on a body, it will cause the body to accelerate in the direction of the resultant force. The acceleration of the body will be directly proportional to the resultant force and inversely proportional to the mass of the body. The mathematical representation is:$\vec{F}_{net} = m\vec{a}$
The total mass of a lift together with its load is $$\text{1 200}$$ $$\text{kg}$$. It is moving downwards at a constant velocity of $$\text{9}$$ $$\text{m·s^{-1}}$$.
What will be the magnitude of the force exerted by the cable on the lift while it is moving downwards at constant velocity? Give an explanation for your answer.
We take upwards as the positive direction.
The acceleration is $$\text{0}$$ since the lift is moving at constant velocity.
\begin{align*} \Sigma F & = T - W \\ (0)(m) & = T - W \\ \therefore T & = W \\ & = (\text{9,8})(\text{1 200}) \\ & = \text{11 760}\text{ N} \end{align*}
The lift is now uniformly brought to rest over a distance of $$\text{18}$$ $$\text{m}$$.
Calculate the magnitude of the acceleration of the lift.
\begin{align*} v_{f}^{2} & = v_{i}^{2} + 2a \Delta x \\ 0 & = (9)^{2} + 2a(18) \\ 36a & = -81 \\ a & = \frac{-\text{81}}{\text{36}} \\ & = -\text{2,25}\text{ m·s$^{-2}$} \end{align*}
Calculate the magnitude of the force exerted by the cable while the lift is being brought to rest.
\begin{align*} T - W & = ma \\ T & = ma + mg \\ & = (\text{1 200})(-\text{2,25}) + (\text{9,8})(\text{1 200}) \\ & = \text{9 060}\text{ N} \end{align*}
A driving force of $$\text{800}$$ $$\text{N}$$ acts on a car of mass $$\text{600}$$ $$\text{kg}$$.
Calculate the car's acceleration.
\begin{align*} F & = ma \\ a & = \frac{F}{m} \\ & = \frac{\text{800}}{\text{600}} \\ & = \text{1,33}\text{ m·s$^{-2}$} \end{align*}
Calculate the car's speed after $$\text{20}$$ $$\text{s}$$.
\begin{align*} v_{f} & = v_{i} + at \\ v_{f} & = 0 + (20)(\text{1,33}) \\ & = \text{26,7}\text{ m·s$^{-1}$} \end{align*}
Calculate the new acceleration if a frictional force of $$\text{50}$$ $$\text{N}$$ starts to act on the car after $$\text{20}$$ $$\text{s}$$.
\begin{align*} a_{2} & = a_{1} - \frac{F_{f}}{m} \\ & = \text{1,33} - \frac{\text{50}}{\text{600}} \\ & = \text{1,25}\text{ m·s$^{-2}$} \end{align*}
Calculate the speed of the car after another $$\text{20}$$ $$\text{s}$$ (i.e. a total of $$\text{40}$$ $$\text{s}$$ after the start).
\begin{align*} v_{f2} & = a_{1}t + a_{2}t \\ v_{f2} & = (20)(\text{1,33}) + (20)(\text{1,25}) \\ & = \text{51,6}\text{ m·s$^{-1}$} \end{align*}
A stationary block of mass $$\text{3}$$ $$\text{kg}$$ is on top of a plane inclined at $$\text{35}$$$$\text{°}$$ to the horizontal.
Draw a force diagram (not to scale). Include the weight of the block as well as the components of the weight that are perpendicular and parallel to the inclined plane.
Determine the values of the weight's perpendicular and parallel components.
The perpendicular component is:
\begin{align*} W_{\perp} & = F \sin(\theta) \\ & = mg \sin (\theta) \\ & = (3)(9.8) \sin(35) \\ & = \text{16,86}\text{ N} \end{align*}
The parallel component is:
\begin{align*} W_{||} & = F \cos(\theta) \\ & = mg \cos (\theta) \\ & = (3)(9.8) \cos(35) \\ & = \text{24,01}\text{ N} \end{align*}
A crate on an inclined plane
Elephants are being moved from the Kruger National Park to the Eastern Cape. They are loaded into crates that are pulled up a ramp (an inclined plane) on frictionless rollers.
The diagram shows a crate being held stationary on the ramp by means of a rope parallel to the ramp. The tension in the rope is $$\text{5 000}$$ $$\text{N}$$.
Explain how one can deduce the following: “The forces acting on the crate are in equilibrium”.
The sum of the forces is equal to the mass times the acceleration. Since the acceleration is $$\text{0}$$, the sum of the forces is $$\text{0}$$. This means that the forces acting on the crate are in equilibrium.
Draw a labelled free-body diagram of the forces acting on the elephant. (Regard the crate and elephant as one object, and represent them as a dot. Also show the relevant angles between the forces.)
The crate has a mass of $$\text{800}$$ $$\text{kg}$$. Determine the mass of the elephant.
\begin{align*} W_{E ||} & = W \sin \theta \\ \text{5 000} & = W \sin(15) \\ W & = \frac{\text{5 000}}{\sin (15)} \\ & = \text{19 318,52}\text{ N} \end{align*} \begin{align*} m_{E} & = \frac{W}{g} - m_{\text{crate}} \\ & = \frac{\text{19 318,52}}{\text{9,8}} - \text{800} \\ & = \text{1 171,28}\text{ kg} \end{align*}
The crate is now pulled up the ramp at a constant speed. How does the crate being pulled up the ramp at a constant speed affect the forces acting on the crate and elephant? Justify your answer, mentioning any law or principle that applies to this situation.
There will be no changes to the forces since the acceleration is $$\text{0}$$.
Car in Tow
Car A is towing Car B with a light tow rope. The cars move along a straight, horizontal road.
Write down a statement of Newton's second law of Motion (in words).
If a resultant force acts on a body, it will cause the body to accelerate in the direction of the resultant force. The acceleration of the body will be directly proportional to the resultant force and inversely proportional to the mass of the body.
As they start off, Car A exerts a forwards force of $$\text{600}$$ $$\text{N}$$ at its end of the tow rope. The force of friction on Car B when it starts to move is $$\text{200}$$ $$\text{N}$$. The mass of Car B is $$\text{1 200}$$ $$\text{kg}$$. Calculate the acceleration of Car B.
The total force on car B is $$\text{600}\text{ N} - \text{200}\text{ N} = \text{400}\text{ N}$$.
\begin{align*} F & = ma \\ a & = \frac{\text{400}}{\text{1 200}} \\ & = \text{0,33}\text{ m·s$^{-2}$} \end{align*}
After a while, the cars travel at constant velocity. The force exerted on the tow rope is now $$\text{300}$$ $$\text{N}$$ while the force of friction on Car B increases. What is the magnitude and direction of the force of friction on Car B now?
The frictional force is equal to the force exerted on the tow rope but in the opposite direction. $$F_{f} = \text{300}\text{ N}$$.
Towing with a rope is very dangerous. A solid bar should be used in preference to a tow rope. This is especially true should Car A suddenly apply brakes. What would be the advantage of the solid bar over the tow rope in such a situation?
Newton's first law states that an object will keep on moving in a straight line unless acted on by a force. A solid bar will stop car B from moving if car A breaks while a rope will not.
The mass of Car A is also $$\text{1 200}$$ $$\text{kg}$$. Car A and Car B are now joined by a solid tow bar and the total braking force is $$\text{9 600}$$ $$\text{N}$$. Over what distance could the cars stop from a velocity of $$\text{20}$$ $$\text{m·s^{-1}}$$?
The acceleration is:
\begin{align*} F & = ma \\ a & = \frac{\text{9 600}}{2(\text{1 200})} \\ & = \text{4} \end{align*}
So the stopping distance is:
\begin{align*} v_{f}^{2} & = v_{i}^{2} + 2a \Delta x \\ 0 & = (\text{20})^{2} + 2(-4)\Delta x \\ 8 \Delta x & = 400 \\ \Delta x & = \text{50}\text{ m} \end{align*}
Testing the Brakes of a Car
A braking test is carried out on a car travelling at $$\text{20}$$ $$\text{m·s^{-1}}$$. A braking distance of $$\text{30}$$ $$\text{m}$$ is measured when a braking force of $$\text{6 000}$$ $$\text{N}$$ is applied to stop the car.
Calculate the acceleration of the car when a braking force of $$\text{6 000}$$ $$\text{N}$$ is applied.
\begin{align*} v_{f}^{2} & = v_{i}^{2} + 2a \Delta x \\ 0 & = (\text{20})^{2} + 2(30)a \\ 60a & = -400 \\ a & = -\text{6,67}\text{ m·s$^{-2}$} \end{align*}
Show that the mass of this car is $$\text{900}$$ $$\text{kg}$$.
\begin{align*} F & = ma \\ \text{6 000} & = (\text{6,67})m \\ m & = \text{900}\text{ kg} \end{align*}
How long (in s) does it take for this car to stop from $$\text{20}$$ $$\text{m·s^{-1}}$$ under the braking action described above?
\begin{align*} v_{f} & = v_{i} + at \\ 0 & = \text{20} + (-\text{6,67})t \\ \text{6,67}t & = 20 \\ t & = \text{3,0}\text{ s} \end{align*}
A trailer of mass $$\text{600}$$ $$\text{kg}$$ is attached to the car and the braking test is repeated from $$\text{20}$$ $$\text{m·s^{-1}}$$ using the same braking force of $$\text{6 000}$$ $$\text{N}$$. How much longer will it take to stop the car with the trailer in tow?
\begin{align*} F & = ma \\ \text{6 000} & = (900 + 600)a \\ a & = \text{4} \end{align*} \begin{align*} t & = \frac{-\text{20}}{\text{4}} \\ t & = \text{5}\text{ s} \end{align*}
It will take $$\text{2}$$ $$\text{s}$$ longer.
A box is held stationary on a smooth plane that is inclined at angle θ to the horizontal.
$$F$$ is the force exerted by a rope on the box. $$w$$ is the weight of the box and $$N$$ is the normal force of the plane on the box. Which of the following statements is correct?
1. $$\tan\theta =\frac{F}{w}$$
2. $$\tan\theta =\frac{F}{N}$$
3. $$\cos\theta =\frac{F}{w}$$
4. $$\sin\theta =\frac{N}{w}$$
[IEB 2005/11 HG]
$$\sin\theta =\frac{N}{w}$$
As a result of three forces $${F}_{1}$$, $${F}_{2}$$ and $${F}_{3}$$ acting on it, an object at point P is in equilibrium.
Which of the following statements is not true with reference to the three forces?
1. The resultant of forces $${F}_{1}$$, $${F}_{2}$$ and $${F}_{3}$$ is zero.
2. Force $${F}_{1}$$, $${F}_{2}$$ and $${F}_{3}$$ lie in the same plane.
3. Force $${F}_{3}$$ is the resultant of forces $${F}_{1}$$ and $${F}_{2}$$.
4. The sum of the components of all the forces in any chosen direction is zero.
[SC 2001/11 HG1]
Force $${F}_{3}$$ is the resultant of forces $${F}_{1}$$ and $${F}_{2}$$.
A block of mass M is held stationary by a rope of negligible mass. The block rests on a frictionless plane which is inclined at $$\text{30}$$$$\text{°}$$ to the horizontal.
Draw a labelled force diagram which shows all the forces acting on the block.
Resolve the force due to gravity into components that are parallel and perpendicular to the plane.
\begin{align*} W & = Mg \\ W_{||} & = Mg \sin \theta \\ W_{\perp} & = Mg \cos \theta \end{align*}
Calculate the weight of the block when the force in the rope is $$\text{8}$$ $$\text{N}$$.
\begin{align*} W_{||} & = Mg \sin \theta \\ 8 & = M(\text{9,8}) \sin (\text{30}) \\ M & = \text{1,63}\text{ N} \end{align*}
A heavy box, mass m, is lifted by means of a rope R which passes over a pulley fixed to a pole. A second rope S, tied to rope R at point P, exerts a horizontal force and pulls the box to the right. After lifting the box to a certain height, the box is held stationary as shown in the sketch below. Ignore the masses of the ropes. The tension in rope R is $$\text{5 850}$$ $$\text{N}$$.
Draw a diagram (with labels) of all the forces acting at the point P, when P is in equilibrium.
By resolving the force exerted by rope R into components, calculate the...
magnitude of the force exerted by rope S.
\begin{align*} S & = R \cos \theta \\ & = (\text{5 850}) \cos (\text{20}) \\ & = \text{5 497,2}\text{ N} \end{align*}
mass, m, of the box.
\begin{align*} mg & = R \sin \theta \\ \text{9,8}m& = (\text{5 850}) \sin (\text{20}) \\ m & = \text{204,17}\text{ kg} \end{align*}
A tow truck attempts to tow a broken down car of mass $$\text{400}$$ $$\text{kg}$$. The coefficient of static friction is $$\text{0,60}$$ and the coefficient of kinetic (dynamic) friction is $$\text{0,4}$$. A rope connects the tow truck to the car. Calculate the force required:
to just move the car if the rope is parallel to the road.
\begin{align*} F_{s} & = \mu_{s}F \\ F_{s} & = \mu_{s} mg \\ & = (\text{0,6})(\text{400})(\text{9,8}) \\ F_{s} & = \text{2 352}\text{ N} \end{align*}
to keep the car moving at constant speed if the rope is parallel to the road.
\begin{align*} F_{f} & = \mu_{k}F \\ & = \mu_{k} mg \\ & = (\text{0,4})(\text{400})(\text{9,8}) \\ & = \text{1 568}\text{ N} \end{align*}
to just move the car if the rope makes an angle of $$\text{30}$$$$\text{°}$$ to the road.
\begin{align*} F_{s} \cos \theta & = \mu_{s}F \\ F_{s} & = \frac{\text{2 352}}{\cos\text{30}} \\ & = \text{2 715,9}\text{ N} \end{align*}
to keep the car moving at constant speed if the rope makes an angle of $$\text{30}$$$$\text{°}$$ to the road.
\begin{align*} F_{f} \cos \theta & = \mu_{f}F \\ F_{f} & = \frac{\text{1 568}}{\cos\text{30}} \\ & = \text{1 810,6}\text{ N} \end{align*}
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.999161958694458, "perplexity": 555.082374877282}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875144058.43/warc/CC-MAIN-20200219061325-20200219091325-00029.warc.gz"}
|
http://mathhelpforum.com/advanced-statistics/119491-one-2-tail-test-hypothesis-cobb-douglas-prod-function.html
|
# Thread: One or 2 Tail test hypothesis (cobb douglas prod. function)
1. ## One or 2 Tail test hypothesis (cobb douglas prod. function)
Considering the Cobb Douglas production function as the initial function to start with.
We regress that to Ln y = ln a + ln b1 k + ln b2 L
Now we state in the hypothesis that ( h0; ) b1=b2=b3 = 0 is equal to zero
And the alternative hypothesis is that it's not equal to zero.
Considering the ' not equal to zero ' u might expect a two tail test because it can be either positive or negative. But since were' talking about Capital and Labor, we expect those 2 variables to have a positive! effect on output.
Can we therefore say that it's better to use a 1 - tail test?
2. Originally Posted by Massachusetts Cowboys
Considering the Cobb Douglas production function as the initial function to start with.
We regress that to Ln y = ln a + ln b1 k + ln b2 L
Now we state in the hypothesis that ( h0; ) b1=b2=b3 = 0 is equal to zero
And the alternative hypothesis is that it's not equal to zero.
Considering the ' not equal to zero ' u might expect a two tail test because it can be either positive or negative. But since were' talking about Capital and Labor, we expect those 2 variables to have a positive! effect on output.
Can we therefore say that it's better to use a 1 - tail test?
Negative Capital is rare occurences, but it happens sometimes, Goldman Sachs is one example.
Negative Labor is an indication that the company needs to hire more labor to catch up with intensive production.
There is nothing wrong in using two-tailed test in this case. If you prefer not to use two-tailed test, you must modify your null hypothesis. Hypotheses can always be tailored to you personal preference. You may set it up to accept or to reject or to postpone decisions.
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8473897576332092, "perplexity": 1392.3817630343383}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257645824.5/warc/CC-MAIN-20180318145821-20180318165821-00043.warc.gz"}
|
https://wiki.opencog.org/w/Real_World_Reasoning
|
# Real World Reasoning
Real-World Reasoning:
The Challenge of Scalable, Uncertain Spatiotemporal, Contextual and Causal Inference
Ben Goertzel, Nil Geisweiller, Cassio Pennachin, Lucio Coelho, Predrag Janicic
# Introduction
The general problem addressed in this book is how to effectively carry out reasoning, knowledge discovery and querying based on huge amounts of complex information about real-world situations. Specifically we conceive “real-world reasoning” here mainly as “massively scalable reasoning involving uncertainty, space, time, cause and context.” Of course there are other important aspects to reasoning about the real world we live in, e.g. the hierarchical structure of much of the human world, and we will briefly touch on some of these here as well. But for the purposes of this book, when we mention “real-world reasoning” or RWR, we’re mostly talking about uncertainty, spacetime, cause, context and scalability.
The RWR problem is critical in at least two respects: as part of the broader pursuit of artificial general intelligence (AGI) [Goertzel & Pennachin, 2006; Goertzel & Wang, 2006a; Goertzel & Bugaj, 2008; Goertzel & Hart, 2008]., and in terms of the practical information processing needs that have arisen in current society.
On the AGI side, it is obvious that every human brain ingests a huge amount of knowledge each waking hour, and somehow we manage to query and analyze our huge, dynamic internal data stores. No AGI design can possibly succeed without some way to effectively carry out intelligent judgment and discovery based on these data stores. AGI also has other aspects, e.g. procedure learning and goal refinement (to name just two), but RWR is certainly a huge part of the puzzle.
On the practical information processing side, anyone who lives in a developed country these days is aware of the tremendous amount of data continually being gathered about all manner of aspects of the human and natural worlds. Much of this data is discarded shortly after it’s gathered, but much of it is retained in various repositories. However, even when the data is retained, it is rarely utilized to anywhere near the full extent possible, because our state-of-the-art technologies for storing, querying, mining and analyzing very large data stores are still very primitive and simplistic (not only compared to what is in principle possible, but compared to what we know to be possible based on contemporary mathematics and computer science).
In these pages we review a class of approaches to handling these RWR problems using uncertain, spatiotemporal, contextual and causal logic. Uncertain logic is not the only possible approach to the RWR problem, but we believe it’s one very promising approach, and it’s our focus here. While the first RWR-capable logic system has yet to be constructed, we make an argument, via detailed review of the literature and the state of the art and suggestion of some original ideas, that the time is ripe for their construction.
The book is intended to serve two purposes: to provide a reasonably accessible overview of the RWR problem and the available technologies and concepts for its solution; and to provide a sketch of one possible avenue toward solution.
Toward the “overview” goal, we review a number of concepts and technologies – some recently developed, some more classical -- that address aspects of the RWR problem. While our treatment centers on formal logic, we also introduce material from other areas such as graph databases, probability theory, cognitive architecture and so forth as appropriate.
After reviewing a variety of other logical approaches, we present our own approach to real-world reasoning, which is based on the Probabilistic Logic Networks (PLN) framework [Goertzel et al, 2008]; and give some detailed suggestions regarding how one might address the scalable real-world inference problem effectively via integrating PLN with other ideas and technologies described. Our goal in this regard is not to propose a particular highly-specific technical solution, but rather to describe a class of possible solutions that might be described as “scalable spatiotemporal uncertain logic systems”. In this vein, in the later chapters we give a number of detailed examples showing the kinds of results one might expect to obtain by approaching a large knowledge store containing information about everyday human activities with the Probabilistic Logic Networks inference framework that we have developed in prior publications.
## The Advantages of a Logical Approach
There are many advantages to the logic-based approach relative to others, some of which will be alluded to as the text progresses, but perhaps the largest advantage is its relative representational transparency. That is, if the knowledge stored in a knowledge base, and the patterns recognized in this knowledge base, are represented in a logical format, then it is reasonably tractable for humans to inspect this knowledge and these patterns. This is a major practical advantage in terms of allowing hybridized human/artificial intelligence – and, given the comments made above about the interesting but erratic performance of AI algorithms in our domain, this seems a very important point.
Given the advantage of logic-based approaches in terms of representational transparency, the only reason to choose an opaque approach over a logic-based approach would be if the opaque approach were dramatically superior in its capabilities. However, this currently seems not to be the case: in fact the evidence so far seems to indicate that logic-based approaches are the most powerful ones in this sort of context.
Some theorists have argued against logic-based approaches to real-world data on the grounds that there are problems with “grounding” logical symbols in real-world data (the so-called “symbol grounding problem” [Goertzel et al., 2006a]). However, these objections do not hold up to scrutiny. It is true that logic-based approaches cannot function adequately for real-world applications unless the logical symbols used are explicitly associated with observed data-patterns, but there are well-understood technologies for making such associations. Historically, many logic-based AI systems have been used in an “ungrounded” way, not containing components that directly connect the logical terms used with real-world observations – but this is a problem of poor system architecture, not a flaw of the logic-based approach in itself.
## Main High-Level Conclusions
To give a small hint at what is to come, the main conclusions at the end of our investigation are that
• the logic-based approach has the in-principle power to solve the problem of querying and analyzing very large scale spatiotemporal knowledge bases, in a manner respecting the contextual and causal knowledge contained therein
• there is a significant amount of scientific and technological knowledge in the literature regarding nearly every aspect of the application of logic-based technology to this problem
• the Achilles heel of current relevant logic-based technology is scalability
• the keys to achieving scalability in this context are conceptually understood -- adaptive inference control and attention allocation – but have not been explored nearly as thoroughly as they need to be
• it seems likely that special techniques may be useful for adaptively controlling real-world scalable inference as opposed to inference in other domains (e.g. mathematical theorem proving)
• one viable way to achieve scalable real-world reasoning may be to use the Probabilistic Logic Networks framework, perhaps within an integrative AGI design like OpenCog which provides flexible means for adaptive inference control
We thus suggest that a critical focus of research should be on the development of methods for exploiting the specific statistical structure of real spatiotemporal data, to adaptively guide logical inference methods in performing query and analytical processing.
## Summary
We now briefly review the chapters to follow, summarizing the main themes and ideas to be introduced.
### Part I: Representations and Rules for Real-World Reasoning
Part I of the book reviews a host of approaches described in the literature for representing and reasoning about real-world knowledge, including temporal, spatial, contextual and causal knowledge.
Chapter Two reviews many of the varieties of formal logic that have been developed during the last century, with a focus on those approaches that appear most relevant to the large-scale information-management problem. We begin with a basic review of predicate and term logic, and then move on to subtler variations such as modal logic (the logic of possibility) and deontic logic (the logic of obligation). We also discuss the methods that logic systems use to actually draw logical conclusions based on the information provided to them: forward chaining, in which information items are combined exploratorily to come to new conclusions; and backward chaining, in which a question is posed to the system and it then seeks to find the answer using multiple logical inference steps based on the information at its disposal.
Chapter Three considers various methods of handling uncertainty in formal logic, including fuzzy sets and logic, possibility theory, probability theory, and imprecise and indefinite probabilities. Uncertainty management is critical to our target application, because a great percentage of real-world data is uncertain, and most of the conclusions one can draw based on real-world data are also uncertain. So, logic systems that only deal with absolute truth or falsehood are not going to be very useful for our target application because a great percentage of real-world data is uncertain, and most of the conclusions one can draw based on real-world data are also uncertain. So, logic systems that only deal with absolute truth or falsehood are not going to be very useful for our target application. But, the literature contains a huge number of different methods for dealing with uncertainty – and one of our conclusions is that there isn’t necessarily a single best approach. Rather, a practical solution may integrate more than one approach, for instance using both fuzzy and probabilistic methods as appropriate. The following figures from Chapter Three illustrate several of the possible methods for representing time within logic:
Chapter Four grapples with the various ways logicians and computer scientists have devised to represent time within logic. This is a core issue for our current pursuit, because a large percentage of real-world knowledge involves time. The most standard method for handling time within logic is Allen’s interval algebra, which treats time-intervals rather than points as the atomic temporal entities, and enumerates a set of rules for combining and reasoning about time-intervals; but it suffers the deficit of being crisp rather than explicitly handling uncertainty. So we review several methods of extending interval algebra to deal with uncertainty, including methods involving fuzziness, probability, and combinations of the two. The following figure from Chapter Four illustrates the logical relationships between time intervals specified by Allen’s interval algebra:
And the following figure, also from Chapter Four, is a graphical representation of some temporal relationships between events, using a probabilistic variation of Allen’s interval algebra:
Continuing the theme of its predecessor, Chapter Five deals with temporal inference, reviewing the multiple methods presented in the literature for incorporating time into logic. These include methods that simply treat time like any other logical information, and also methods that give time a special status, including reified and modal techniques. We conclude that methods giving time a special status are likely to be dramatically more efficient, and express a particular favor for reified techniques compatible with Allen’s interval algebra (discussed above) and its variations. We give some concrete examples of temporal inference regarding peoples’ daily activities.
For instance, one of the example problems we consider involves a query regarding “which people were in the same place as Jane last week,” and a knowledge base with the following information
• Susie and Jane use the same daycare center, but Jane uses it everyday, whereas Susie only uses it when she has important meetings (otherwise she works at home with her child).
• Susie sends a message stating that Tuesday she has a big meeting with a potential funder for her business.
Given this information, inference is needed to figure out that on Tuesday Susie is likely to put her child in daycare, and hence (depending on the time of the meeting!) potentially to be at the same place as Jane sometime on Tuesday. To further estimate the probability of the two women being in the same place, one has to do inference based on the times Jane usually picks up and drops off her child, and the time Susie is likely to do so based on the time of her meeting. We show in detail how temporal inference methods can be used to carry out this commonsense inference, and other similar examples.
Chapter Six builds on the treatment of time and presents an analogous discussion of a more complex subject, space (critical to our core theme as a substantial percentage of real-world knowledge involves spatial as well as temporal information). We review the Region Connection Calculus, which models the logic of space in terms of a fixed set of logical relationships between logical terms that correspond to spatial regions. As this is a simple but limited technique, we then consider more complex approaches to representing space in logic, including directional calculus, and occupancy grids as utilized in robotics (which are extremely general yet also resource-intensive, and so should only be used when simpler methods fail). The following diagram, drawn from Chapter Six, depicts the relationships between various spatial regions and spatially distributed phenomena (NTPP stands for NonTangentialProperPart, and O stands for Overlapping; these are spatial-relationship predicates drawn from the Region Connection Calculus formalism):
Next, as well as time and space, another critical aspect of real-world reasoning is context. Nearly all real-world knowledge implicitly or explicitly gains its meaning from the specific context in which it is understood by human knowledge producers and consumers to exist. So if logical methods are to be applied effectively to real-world data, it is important that they explicitly represent contextuality. In Chapter Seven, we review a number of approaches to representing contextuality in logic, and give detailed examples of several. We also consider one example of context representation that is particularly acutely relevant to our application area: the use of contextual logic to handle user modeling. If different users of an information system have different biases and interests, then a logic based system can pay attention to this and give them different information via treating each user as a separate context and then doing contextually-biased reasoning.
In addition to context representation, Chapter Seven treats contextual inference, reviewing a number of techniques presented in the literature, and again finding favor in those methods that explicitly represent context as a special relationship within the base logic. We give a concrete examples of contextual inference applied to practical problems regarding people and their interrelationships. One example we consider involves the following assumptions:
• Alison is an accountant who is also a musician. Alison is emotional in the context of music, but not in the context of accounting. She frequently mentions Canadian place names in the context of music (maybe she's a Canadian music fan), but not in the context of accounting.
• Bob is in a similar situation, but he frequently mentions Canadian related stuff in both the music and accounting contexts.
• Clark is also in a similar situation, but he frequently mentions Canadian related stuff only in the accounting context, not the music context.
• People who have a lot to do with Canadian people, and a lot to do with money, have a chance of being involved in suspicious money-laundering activities.
We then show how contextual inference methods can be used to estimate the probability that Clark may be involved with money-laundering.
Chapter Eight turns briefly to causal reasoning, reviewing the multiple formalisms used to represent the notion of causality, and connecting causation to probabilistic and inductive reasoning.
### Part II: Acquiring, Storing and Mining Logical Knowledge
Our focus in this book is doing logical reasoning on real-world knowledge, and this is a large and critical topic – but, once one has a large store of real-world knowledge in logical format, reasoning per se is not the only thing that must be done with it. Part II, a brief interlude at the center of the book, consists of three short chapters which lightly touch three other important issues to do with large stores of logical knowledge: acquiring logical knowledge via transforming real-world data, storing and querying large volumes of logical knowledge, and mining patterns from large logical knowledge stores. Each of these topics could be a book in itself, and here we only roughly sketch the main problems involved and give some pointers into the literature.
Chapter Nine very briefly reviews existing relevant literature, discussing the use of natural language processing technology to map text and voice into sets of logical relationships; and the use of image processing and heuristic algorithms to create logical relationships out of tables, graphs and diagrams. For instance, the following diagram drawn from Chapter Six shows some logical relationships that current NLP technology can extract from the simple sentence “Gone for dinner with Bob”:
Another key question that must be answered if logic-based methods are to be applied to massive real-world data stores is: how can a huge amount of logical knowledge be stored and manipulated? This is not a question about logic per se, it’s a question about modern computer systems, database and database-like technologies, and so forth. In Chapter Ten, we review a number of current technologies, including relational databases, RDF triple-stores, object databases, and hypergraph and graph databases. Our conclusion is that at present the latter form the best option, and we give some specific examples of how to translate complex logical knowledge into the specific format required for a graph database. The following table, drawn from Chapter Ten, summarizes some of our findings in more depth:
Technology Pluses Minuses Leaders Relational DBs Mature, enterprise grade solutions Ease of integration with other systems Bad conceptual fit for graph datasets Poor scalability performance Rigid data model Oracle, IBM, Microsoft, Sybase, MySQL Object Oriented DBs Better conceptual fit than relational DBs Performance and scalability Mature, enterprise grade solution Single data model Not designed for mining and analysis Objectivity, Versant Graph DBs Flexible data model Performance and scalability Designed with mining and analysis in mind Less mature solution than relational or OO Neo, Cogito, Open Source Hypergraph DBs All the advantages of graph DBs, but with an even more flexible data model Only currently exists at the alpha stage HGDB (Open Source) RDF Triplestores Semantic web friendly Immature solutions Rigid data model Niche players and open source
Chapter Ten turns to one of the most important applications desirable to carry out on large data stores -- “data mining” (also known as “information exploitation”, “pattern discovery”, etc.). Most existing datamining techniques are either specialized for relational databases, or don’t scale beyond small knowledge stores. We review here some specific datamining algorithms in depth. One conclusion drawn is that, for datamining to really be effective in this context, it will need to be hybridized with inference. Datamining technology, in itself, will always find too many potentially interesting patterns for any human user to want to explore. So logical inference technology is needed to filter the results of datamining, either via interaction with the datamining process, or via postprocessing.
### Part III: Real World Reasoning Using Probabilistic Logic Networks
The second major of the book provides a detailed exploration of the applicability of one particular logical framework, Probabilistic Logic Networks, to real-world reasoning problems. This part is different from the previous ones, in that it comprises primarily original work, rather than literature survey and summary.
Chapters Twelve and Thirteen summarize Probabilistic Logic Networks (PLN), the particular uncertain logic system called which several of the authors (Goertzel and Pennnachin and Geisweiller) and their colleagues have developed over the last years (and published extensively on elsewhere). We outline the basic mechanisms via which PLN deals with a variety of aspects of inference, including term and predicate logic, extensional and intensional inference, and contextual, causal, spatial and temporal inference.
Chapter Fourteen turns to the specific problem of inference about changes in large knowledge bases. We consider several concrete examples including the following causal inference scenario:
• Before March 2007, Bob never had any Canadian friends except those who were also friends of his wife.
• After March 2007, Bob started acquiring Canadian friends who were not friends of his wife.
• In late 2006, Bob started collecting Pokemon cards. Most of the new Canadian friends Bob made between March 2007 and Late 2007 are associated with Pokemon cards
• In late 2006, Bob started learning French. Most of the new Canadian friends Bob made between March 2007 and Late 2007 are Quebecois.
We show in detail how a PLN inference engine, combining temporal inference with causal inference and numerous other aspects, can attempt to answer the question: What is the probable cause of Bob acquiring new Canadian friends who are not also friends of his wife?
Chapter Fourteen also considers spatial inference in the context of change analysis, giving particular attention to the incorporation of the Region Connection Calculus (RCC) into PLN. It is shown how a fuzzy/probabilistic version of RCC may be used together with a fuzzy/probabilistic version of Allen’s interval algebra to carry out commonsense inferences about the causes of peoples’ activities and relationships, based on knowledge involving time and space. To exemplify the practical use of these ideas, we extend the example of Bob and his Pokemon cards, from the previous chapter, to include the case where some of Bob’s friends live near Canada but not actually in Canada, and the inference system has to deal with the notion of “fuzzy Canadian-ness” as related to spatial geometry. The following figure illustrates the fuzzy spatial membership function corresponding to Canada, used in the example inference:
Finally (before Chapter Sixteen which is a brief conclusion), Chapter Fifteen confronts the thorny conceptual and algorithmic issue of inference control: determining which inference steps to take, in which order, in order to answer a question, filter a datamining results list, or carry out an analysis. Far from being “merely an efficiency issue,” inference control actually hits many of the deepest issues of AI, including the “frame problem” (briefly stated, that AI systems tend to lack tacit background knowledge about what questions not to bother asking because their answers are supposed to be obvious, or are irrelevant). We discuss a number of specific techniques that may be able to achieve effective inference control in the context of inference on large stores of spatiotemporal logical knowledge, including techniques that hybridize logic with other AI methods such as activation spreading. Here the discussion broadens from logic per se to the topic of “cognitive architectures” and general AI systems, the point being made that the integrative architectures underlying many such systems exist largely in order to provide effective, scalable inference control. As an example, the OpenCog cognitive architecture in which the PLN inference system is embedded is briefly considered.
'Part I: Representations and Rules for Real-World Reasoning'
# 2 Knowledge Representation Using Formal Logic
Now we begin to dig into the nitty-gritty of our subject matter. Before discussing querying and analysis of complex, heterogeneous spatiotemporal and contextual knowledge, we must discuss representation of temporal knowledge (as well as, to a certain extent, spatial knowledge)… and before that, we must address knowledge representation in general.
In the course of our investigation we must work through a number of difficult questions regarding knowledge representation, including:
• Which of the many species of uncertain logic to use as the basis for our knowledge representation
• How specifically to represent temporal knowledge?
• How specifically to represent spatial knowledge?
• What is the best low-level (e.g. graph) representation of logical knowledge for efficient storage and processing?
“Logic” itself is not a monolithic entity; it comes in many different flavors. At the highest level, there is the dichotomy between predicate logic and term logic (and there are also systems that hybridize the two, e.g. [Goertzel et al., 2008; Wang, 2006a]). There are also many types of logical system within each of these broad categories, some of which will be reviewed later on.
The material in this chapter becomes somewhat formal and technical, for which we apologize to the reader who lacks the relevant taste or experience; but which unfortunately seems unavoidable if we are to give a serious treatment of our topic. The reader lacking appropriate expertise may either consult relevant background material [Copi & Cohen, 1998], or less ideally, skim this material and proceed to the later chapters, some of which will be quite clearly comprehensible without grasp of these preliminaries, some less so.
## Basic Concepts of Term and Predicate Logic
Term logic, or traditional logic, was founded by Aristotle and was the dominating logical framework until the late nineteen century. Term logic uses subject-predicate statements of the form “S is P” (for instance, “Socrates is a man”). There are singular and universal terms (the former correspond to unique subjects). There are just four forms of propositions in term logic:
• Universal and affirmative (e.g. "All men are mortal")
• Particular and affirmative (e.g. "Some men are philosophers")
• Universal and negative (e.g. "No philosophers are rich")
• Particular and negative (e.g. "Some men are not philosophers").
New conclusions are derived from premises by syllogisms. Aristotle introduced fourteen syllogisms, of which we will give just two here for illustrative purposes:
• (Barbara) If every M is L, and if every S is M, then every S is L. (for instance, “if every man is mortal, and if every philosopher is a man, then every philosopher is mortal”)
• (Celarent) If no M is L, and if every S is M, then no S is L. (for instance, “if no philosopher is rich and if every poet is a philosopher, then no poet is rich”).
Syllogisms provide a method for deduction – deriving new facts from already proved facts.
In addition there are rules for induction and abduction:
• (Induction) If every M is L, and if every M is S, then every S is L. (for instance, “if every poet is mortal, and if every poet is a philosopher, then every philosopher is mortal”)
• (Abduction) If every L is M, and if every S is M, then every S is L. (for instance, “if every poet is mortal, and if every philosopher is mortal, then every philosopher is poet”)
Notice that the induction and abduction rules do not neccesarily derive true statements. Nevertheless these are important forms of inference in the face of insufficient evidence, in modern AI reasoning systems as well as in classical Aristotelian term logic [Dimopoulos & Kakas, 1996]. Induction and abduction are omnipresent in human commonsense inference.
Put simply, induction aims at generalization. In the above example (“if every poet is mortal, and if every poet is a philosopher, then every philosopher is mortal”), the first premise yields that all philosophers that are also poets are mortal, but then it is generalized to conclude that all philosophers are mortal. Yet, it is possible that there are some philosophers that are not poets, so potentially not mortal, so the above generalization rule does not neccesarily lead to true conclusions.
Similarly, abduction aims at explanation. In the above example, the explanation for the fact that “every philosopher is mortal” may be that it is because “every philosopher is a poet”.
In the late nineteenth century, classical term logic was the subject of criticism, for its weak expressive power and the limited forms of reasoning it permitted. For example, in classical term logic from "every car is a vehicle" one cannot infer "every owner of a car is an owner of a vehicle." In that period, predicate logic was designed, and it still serves as a basis for most mathematical and philosophical formal reasoning. However, modern theorists have extended classical term logic in various ways [Englebretsen & Sommers, 2000; Wang, 2006b], so that there are now term logics which equal predicate logic in expressive power. There are also systems that hybridize term and predicate logic, such as our own Probabilistic Logic Networks framework [Goertzel et al,, 2008], which will be discussed below. Advocates of term logic often argue that it more closely matches the patterns of human commonsense reasoning.
In standard predicate, or first-order logic, statements have arbitrary propositional form (involving conjunctions, disjunctions, negations, …) and arbitrary use of quantifiers (for instance, “for every man, there is a woman, such that for every man, …”). Modern variants of term logic provide this same expressive flexibility.
Pure predicate logic is a framework in which one can describe other theories. This framework is defined by the set of axioms and the set of inference rules (such as “if P and if P yields Q, then Q”). The proofs are sequences of derivation steps based on these axioms and rules. For example, one can represent in predicate logic and derive “not every man is a philosopher if and only if there is a man such that it is not a philosopher”. For first-order logic there are also inductive and abductive rules, not used in mathematical theorem-proving, but for uncertain reasoning, most often in AI. First order logic is also used as a basis for many specific logics, including modal, deontic, temporal and spatial logics as will be discussed below.
## Review of Propositional Logic
In order to explain predicate logic in more depth, we must begin with a simpler variant called “propositional logic.” Propositional logic can express simple facts while first-order logic (or predicate logic) also involves quantification and more complex statements. In this sense, first-order logic subsumes propositional logic. Both propositional and first-order logic have many practical applications beyond the ones considered here, in describing different processes and concepts. Most important perhaps are applications in computer science, ranging from chip design [Aehlig & Beckmann, 2007] to natural language processing [Meulen, 2001].
Both propositional logic and first-order logic have three important aspects:
• syntax – describing well-formed formulae and their basic properties;
• semantics – describing meaning of well-formed formulae;
• deduction – describing systems for syntactically deriving new formulas from other formulas (with no respect to their meaning).
We now give a brief mathematical exposition of these three aspects.
First, syntax. Let P be an infinite, but countable set, whose elements will be called propositional letters or atomic propositions. The set of propositional formulas is defined by the following rules:
• all elements of P are propositional formulas;
• ⊥ and ⊤ are propositional formulas (which as we will see below, are normally taken to semantically represent False and True)
• if A is a propositional formula, then so is ¬A (which is normally taken to represent the negation of A)
• if A and B are propositional formulas, then so are A∧B , A∨B , A⇒B , A⇔B (which are normally taken to represent And, Or, and implication and equivalence)
• each propositional formulas is obtained by a finite number of applications of the above rules.
Next, semantics. A valuation v is defined as a mapping from P to the set {0,1}. That is, a valuation assigns either 0 or 1 to any propositional letter. An interpretation Iv is an extension of a valuation v, mapping propositional formulas to the set {0,1}, defined in the following way:
• Iv (⊥)=0;
• Iv (⊤)=1;
• Iv (p)=v(p), if p belongs to P;
• Iv (¬A)=1- Iv (A);
• Iv (A∧B)=min(Iv (A), Iv (B));
• Iv (A∨B)=max(Iv (A), Iv (B));
• Iv (A⇒B)=max(1-Iv (A), Iv (B));
• Iv (A⇔B)=1, If Iv (A)= Iv(B) and Iv (A⇔B)=0 otherwise.
The above semantics is referred as to Tarski’s semantics (that he introduced in the 1930’s) [Tarski 1994]. Put simply, this allows the interpretation of propositional formulas like
A∧(¬B∨(C ⇒ A))
as having truth values drawn from the set {0,1} (given the truth values for A, B and C), with 0 usually interpreted as meaning False and 1 as meaning True.
A formula A is satisfiable if there is a valuation v such that Iv (A)=1 (otherwise, it is unsatisfiable or inconsistent, aka self-contradictory). A formula A is a tautology if for an arbitrary valuation v Iv (A)=1. If a formula A is a tautology, then we denote that by ⊨ A.
For example, it holds that Iv(p∨q)=1 in a valuation v such that v(p)=1, v(q)=0 – so the formula p∨q is satisfiable. On the other hand, the formula p∨¬p is tautology.
The problem of checking the satisfiability of a formula made of conjunctions of disjunctions of literals (variables or negations of variables) is decidable (there is an effective algorithm for solving it) and is called SAT. SAT is on of the most important NP-complete problems. Programs for testing satisfiability are called SAT solvers. There are different methods for checking satisfiability of a formula: the simplest is based on truth-tables (i.e. tabular enumeration of all possible combinations of values for the formula’s variables, and evaluation of the truth value of the formula for each combination). Other include Davis-Putnam-Logmann-Loveland (DPLL) procedure, and modern solvers – DPLL-based, resolution-based solvers, tableaux-based solvers etc. Modern SAT solvers can decide propositional formulas with thousands variables and clauses [Lynce & Marques-Silva, 2002].
### Deduction in Propositional Logic
There is a number of inference systems for propositional logic. Most of them are actually restrictions of inference systems for first-order logic. Hilbert-style inference system consists of the following axiom schemes [Mendelson, 1997]:
(A1) A ⇒ (B ⇒ A)
(A2) (A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))
(A3) (¬B ⇒¬A) ⇒ ((¬B ⇒ A) ⇒ B)
and the inference rule modus ponens: A, A⇒B ⊢ B.
A proof or a derivation in a Hilbert system is a finite sequence of formulas such that each element is either an axiom or follows from earlier formulas by the rule of inference. A proof of a derivation from a set S of formulas is a finite sequence of formulas such that each term is either an axiom, or is a member of S, or follows from earlier formulas by the rule of inference.
If there is a proof for A, then A is a theorem and we denote that by ⊢ A. For example, it can be proved that A ⇒ A is a theorem, as follows:
1. (A ⇒ ((A ⇒ A) ⇒ A)) ⇒ ((A ⇒ (A ⇒ A)) ⇒ (A ⇒ A)) (instance of A2)
2. A ⇒ ((A ⇒ A) ⇒ A) (instance of A1)
3. (A ⇒ (A ⇒ A)) ⇒ (A ⇒ A) (from 1 and 2, by MP)
4. A ⇒ (A ⇒ A) (instance of A1)
5. A ⇒ A (from 3 and 4, by MP)
This may seem like a lot of work to prove that “A implies A”, but that’s the nature of formal logic systems! Derivations are broken down into extremely small steps that are rigorously mathematically justified. In human commonsense inference we tend to proceed in large leaps instead, at least on the conscious level – but unconsciously, our brains are carrying out multitudes of small steps, though the analogy between these small steps and the small steps in logical proofs is a subject of debate in the AI and cognitive science community.
There is a link between the semantics of propositional logic and the above Hilbert-style system stating that the system is sound and complete: every theorem is tautology, and every tautology is theorem, i.e., ⊨ A if and only if ⊢ A.
While the above is a standard and workable, approach, there are also other inference systems for propositional logic, including the one obtained as restrictions of Gentzen natural deduction and sequent calculus (see the section on first-order logic). There are also variants for classical and for intuitionistic propositional logic: in the former A∨¬A is a theorem, and in the latter it is not (the above Hilbert-style system is classical).
## Review of Predicate Logic
Standard, first-order predicate logic builds on propositional logic as defined above. We will review it using the same categories of syntax, semantics and deduction.
Firstly, syntax. Let Σ be a finite or a countable set, its elements will be called function symbols. Let Π be a finite or a countable set, its elements will be called predicate symbols. Let arity be a function that maps elements of Σ and Π to natural numbers. The triple (Σ,Π,arity) is called a signature. The set of terms over a signature (Σ,Π,arity) and a countable set of variables V is defined in the following way:
• all elements of V are terms;
• if f is a function symbol and arity(f)=0, and then f is a term;
• if f is a function symbol and arity(f)=n, and if t1 , …, tn are terms, then f(t1 , …, tn) is a term.
• each term is obtained by a finite number of applications of the above rules.
The set of atomic formulas over a signature (Σ,Π,arity) and a countable set of variables V is defined in the following way:
• ⊥ and ⊤ are atomic formulas;
• if p is a predicate symbol and arity(p)=n, and if t1 , t2 , …, tn are terms, then p(t1 , …, tn) is an atomic formula.
• each atomic formula is obtained by a finite number of applications of the above rules.
The set of formulas over a signature (Σ,Π,arity) and a countable set of variables V is defined in the following way:
• each atomic formulas is a formula;
• if A is a formula, then so is ¬A;
• if A and B are formulas, then so are A∧B , A∨B , A⇒B , A⇔B;
• if A is a formula and v is a variable, then (∀x)A and (∃x)A are formulas;
• each formula is obtained by a finite number of applications of the above rules.
Next, semantics. The meaning of a formula is defined with respect to a pair (D,I), where D is a non-empty set, called the domain, and I is a mapping such that:
• To each function symbol of arity 0, I associates an element c from D;
• To each function symbol of arity n>0, I associates a total function fI from Dn to D;
• To each predicate symbol of arity n>0, I associates a total function pI from Dn to {0,1}.
A valuation v, in this context, is defined as a mapping from the set of variable V to the set D. An interpretation Iv of terms, with respect to a pair (D,I) and a valuation v is defined in the following way:
• Iv(t)=v(t), if t is an element V;
• Iv(t)=I(t), if t is a function symbol and arity(t)=0;
• Iv(f(t1 , …, tn))= fI ( Iv(t1), …, Iv(tn))) where fI=I(f).
An interpretation Iv of formulas, with respect to a pair (D,I) and a valuation v is defined in the following way:
• Iv(⊥)=0 and Iv(⊤)=1
• Iv(p(t1 , …, tn))= pI ( I(t1), …, I(tn))) where pI=I(p).
• Iv (¬A)=1- Iv (A);
• Iv (A∧B)=min(Iv (A), Iv (B));
• Iv (A∨B)=max(Iv (A), Iv (B));
• Iv (A⇒B)=max(1-Iv (A), Iv (B));
• Iv (A⇔B)=1, If Iv (A)= Iv(B) and Iv (A⇔B)=0 otherwise.
• Iv((∀x)A)=1, if for every valuation w that is identical to v, with a possible exception of x, it holds Iw(A)=1. Otherwise, Iv((∀x)A)=0;
• Iv((∃x)A)=1, if there is a valuation w that is identical to v, with a possible exception of x, such that it holds Iw(A)=1. Otherwise, Iv((∃x)A)=0.
If there is a pair (D,I) and a valuation v, such that Iv(A)=1, then A is satisfiable, otherwise it is unsatisfiable, or inconsistent. If for a fixed pair (D,I) it holds that Iv(A)=1 for arbitrary valuation v, then A is valid with respect to (D,I). If it holds that Iv(A)=1 for arbitrary pair (D,I) and for arbitrary valuation v, then A is valid and we denote that by ⊨ A.
For instance, if the domain D is the set of natural numbers and if I maps p to the relation ≤, then Iv((∀x)p(x,x))=1 in every valuation v, hence the formula (∀x)p(x,x) is valid with respect to (D,I).
### 2.3.1 Deduction in First-Order Logic
Deduction in first-order logic is similar conceptually to its analogue in propositional logic, but more complex in detail due to the presence of quantified variables. There are several different deductive systems available; one of the first was developed by Hilbert in the early 20th century and we will describe it now. In Hilbert’s systems, formulas are built using only the connectives ⇒ and ¬, and the quantifiers ∀ (“for all”) and ∃ (“there exists”). The system consists of the following axiom schemes:
(A1) A ⇒ (B ⇒ A)
(A2) (A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))
(A3) (¬B ⇒¬A) ⇒ ((¬B ⇒ A) ⇒ B)
(A4) (∀x)A ⇒ A[x → t], while the term t is free for x in A
(A5) (∀x)(A ⇒ B) ⇒ (A ⇒ (∀x)B), while A does not involve free occurrences of x
and the following inference rules:
Modus ponens: A, A⇒B ⊢ B
Gen: A ⊢ (∀x)A
A proof or a derivation in a Hilbert system is a finite sequence of formulas such that each element is either an axiom or follows from earlier formulas by one of the rules of inference. A proof of a derivation from a set S of formulas is a finite sequence of formulas such that each formula is either an axiom, or is a member of S, or follows from earlier formulas by one of the rules of inference. If there is a proof for A, then A is a theorem and we denote that by ⊢ A.
There is a link between the semantics of first order logic and the above Hilbert-style system stating that the system is sound and complete: every theorem is valid formula, and every valid formula is theorem, i.e., ⊨ A if and only if ⊢ A.
In Hilbert-style systems, even for trivial statements, proofs can be non-trivial and rather unintuitive. However, although this kind of system is very demanding for practical use, it is very suitable for formal analyses of formal logic (since it has just a few axioms and inference rules). On the other hand, Gentzen constructed (in the 1930’s) a “natural deduction” system, that better reflects usual mathematical reasoning. The price for this increased naturalness is a larger set of inference rules – one for eliminating and one for introducing each logical connective (¬, ∧, ∨, ⇒,⇔) and quantifier (∀, ∃), and one for eliminating the logical constant ⊥ (13 rules altogether). On the other hand, there is just one axiom scheme (A∨¬A), which is not even needed if one adopts the intuitionistic version of Gentzen’s logic.
Some of the rules in Gentzen’s deductive system are:
$\displaystyle \frac{A}{A\vee\,B}$ (introducing $\displaystyle \vee\,$ )
$\displaystyle \frac{\neg\,A}{\bot\,}$ ( eliminating $\displaystyle \neg\,$ )
$\displaystyle \frac{A\Rightarrow\,B} {b}$ (eliminating $\displaystyle \Rightarrow\,$ )
and so forth.
Proofs in Gentzen’s natural deduction are usually represented as trees with the statement to be proved in the root (at the bottom), and with axioms or assumptions in leaves (all these assumptions have to be eliminated along the proof). Somewhat different in nature from usual mathematical proofs, is the Gentzen’s sequent calculus, suitable for formal analyses and for automation. The elementary object in this system is a sequent, a construct of the form A1, A2, …, An ⊢ B1, B2, …, Bm. The calculus itself consists of the inference rules for introducing logical connectives and quantifiers on both sides of sequents, for instance:
$\displaystyle \frac{\Gamma\,\vdash\,\Theta\,,A} {\neg\,A,\Gamma\,\vdash\,\Theta\,}$ (introducing $\displaystyle \neg\,$ left)
There are also structural rules for dealing with formulas within one side of a sequent, for instance:
$\displaystyle \frac{\Gamma\,\vdash\,\Theta\,} {D,\Gamma\,\vdash\,\Theta\,}$ (weakening)
There are many variations of the above inference systems, both for first order logic and for other theories. The above systems can also be used as inference systems for propositional logic – it suffices to omit all axioms and rules involving quantifiers.
Finally, it is worth noting that first-order logic is semi-decidable, which means that there can be an algorithm that can always confirm a valid formula is indeed valid (i.e, is a theorem), but cannot always detect that a non-valid formula is not valid. Things are simpler in propositional logic which is decidable, meaning that there is an algorithm which can always decide whether a given propositional formula is valid or not.
### First-order Theories
Next, one can extend basic predicate logic by defining “theories” within it, which extend the basic axioms by adding other specialized axioms. We will be doing a lot of this in the present book, in the context of specialized theories about time and space.
Mathematically, we say that each such theory has a certain “signature,” consisting of specific sets Σ and Π of function and predicate symbol (with certain arities) extending the basic ones used in predicate logic. Beside the new symbols, to create a theory one has to provide a list of axioms (in the described language). These axioms are then used within a selected deductive framework (e.g., Hilbert’s system, Gentzen’s natural deduction, etc.) as additional axioms of the system.
If a theory T is defined by a signature (Σ, Π,arity), and within a deductive system, then we often write T ⊢ F or ⊢T F to denote that the formula F can be derived in the theory T (i.e., F is a theorem of T).
As an example within mathematics, the branch of math called “group theory” can be constructed easily as an extension of “pure” predicate logic. In this formalization, the signature of theory of group consists of:
• Functional symbol 0 (of arity 0);
• Functional symbol + (of arity 2);
• Functional symbol - (of arity 1);
• Predicate symbol = (of arity 2),
And, the axioms of the theory of groups are:
(∀x)(x = x)
(∀x)(∀y) (x = y ⇒ -x = -y)
(∀x1)(∀x2)(∀y1)(∀y2)(x1 = y1 ∧ x2 = y2 ⇒ x1+x2 = y1+y2)
(∀x1)(∀x2)(∀y1)(∀y2)(x1 = y1 ∧ x2 = y2 ⇒ (x1=x2 ⇒ y1=y2))
(∀x)(∀y)(∀z)(x + (y + z) = (x + y) + z)
(∀x)(x+0 = 0+x = x)
(∀x)(x+(-x) = (-x)+x = 0)
These axioms, added on to the regular axioms of predicate logic, tell you how to use the entities {0, + , - , =} in the manner required in group theory. For instance, it can be proved that the following formula is theorem of the theory of groups:
(∀x)(∀y)(∀z)(x + z = y + z ⇒ x=y)
What we’ll see later on in this book are similar theories that involve, not {0, + , - , =} but rather entities such as time intervals, spatial regions, and relationships between them.
Finally, although we won’t make use of this here, it’s worth noting that, apart from the axiomatic approach to defining theories, theories can be also defined semantically. For a given signature L and a corresponding pair (D,I), a theory of a structure is the set of all sentences over L that are true in (D,I).
### Forward and Backward Chaining
There are two basic search strategies for using inference rules in theorem-proving, and other related AI areas:
• Forward chaining
• Backward chaining
Both strategies are applied to tasks of the same sort: given a set of facts (axioms) and a set of inference rules, the task is to check whether some given fact (formula) F can be derived. The difference between the two strategies is the direction of their search. Namely, forward chaining (also known as data-driven search) starts with the available facts and derive all facts that can be derived, until the given fact F is reached. On the other hand, backward chaining (also known as goal-driven search) starts with the given goal F and apply inference rules in opposite direction, producing new subgoals and trying to reach the facts already available.
Let us consider the following example. Let there be given facts:
1. Derek will go out for lunch.
2. If Alison goes out for lunch, then Bob will go out for lunch as well.
3. If Bob goes out for lunch, then Clark will go out for lunch as well.
4. If Derek goes out for lunch, then Ellen will go out for lunch as well.
5. If Ellen goes out for lunch, then Clark will go out for lunch as well.
and assume the (only) inference rule is modus ponens: if X and X⇒Y, then Y. The goal to be proved is “Clark will go out for lunch”.
Forward chaining will try to match any two facts with X and X⇒Y in order to apply modus ponens and to derive new facts. One (and the only at this step) option is to use the facts 1 and 4 and derive the fact “Ellen will go out for lunch”. Then, in next step, this new fact and the fact 5 yield “Clark will go out for lunch”, which was required.
Backward chaining starts from the goal – from the fact “Clark will go out for lunch”. If it is matched with Y in modus ponens, then the new subgoals will be X and X⇒Y, where X can be anything. The subgoal X⇒Y matches the fact 3, so X is matched with “Bob will go out for lunch”. When proving this subgoal, the fact 2 is used and a new subgoal “Alison will go for lunch” should be proved. However, this leads to failure. Another option is the following: the subgoal X⇒Y matches also the fact 5, so X is matched with “Ellen will go out for lunch”. When proving this subgoal, the fact 4 is used and a new subgoal “Derek will go out for lunch” should be proved. This is trivially true, by the fact 1. So, it was proved that “Clark will go out for lunch”, as required.
The above description of forward and backward chaining is simplified. Typical applications of forward and backward chaining involve different methods of directing the search. Namely, the main problem that both strategies face is search expense, due to typically huge numbers of combinatorial options that have to be considered. Neither of these approaches is superior to the other one, and there are domains in which one if more appropriate than another. There are also hybrid approaches that combine forward and backward search.
### Decidability and Decision Procedures
In this section we briefly introduce a distinction that may be important in practical large-scale logic-based systems: the distinction between proof procedures and decision procedures. Put simply:
• a proof procedure finds a specific series of steps for getting to a certain conclusion from the axioms of a logical theory
• a decision procedure checks whether a certain conclusion can be obtained from the axioms of a logical theory, without necessarily directly supplying the proof (the series of steps)
In practical cases, given the outcome of a decision procedure, plus knowledge of the algorithm used to carry out the decision procedure, it is in principle possible to construct a proof. But this may be quite laborious. In some situations, if one just needs to know whether something is true or not in a certain formal theory, it may be easier to use a decision procedure than to find a specific proof.
In the formal lingo of logic, one says a theory T is decidable if there is an algorithm that for any sentence F of T it can decide whether F is theorem of T or not. Such an algorithm is called a decision procedure. An example of a decision procedure for propositional logic is the DPLL (Davis-Putnam-Logemann-Loveland) procedure, which lies at the core of all modern SAT solvers. A SAT solver is a program that checks if a large propositional-logic formula can possibly be satisfied by any assignment of values to variables or not; and doing this via a decision procedure rather than a proof procedure is vastly more computationally efficient. SAT solvers are used in a huge variety of practical applications nowadays, including circuit analysis, natural language parsing, and all manner of large-scale discrete optimization problems. If one had to approach these problems using direct theorem proving in propositional logic, then one would run into terrible combinatorial explosions and quite possibly formal logic would need to be set aside in favor of some different analytical approach.
There are many widely applied decision procedures for first-order theories, including decision procedures for linear arithmetic over reals (involving only addition, not multiplication), multiplicative arithmetic over reals (involving only multiplication, not addition), theory of lists, theory of arrays, etc. [Barrett, 2009]. There is a family of modern decision procedures for first-order theories, heavily using propositional reasoning and SMT (satisfiability modulo theory) techniques. Decision procedures and SMT solvers are widely applied in software and hardware verification.
Mathematically speaking, there are many important theories that are not decidable, such as arithmetic over natural numbers, the theory of groups, etc. Hence, for these theories there can be no decision procedures, only heuristics that can prove/disprove certain classes of formulas. However, undecidability always involves infinity in some form; if one restricts attention to a finite set of data items (as is always the case in practical applications), then it is not an issue.
In practical applications, the use of decision procedures is sometimes unavoidable. However, developing an efficient decision procedure is not an easy task – it requires specific knowledge about the theory. There is no generic (decidable) method for constructing efficient decision procedures. In the case of temporal and spatial logics such as we will discuss here, scientists have not yet created appropriate specialized decision procedures, but this seems a highly worthy area of investigation.
## Simple Examples of Formal Logical Inference
We’ve been discussing logic in an extremely abstract and mathematical way – but our main goal here is real-world reasoning, not mathematical reasoning. So before progressing further in the direction of elaborating abstract logic systems, we will digress a bit to clarify the relationship between logic and commonsense inference. This is after all where logic started: formal logic originated, not as a branch of math, but from the motivation of formalizing everyday human thinking. Many real world problems can be formulated in terms of propositional or first order logic. It is only during the last century and a half that logic has become a sophisticated branch of mathematics.
### Example of Kings and Queens
Then the orders can be formulated as follows:
• King’s order: p∨q∨r
• Queen’s order: (q∧r) ⇒ p
• Prince’s order: r ⇒ p∨q
Or, equivalently:
• King’s order: p∨q∨r
• Queen’s order: ¬q∨¬r∨p
• Prince’s order: ¬r∨p∨q
To solve a problem, the head of protocol has to check whether the formula
(p∨q∨r)∧(¬q∨¬r∨p)∧(¬r∨p∨q)
is satisfiable. He can use a SAT solver for that (and he can find that he can meet all orders by inviting only the ambassador of Peru). Or, as this is a simple case, using any reasonable propositional logic theorem prover would also work fine.
### Example of Minesweeper
Or consider, as an another example, the popular computer game of Minesweeper, as depicted in Figure 2.1:
Figure 2.1: A screenshot of a Minesweeper GUI
Let us suppose that we open the position p, and there is no mine. Let us also suppose that we got the number 1 for the position p. If we open a position q and we get the answer 1 again, we can safely open the positions r and u. The explanation is as follows. We associate a propositional variable to each position – the variable is true if there is a mine on the position, and the variable is false otherwise. Since there is no mine on the position p, it follows that:
¬p
Since the position p has the associated value 1, there is one mine on either q, s, or t. We can encode this as follows:
q∨s∨t
q ⇒ ¬s∧ ¬t
s ⇒ ¬q ∧ ¬t
t ⇒ ¬q ∧ ¬s
Since there is no mine on the position q, it follows that:
¬q
Since the position q has the value 1, it means that there is one mine on either p, s, t, r, or u. We can encode this as follows:
p∨s∨t∨r∨u
p ⇒ ¬s∧ ¬t ∧¬r ∧¬u
s ⇒ ¬p∧ ¬t ∧ ¬r ∧¬u
t ⇒ ¬p∧ ¬s ∧ ¬r ∧¬u
r ⇒ ¬p∧ ¬s ∧ ¬t ∧¬u
u ⇒ ¬p∧ ¬s ∧ ¬t ∧ ¬r
The question is whether one can safely open the position r, i.e., whether r is consistent with the above formulas (i.e., is it possible that there is a mine on r). A SAT solver (applied to the above formulas) would easily find that such a set of formulas is not consistent, so there cannot be a mine on the position r (and, similarly, there cannot be a mine on the position u).
### Example of Socrates
The above examples all involve propositional rather than predicate logic. Concerning first-order logic, consider the following classical example. One is given the following two facts:
• Socrates is a man.
• All men are mortal.
We can encode this in first-order logic as follows:
• man(Socrates)
• (∀x)(man(x) ⇒ mortal(x))
Applying deduction allows us to conclude that, since Socrates is a man, he must be mortal. For instance, a very simple proof procedure can verify that mortal(Socrates) is a consequence of the above facts by verifying that ¬mortal(Socrates) is inconsistent with the above formulas.
## Modal logic
Propositional logic and first-order logic are sometimes not expressive enough for some sorts of common-sense reasoning. For instance, one may wish to extend them using so-called modal statements including qualifiers like necessary, possibly, future, past, etc. For such cases, specific logics are used, such as modal and temporal logics. These logics are not contradictory to first-order logic, and in fact could be viewed as specialized theories constructed within first-order logic; but this is not always the most useful way to look at them. It is often more helpful to think about them as alternative formalizations of logic.
Modal logic describes logical relations of modal notions. These notions may include metaphysical modalities (necessities, possibilities, etc.), epistemic modalities (knowledge, belief, etc.), temporal modalities (future, past, etc.), and deontic modalities (obligation, permission, etc.). Modal logics are widely used in philosophy, artificial intelligence, database theory, and game theory. Modern work on modal logics started in 1918. with the monograph A Survey of Symbolic Logic by C. I. Lewis.
The main metaphysical modal notions are necessity and possibility; and the modal logic that describe logical relations over them is called alethic modal logic. In alethic modal logic, one can express a statement like “It is possible that Bob quit his job”. The modailities are represented by modal operators. Formulas are built using propositional connectives and these modal operators. The basic unary modal operators are usually written $\displaystyle \Box$ (or L) for necessarily and $\displaystyle \Diamond\$ (or M) for possibly. These two operators are linked in the following way:
$\displaystyle \Diamond\ P \Leftrightarrow \neg \Box \neg P$
$\displaystyle \Box P \Leftrightarrow \neg \Diamond \neg P$
There is a number of modal extensions of some underlying logics, including the extension of first-order logic, called modal predicate logic.
The standard semantics of modal logics is Kripke semantics. The concept of the Kripke semantics of propositional modal logic includes:
• A non-empty set W – a set of possible worlds;
• A two-place relation R on elements from W– the accessibility relation between worlds, which represents the possible worlds that are considered in a given world, i.e., if we consider a world w0, every world v such that it is in relation R with w0 represents a possibility that is considered at a world w0;
• A frame – a tuple (W, R);
Given a frame (W, R), a model M is a tuple (W, R, V) where V is a map that assigns to a world a valuation on propositional variables, i.e. for a given world w, V(w) is a function from the set of propositional variables to {0, 1}. Interpretation of a formula F with respect to M is defined in the following way (the fact that F is true at a world w in a model M” is denoted by M,w F.):
M,w p iff V(w)(p)=1 (where p is a propositional variable)
M,w ⊨ F ∧ F’ iff M,w ⊨ F and M,w ⊨ F’.
M,w ⊨ ¬F iff not M,w ⊨ F
M,w ⊨□F iff, for every world w’ such that it is in relation R with W it holds that, M,w F.
The semantics of other propositional connectives and the operator ◊ are implied by the above definition. A formula is then defined to be valid in a model M if it is true at every possible world in M. A formula is valid if it is valid in all frames (or every model).
For the given semantics, there is a sound and complete inference system for propositional modal logic – the system K. In addition to underlying propositonal axioms and inference rules, there is the axiom
$\displaystyle \Box (F \Rightarrow F') \Rightarrow (\Box F \Rightarrow \Box F')$
and the rule
if $\displaystyle \vdash F, then \vdash \Box F$
There are various inference systems for propositonal modal logic obtained by adding extra axioms to K.
There are also fuzzy versions of modal logics [Thiele, 1993, Ying, 1988].
## Deontic logic
An interesting specialization of modal logic, which is potentially useful for doing logical inference about human behaviors and motivations, is deontic logic -- which is concerned with the ideal and actual behavior of social agents, and involves notions like permissible, impermissible, obligatory, gratuitous, optional, etc. Deontic logic has many analogies with alethic modal logic. In addition to theoretical interest, deontic logics are used for formalizing different real-world concepts and problems such as morality, normative law, legal analysis, social and business organizations and security systems, computer security, electronic commerce, or legal expert systems.
A survey of applications of deontic logic in computer science can be found in [Wieringa & Meyer, 1993], which supplies the following systematization of applications of deontic logic in computer science:
1. Fault-tolerant computer systems.
2. Normative user behavior.
3. Normative behavior in or of the organization.
1. Policy specification.
2. Normative organization behavior (e.g. contracting).
4. Normative behavior of the object system.
1. The specification of law.
2. The specification of legal thinking.
3. The specification of normative rules as deontic integrity constraints.
4. Other applications, not discussed above.
The first formalization of deontic logic was given by E. Mally in 1926. More details on deontic logic can be found in [McNamara & Prakken, 1999].
In the “Traditional scheme” for deontic logic, there are five normative statuses considered:
• it is obligatory that (OB)
• it is permissible that (PE)
• it is impermissible that (IM)
• it is gratuitous that (GR)
• it is optional that (OP)
The first one of the above can be used as a basis, while the remaining ones can be defined in the following way:
PEp ⇔ ¬OB¬p
IMp ⇔ OB¬p
GRp ⇔ ¬OBp
OPp ⇔ (¬OBp ∧ ¬OB¬p)
Standard Deontic Logic (SDL) is the most studied deontic logic. It extends propositional logic by the (one-place) OB deontic operator. Formulas are built in the standard modal-logic way. The semantics of SDL is usually given in Kripke-style. The inference system for SDL consists of axioms for (classical) propositional calculus and inference rules, the additional inference rule “if $\displaystyle \vdash\$ P then $\displaystyle \vdash\$ OBp” and the following axioms:
OB(p ⇒q) ⇒ (OBp ⇒OBq)
OBp ⇒¬OB¬p
Consider the following simple example. Let us assume that the hypotheses are:
It ought to be the case that Alison does the paperwork.
If Alison does the paperwork, then Alison leaves the office late.
Let us denote Alison does the paperwork by p and Alison leaves the office late by q:
Then, we can prove It ought to be the case that Alison leaves the office late (i.e., OBq) as follows:
(C1) OBp (hypothesis)
(C2) p⇒q (hypothesis)
(C3) OB(p ⇒q) (deontic inference rule, from (C2))
(C4) OB(p ⇒q) ⇒ (OBp ⇒OBq) (deontic axiom)
(C5) OBp ⇒OBq (Modus ponens, from (C3) and (C4))
(C6) OBq (Modus ponens, from (C1) and (C5))
There is a number of variants of the SDL inference system and there are interesting logical and philosophical considerations for each of them.
### Fuzzy deontic logic
While there is a number of approaches to fuzzy modal logic (see the next chapter for a recall of fuzzy logic), there is a very limited literature on fuzzy deontic logic. One version of fuzzy deontic logic was introduced and discussed by [Gounder & Esterline, 1998]. In their framework, given the statements
q = Person p is 18 or older.
r = Person p is an employee of company c.
s = Person p is over 80 years old.
t = Person p is under 20 years old.
u = Company c gives its employees a bonus.
v = The employees of company c arrive at work not more than ten minutes late.
one can consider the following interesting deontic statements:
OB(p ⇒ q)
OBp ⇒ OBq
r ⇒ OB(¬s ∧ ¬t)
u ⇒ OBv
These statements need not be crisp, but can be in a permissible range which is given in the fuzzy truth value in the interval [0, 1] and one obligation may lead to another obligation.
Concerning reasoning in this theory, as said in [Gounder & Esterline, 1998], since fuzzy logics work with numerical measures, axiomatic systems are not appropriate. Instead, fuzzy versions of the semantic properties exist and can be shown to correspond to some of the axioms for the crisp systems in special ways that support dependency among assertions in a modal domain.
## The frame problem
A final issue that must be discussed, in the context of knowledge representation using formal logic, is the “frame problem,” as originally recognized and named in [McCarthy & Hayes, 1969]. Put most simply, this is the problem of representing the effects of action without having to represent explicitly a large number of intuitively obvious non-effects (i.e., properties that are not affected by the action). The frame problem also has a wider epistemological importance and it considers whether it is possible, in principle, to limit the scope of the reasoning required to derive the consequences of an action. The name "frame problem" was derived from a common technique used in producing animated cartoons where the currently moving parts of the cartoon are superimposed on the "frame", which depicts the non changing background of the scene.
While the frame problem is a major issue for certain approaches to logic-based AI, we don’t see it as an objection to scalably deploying logic-based technology to draw inferences based on large spatiotemporal knowledge bases (nor to other scalable real-world inference applications). Rather, we see it as an objection to embedding logical inference engines in overly simplistic cognitive architectures. We believe the frame problem can be bypassed via judicious inference control heuristics, including some that are implicit in the ideas discussed in the previous section, and some others that will be discussed here.
### Review of the Frame Problem
To elaborate the frame problem a little more fully, suppose we have the following knowledge:
• Alison is in her office and she wears a blue suit.
• If Alison moves from her office, then she is in the loby.
If we represent the above in classical first-order logic, using some suitable formalism for representing time and action (e.g., CTL logic, or an appropriate subset of PLN, both discussed below), we can derive that after Alison moves from her office she will be in the lobby. However, we will not be able to derive that Alison’s suit is still blue. Namely, the knowledge given above does not rule out the possibility that the color of Alison’s suit changes when she moves out of her office. A straightforward solution for this is to add rules that explicitly describe the non-effects of each action (e.g., “when Alison moves from her office, the color of her suit does not change”). Such formulae are called frame axioms. However, this is not a satisfactory solution. Namely, since most actions do not affect most properties of a situation, in a domain comprising M actions and N properties we will, in general, have to write out MN frame axioms which would make any reasoning process impractical.
In [Shanahan & Baars, 2005] there is a more detailed account on the frame problem, including a brief description of Dennett’s memorable example:
…consider the challenge facing the designers of an imaginary robot whose task is to retrieve an object resting on a wagon in a nearby room. But the room also contains a bomb, which is timed to explode soon. The first version of the robot successfully works out that it must pull the wagon out of the room. Unfortunately, the bomb is on the wagon. And although the robot knows the bomb is on the wagon, it fails to notice that pulling the wagon out brings the bomb along too. So the designers produce a second version of the robot. This model works out all the consequences of its actions before doing anything. But the new robot gets blown up too, because it spends too long in the room working out what will happen when it moves the wagon. It had just finished deducing that pulling the wagon out of the room would not change to color of the room’s walls, and was embarking on a proof of the further implication that pulling the wagon out would cause its wheels to turn more revolutions than there were wheels on the wagon—when the bomb exploded. So the robot builders come up with a third design. This robot is programmed to tell the difference between relevant and irrelevant implications. When working out the consequences of its actions, it considers only the relevant ones. But to the surprise of its designers, this version of the robot fares no better. Like its predecessor, it sits in the room “thinking” rather than acting. “Do something!” they yelled at it. “I am,” it retorted. “I’m busily ignoring some thousands of implications I have determined to be irrelevant. Just as soon as I find an irrelevant implication, I put it on the list of those I must ignore, and.” the bomb went off.
It is obvious that the human brain incorporates a solution to the frame problem and does not suffer from overwhelming information when deriving new conclusions. When we say that Alison left her office, we don’t need to state explicitly that her suit hasn’t changed its color, that Bob’s cat hasn’t changed its sex, that the Sun continues to shine, etc. Such information is taken for granted by common sense. In mathematical logic, however, nothing is taken for granted and in classical logic it is necessary to represent explicitly all the things that change and all the things that do not change by some action.
### Working around the Frame Problem
Perhaps the best-known attempt to work around the frame problem, within the scope of logic-based AI, begins from the observation that the inference process in classical logic is monotonic, meaning that the set of conclusion can only grow when we add new premises (we do not retract some conclusions if we are presented with some new premise). But, it would be nicer if one could infer that Alison’s suit is, generally, of the same color when she leaves her office and, in addition, it would be suitable, to add some exceptions, stating otherwise (“If Alison spilled coffee on her suit, then she has to change her suit before she leave her office”). In other words, one would like to be able to declare the general rule that an action can be assumed not to change a given property of a situation unless there is evidence to the contrary. Such reasoning is possible within non-monotonic logics. Despite the fact that there is also a number of problems with the frame problem when addressed by non-monotonic logics, it can be considered that they provide a satisfactory solution. In artificial intelligence, there are also some other approaches, that handle different incarnations of the frame problem with more or less success. The frame problem is still making an influence on issues in cognitive sciences, philosophy, psychology, etc.
Propositional and first-order logic as defined are monotonic. This means that the set of facts that can be derived from S increases when S increases. However, again, this is not appropriate for some sorts of common-sense reasoning. For example, if we are given a fact that Tweety is a bird, we by default derive the fact that Tweety flies. But, if we are given an additional fact that Tweety is a penguin, then we retract our conclusion and derive the new one – that Tweety does not fly. There is a family of logics following this motivation and trying to model common-sense reasoning, summarized for instance in [Reiter, 1980; Delgrande & Schaub, 2003]
However, nonmonotonic logic is not the only route for circumventing the frame problem; for instance in our own work with PLN, we have taken a significantly different approach, to which we will return in the final chapter of this book.
# Quantifying and Managing Uncertainty
Another major issue in formal logic systems is the quantification and incorporation of uncertainty. It is of course possible to represent uncertainty using theories within first-order logic, but many have argued that uncertainty is sufficiently basic to commonsense inference that it should be introduced into formal logic at a foundational level.
With this in mind, there are many different ways of representing uncertainty within logical formalisms. Some of the leading approaches may be summarized as:
• Fuzzy
• Imprecise / Indefinite Probabilistic
We have touched some of these briefly above, but we will now review them slightly more systematically.
## Fuzzy logic
Fuzzy logic [Zadeh, 1965] extends the notion of boolean truth (true vs. false), to encompass degrees of truth varying between 0 and 1; and generalizes the standard boolean operators {¬, ∧, ∨} by “fuzzified” counterparts. These generalized operators are usually (but not always) defined as follows:
$\displaystyle \neg x = 1-x$
$\displaystyle x\wedge y = min(x, y)$
$\displaystyle x\wedge y = max(x, y)$
It appears that for various instances of commonsense reasoning, fuzzy logic is more immediately appropriate than boolean crisp logic; for instance the fuzzy predicate:
$\displaystyle (strong(x) \wedge healthy(x)) \wedge intelligent(x)$
allows one to characterize the degree to which a person is both strong and healthy, or intelligent; while its boolean interpretation can only draw two crude categories for each predicate, which does not fit how one would think about the proposition with our common interpretation of the concepts strong, healthy and intelligent.
## Possibility theory
In fuzzy logic, values associated to facts or propositions range from 0 to 1 but represent truth values. In possibility theory [Zadeh, 1978], on the other hand, a proposition may be boolean but what is considered instead is its degree of belief by an agent. That degree of belief is represented by two values, the necessity denoted nec(p) and the possibility denoted pos(p) representing the extent to which an agent considers p to be necessary and possible. For instance pos(earth_flat)=0.9 and nec(earth_flat)=0.2 represent respectively how much an agent believes the earth is possibly and necessarily flat.
The possibilities pos(p)=1 and pos(¬p)=1 represent a total state of ignorance about p, or equivalently nec(p)=0 and nec(¬p)=0. Having pos(p)=pos(¬p) does not contradict the principle of bivalence because this is the degree of belief of p which is considered not its truth. Thus possibility and necessity are not self-dual, that is it is not the case that pos(p)=1-pos(¬p), however there are mutually-dual as formalized by the following axiom:
$\displaystyle nec(p)=1-pos(\neg p)$
Other axioms permit one to determine the possibility and necessity of a formula based on the possibilities and necessities of its components -- but not always. For instance nec(p ∧ q) equals to min(nec(p), nec(q)) like in fuzzy logic, but pos(p ∧ q) is not generally equal to min(pos(p), pos(q)); however one can always bound the latter value using necessity and possibility measures combined:
$\displaystyle min(nec(p), nec(q)) \leq pos(p \wedge q) \leq min(pos(p), pos(q))$
Finally, although fuzzy logic and possibility theory are built around two different semantics, if nec=pos then possibility theory amounts axiomatically to fuzzy logic (with the min/max interpretation for $\displaystyle \wedge$ / $\displaystyle \vee$ ); and for that reason possibility logic is sometimes referred as an extension of fuzzy logic.
As probabilistic methods have become very popular in the AI field lately, there now exists a wide range of methods and theories regarding probabilistic inference and logic; we will only describe a handful of the most relevant ones here. To understand the following sections the reader needs to be familiar with probability theory.
### Bayesian Inference
Methods to reason about probabilities include Bayesian inference and Bayesian networks, both called so because they mainly rely on Bayes' theorem, a formula from elementary probability theory recalled below:
P(Y|X) = P(X|Y)P(Y)/P(X)
Bayesian inference is a mean to infer or revise the probability of an hypothesis knowing a set of observations (for instance determining the probability of a disease in the presence of symptoms). Bayesian networks are graphical models, formally DAGs (directed acyclic graphs), representing dependencies and independencies between random variables. We will describe Bayesian networks in some detail here, as they will arise in later chapters when we discuss their applicability to mining causal relationships from large knowledge stores.
### Bayesian Networks
A Bayesian network is a graphical model that represents a probabilistic joint distribution over a set of variables represented as nodes and direct dependences represented by arcs between nodes. More formally a Bayesian network is a DAG (directed acyclic graph), where each node contains a variable and a function representing a conditional probability of that variable knowing its parents, which are all variables that have an arc pointing to that given node. If the node has no parent then the function represents a marginal probability. See Figure 3.1 for an example of a Bayesian network with 4 variables.
Figure 3.1: A small Bayesian network
Usually the probabilities, conditional and marginal, are coded into matrices. Figure 3.2 represents possible probabilities coded in matrices of the Bayesian network of Figure 3.1. They also can be coded in any other appropriate structures, like decision trees, decision graphs and such.
Figure 3.2: Probability matrices associated with the network nodes
As mentioned above, direct dependences are represented with arcs between variables; additionally it is possible to assess the conditional dependence and independence of any group of variables by applying the notion of d-separation introduced by Pearl [Pearl85]. When a triplet of variables X, Y, Z are d-separated then we can conclude that X and Y are independent conditionally to Z ; and conversely if there are independent then they are d-separated (actually that equivalence between d-separation and independence does not always hold, only when the distribution is DAG-faithful; we won't recall what DAG-faithful is here, however one should note that most of practical real life distributions are DAG-faithful or close to it). We also won't recall the detailed definition of d-separation here, but what is most important to note here is that the d-separation criterion is a graphical one. This means that one can assess the dependences of variables solely based on the topology of the network, which in itself is a useful thing. Figure 3.3 and 3.4 display two examples of d-separated variables.
Figure 3.3: d-separation in sets of variables
Figure 3.4: X8 and X9 are d-separated conditionally to {X3, X4, X7} in this example
Given a Bayesian network one can complete the joint distribution of a set of variables X1,...,Nn by applying the following formula:
P(X1,...,Xn)=$\displaystyle \prod_{i \mathop =1}^n$ P(Xi $\displaystyle \mid$ parents (Xi))
Where (Xi) is the set of variables with outgoing arcs pointing to Xi.
It is worth noting that any distribution can be represented by a Bayesian network (see Figure 3.5 for an example of how that can be done). And there are usually many possible Bayesian networks to represent a given distribution, see Figure 3.6 for an example.
Figure 3.5: Bayesian networks so that parent (Xi)=X!,…,Xi-1
Figure 3.6: Two Bayesian networks representing the same distribution (bottom)
### Bayesian Causal Networks
Next, a Bayesian causal network is a Bayesian network where the arcs represent causal relationships. A probability distribution alone is usually not enough to determine causality, for reasons we reviewed earlier. And one needs additional knowledge either given by an expert (stating for instance that weather may influence ice cream sales but not the opposite), or coming from additional assumption, for instance the knowledge that if event A occurs before event B then A can be a cause of B but B cannot be a cause of A. Note however that this sort of assumption involving time must be used cautiously because when A occurs before B, both A and B may actually be the result of a common cause C; in this case, C is called a confounding variable. Real life is full of confounding variables!
Many techniques available for Bayesian networks apply with little or no modification to Bayesian causal networks. For instance Bayesian inference, described in the next section, works exactly the same for a causal or non causal Bayesian network. On the other hand, network learning techniques may diverge more significantly because they need to take into account additional background knowledge in order to decide whether or not a given causal relationship is authorized.
### Bayesian Inference
Given a Bayesian network one can use it to perform various probabilistic inferences. In this context, inference means calculating joint marginal and conditional probabilities.
For instance, let us consider the Bayesian causal network of Figure 3.7.
Figure 3.7: Bayesian casual network
Let I, P, O and R be binary random variables respectively representing Internet connection working, Pay-check arrived, On-line purchases and Router overloaded.
One may want for instance to know the probability that a particular individual is going to make on-line purchases knowing that his router is overloaded and he hasn't received his pay-check yet. That is, we want to compute the conditional probability:
P(O=1 $\displaystyle \mid$ R=1,P=0)
The basic method to compute the above is called variable elimination. That is, one eliminates by summation the variables which are absent of the conditional probability of interest, in this example the variable I. This permits one to compute any partial joint marginal probability and the conditional probability is obtained by normalization.
So, in the example, one needs to compute the joint marginal probabilities:
P(O = 1 $\displaystyle \med$ R = 1,P = 0)
And the conditional probability is obtained by normalization:
P(O = 1 $\displaystyle \mid$ R = 1,P = 0) = P(O = 1,R = 1,P = 0)/P(R = 1,P = 0)
Let’s first compute
P(O = 1,R = 1,P = 0)
=$\displaystyle \sum_{i=0,1}$ P(I =i,O = 1,P = 0)
=$\displaystyle \sum_{i=0,1}$ P(R = 1)P(I = i $\displaystyle \mid$ R = 1)P(P = 0)P(O = 1 $\displaystyle \mid$ p = 0,I = i)
=P(R = 1)P( P = 0)$\displaystyle \sum_{i=0,1}$ P(I = i $\displaystyle \mid$ R = 1)P(O= 1 $\displaystyle \mid$ P = 0,I = i)
=.2 $\displaystyle \times$ .8 $\displaystyle \times$ (.4 $\displaystyle \times$ 0 + .6 $\displaystyle \times$ .1)
=0.0096
Then we compute P(R=1,P=0),but since R and P are independent, according to the d-separation criteria, we can directly get P(R = 1) $\displaystyle \times$ P(P = 0) = .2 $\displaystyle \times$ .8= .16
so the conditional probability of making on-line purchases is:
P(O = 1 $\displaystyle \mid$ R = 1,P = 0) =.0096/.16 = .06
There exists a variety of algorithms to perform these sorts of inferences, which include various optimizations, like caching and reusing intermediate results and using d-separation criteria (as we have done in the example above) to skip unnecessary summations.
### Markov Logic Networks
Next, Markov Logic Networks [Richardson & Domingos, 2006] involve a combination of First Order Logic (FOL) and Markov Network. This constitutes an elegant way to define probability distributions over valuations (defined earlier in Section 2.2). Markov Networks are very similar to Bayesian Network but are represented by an undirected graph – instead of a DAG like in Bayesian Networks. As a Bayesian Network, a Markov Network can model any joint distribution; but it gives more compact models for certain classes of distributions. The notion of conditional independence has a simpler representation with Markov Networks. However traditional Bayesian Networks can explicitly represent causality while Markov Networks cannot.
In probability theory terms, the full joint distribution of a Markov Network is defined by composing together a set of partial joint distributions over a group of variables that are all directly dependent on each other (i.e., a clique in the usual graph-theoretic terminology). Logic-wise, what this means is that one can define a probability distribution over the set of valuations (or possible worlds) given a set of first order logic axioms by building its corresponding Markov network. An valuation is a truth table of all atoms defined in the logic (like for instance follows(Jill, Joel)=true, online(Joel, 11pm)=false, etc), and each atom has a corresponding random variable in the Markov network. Each axiom is associated with a clique (containing the atoms of the formula). Or if the axiom contains variables, as many cliques as instantiated formulas (called ground formulas) of that axiom. Which means that the larger the number of constants in the domain of interpretation is, the bigger the resulting Markov network will be. Informally, the probability of a valuation is proportional to the exponential of the sum of the ground formulas that this valuation satisfies. It is also possible to weight the axioms, so the higher an axiom is weighted the stronger is the constraint of its satisfaction. All ground formulas are weighted identically to their corresponding axioms. Formally
P(V = v) = $\displaystyle \frac{1}{Z}exp$ $\displaystyle \left(\sum_{j}W_j \times f_j(v)\right)$
Where Z is a normalizing factor so that the probabilities over all valuations sum up to 1, Wj is the weight of ground formula j and fj is the evaluation of formula j over the valuation V.
Once the Markov network has been built it can be used to determine the probability of any given valuation, and of course the conditional or marginal probabilities of any combination of atoms, like
$\displaystyle P(online (Joel,11pm) = false)$
or
$\displaystyle P(online (Jill, 11:20pm) \mid online (Joel, 11pm),follows (JIll,Joel)$
Valuations that do not actually satisfy all the axioms can have a non null probability as well – but they'll usually have a lesser strength, due to the fact that they fulfill less axioms than a satisfying valuation (although that actually depends on the weights of the axioms).
Interestingly, one can define any Markov network by listing the right set of axioms in first order logic. And most importantly one can perform probabilistic inferences about formulas using that Markov network, like computing the probability that a formula F2 is satisfiable knowing that F1 is satisfiable.
Let us conclude this section with a small example based on the following axioms
1) $\displaystyle \forall X\neg follows (X,X) <5>$
2) $\displaystyle \forall X,Y (follows (X,Y)\wedgeonline(Y,11pm))\Rightarrow online (X,11:20pm) <1>$
with their respective weights between brackets. What it says is that it is quite true that someone does not follow itself (weight = 5). And it is relatively less true that someone following someone else who is online would be online 20 minutes later (weight = 1). The constants are Jill and Joel. 11pm and 11:20pm can be seen as constants of a different sort, but since they are not relevant for this example they are ignored by the universal quantifier (this would be easily formalized with a typed FOL).
The ground formulas of Axiom 1 are
1) $\displaystyle \neg follows (Jill,Jill) <5>$
2) $\displaystyle \neg follows (Joel,Joel) <5>$
The ground formulas of Axiom 2 are
1) $\displaystyle (follows (Jill, Jill) \wedge online (Jill, 11pm))\Rightarrow online (Jill,11:20pm) <1>$
2) $\displaystyle (follows (Jill, Joel) \wedge online (Joel, 11pm))\Rightarrow online (Jill,11:20pm) <1>$
3) $\displaystyle (follows (Joel, Jill) \wedge online (Jill, 11pm))\Rightarrow online (Joel,11:20pm) <1>$
4) $\displaystyle (follows (Joel, Joel) \wedge online (Joel, 11pm))\Rightarrow online (Joel,11:20pm) <1>$
With their respective weight between brackets. The corresponding Markov network is given in Figure 3.8
Figure 3.8: Markov network corresponding to the axioms given above. A link between two atoms are build whenever the two atoms are both present in a ground formula as they then obviously depends on each other with respect to the probability distribution of valuations.
There are 8 variables and therefore 2^8=256 valuations, let give an example of the calculation of the probability of two valuations, the one where all variables are false and the one where all variables are true. In the case where all variables are false we can see that all axioms are satisfied (neither Jill nor Joel follows itself, and the premise of Axiom 2 is never fulfilled therefore true). The calculation goes as follows
$\displaystyle P(V_false)$ = $\displaystyle \frac{1}{z}(2\times exp(5)+4\times exp(1))$
In the case where all variables are true we can see that only the first axiom is not fulfilled, so the calculation goes as follows
$\displaystyle P(V_t_r_u_e)$ = $\displaystyle \frac{1}{z}(4\times exp(1))$
We still need to calculate the normalizing factor
$\displaystyle Z = \sum\sum exp(w_j \times f_j(v))$
There are 7ˆ2 valuations satisfying ground formulas 1 and 2, and there are 5ˆ2 valuations that do not satisfy ground formula 3 to 6 (when the conclusion is false but the premise is true), therefore Z is
$\displaystyle Z = 2 \times 7^2 \times exp(5)+4 \times (256-5^2) \times exp (1) = 17056$
which lets us with
$\displaystyle P(V_false)= 0.0018$
$\displaystyle P(V_t_r_u_e)= 0.0006$
## Imprecise and indefinite probability
Imprecise probability is a generic term referring to a variety of theories that extend probability theory when the probability measure is partially known or inadequate. The most common class of imprecise probabilities uses upper and lower probabilities [Walley, 1991]. That is, an event may be characterized by two values, its upper and lower probabilities, instead of one probability; and this interval is interpreted to delimit the means the probability distributions lying in a certain envelope, and constructed according to certain distributional assumptions.
Other theories for dealing with imprecise probability use the notion of meta-probabilities (that is probability distributions over probabilities). There are debates amongst statisticians whether or not to allow meta-probabilities because that meta-level often cannot be subject to repetitive experiments. Therefore some prefer to consider a subjective interpretation of probability (as opposed to a frequentist interpretation), which may affect the choice of the assumptions regarding how to model imprecise probability.
One well developed imprecise probability theory using meta-probabilities is called Indefinite Probability [Goertzel et al., 2006c]; it uses lower and upper probabilities plus a degree of confidence, more formally any event w has an indefinite truth value consisting of a quadruplet <[L, U], b, k> which roughly means that after k more observations there is a meta-probability b that the probability of w lies within [L, U]. This method of quantifying probabilities is used in the Probabilistic Logic Networks approach to inference, which is detailed in the following chapter.
# Representing Temporal Knowledge
We’ve discussed the basics of logical reasoning and logical knowledge representation, in a general way. It’s now time to get more specific and talk about one of the applications of logic that’s most central to our topic: logical representation of time.
The natural way to present this material is to start with temporal representation and then move to temporal reasoning. However, we will first make a few brief comments about temporal reasoning, with the goal of motivating our choices in temporal representations. This should be expectable, because in large part it’s the requirements of reasoning that determine what kinds of knowledge representation are appropriate.
Temporal reasoning, broadly construed, is the process of inferring new relations between events localized in time based on known facts and relations about those events. As such, temporal reasoning can be divided in two main branches, depending on how time is represented:
• Quantitative: in this variant, information regarding absolute, numeric temporal labels - or time stamps, using a computational jargon - is important for reaching conclusions about events, and therefore it is used as part of the modeling. Quantitative temporal reasoning will work with events specified in a temporally hard way, such as “event A begins at 01:31 and ends at 02:03”.
• Qualitative: this variant is not concerned at all with absolute time stamps; instead, only relative relations between events - such as event A happens before event B, event C happens during event B, and so on - are relevant for producing inferences on the known temporal facts.
The quantitative approach is mainly applicable (and relevant) in applications where both accurate timing data is available and extreme time precision is necessary, such as reasoning about the functioning and performance of real-time systems. On the other hand, the qualitative approach is more appropriate to the analysis of data arising from human systems, or from noisy sensors as possessed by biological organisms or robots), and therefore we will concentrate here mainly on the qualitative approach.
Within the qualitative approach, there are two main issues to be confronted:
• how to represent basic units of time (how to “quantify time”)
• how to represent relationships between basic units of time
A number of different approaches exists to both of these problems, and we will now review these.
## Approaches to Quantifying Time
Figure 4.1 presents a simple ontology of the different approaches that have been taken to the representation of basic time-units.
Figure 4.1: Qualitative relations between approaches for temporal representation
As depicted in Figure 4.1, the essential aspect of time is its ordered nature: for any given two temporal units, we can judge whether one is before or after the other, or whether the two are simultaneous. All methods of quantifying time incorporate this ordering aspect. The simplest mean of ordering time, the point, has basically no qualitative aspects other than ordering. Two time-points can be compared as to their ordering, and there’s nothing else to do with them, except to get quantitative and measure the distance between two time-points.
On the other hand, we can also consider models of the time-unit that have a richer set of relationships amongst them, such as intersection, adjacency and so forth. We call models like these “topological,” and the most common approach here is to represent time using intervals. There is an argument that this is a more psychologically natural approach than time-points – a single, indivisible, instantaneous point being a kind of mathematical abstraction. The topological relationships between time-intervals are most commonly treated using Allen’s interval algebra, to be discussed shortly below.
In addition to simple time intervals, there are other related approaches such as
• fuzzy intervals (the areas near the edge of the interval have less membership in the time-unit than the ones near the center, as determined by some fuzzy membership function)
• probabilistic intervals (with the meaning that each subinterval either does or does not belong to the event associated with the time-unit, but the subintervals near the edge of the interval have a smaller chance of belonging to it), which may take the form of treating an interval as a confidence interval, or of utilizing a whole probability distribution instead of an interval
• distributional fuzzy intervals, which use a probability distribution of fuzzy sets (meaning that each subinterval has a certain probability distribution of membership degrees in the event associated with the time-unit)
These various possibilities are depicted in Figure 4.2 and 4.3.
Figure 4.2: Point-like, interval and fuzzy representations of an event
Figure 4.3: Probabilistic and distributional representations of the same event
# Allen’s Interval Algebra
The most traditional and well-developed theoretical framework for systematizing the qualitative relationships between time-intervals is Allen’s Interval Algebra, or simply Interval Algebra (IA), formalized for the first time in [Allen, 1983]. IA, in its original and simplest form, models temporal events as intervals, that is, processes that have a beginning and an end in time. Based on that, thirteen temporal relations may exist between any given pair of events. Figure 4.4 illustrates those relations with a simple bar representation for intervals:
Figure 4.4: Possible relations between two intervals according to Allen’s Algebra
There is a calculus that defines possible relations between time intervals and provides a composition table (see the next chapter for an example of composition table) that can be used as a basis for reasoning about temporal descriptions of events. Time intervals are not necessarily represented by precise endpoints, but rather by their relationships. So, in this framework without metric one can express that a time interval is contained by another time interval etc.
The following table shows many of the base relations in a different notation, explicitly typing them in with logical constraints (X- denotes the left end of the time interval X and X+ denotes the right end of X):
Relation Name Constraint < > X Before Y Y After X X+ < Y- m mi X Meets Y Y Met by X X+ = Y- o oi X Overlaps Y Y Overlaps by X X-Y- ∧ X+=Y+ = X Equals Y X-=Y- ∧ X+=Y+
Constraints in Allen’s algebra are of the form
I1(rel1,..., reli)I2,
where I1 and I2 are time intervals, and rel1,..., reli are some of the above 13 relations, with the meaning that at least one of them holds. Consider the example, "John was not in the room when I touched the switch to turn on the light". Let
A be the time John was in the room,
B be the time I touched the light switch, and
C be the time the light was on.
Then we can say
A { p, m, mi, pi, } B,
that is, A precedes, meets, is met by, or is preceded by B; and B { m, o } C, that is, B meets or overlaps C.
Similarly, the sentences
Afterwards, she goes to her office.
may be formalized in Allen's Interval Algebra as follows:
AlisonReadsNewspaper { d, s, f } AlisonIsHavingLunch
AlisonIsHavingLunch { <, m } AlisonGoesToOffice
For instance, the notation {d,s,f} refers to three relations in the above table, and indicates their disjunction; so it means “during or starts or finishes.”
It is clear that all of the above relations can be defined from the three binary relations <, =, and > applied to the bounds of two intervals to be located w.r.t each other. For instance, the assertion X overlaps Y corresponds to
X-<Y- ∧ X+<Y+ ∧Y-<X+
as shown in the above table.
The basic relations describe relations between definite, certainly known intervals. Uncertainly known intervals may be described by a set of all the basic relations that may apply. We call such a set of basic relations a general Allen relation, or just an Allen relation.
There is a general relation for every combination of the thirteen basic relations: 213 or 8192 of them. Each of the basic relations is a relation, of course, as are all their combinations. The full relation holds between two intervals about whom nothing is known. The empty relation {} has no meaning in terms of relations between actual intervals, but is the result of some operations on interval relations.
The “satisfaction problem” for Allen's interval algebra is determining, for a particular collection of relations on indefinite intervals, whether there is any set of specific time values for the intervals such that all the relations in the collection are true. For example, a collection of relations and intervals that is not satisfiable is three intervals A, B, and C such that A { p } B, B { p } C, and C { p } A (each precedes the next, and the last precedes the first). There are no definite intervals for which all these relations can hold. The satisfaction problem is shown to be NP-complete [Vilain et al., 1989]. However, there are subclasses of the problem that are tractable and that permit polynomial-time decision procedures.
### Allen Algebra in the Twitter Domain
In order to better understand the practical uses of IA, we will now explain how to apply logical formalism to a specific domain, consisting of messages in the Twitter microblogging service. We will make use of this same example domain in some other examples in the following, and so take this opportunity to briefly summarize the domain before using it to exemplify interval algebra.
Twitter is a free web service allowing individuals to post brief public or private messages (“tweets”). Each “tweet” deals with specific concepts, entities and sentiments, and has a specific author. Further, there is a social network of tweet authors; tweets may be geographically localized via IP address; and information about tweets is publicly available via the Twitter software API. See Figure 4.5 for a screenshot of the Twitter interface.
Figure 4.5: Screenshot of the Twitter.com web service
In order to develop some of our future examples using this domain, we will need to introduce some special “primitive” logical term and relationship types (i.e. a “signature” for our theory relating Twitter entities). The primitive terms and relationships we introduce are depicted in Figure 4.6 and 4.7.
Figure 4.6: Primitive terms for a logical formalismtargeting the Twitter domain
Figure 4.7: Logical relationships in the Twitter formalism
The commonsense semantics of these concepts and relationships should be fairly obvious; however, the right way to rigorously formalize some of these relationships (such as the spatial and temporal ones) is a complex issue, and will occupy a considerable percentage of this book!
Returning to interval algebra, Figure 4.8 shows schematically an example sequence of Twitter entries and the events that can be inferred from them.
Figure 4.8: Tweets as temporal events, associated with time intervals. Note that in the figure, more recent events occur towards the top of the figure.
Referring to Figure 4.8, note that many relations between the assigned events can be inferred. For instance:
“Peter watches debate” occurs during “Jane plays game”.
“Jane plays game” overlaps “Bill writes report”.
“Bill makes coffee” occurs during “Bill writes report”.
“Peter watches debate” precedes “Bill writes report”.
Temporal relations such as those listed above can be represented in the form of a graph, forming a type of graph that we call a Temporal Interval Network. Those are methodologically interesting in the sense that they transform many problems of IA into graph problems. Figure 4.9 shows the equivalent TI network for many of the relations between the events in Figure 4.8:
Figure 4.9: Time interval network for the events of figure 4.8
## Uncertain Interval Algebra
As noted above, IA in its original form presumes the existence of hard instants in time where events start and begin. However, in the real world, the actual start and end times of many events may be hard to define; and furthermore, in the specific case of the Twitter interface the imprecision of events in personal lives is taken into account by the use of a time scale of varying discretization (jumping from half an hour to hourly marks, for instance) as well as the use of words denoting imprecision (“about”).
In the fuzzy version of IA, relations between events are defined in terms of degrees of truth. Putting it simply, the truth value of a relation is the degree of “existence” of that relation, varying from 0 to 1. For instance, one may estimate that event A precedes event B with a degree of truth of 0.7 . From now on, we will denote such assignments with a functional notation - for instance A precedes B with a truth value becomes precedes(A,B,0.7).
Interestingly, the use of fuzzy logic allows the existence of multiple relations that are mutually exclusive in the original, rigid IA. For instance one can say that at the relations precedes(A,B,0.7) and overlaps(A,B,0.3) are both valid under a fuzzy modeling. (The previous numeric also illustrates that the sum of the truth values of all relations between two events has to be 1.0, “full existence” so to speak.) Figure 4.10 shows another TI, this time fuzzy (with truth values attached to edges), showing some of those “ambiguously” modeled relations for the events of Figure 4.8. As one can see, now an edge between any two events is defined by a tuple that may be composed of multiple truth values for different relations.
Figure 4.10: Yet another variation of a TI network modeling imprecision
Finally, in the probabilistic version of IA, a value represents the probability that a given relation between two events is assigned. Multiple probabilistic relations between two events that would have only one relation in the classic version of IA are also possible. At first, that may sound very similar to the fuzzy version, specially under the numerical viewpoint. However, conceptually they are completely different. Speaking in terms of TI networks, in the fuzzy case a given edge can be multiple types of relations at the same time, in different degrees. In the probabilistic case, a given edge can have the possibility of being multiple relations, with different probabilities, but that is a modelling of uncertainty on the nature of an edge that in the real world fits into just one of the options.
One aspect of probabilistic IA that may clarify its conceptual difference relative to the fuzzy version is the problem of determining the most likely subnetwork. An example of that is illustrated in Figure 4.11, where a very small subset of events from the example Twitter flow is used to compose a probabilistic IA network. For sake of clarity, different relations (with assigned probabilities) between two events are represented by different edges. The dotted edges form the less likely subnetwork, while the solid edges represent the most likely network. In the example, that hints at the sequence of events that most likely happened in reality.
While the possibility has not been explored in the literature, one may also conceive fuzzy probabilistic IA, involving intervals interpreted as confidence intervals of centers of fuzzy membership functions. Various other related options also exist, offering a great deal of modeling flexibility.
The problem of finding the most likely subnetwork brings to mind computational considerations on dealing with IA, which will be briefly examined here. As already mentioned (and as the most likely network example already illustrates) most if not all IA problems can be modelled in terms of graph problems, and finding the optimal solutions for many of them is combinatorially explosive, and not viable for realistically large IA networks. However, the theoretical work on IA has found subsets of it that are computationally tractable, as well as heuristics that significantly reduce computational time at the cost of finding a solution that is not guaranteed to be optimal [van Beek and Manchak, 1997]. And much of that is just applied graph theory, since many of the relevant problems in IA are reducible to classical graph problems. That is, IA networks (classic or otherwise) are in principle manageable by an extremely diverse plethora of optimization methods, as in the case of any graph problem.
# Temporal Reasoning
In principle, logical inference about propositions involving events embedded in time and space could be treated the same as any other kind of logical inference, but experience shows that this is not an optimal approach, mainly because such an approach leads to extreme inefficiencies of inference control. Thus, in this section and the following ones we will consider specific logical formalisms recently proposed for reasoning about time and space.
A variety of logical formalisms have been proposed for reasoning about time, including
• Standard first-order logic with temporal arguments
• Reified temporal logic
• Modal temporal logic
(all of which we will review in detail below) and others. A large percentage of work in the area of temporal logic appears to fall prey to one of the pitfalls of
• Unrealistic simplicity
• Excessive complexity
• Computational intractability
The creation of a scalable, reasonably general temporal logic remains a difficult research problem, and one of our tasks here is to characterize this research problem and explain exactly which of its aspects are most pertinent and most worth attacking at this juncture. In order to apply temporal logics such as the above to reasoning on large, complex real-world problems, several issues have to be addressed:
• the logic should be simple – if a logic is overcomplicated, crafting a workable, scalable inference control strategy is very unlikely to be feasible; and integrating the logic with other system components like query languages and knowledge bases will likely be infeasible
• the logic should not be too simple – for instance, for most serious real-world applications simple Prior’s logic is too simple and inadequate for applications such as reasoning about real-time behaviour of software. Also, a logic that achieves simplicity at the expense of needed expressiveness will be impractical because the formulas describing real algorithms will be too long and complicated to understand.
• the corresponding deduction procedures should be efficient
It seems feasible to construct many different approaches satisfying these criteria (including perhaps a PLN based approach as will be exemplified in Part III below); but there is not yet any approach that has been convincingly validated by practical applications to satisfy them.
## The Challenge Time Presents to Classical Logic
The essential challenge temporality poses to classical logic is that, in the latter, formulas are evaluated within a single fixed world and have fixed truth value, that does not change over time. However, most real-world systems are dynamic.
Temporal statements, statements involving temporal information, may have truth value that changes over time. For instance, the statements “it is Monday” or “I am at a meeting” have constant meaning, but their truth value can vary in time (but, of course, neither statement is ever true and false simultaneously; unless one is dealing with paraconsistent logic, which is beyond the scope of this book). Thus, “it is Monday” may be satisfied in some contexts, but not in others. Examples of temporal statements are also “I am always too busy”, “I will eventually be at my office ”, or “I will be at my office until John give me a phone call”. In temporal logics, evaluation takes place within a temporal context.
Temporal relationships and reasoning deal with issues like change, actions, causality, ontology of time, underlying logic, temporal constraints used, reasoning algorithms employed, etc. A change describes moving from one state or condition to another one. This relation is temporal (either implicit or explicit) and requires an appropriate representation. Representing the temporal aspect of the knowledge adds the time dimension to the truth of the information. A formalism for representing temporal information has to provide a way for establishing a link between an atemporal assertion and a temporal reference. There is a number of approaches for representing temporal relationships, each of them requiring specific forms of reasoning. Most of them are influenced by studies by Aristotle and the Megarian and Stoic schools in ancient Greece.
Temporal systems can be classified according to different aspects [Emerson, 1991; Pani & Bhattacharjee, 2001]:
• propositional versus first-order;
• global versus compositional;
• time points versus time intervals;
• discrete versus dense time;
• bounded versus unbounded time;
• branching versus linear versus paralel versus circular time;
• past versus future tense.
These aspects determine expressiveness, and also computational demands.
In the rest of this section we review three main families of modern approaches to temporal representation (more details can be found, for instance, in surveys [Pani & Bhattacharjee, 2001] and [Emerson, 1991]. Approaches within these families may have different properties according to the list of aspect given above.
### First order logic: temporal arguments approach
The simplest approach to handling time within logic is to represent time by numerical values. This representation can be applied within first order logic (FOL), with reals or integers as the intended domain for time variables and constants. There is a number of efficient methods for dealing with first order statements and reasoning with them, so this approach starts from a well established grounds. This approach is often referred as to temporal arguments approach. Within FOL, we can simply express temporal facts like “Alison is working on the same project as Bob at time t”, for instance, in the following way:
work_on_same_project (Alison,Bob,t)
Using sorted first order logic may be more suitable and, in this case, there is a distinguished sort for temporal arguments. For instance, the classical example „You can fool all the people some time and you can fool some people all the time, but you cannot fool all the people all the time“ can be represented in sorted first order logic in the following way:
$\displaystyle \forall$ x:H. $\displaystyle \exists$ t:T. youcanfool(x,t) $\displaystyle \wedge$ $\displaystyle \exists$ x:H. $\displaystyle \forall$ t:T. youcanfool(x,t) $\displaystyle \wedge$ $\displaystyle \neg$ $\displaystyle \forall$ x:H. $\displaystyle \forall$ t:T. youcanfool(x,t)
with an intended semantics that links the sort H to the set of humans and the sort T to the set of all time values. By this, time is given a special syntactic and semantic status, but can still share much of the treatment of other sorts of values. If the theory is equiped with axioms on ordering, one can also reason (using general first-order reasoning frameworks) about statemets like the following one:
$\displaystyle \exists t_0:T. \forall t:T.(t>t_0 \Rightarrow f(t)).$
In a similar way, one can use arithmetic operators over time values.
Despite its simplicity, well-foundness, and developed proof theory and methods, this approach has severe limitations. For instance, without adding a host of specialized additional predicates, one cannot model aspectual distinctions between for example, states, events and processes, and cannot represent notions used in natural language like now, then, since, while, until, nor notions like a bit later and similar.
To get around these problems, there are different formalisms for representing temporal information within first order logic – a line of thinking that eventually leads back towards the more sophisticated approaches to time representation that we considered earlier.
For instance, Interval Temporal Logic (ITL) is a temporal logic for both propositional and first-order reasoning about periods of time found in descriptions of hardware and software systems, but also in artificial intelligence and linguistics. ITL can handle both sequential and parallel composition and offers powerful and extensible specification and proof techniques for reasoning about properties involving safety, liveness and projected time [Moszkowski, 1994], It is not identical to Allen’s interval algebra, but has a great deal in common with the latter.
### Reified temporal logic
Alternately, rather than the encoding of temporal information within FOL, there is a “reified” approach relating atemporal terms to temporal terms by specific temporal contexts. In the reification approach to temporal reasoning, the truth of temporal assertions is considered while keeping atemporal parts of assertions unchanged within a specific temporal context. A good overview of reified temporal logics can be found in [Ma & Knight, 2001]. One of the most influential, formal approaches to reified logic was presented by Shoham in 1987 [Shoham, 1987].
Reifying a logic means moving into a meta-language where an assertion in an initial language (usually FOL, but a modal logic can also be used), becomes a term in a new language. In this new logic, one can reason about truth values of expressions in an object language through use of the truth values of expressions built with the operator like TRUE and with temporal object as arguments. Thus,
TRUE(atemporal expression, temporal qualification)
represents a statement whose intended meaning is that the first argument is true “during” the time denoted by the temporal qualification. TRUE is not a relation in the logic, nor it is a modal operator, rather it is a reifying context. For instance, the sentence “Alison is on a meeting with Bob between 11am and 12am” can be expressed as
TRUE(MEETING(Alison,Bob),(11am,12am)).
The truth predicates are used to express not only the time when an expression is true but also the patterns of its temporal occurrence. So, in general, the pattern of temporal occurrence for the atemporal expression admits many interpretations other than during.
And this brings us back, finally to Allens’s interval algebra as discussed above, which naturally leads to a form of reified temporal logic. In Allen’s variant of reified temporal logic [Allen, 1984], temporal incidence is expressed using the operators HOLDS, OCCURS, and OCCURING in order to express distinctions between states, events and processes.
Temporal reification has several advantages. On one hand, this logic gives a special status to time. On the other, it allows one to flexibly predicate and quantify over propositional terms. Therefore, it gives the expressive power to discuss relationships between propositional terms and temporal aspects with a high level of generality. Due to these qualities, the reified approach has enjoyed considerable popularity in AI. However, it has also been a subject of criticism and attacks. A number of authors have argued that reified temporal logics are unnecessarily complicated and imply a philosophically suspect ontology of time. However, it seems to us that these objections really pertain to specific, simple versions of reified temporal logic, rather than to reified temporal logic in general.
For instance, a major problem with simple reified logics is that there is no straightforward way of referring to one temporally referenced object within the context of another temporal interval, such as “The leader of the Konrad project in 2003 left our company in 2006“. Rather, in simple reified logics, all non-temporal terms have to be evaluated with respect to the same temporal terms, i.e., those given in the TRUE context. So, in a simple reified logic, the given example could be expressed in the following inelegant way:
$\displaystyle \forall$ x$\displaystyle ($ TRUE$\displaystyle ($ project_leader$\displaystyle ($ Konrad$\displaystyle )$ = x, $\displaystyle ($ 2003,2004$\displaystyle )$ $\displaystyle )$ $\displaystyle \Rightarrow$ TRUE$\displaystyle ($ left_job$\displaystyle ($ x$\displaystyle )$ , $\displaystyle ($ 2006,2007$\displaystyle )$ $\displaystyle )$ $\displaystyle )$
In addition, in Shoham’s reified temporal logic, there are no temporal predicates, except ≦ and =, which makes expression of some temporal phenomena awkward.
But in the '''BTK''' approach (named after the initials of its authors), presented by Bacchus et al in 1991 [Bacchus et al., 1991], each predicate and function symbol can take any number of temporal arguments. For instance, the above example can be represented in a much simpler way:
left_job$\displaystyle ($ project_leader$\displaystyle ($ 2003,Konrad$\displaystyle )$ , 2006$\displaystyle )$
.
This is also the approach adopted in PLN, as will be discussed a little later.
Alternatively, Galton proposed a method of unreification based on incorporating tokens [Galton, 1991]. Galton's event token is basically the occurrence of some event at some point m in time Thus,“Alison is having a lunch at 3 00pm” is an event token. Event tokens act on the one hand as additional parameters to predicate and function symbols, while on the other they can be used in the temporal occurrence predicates.
### Modal temporal logic
Modal logic, as discussed above, is a formal logic system that deals with modalities, like possibility and necessity. In modal logic, one can express a statement like “It is possible that Bob quit his job”. The modalities are represented by modal operators. The basic unary modal operators are usually written □ (or L) for necessarily and ◊ (or M) for possibly. In a classical modal logic, these two operators are linked in the following way:
$\displaystyle \Diamond P \Leftrightarrow \neg\Box\neg P$
$\displaystyle \Box P \Leftrightarrow \neg\Diamond\neg P$
Tense logic is a kind of modal logic-based system, introduced first by Prior in 1955 following the idea that tense is a sort of modality, to be set alongside the ordinary modes of necessity and possibility. Tense logic has two sets of modal operators, one for the past and one for the future (while ordinary modal logic has only one). In Prior's notation,
• Pp stands for "It has at some time been the case that p"
• Fp stands for "It will at some time be the case that p"
• Hp stands for "It has always been the case that p"
• Gp stands for "It will always be the case that p".
For instance, if p stands for “Alison’s department moves to China” and if q stands for “Alison and Bob have communication problems”, then
G$\displaystyle ($
p$\displaystyle \Rightarrow$ Gq$\displaystyle )$ stands for "It will be the case that if Alison’s department moves to China, then Alison and Bob will always have communication problems." Tense operators can express standard modal operators $\displaystyle \Box$ and $\displaystyle \Diamond$ :
$\displaystyle \Box P$ can be expressed as $\displaystyle Pp \vee p \vee Fp$
$\displaystyle \Diamond P$ can be expressed as $\displaystyle Hp \wedge p \wedge Gp$
Tense logic is obtained by adding the tense operators to an an underlying logic, for instance propositional logic, or first-order logic. For instance, in tense logic built over first-order logic, the statement “A woman will be a CTO” can be represented as ∃x(woman(x) ∧ F cto(x)).
The standard model-theoretic semantics of tense logic is defined with respect to temporal frames. A temporal frame consists of a set of times together with an ordering relation < on it. This defines the "flow of time" and the meaning of the tense operators. An interpretation of the tense-logical language assigns a truth value to each atomic formula at each time in the temporal frame. For instance, Pp is true at t if and only if p is true at some time t′ such that t′ $\displaystyle <$ t (the semantics of other tense operators is defined by analogy). A tense logic formula is valid if it is true at all times under all interpretations over all temporal frames. Prior developed several versions of a calculus of tenses in which one can derive, for instance, $\displaystyle GP \Leftrightarrow \neg F\neg P,$ and also $\displaystyle (\neg p \wedge \neg Fp) \Rightarrow \neg\Diamond p$ (“what neither is true nor will be true is not possible”), linking standard modal operators with Prior tense operators in an elegant way.
While in typical first order representation time is absolute, in modal temporal logic time is relative and statements refer to the present or to other events. There are variants of modal temporal logic for dealing with absolute precise times. Prior’s system does not specify the nature of time (points or intervals), but in this approach understanding temporal elements as points of time is more natural.
Apart from tense logic, there is a whole family of other modal temporal logic systems. The modal μ calculus is a powerful class of temporal logics with a least fixpoint operator μ [Scott & De Bakker, 1969; Kozen, 1983]. It is used for describing and reasoning about temporal properties of labelled transition systems. The modal μ calculus subsumes dynamic and temporal logics like CTL and LTL.
In linear temporal logic (LTL), time is not branching, and one can encode formulas about the future of paths, such as that a condition will eventually be true, that a condition will be true until another fact becomes true, etc. In linear temporal logic, there are operators for next (X), globally (G), finally (F), until (U), and release (R). For instance, if p stands for „There is electric power“, and q is „I work on my computer“, G(¬p⇒X¬q) stands for „It always hold that if there is no electric power, in the next moment I will not work on my computer”. In the simplest formulation of LTL, the U (until) operator is used to derive all the others.
On the other hand, computational tree logic (CTL) is a branching-time logic, meaning that its model of time is a tree-like structure in which the future is not determined; there are different paths in the future, any one of which might be an actual path that is realised. Some of the CTL operators are quantifiers over paths (for instance, Ap stands for “p has to hold on all paths starting from the current state”) and path-specific quantifiers (for instance, Xp stands for “p has to hold at the next state”, Gp stands for “p has to hold on the entire subsequent path”). For example, if p mean "I give a presentation", in CTL one encode “I will give a presentation, whatever happens” by AGp. There are also other variants of branching-time logic, such as Branching Temporal Logic which also generalizes LTL, and differs from CTL in subtle ways [Kontchakov, 2007].
Modal temporal logics are, in a sense, more expressive than FOL. They are also suitable for their modularity. They can directly and neatly be combined with other modal qualifications like belief, knowledge etc. For certaing sorts of modal temporal logics there are efficient automated theorem provers.
Concerning applications, Prior used his tense logic to build theories about the structure and metaphysics of time, while now it has numerous other applications. The notational efficiency of modal temporal logics makes them very appealing for applications in natural language understanding, where they have been widely used. There are various areas of applications of temporal logic, such as, for example, controlling operation of a bank of identical elevators in a building [Wood, 1989]. However, the area where modal temporal logics probably found the widest acceptance and success is in computing, including the theory of programming, database management, program verification, and commonsense reasoning in AI. In applications in program verification, following the influential work of Pnueli [Pnueli, 1977], the main idea is to formally specify a program and to apply deduction methods for proving properties of the program like correctness, termination, and possibility of a deadlock. In describing properties to be proved, one can use modal operators, for instance, to state that whenever a request is made, access to a resource is eventually granted, but it is never granted to two requestors simultaneously.
One of the very successful temporal logics used in verification is temporal logic of actions (TLA), developed by Leslie Lamport, which combines temporal logic with a logic of actions [Lamport, 1994]. It is a logic for specifying and reasoning about concurrent systems. Systems and their properties are represented in the same logic. Syntax and semantics of TLA are very simple, yet it is very powerful, both in principle and in practice. It has been used to study and verify cache coherence protocols, memory models, network protocols, database protocols, and concurrent algorithms [Batson & Lamport, 2003].
### Integration of deontic and temporal logic
There are several approaches for combining deontic and temporal logic. Some of them are based on expressing deontic constraints in terms of temporal logic, while in some deontic logic is extended by temporal operators. Some of the most important issues in combining deontic and temporal logic are:
• The temporal approaches considered are usually based on tree-structures representing branching time with the same past and open to the future.
• On top of these tree-structures, temporal deontic logics typically define one modal necessity operator, expressing some kind of inevitability or historical necessity, plus deontic obligation operators.
One family of those logics expresses the modal and deontic operators in temporal terms, while another family introduces temporal operators that can be combined with the modal and deontic operators.
In the approach described in [Dignum & Kuiper, 1997], deontic logic is extended by temporal operators. This approach is used for the specification of deadlines. It uses deontic constraints to specify what is the agent obliged to do and temporal constraints since the obligation is usually to be performed before a certain deadline. The following example from [Dignum & Kuiper, 1997] requires both deontic and temporal reasoning.
• Alison has to pay the mortgage for her house every month.
• Alison borrowed some money from Bob, which she has to repay as soon as she is able to do so.
• The roof of Alison’s house started leaking. It has to be repaired before the October rains start (It is now September).
• Alison wants to go on a midweek holiday. She has an offer to rent a cottage for relatively little money which has to be paid within 30 days after the reservation has been made.
Formulas in the system introduced in [Dignum & Kuiper, 1997] are propositional formulas built in the standard way and, in addition, formulas involving actions, the deontic operator OB, and a preference relation over actions called PREFER. If α is an action and F is a formula, then [α]F is also a formula, and its informal meaning is that “doing α neccessarily leads to a state where F holds”. If F is a formula, then OBF is also a formula, and its informal meaning is “it has to be the case that F holds”. If α1 and α2 are actions, then PREFER(α1,α2) is also a formula and its informal meaning is that the action α1 is preferable to the action α2. The formulas are built also using temporal operators X (unary, for next), U (binary, for until), P (unary, for held), S (binary, for held since), DO (unary, for an action to be performed next), DONE (unary, for an action that was performed last). The semantics of formulas is given in Kripke style.
For example, the statement “If Alison borrowed some money from Bob, she has to repay as soon as she is able to do so.” is modeled by the formula:
DONE(borrow(Alison,Bob)) ⇒
∀β OB(true < repay(Alison,Bob)) < DONE(β) ∧ PREFER(repay(Alison,Bob),β)
where OB(x<y) states that y can not hold true before x is performed
In [Dignum & Kuiper, 1997] there is not a corresponding axiomatization and inference system, nor algorithms for automation of reasoning in the described logic.
Partly following and extending ideas from [Dignum & Kuiper, 1997], in [Broersen et al., 2004], deontic logic SDL is combined with temporal logic CTL and deadline constraints are expressed in this new logic. Automation of reasoning in this logic is not discussed, but there is an expectation that it can be handled by CTL theorem provers.
The approach described in [Cheng, 2006] starts with linear temporal logic (LTL), extend it to state/event LTL (SE-LTL) which takes into account both events and propositions, and finally, extends it to the system SED-LTL with deontic modalities. The semantics of formulas is given in Kripke style, while there is no corresponding axiomatization or reasoning procedures. The system can be used in different computing scenarios involving both temporal and deontic notions. For instance, it can be used for expressing an access control policy in which the permissions depend on time or events. For instance, the following resource monitoring problem can be represented within the proposed logic:
• Useri has the permission to use the resource r for 5 time units continuously, and he must be able to access it 15 time units after asking, at the latest;
• Useri has always the permission to use the resource r, and he has to release it after 5 time units of utilization;
• If useri is asking for the resource and he has the permission to use it, then the system has the obligation to give it to him before 5 time units;
• If useri uses the resource without the permission, he must not ask for it during 10 time units.
Alternately, in the approach for combining temporal and deontic modalities proposed in [Åqvist, 2005], the “absolute” temporal operator Rt (“it holds at the time t that”) is used, instead of the “relative” temporal operators X and U, which are more expressive. There is also a variant of the system with the temporal operator Rth (“it is realized at time t in history h that”). In semantics of deontic operators, frames that are used are considered as (finite) two-dimensional coordinate systems, where it is possible to distinguish between the longitude (i.e., x-value) and the latitude (i.e., y-value) of any point in such a coordinate system. Times are interpreted as longitudes, and worlds, or histories, as latitudes. There is a corresponding, sound and complete, axiomatization and inference system, but automation of reasoning is not considered.
Despite the fact that all combinations of deontic and temporal logic have been constructed with applications in mind, it seems that there are still no large-scale applications of these techniques for real-world problems. Hopefully this situation will change in the near future.
## Inference systems for temporal logic
As noted above, there are three main approaches to temporal representation and reasoning. Here we continue the discussion and give some simple examples of inference in each of them.
We will focus here on inference systems for temporal logic, but wish to also note that there are decision procedures for some of the logics discussed here [Emerson, 1995]. Recall that for a given formula F, a decision procedure only gives a yes or no answer whether F is valid/theorem, without providing deductive argument for that. Most decision procedures use model-theoretic, semantic arguments for obtaining the final result. In some cases, these arguments can be turned to deductive arguments, but that can be very expensive.
### Inference in the Simple First Order Logic Approach
In the approach based on first order logic and expressing temporal information as first-order formulas, giving each predicate an extra argument place that corresponds to the temporal dimension. For instance,
work_on_same_project (Alison,Bob,2007)
If the first-order signature is extended with a binary infix predicate < denoting the temporal ordering relation "earlier than", and a constant "now" denoting the present moment, then the tense operators can be simulated by means of first-order logic, in the following way (p(t) denotes the result of introducing an extra temporal argument place to the predicate p):
Pp (informally: p held at some point in past): ∃t(t<now ∧ p(t))
Fp (informally: p will hold at some point): ∃t(now<t ∧ p(t))
Hp (informally: p always held): ∀t(t<now ⇒ p(t))
Gp (informally: p will always hold): ∀t(now<t ⇒ p(t))
In this framework, temporal information is completely expressed by means of first order logic, therefore standard classical first-order inference systems can be used as a reasoning vehicle. For instance, let us assume the following statements:
• Year 2007 was in the past.
• Alison and Bob worked together on the same project in 2007.
• If Alison and Bob worked together on the same project, they will always be friends.
One can represent these statements in the following way:
• 2007<now
• work_on_same_project (Alison,Bob,2007)
• Pwork_on_same_project (Alison,Bob) ⇒ Gfriends(Alison,Bob)
According to the above definitions, this can be rewritten as:
• 2007<now
• work_on_same_project (Alison,Bob,2007)
• ∃t(t<now ∧ work_on_same_project (Alison,Bob,t)) ⇒ ∀t(now<t ⇒ friends(Alison,Bob,t))
From 2007<now and work_on_same_project (Alison,Bob,2007), it follows that
2007<now ∧ work_on_same_project (Alison,Bob,2007)
and further,
∃t (t<now ∧ work_on_same_project (Alison,Bob,t)),
which, together with the implication from above, yields
∀t(now<t ⇒ friends(Alison,Bob,t))
i.e., Alison and Bob will always be friends (all three inference steps in this derivation can be simply justified in any reasonable inference system for classical first order logic).
### Reified temporal logic
In Allen’s variant of reified temporal logic [Allen, 1984], discussed above, temporal incidence is expressed using relational predicates HOLDS, OCCURS, and OCCURING, as for example:
HOLDS(MEETING(Alison,Bob),(11am,12am)).
OCCURS(DRIVES(Alison,Home,Office),(8am,8.45pm))
where terms of the form (t,t′) denote time intervals. These two predicates ensure distinctions between states and events.
Statements about states have homogeneous temporal incidence, i.e., they must hold over any subintervals of an interval over which they hold (e.g., the meaning of HOLDS(MEETING(Alison,Bob),(11am,12am)) is that Alison and Bob are at a meeting at any time interval between 11am and 12am).
On the other hand, statements about events have inhomogeneous temporal incidence, i.e., such a statement is not true at any subinterval of an interval of which it is true (e.g., if Alison drives from her home to her office from 8am to 8.45am, then in any subinterval of that interval, she does not drive from her home to her office, but between some other points).
These features of reified temporal logic (homogeneity of states and inhomogeneity of events) are ensured in inference systems by axioms such as
∀s,i,i′(HOLDS(s,i) ∧ In(i′,i) HOLDS(s,i′))
∀e,i,i′(OCCURS(e,i) ∧ In(i′,i) ¬OCCURS(e,i′))
where "In" denotes the proper subinterval relation.
Therefore, given HOLDS(MEETING(Alison,Bob),(11am,12am)), the above axiom (along with the underlying first-order style inference system) enables deriving HOLDS(MEETING(Alison,Bob),(11.15am,11.20am)). Indeed:
1. HOLDS(MEETING(Alison,Bob),(11am,12am))
(assumption)
2. In((11.15am,11.20am),(11am,12am))
(by means of interval arithmetic)
3. HOLDS(MEETING(Alison,Bob),(11am,12am)) ∧
In((11.15am,11.20am),(11am,12am))
(by propositional calculus)
4. HOLDS(MEETING(Alison,Bob),(11am,12am)) ∧
In((11.15am,11.20am),(11am,12am))
HOLDS(MEETING(Alison,Bob),(11.15am,11.20am))
(by homogeneity axiom and by Hilbert’s axiom A4)
5. HOLDS(MEETING(Alison,Bob),(11.15am,11.20am))
(from 3 and 4, by modus ponens)
### Modal temporal logic
As discussed earlier, Prior’s Tense Logic contains, in addition to the usual truth-functional operators, four modal operators P, F, H, G. P and F are known as the weak tense operators, while H and G are known as the strong tense operators. There is an intended correspondence between pair of these operators:
Pp ≡ ¬H¬p
Fp ≡ ¬G¬p
If this correspondence is ensured, then one pair of the operators is redundant.
Post worked on different variants of the inference systems for Tense Logic, trying to build an elegant system that has as theorems all formulas that are true following his intended semantics. Some of those formulas are:
Gp⇒Fp : "What will always be, will be" Fp⇒FFp : "If it will be the case that p, it will be — in between — that it will be“ ¬Fp⇒F¬Fp : "If it will never be that p then it will be that it will never be that p" (¬p ∧ ¬Fp) ⇒ ¬◊p : "What neither is true nor will be true is not possible” Fp→□Fp : "What will be, will necessarily be"
Of particular significance is the system of Minimal Tense Logic Kt for classical propositional tense logic which consists of Hillbert’s axioms, the modus ponens inference rule, the following axioms:
(1) p⇒HFp ("What is, has always been going to be")
(2) p⇒GPp ("What is, will always have been")
(3) H(p⇒q) ⇒ (Hp⇒Hq) ("Whatever always follows from what always has been, always has been")
(4) G(p⇒q) ⇒ (Gp⇒Gq) ("Whatever always follows from what always will be, always will be")
(5) Gp⇒ GGp ("What will always be will always will always be")
and the following rules of inference:
(RH) p ⊢ Hp
(RG) p ⊢ Gp
(US) p ⊢ p[φ→ψ]
The system Kt is sound and complete, i.e., its set of theorems is the set of all valid tense logic formulas. In addition, it is decidable whether a tense logic formula is theorem. The theorems of Kt are all the properties of the tense operators not relying on the temporal order.
Let us consider the following statements:
“If our company has small administrative overhead, then it has big profit”
“It will always be the case that our company has small administrative overhead”.
From the above assumptions, one can derive the fact “It will always be the case that our company has big profit”. Let us denote by p “our company has small administrative overhead” and by q “our company has big profit”. The proof of Gq is as follows:
1. p⇒q (assumption) 2. Gp (assumption) 3. G(p⇒q) (from 1, by RG) 4. G(p⇒q) ⇒ (Gp⇒Gq) (instance of the axiom scheme (4)) 5. Gp⇒Gq (from 3 and 4, by modus ponens) 6. Gq (from 2 and 5, by modus ponens)
### Computational Tree Logic
Computational tree logic (CTL), briefly described earlier, has many applications in describing and verifying computing processes. CTL uses the following temporal operators:
• Ap – stands for „all futures“, i.e., “p has to hold on all paths starting from the current state”
• Ep – stands for „some future“, i.e., “p has to hold on some path starting from the current state”
Temporal operators have to be followed by one of the following linear temporal operators:
• G –always
• F – sometime
• X – next time
• U – until
The semantics of p U q is that p is true in all states preceding the state where q holds and q will eventually hold. The semantics of other operators are intuitive. The syntax of CTL is given as follows. There are state formulas and path formulas defined in the following way:
• Each atomic proposition P is a state formula.
• If p, q are state formulas, then so are p∧q, ¬p.
• If p is a path formula, then Ep and Ap are state formulas.
• Each state formula is also a path formula.
• If p, q are path formulas, then so are p∧q, ¬p.
• If p, q are path formulas, then so are Gp, Fp, Xp, pUq.
The following system is a complete and sound deductive system for CTL [Emerson, 1991]:
Axiom schemes:
All propositional tautologies.
EFp ⇔ E[ true U p]
AGp ⇔ ¬EF¬p
AFp ⇔ A[ true U p]
EGp ⇔ ¬AF¬p
EX(p∨q) ⇔ EXp ∨ EXq
AXp ⇔ ¬E¬p
E(p U q) ⇔ q ∨ (p ∧ EXE(p U q))
A(p U q) ⇔ q ∨ (p ∧ AXA(p U q))
EXtrue ∧ AXtrue
AG(r ⇒ (¬q ∧ (p⇒EXr))) ⇒ (r ⇒ ¬A(p U q))
AG(r ⇒ (¬q ∧ EXr)) ⇒ (r ⇒ ¬AFq)
AG(r ⇒ (¬q ∧ (p⇒AXr))) ⇒ (r ⇒ ¬E(p U q))
AG(r ⇒ (¬q ∧ AXr)) ⇒ (r ⇒ ¬EFq)
AG(p ⇒ q) ⇒ (EXp ⇒ EXq)
Inference rules:
If ⊢ p then ⊢ AGp (Generalization)
If ⊢ p and ⊢ p⇒q then ⊢ q (Modus ponens)
For example, it can be proved that ⊢ AGp yields ⊢ p, and if F1 and F2 are logically equivalent propositional formulas, then ⊢ AG F1⇔ AG F2 (these statements will be used in the example in further text).
### Branching Temporal Logic
An interesting variant of CTL is Branching Temporal Logic [Kontchakov, 2007], which also extends linear temporal logic.
In BTL, one defines a tree as a flow of time F = (W, <) containing a root point r, for which W = {v | r < v} ∪ {r}, and such that for every w ∈ W , the set {w | v < w} is well-founded and (strictly) linearly ordered by <. A history in F is then defined as a maximal linearly <-ordered subset of W . Finally, an ω-tree is a tree of this sort in which every history is order isomorphic to (N,<).
A “branching time model” is then defined as a structure B = (F,H,pB0 ,pB1 ,...), where F = (W,<) is an ω-tree, H is a set of histories in F (conceived as the set of possible flows of time in the model), and pBi⊆ W for all i. A logical formula is then evaluated relative to a pairs (h, w) consisting of an actual history h ∈ H and a time point w ∈ h – i.e. formulas are evaluated at time-points in the context of their histories. In such a pair (h, w), the temporal operators are interpreted along the actual history h as in the linear time framework, while the operators E and A quantify over the set of all histories coming through w. We say that a BT L-formula φ is satisfiable if there exists a branching time model B such that (B,h,w) implies φ for some history h ∈ H and some time point w ∈ h. The practical upshot is similar to CTL, but the formal properties are subtly different.
## Examples of Temporal Inference in the Twitter Domain
In this section we give a concrete exampleof temporal inference in our target domain. Specifically, we illustrate the deductive system for CTL logic on one common-sense example. Let us suppose we have:
• Bob and Clark always have lunch together.
• If Alison is having lunch with Bob, and Bob is having lunch with Clark, then Alison is having lunch with Clark.
• In all circumstances, it is true that: if Alison has lunch with Clark, then he will not forget her name that day and potentially the next day Alison will have lunch with Clark again.
And let us assume that we want to derive:
If Alison has lunch with Bob, then potentially Clark will never forget her name.
Let us denote
Alison has lunch with Bob today by p
Bob has lunch with Clark today by q
Alison has lunch with Clark today by r
Clark forgets Alison’s name today by s
Then, the assumptions can be represented as:
AGq
p∧q ⇒ r
AG(r ⇒ (¬s ∧ EXr))
And the conjecture as:
p ⇒ EG¬s
We will use the following metatheorem of propositional logic: if ⊢ a ⇒ b and ⊢ b ⇒ c, then ⊢ a ⇒ c.
One proof of p ⇒ EG¬s then looks like the following:
(1) AG(r ⇒ (¬s ∧ EXr)) (hypothesis) (2) AG(r ⇒ (¬s ∧ EXr)) ⇒ (r ⇒ ¬AFs) (axiom 12) (3) r ⇒ ¬AFs (1,2,modus ponens) (4) ¬AFs ⇒¬AF¬¬s (3, propositional logic) (5) r ⇒ ¬AF¬¬s (3,4, propositional logic) (6) ¬AF¬¬s ⇒ EG¬s (axiom 5) (7) r ⇒ EG¬s (5,6,propositional logic) (8) AGq (hypothesis) (9) q (8, theorem) (10) p ⇒ p ∧ q (9, propositional logic) (11) p ∧ q ⇒ r (hypothesis) (12) p ⇒ r (10,11,propositional logic) (13) p ⇒ EG¬s (7,12,propositional logic)
# Representing and Reasoning On Spatial Knowledge
A large percentage of real-world knowledge pertains to space as well as time. Humans possess a great deal of evolved and learned common sense regarding representing and reasoning about space, but, this is not innate to software programs so we need to explicitly think about how to make our software systems space-friendly. From a logic point of view, space and time can in principle be treated just like any other sorts of relationships – but this turns out not to be an optimal point of view in terms of either human-friendliness or computational efficiency of logic systems. Rather, with space as with time, it is worthwhile to invest the effort to develop specialized techniques for handling spatial and spatiotemporal knowledge.
As in the case of time, we will not be able to entirely disentangle spatial representation from spatial reasoning; but we will endeavor to do so as much as possible, and in this chapter we will focus mainly on representation (introducing reasoning only when really needed for motivational purposes) and defer reasoning for later.
Most generally conceived, spatial reasoning [Cohn et al 2008] is the capacity to infer relations of direction, morphology, topology, distance, etc, about entities existing in a given space (typically 2D or 3D). There are many combinations of space representation/abstraction and inferential mechanics for dealing with spatial reasoning, usually focusing on different aspects of spatial representations: topology, morphology, direction, etc. In this chapter, we will describe in some detail the main representatives among them, and also discuss ways of extending and integrating those techniques.
## An Example Scenario for Spatial Representation and Reasoning
In order to provide a common ground for explaining the different methodologies used for logically representing spatial knowledge, we will articulate an example scenario wherein various sorts of spatial reasoning can be performed.
In order to accommodate several types of spatial reasoning tasks, the example presented here is composed of layered data produced by multiple sources. The use of such multi-layered data portrays a situation possible and maybe even frequent in a real-world, large-scale, heterogeneous store of spatiotemporal knowledge.
The first layer of data is of linguistic nature - a stream of Twitter entries, from which semantic interpretation can extract many elements of spatial and temporal information, as we can see in Figure 6.1.
One of the spatial inferences that can be made from the Twitter stream is that the three characters involved - Pam, Sam and Tim - appear to be in the same city or same metro area, which is called simply “Metro City” by Pam. That brings the second layer of data, shown in Figure 6.2, which presents a more evident spatial nature: it is a simplified map of Metro City, showing many city features - neighborhoods, bridges, canals, rivers, islands - including those mentioned in the Twitter stream. Many of those features are of an imprecise nature - for instance often it is difficult to tell where a neighborhood ends and other one begins.
The stream also mentions a weather phenomenon - rain - that may affect the plans of the characters involved in the Twitter stream. That brings the third layer of data: a (admittedly simplistic) weather pattern map for Metro City, shown in Figure 6.3, portraying the conditions of the city at the time of the first message. Arrows in the picture show direction of the wind and therefore the overall tendency of the dislocation of the weather patterns. It is interesting to note that this third layer is intrinsically uncertain, as the arrows point to tendencies and weather forecast is essentially probabilistic.
Now that the common scenario is described and illustrated, the following sections will show how several approaches for representation oriented to spatial reasoning “see” that scenario, and how they may allow the inference of new information based on that scenario.
Figure 6.1: Twitter stream containing spatial and temporal assertion
Figure 6.2: High-Level map of Metro City
Figure 6.3: weather patterns over Metro City
## Topological Representation
The first mode of spatial representation we will discuss here is “topological” representation. To make a rather crude and concise definition, topology deals with connection and inclusion of regions. For instance, telling if something is an island or a lake inside an island is a topological problem.
A spatial representation and reasoning technique that deals with both connection and inclusion at the same time is the Region Connection Calculus (RCC) [Cohn et al., 1997]. RCC may be interpreted as a possible extension of Allen’s Interval Algebra into the spatial domain. The basic entities modeled by RCC are regions (as opposed to points), assumed to be embedded in the same space of some dimensionality. The basic relations in RCC are defined in terms of connection or inclusion between two regions of the analyzed space. The number of possible relations varies according to the “granularity” of the RCC variety being used. The “coarser” variety, RCC8, is thus called due to the set of just 8 primitive relations that define it. Those are illustrated in Figure 6.4.
Figure 6.4: Connection-inclusion relations in RCC8
A set of RCC relations stating spatial facts about a group of entities can be used to infer new spatial facts. For instance, if A is a non-tangential proper part of B and B is a non-tangential proper part of C, then A is a non-tangential proper part of C. If on the other hand A is a tangential proper part of B and B is a tangential proper part of C, then A can be either a non-tangential or a tangential proper part of C. Or, showing the same inferences predicate-wise:
nonTangentialProperPart(A,B) ^ nonTangentialProperPart(B,C)
nonTangentialProperPart(A,C)
tangentialProperPart(A,B) ^ tangentialProperPart(B,C)
tangentialProperPart(A,C) XOR nonTangentialProperPart(A,C)
(Please note that an exclusive or is used at the right side of the logical conditional in the last inference, for an entity cannot be at the same time a tangential and non-tangential proper part of another in the crisp formulation of RCC8.)
All such inferential combinations between three entities are summarized in the composition matrix for RCC8, shown in Table 6.1. For sake of space the usual abbreviations for RCC8 operations are used. A set of operations in a given cell means that if two entities A and B are related by the operation corresponding to the row and B is related to a third entity C by the operation in the column, then A and C are mapped by the operations in the cell. For instance one can find the composition corresponding to the last inference above by checking the cell at row TPP, column TPP. The symbol “*” in the table denotes the “universal relation”, meaning that any RCC8 relation between A and C is possible in the corresponding case.
o DC EC PO TPP NTPP TPPi NTPPi EQ DC * DC EC PO TPP NTPP DC EC PO TPP NTPP DC EC PO TPP NTPP DC EC PO TPP NTPP DC DC DC EC DC EC PO TPPi NTPPi DC EC PO TPP TPPi EQ DC EC PO TPP NTPP EC PO TPP NTPP PO TPP NTPP DC EC DC EC PO DC EC PO TPPi NTPPi DC EC PO TPPi NTPPi * PO TPP NTPP PO TPP NTPP DC EC PO TPPi NTPPi DC EC PO TPPi NTPPi PO TPP DC DC EC DC EC PO TPP NTPP TPP NTPP NTPP DC EC PO TPP TPPi EQ DC EC PO TPPi NTPPi TPP NTPP DC DC DC EC PO TPP NTPP NTPP NTPP DC EC PO TPP NTPP * NTPP TPPi DC EC PO TPPi NTPPi EC PO TPPi NTPPi PO TPPi NTPPi PO TPP TPPi EQ PO TPP NTPP TPPi NTPPi NTPPi TPPi NTPPi DC EC PO TPPi NTPPi PO TPPi NTPPi PO TPPi NTPPi PO TPPi NTPPi PO TPP NTPP TPPi NTPPi EQ NTPPi NTPPi NTPPi EQ DC EC PO TPP NTPP TPPi NTPPi EQ
Table 6.1: Composition matrix for RCC8 relations
There are “finer-grained” versions of RCC like RCC15 and RCC23. Also, RCC can receive extensions (for instance a primitive detecting concavity) that also deal with morphology to some extent. Those are extensively discussed in [Cohn et al., 1997]. With RCC-8, though, it is already possible to see how space can be represented in terms of region connection in our example scenario. (And the relations in RCC15 and RCC23, as well as their numerous compositions, are probably overcomplex for the didactic examples intended here.)
Looking at the map in Figure 6.2, one can make the following assertions using the RCC8 predicates:
nonTangentialProperPart(“Center Island”, “Kidney Bay”)
externallyConnected(“Clarice Heights”, “Seaside District”)
tangentialProperPart(“Bridge Town”, “Seaside District”)
overlaps(“East Side”, “Clarice Heights”)
disconnected(“West Side”, “Center Island”)
nonTangentialProperPart(“L Park”, “Center Island”)
externallyConnected(“East Side”, ”Snail River”)
externallyConnected(“West Side”, ”Snail River”)
overlaps(“Snail River”, “Kidney Bay”)
According to the rules of Region Connection Calculus, it is already possible to make some topological inferences about Metro City from the assertions above. For instance:
nonTangentialProperPart(“Center Island”, “Kidney Bay”),
nonTangentialProperPart(“L Park”, “Center Island”) →
nonTangentialProperPart(“L Park”, “Kidney Bay”)
Inferences like the one above can be seen as the construction of new links in a RCC network. As often happens in reasoning, problems represented through RCC can also be seen as a graph - the aforementioned RCC network -, and the reasoning problem is accordingly reduced to a graph problem with a plethora of methodologies and tools for dealing with it. In order to better illustrate that, Figure 6.5 shows the assertions in the list above in the form of a RCC network, with the inferred relation nonTangentialProperPart( “L Park”, “Kidney Bay” ) represented as a dotted edge. (Due to space restrictions relation names are depicted in the figure as acronyms made from their initials, as traditionally done in RCC notation.)
Figure 6.5: RCC network representing some topological relations among Metro City regions
Even in the reasonably crisp map of Metro City, it is possible to see that some of the classic RCC8 relations shown in Figure 6.5 are in fact ambiguous due to the intrinsically imprecise nature of many geographic concepts. For instance, is Bridge Town a tangential proper part of the Seaside District or are they externally connected or perhaps overlapping? Are the East and West Side really disconnected or should the Old Bridge be considered a connection between them?
That imprecision and ambiguity of some geographical features in the RCC point of view can be better modeled by adding fuzziness to the representation (Schockaert et al. 2008). In fuzzy logic, two entities can be related by multiple predicates, even ones that may sound mutually exclusive in classic logic. Those contradictions and ambiguities are conciliated by the use of truth values indicating the degree to which the relation applies between the two entities. In the notation used here, we define a new logic predicate truthValue(R(A,B),v), indicating that the (fuzzy, in this case) truth value of relation R(A,B) between entities A and B is v. With that in mind, many fuzzy relations can be outlined among the entities shown in Metro City map, as exemplified below:
truthValue(overlaps("Bridge Town", "Sea Side District"),0.3)
truthValue(externallyConnected("Bridge Town", "Sea Side District"),0.3)
truthValue(tangentialProperPart("Bridge Town", "Sea Side District"),0.4)
truthValue(disconnected("East Side", "West Side"),0.9)
truthValue(externallyConnected("East Side", "West Side"),0.1)
truthValue(disconnected("Bridge Town", "Channel Beach"),0.9)
truthValue(externallyConnected("Bridge Town", "Channel Beach"),0.1)
truthValue(overlaps("Channel Beach", "West Side"),0.3)
truthValue(externallyConnected("Channel Beach", "West Side"),0.3)
truthValue(tangentialProperPart("Channel Beach", "West Side"),0.4)
As in the case of classic RCC, those relations can also be translated to a graph form. Figure 6.6 shows the fuzzy RCC8 network corresponding to the assertions above. The main difference visible between that graph and the one in Figure 6.5 is the presence of labels in the edges - each one is assigned to a tuple indicating the kind of relation and the truth value of it. (Due to space limitations, the unambiguous though highly verbose logic predication used in the list above is avoided.) Another difference is the presence of multiple edges (relations) between two nodes. As an example of a possible interesting subgraph that could be derived from that one, the most intense relations (those with higher truth values) between all connected pairs of nodes are shown as solid edges.
Figure 6.6: A fuzzy RCC network representing some topological relations among Metro City regions
Finally, some types of spatial information are better modeled neither in crisp nor in fuzzy ways - they are better modeled in terms of probability. Probabilistic spatial relations are conceptually different from the fuzzy ones in the sense that they do not say that two entities have a set of relations at the same time in different degrees. Instead, a set of probabilistic relations between two entities assumes that the actual relation between the two entities is unknown/uncertain, but assigns different probabilities to the existing possibilities.
Weather forecast based on the map shown in Figure 6.3 is a perfect target for probabilistic modeling, as it represents intrinsically uncertain information. Here we will try to model through RCC an extrapolation of the weather situation on Center Island at the time of the last Twitter message, based on the patterns and tendencies shown in Figure 6.3. In order to accommodate the uncertainty inherent to such a “weather forecast”, we introduce a new predicate probability(R(A,B),p), indicating that a given spatial relation R between entities A and B has a probability p of being real. The list below suggests some probabilities for relations extrapolated by the forecast:
probability(nonTangentialProperPart(“Center Island”, “Sunny”), 30%)
probability(overlaps(“Center Island”, “Sunny”), 30%)
probability(overlaps(“Center Island”, “Rain”), 30%)
probability(overlaps(“Center Island”, “Hail”), 20%)
probability(overlaps(“Center Island”, “Fog”), 10%)
probability(nonTangentialProperPart(“L Park”, “Center Island”),100%)
probability(nonTangentialProperPart(“L Park”, “Sunny"),40%)
probability(nonTangentialProperPart(“L Park”, “Rain”),30%)
probability(nonTangentialProperPart(“L Park”, “Hail”),20%)
probability(nonTangentialProperPart(“L Park”, “Fog”),10%)
That list of assertions can also be represented as a RCC network, this time of probabilistic nature, as shown in Figure 6.7.
Figure 6.7: A probabilistic RCC network representing some relations among Metro City regions and weather pattern zones
## Directional Reasoning
Although qualitative topological reasoning such as that provided by RCC can encompass many spatial relations between entities of many kinds, as illustrated in the previous section, it is by no means without limitations. For instance, RCC is inadequate for representing directional assertions.
Directional spatial relations assume that the space is divided into a set of directions - or angular alignments - that are applicable to all entities. Directional relations are often used in everyday descriptions such as “Canada is north of the United States” or “the hurricane is heading Northwest” or “in order to get to the Main Plaza, turn right, follow two blocks and then turn left”.
Qualitative directional reasoning is represented in a broader way by the STAR Calculus [Mitra, 2004], which by its turn can be considered a generalization of Frank’s Cardinal Calculus [Frank, 1991]. Cardinal Calculus used a system of four or eight directions (as in the conventional cardinal system used in geography, although they did not need a direct correspondence to those), while in the STAR calculus that is generalized to n directions forming n sectors of equal angular amplitude across the 360 degrees of the circle. (The mention of the circle bespeaks the fact that these forms of directional reasoning are primarily concerned with 2D spatial reasoning. However, it is not difficult to see that similar concepts can be produced in higher dimensional spaces with some work.) The two systems (hereinafter referenced collectively as “directional calculus”) have still an additional special “Null” direction meaning no direction at all, or in other words meaning that the two entities directionally compared are “in the same place”, that is, so close that any statement about directions is meaningless.
The basic operations of directional calculus are inversion and composition. Inversion allows commonsensical inferences such as
direction(A, B, North) → direction(B, A, South)
or “if B is North of A, then A is South of B”, in natural language. Composition on the other hand combines known directional relations between entities in order to infer new ones, such as in
direction(A, B, North), direction(B, C, East) → direction(A, C, Northeast)
Indeed, similarly to the case of RCC, all possible compositions in directional calculus can also be mapped into a composition matrix. As an example, Table 6.2 shows a composition table for the classic subdivision of geographic 2D space into eight cardinal directions (plus the null direction), adapted from [Frank, 1991] (where other variations are also discussed). Indeed, the same case convention of that publication is used: upper case letters for “exact” directions and lower case for “Euclidean approximates”. (The issue of “exact” and “approximate” directions is further discussed below as a prologue to fuzzy and probabilistic formulations of directional calculus.)
N NE E SE S SW W NW null N N n ne null null null nw n N NE n NE ne e null null null N NE E ne ne E e se null null null E SE null e e SE se s null null SE S null null se se S s sw o S SW null null null s s SW sw w SW W nw null null null sw sw W w W NW n n null null null w w NW NW null N NE E SE S SW W NW null
Table 6.2: Composition matrix for the operations corresponding to the eight cardinal directions
Using the directional predicates defined above, we can again describe the example scenario in terms of a symbolic spatial reasoning representation. Here is an example list of predicates taking into account directly information shown in Figure 6.1 and 6.2:
direction(“Sam”, “Center Island”, Null)
direction(“Pam”, “Bridgetown”, Null)
direction(“Center Island”, “Bridgetown”, South)
direction(“Sam”, “Pam”, South)
In the case of Tim’s Twitter entry, a human looking at Figure 6.2 can easily suspect that he is somewhere in the West Side, since he apparently has to cross the South Bridge to reach Sam, who lives in Center Island. An integrated spatial reasoning system taking into account both topological as well as directional reasoning could mount a graph of the connection relations in Metro City (like many of those exemplified here) and also reach the conclusion that Tim is likely on the West Side. (The system would of course have to reason on the assumption that people would choose the shortest path from one place to another, or make some related heuristic assumption, in order to rule out awkward possibilities like Tim being in the West Side, crossing Old Bridge, going South and then entering Center Island through South Bridge.) Therefore, we can also add the relations:
direction(“Tim”, “West Side”, Null)
direction(“West Side”, “Center Island”, East)
direction(“Tim”, “Sam”, East)
Using composition on the assertions above, we can make the inference
direction(“Tim”, “Sam”, East), direction(“Sam”, “Pam”, South) →
→ direction(“Tim”, “Pam”, Southeast)
Again, all those facts can be represented (and reasoned upon) in the form of a graph, as shown in Figure 6.8.
Figure 6.8: A directional network showing spatial relations between the entities shown in Figure 6.1 and 6.2
Here again the question of imprecision and uncertainty comes into play. In the case of topological reasoning some relations are naturally crisp (for instance, an island is surely a non-tangential proper part of a body of water, by definition). In the case of directions, on the other hand, arguably there are no crisp relations in almost all practical cases. For instance in the crisp example, by the assertions above it is declared that Center Island is North of Bridgetown, although perhaps one could also say that it is northeast of Bridgetown. The very definition of directions in directional calculus states them in terms of broad sectors of the circle instead of sharp lines, hinting already at a non-crisp nature for directions.
Therefore, a fuzzy modeling for directional calculus comes naturally; it is basically a way of telling that entity A is more in this or that direction than others, relatively to entity B. Following the example case, the same geographical entities in the example above can have directional relations re-stated in fuzzy form as suggested below:
truthValue(direction(“Center Island”, “Bridgetown”, Southwest), 0.5)
truthValue(direction(“Center Island”, “Bridgetown”, South), 0.5)
truthValue(direction(“Center Island”, “West Side”, West), 0.8)
truthValue(direction(“Center Island”, “West Side”, Southwest), 0.2)
The directional relations above, being between extensive regions, naturally look intrinsically fuzzy. In the case of the characters in the Twitter stream, though, we are talking about directional relations involving point-like entities. In that case, a probabilistic modeling seems to capture better the conceptual setting at play. For instance we can infer that Tim most likely is somewhere in the West Side, but since the West Side is rather extensive relative to Sam’s location in Center Island Tim may be West, Southwest or Northwest of Sam. However, we cannot say that Tim is at the same time West and Southwest of Sam (as in a fuzzy statement), but we can say that Tim is West of Sam or Southwest of Sam with different probabilities for each alternative. The list below suggests a probabilistic reformulation of the directional relations between the Twitter characters.
probability(direction(“Sam”, “Tim”, West), 70%)
probability(direction(“Sam”, “Tim”, Southwest), 20%)
probability(direction(“Sam”, “Tim”, Northwest), 10%)
probability(direction(“Sam”, “Pam”, South), 60%)
probability(direction(“Sam”, “Pam”, Southwest), 40%)
probability(direction(“Pam”, “Tim”, Northwest), 40%)
probability(direction(“Pam”, “Tim”, North), 60%)
As seen in previous approaches, fuzzy and probabilistic relations can also be represented as networks and thus be treated by many techniques for solving graph problems (analogously to the use of graph theory in analyzing biological networks, see e.g. [Przulj, 2005], [Grindrod and Kibble, 2004]). The graph in Figure 6.9 shows at the same time part of the crisp, fuzzy and probabilistic directional relations mentioned above. (This breakdown from the traditional approach of showing just “homogeneous” graphs used so far is in a way a preview of the following section, where we discuss a modeling tool that might allow the fusion of the approaches discussed so far.)
Figure 6.9: A mixed network showing probabilistic, fuzzy and crisp directional relations between the example entities
## Occupancy Grids: Putting It All Together
The previous sections show that available representation models for spatial reasoning are somewhat compartmentalized. Region Connection Calculus is appropriate for reasoning in topological (and even morphological) terms on entities with non-negligible extension. STAR calculus is interesting for dealing with directional relations between entities treated as dimensionless points. Both approaches are qualitative and therefore they do not deal explicitly with distances and measurements (although it seems unproblematic to add these to the reasoning framework at least in the case of the directional models). A natural question is whether there is a third approach able to integrate most or all of the strong points of the previous approaches, perhaps with a few additional advantages?
A positive answer to this question is provided by a modeling construct originally developed not with reasoning explicitly in mind, but rather as a probabilistic spatial representation the distribution of entities in space. This is the formalism of so-called Occupancy Grids [Martin & Moravec, 1996].
Occupancy grids (also called Evidence Grids in their more complex and generic forms) were developed for Mobile Robotics and to date that remain their main field of practical application. The initial goal was simply to represent space and obstacles on it, in a probabilistic way that could accommodate noisy, incomplete or ambiguous sensor data gathered by a robot while navigating through an unknown terrain, thus providing an adequate data framework for typical tasks such as obstacle avoidance and path planning. However, it has been demonstrated that spatial data represented in the form of occupancy grids can be used to achieve topological, morphological and directional inferences about the work environment of a robot [Szabó 2004]. Moreover, the original purely probabilistic version can also incorporate fuzziness [Bloch & Saffiotti, 2003]. Therefore, occupancy grids (and their variations) look like the most promising candidates for a common framework able to bridge and integrate all the aspects of spatial reasoning, even though (or perhaps because) they are a paradigm shift from the more symbolic, higher-level representations traditionally used in reasoning.
The original model of occupancy grids divides space into a regular lattice of cells, in which each cell has an associated probability of being occupied or not. Usually the grid is orthogonal and the space is 2D, although versions using other grid topologies (e.g., hexagonal) and higher-dimensional spaces are easy to devise. From the point of view of a conventional wheeled robot, “occupancy” is the presence of nondescript impassable solid obstacles. In order to accommodate versatile spatial reasoning about multiple entities, some complexity has to be added to that basic model. For instance the occupancy of a grid cell should not be nondescript. Rather, it should contain the probabilities of presence or fuzzy presences of the entities involved in the reasoning problem. In order to better illustrate that, Figure 6.10 and 6.11 show fuzzy and probabilistic occupancy grids representing data related to the example scenario.
Figure 6.10 shows a spatial discretization of Metro City land area in terms of an occupancy grid. Different regions of Metro City are shown in different colors. The representation is fuzzy in order to capture the imprecision inherent in dividing neighborhoods of a city. That fuzziness is color coded, and explains why for instance the East Side is green and Clarice Heights is red, but the border zone between them is a gradation of yellow (red plus green in the additive system), representing the variation from red to green and vice-versa as a given cell seems to be more part of Clarice Heights than the East Side, or the other way around. Color mixtures observing the same convention can be observed along other neighborhood boundaries of the city.
Figure 6.10: A fuzzy occupancy grid representing different regions of Metro City
Figure 6.11 on the other hand, is a probabilistic representation of the location of Tim, Sam and Pam. In terms of occupancy grids, those point-like individuals actually become clouds of probability. That can be an uniform cloud, as in the case of Sam - from the Twitter stream it is impossible to tell where exactly he is in Center Island, and so his probability cloud is homogeneous across the whole extension of the landmass. In the case of Tim, on the other hand, the grid cells of the West Side closer to the Old Bridge are less likely to contain him, since if Tim were in that location presumably he could opt to cross Old Bridge and then North Bridge instead of using South Bridge. (An option that would look particularly attractive if Sam lives on the North part of Center Island.) Finally Pam’s probability cloud coincides with the fuzzy cloud from Figure 6.10 representing Bridgetown.
Figure 6.11:a probabilistic occupancy grid showing the location clouds of Tim, Pam and Sam.
Both figures above point in an intuitive way to a set of methods -- often borrowed from digital image processing and computer vision -- that allow the “translation” of information in occupancy grids to more symbolic representation frameworks, such as RCC and directional calculus. For instance the color gradations representing neighborhoods in Figure 6.10 can be segmented by a plethora of techniques in order to determine “crisp” neighborhoods, which would be suitable for classic RCC. And the probability clouds in Figure 6.11 could have their centroids determined and submitted to a crisp directional analysis or (perhaps more interestingly) directional probabilities could be computed by considering all the possible pairs of cells from two different clouds, supplying information in a very refined way for a probabilistic directional reasoning process.
Also, throughout this analysis of spatial reasoning we have tried to show the equivalence between many spatial reasoning representations and graph representations. In the case of occupancy grids, that equivalence comes with particular ease, as grids in general are graphs, in the sense that they are sets of cells (which can be viewed as nodes) connected by a regular neighborhood topology (which can be represented by edges). Figure 6.12 illustrates that grid-graph equivalence.
Figure 6.12: occupancy grids as graphs following a lattice topology.
For the purposes of this text, more specifically we are talking about grid cells that can be related to spatial entities being reasoned upon by sets of logic predicates, often of probabilistic or fuzzy nature. Here we define the construct occupies(A, cell(i,j)), indicating that entity A occupies the cell assigned by coordinates i,j. (Many variations of an orthogonal coordinate system can be used, but throughout the examples in this section we will assume that i, j refer to row, column, with numbering starting at 1 for both dimensions.) The predicate occupies is different from the RCC predicates in the sense that it does not inform anything about the relative sizes and states of superposition of A and the cell. A may be a large entity and the cell contains just a tiny portion of it, or A may be a point-like entity that is somewhere in the space mapped by the cell. For instance, the following set of predicates could be related to the cell (20,15), part of the blurred frontier between Bridgetown and the Seaside District in Figure 6.10:
truthValue(occupies(“Bridgetown” ,cell(20,15)),0.63)
truthValue(occupies(“Seaside District” ,cell(20,15)),0.37)
probability(occupies(“Pam”, cell(20,15)),63%)
Of course, an occupancy grid is in a sense a “fine-grained” and regular graph, while in the spatial reasoning representations shown here we have “coarse-grained” graphs (representing higher-level entities such as whole regions and people, instead of individual space cells). However, in principle, topological and directional information about higher-level entities can be extracted from occupancy grids. Indeed, the issue of low-level to high-level conversion raises the awareness that, in order to be used as a universal framework for spatial reasoning, the operation of occupancy grids has to involve translations from and to other forms of spatial representation. The diagram in Figure 6.13 outlines some possible input and output relations between several techniques (not restricted to spatial reasoning) that might play a role in an spatial reasoning engine centered on occupancy grids.
Figure 6.13: A diagram showing inter-relationships between different kinds of processing that may exist in a comprehensive spatial reasoning engine
As one can see, a spatial reasoning engine created along these lines could potentially pass all the data flux through occupancy grids, which could be the main spatial data representation approach. External information sources (which can be anything, from raw sensory data to natural language extraction of spatial information) could write to the occupancy grids through the Visual Creation module. Interestingly, as one may note from the above discussion of mathematical morphology, many techniques for extracting information from occupancy grids can be borrowed from digital image processing and computer vision (grouped under Visual Interpretation), while techniques for adding information to occupancy grids can be borrowed from data visualization and computer graphics (grouped under Visual Creation). Data extracted from Occupancy Grids by Visual Representation may feed both Directional Reasoning and Topological/Morphological Reasoning processes, and those in turn may write their conclusions back to the occupancy grids through Visual Creation as well as output those conclusions to External Information Consumers.
## Handling Change
The examples we've given using the grid-centered system proposed above, as well as higher-level graph representations, so far have involved static situations. (With the possible exception of the example involving weather, where uncertainty is implicit in the dynamics of the phenomena involved; but we did not focus on that aspect.) However, in the greater context of this work, the spatial reasoning approach described in the previous section would most likely work in a dynamic environment, where data from multiple sources and represented in many ways would constantly update the “world picture” maintained by the system. Those updates could be from higher-level entities and cascade down to the low level of occupancy grids. Conversely, change occurring on the lower levels could also cascade up and rearrange the relations between higher-level entities. This last aspect is specially interesting for detecting change between entities (usually high-level ones) important for the problem being handled by the system.
Such a bottom-up cascading event is illustrated by Figure 6.14 and 6.15. Indeed Figure 6.14 shows an updated weather map, representing the weather patterns over Metro city at the instant of the last message from Sam. (And here an occupancy grid representation is used instead of the “weather blobs” of Figure 6.3. Different colors are used to assign the prevalent weather pattern in the grid cell, and the lighter the color the more intense the pattern represented.) We can see from that updated map that it is actually raining heavily over the whole of Center Island. Therefore, the probabilistic RCC network shown in Figure 6.7 must be updated with the new data. The result of such update is shown in Figure 6.15.
Figure 6.14: an updated weather map of Metro City.
Figure 6.15: an updated RCC network relating weather patterns to regions and individuals.
Instance (c) also adds one of the Twitter characters -- Sam -- to the picture, putting him in Center Island and therefore also under rain. The example network is kept small for didactic purposes, but considering the probabilities of Pam and Tim also being in the island, they could also be similarly related to present patterns. That kind of inference could be further used by a broader reasoning system (of which the proposed spatial reasoning engine would be just a component) to infer that perhaps the Twitter characters will make subsequent observations about rain at Sam’s place in their tweets.
## Spatial Logic
Temporal logic is a larger field than spatial logic, but the latter also has a long history and a flourishing body of contemporary work.
As an early example, consider Tarski’s 1929 work in his Geometry of Solids [Tarski, 1956] – this is a second-order logic system, whose variables range not over points, but instead over the regular closed subsets of 3D Euclidean space (those subsets of R3 equal to the closure of their interior). Tarski’s language features two non-logical predicates, corresponding to the binary relation of parthood, and the unary property of being spherical.
Much more recently, Aiello, Pratt-Hartmann and Benthem have reviewed the field of spatial logic in a masterful way in their lengthy book [Aiello, 2007], which draws together research from a variety of different areas under a common spatial logic framework. They conceive spatial logic broadly as “the study of the relationship between geometrical structures and the spatial languages which describe them,” and more precisely as follows:
By a spatial logic, we understand any formal language interpreted over a class of structures featuring geometrical entities and relations, broadly construed. The formal language in question may employ any logical syntax: that of first- order logic, or some fragment of first-order logic, or perhaps higher-order logic. The structures over which it is interpreted may inhabit any class of geometrical ‘spaces’: topological spaces, affine spaces, metric spaces, or perhaps a single space such as the projective plane or Euclidean 3-space. And the non-logical primitives of the language may be interpreted as any geometrical properties or relations defined over the relevant domains: topological connectedness of regions, parallelism of lines, or perhaps equidistance of two points from a third. What all these logics have in common is that the operative notion of validity depends on the underlying geometry of the structures over which their distinctively spatial primitives are interpreted. Spatial logic, then, is simply the study of the family of spatial logics, so conceived.
As Aiello et al note, contemporary work in spatial logic emanates mainly from three application domains, the first two of which are highly relevant to the present work:
• AI, where attempts have recently been made to develop logics of qualitative spatial reasoning.
• The theory of spatial databases, related e.g. to geographical information systems
• Image processing, where it is convenient to describe objects as sets of vectors that can be ‘added’ (taking all linear sums) or ‘subtracted’ (taking all linear differences), and one may proceed to define logics involving these operations
### Extending RCC into a Topological Logic
The spatial logic most relevant to our own work with PLN is the extension of RCC into a topological logic, an excellent example of which is presented in [Kontchakov, 2010].
To see how the RCC8 relations give rise to a spatial logic, let r1, r2 and r3 be regions in R2 that are homeomorphic to the disc, and suppose that r1, r2 stand in the relation TPP, while r1, r3 stand in the relation NTPP. It’s easy to see (from the visual natural of these relationships) that r2, r3 must stand in one of the three relations PO, TPP or NTPP. In formal terms, what this means is that the RCC-8-formula
TPP(r1, r2) ∧ NTPP(r1, r3) → PO(r2, r3) ∨ TPP(r2, r3) ∨ NTPP(r2, r3)
is valid over the spatial domain of disc-homeomorphs in the plane: all assignments of such regions to the variables r1, r2 and r3 make it true. Similar experimentation shows that, by contrast, the formula
TPP(r1, r2) ∧ NTPP(r1, r3) ∧ EC(r2, r3),
is unsatisfiable: no assignments of disc-homeomorphs to r1, r2 and r3 make this formula true.
A similar analysis may be done in Euclidean 3D space, and more generally. Suppose one has a formal language L, whose variables refer to spatial “regions” that are subsets of some geometric space. Given a subset K of regions, the notion of the satisfaction of an L-formula by a tuple of regions, and thus the notions of satisfiability and validity of an L-formula with respect to K, can be understood in the usual way. According to the approach of [Kontchakov, 2010], the pair (L,K) may then be called a spatial logic. If all the primitives of L are topological in character—as in the case of RCC8— then we have a topological logic. Otherwise one might have, say, a metric logic. For languages featuring negation, the notions of satisfiability and validity are dual in the usual sense.
Considering RCC8 as the foundation for a logic immediately makes the limitations of the RCC8 relationship-set apparent. For instance, constraints featuring RCC-8 predicates give us no means to combine regions into new ones. One way to remedy this deficiency is to turn the set of RCC regions into a Boolean algebra with binary operations + and · and a unary operation −. Intuitively, we are to think of r1 + r2 as the agglomeration of r1 and r2, r1 · r2 as the common part of r1 and r2, and −r as the complement of r. Adding these operators to RCC8 yields what is known as BRCC8 (Boolean RCC8). For instance, in BRCC8, the formula
EC(r1 + r2, r3) ↔ EC(r1, r3) ∨ EC(r2, r3),
is valid, whereas the formula
EC(r1, r2) ∧ EC(r1, −r2),
is unsatisfiable.
Taking things one step further, [Kontchakov, 2010a] explores the augmentation of BRCC8 with additional relationships representing connectedness: the unary predicates c and c≤k (for k ≥ 1). Here one may interpret c(r) as “region r is connected” and c≤k(r) as “region r has at most k connected components.”
These directions of research are highly relevant to the PLN approach that will be reviewed in Part III of this book, as PLN combines RCC relationships with Boolean and other operators in a fairly free way.
### Combining Spatial and Temporal Logic
PLN also combines spatial and temporal logic freely, as we shall see in many of the examples presented in Part III. Further, it does so in a manner embracing rich time representations (such as intervals) and uncertainty. However, the full formalization of the combination of spatial and temporal logic is still a subject of current research. One classic paper in this area is titled “Spatial Logic + Temporal Logic = ?” [Kontchakov, 2007], and discusses several ways of combining the two, focusing on combinations of Branching Temporal Logic with spatial logics based on the RCC. This is a big step in the right direction, but doesn’t handle interval representations of time or uncertainty.
# Representing and Reasoning on Contextual Knowledge
One of the critical issues confronted in doing logical inference on large spatiotemporal knowledge bases is the fact that most real-world logical relationships are contextual in nature. Contextuality must be explicitly taken into account to make real-world inference tractable. This is one of the issues with which formal-logic-based AI has traditionally had the most difficulty. In this chapter we briefly explore the notion of contextual inference and the key approaches that have been taken.
Generically speaking, cognitive processes are usually contextual in the sense that they depend on the environment, or context, inside which they are carried out. The notion of context, in different varieties, plays a crucial role in different disciplines including natural language semantics, linguistics, cognitive psychology, and artificial intelligence. Contextual information covers both knowledge representation and reasoning and their interaction.
A completely general representation of a particular piece of commonsense knowledge is impossible in practice, because “common sense” by its nature is not general but has to do with the properties of particular classes of situations. In this sense the only practical way to consider knowledge is as contextual. John McCarthy, one of the founders of the AI field, discussed this topic as follows (referring to axioms pertaining to real-world phenomena, rather than pure mathematics) [McCarthy, 1987]:
“Whenever we write an axiom, a critic can say that the axiom is true only in a certain context. With a little ingenuity the critic can usually devise a more general context in which the precise form of the axiom doesn't hold. Looking at human reasoning as reflected in language emphasizes this point. Consider axiomatizing on so as to draw appropriate consequences from the information expressed in the sentence, 'The book is on the table'. The critic may propose to haggle about the precise meaning of on, inventing difficulties about what can be between the book and the table, or about how much gravity there has to be in a spacecraft in order to use the word on and whether centrifugal force counts. Thus we encounter Socratic puzzles over what the concept means in complete generality and encounter examples that never arise in life. There simply isn't a most general context. Conversely, if we axiomatize at a fairly high level of generality, the axioms are often longer than is convenient in special situations. Thus humans find it useful to say, 'The book is on the table', omitting reference to time and precise identification of what book and what table. [. . . ] A possible way out involves formalizing the notion of context [. . . ]”
Another way to phrase this is to say that, generally, reasoning is local to a subset of all the known facts. One never consider all one knows, but rather a very small subset of it. This small subset, the subset that is used for reasoning about a given goal, determines the context of reasoning. A context encodes an individual's subjective view of some portion or aspect of the world. The individual's complete description of the world is given by the set of all the contexts. There can be different contexts for the same phenomenon and they can describe it at different levels of approximation. The same statement may have different truth values in different contexts. For instance, “Alison is my boss” can be true from John's point of view, but not from Mike's point of view.
As the above quote from McCarthy suggests, the topic of representing contextual information and contextual reasoning has its roots in the early years of artificial intelligence. Also, McCarthy’s was probably the first proposal for rigorously modeling humans’ common sense, the key feature for artificial intelligence. Another important early use of the concept of context in artificial intelligence was [Weyhrauch, 1980], who proposed a mechanized formal contextual reasoning system for Advice taker, a hypothetical computer program, proposed by John McCarthy [McCarthy, 1958] in the 1950’s, that was the first proposal to use formal logic to represent information in a computer.
In the late 80's, interest in related issues grew, and context became a widely discussed issue, with a number of different formalizations and applications in knowledge representation and reasoning. Since the 90’s the concept of context-awareness has increasingly gained importance in the area of computing and distributed systems, due to its promise as a solution for describing mobile computing in ever-changing environments. Most approaches in the literature have focused on addressing the modeling of context with respect to one application or an application class. While some models take the user’s current situation (e.g., “in a meeting”) into account, others model the physical environment (e.g., “in an office”); and some work in this area has taken steps towards a common understanding of context, with respect to location, identity, and time. The objective of this work overall is to develop uniform context models, representation and query languages, and accompanying reasoning algorithms.
In the remainder of this chapter, we give a brief account of some of the major context representation approaches discussed in the literature today, with an emphasis on logic-based approaches. More details and a survey of models of context can be found, for instance, in [Akman & Surav, 1996] and in [Strang & Linnhoff-Popien, 2004], while a survey of applications of context can be found in [Brezillon, 1999]. A broad review of context modeling and reasoning can be found in [Bettini, 2010].
## Logic-Based Context Models
Following [Bouquet et al., 2001], logic based models of context can be divided into two groups: divide-and-conquer and compose-and-conquer.
### The Divide-and-conquer approach
In this approach a context is a way of partitioning (and giving a more articulated internal structure to) a global theory of the world. This global theory has an internal structure, and this structure is articulated into a collection of contexts. Some of the main features of this approach are:
• the facts that are true in a given context c can be isolated and treated as a distinct collection of facts;
• there are hierarchical relations between contexts that allow reasoning to “climb" from a context to a more general context in which the dependence of a fact on a context is explicitly stated;
• there are non-hierarchical relations between facts of different contexts (for example, one would like to be able to represent the relation between facts of the form in_office(person;office_number) and facts of the form in_central_office(person) in a context that specializes the former context to one specific office).
The propositional logic of context (LoC) is one of the theories following this approach. It was first described by McCarthy, and then formalized by Buvac and Mason [Buvac & Mason, 1993]. Some of the most important features of LoC are:
• Any formula can only be asserted in some context (there are no context independent formulae). The fact that a formula p is asserted in a context c is written as c:p.
• Context sequences are used to represent nested contexts. Context sequences allow distinctions, say, between the context of meeting in the context of one company team and the context of meeting in the context of a company.
• There is no outermost context. This means that one can always transcend a context c and move to a more general context in which facts about c can be asserted.
• Statements about a context c are made through the formula ist(c; p), meaning that the formula p is true in the context c.
• some relations between facts belonging to different contexts are stated through the lifting axioms with the general form ist(c; p) ⇒ ist(c0; p0)
For example, if in the context of Alison's office, Alison's computer is close to her chair, then, in the context of their company, Alison's computer is close to her chair.
There is a Hilbert style axiomatization of validity for the logic of context [Buvac & Mason, 1993]. It ensures that all propositional tautologies are valid in every context c.
Next, Dinsmore’s theory of partitioned representations PR [Dinsmore, 1991] is similar to LoC, but in PR there is a special space, BASE, that cannot be transcended and can be considered as a sort of outermost context. In PR, a statement is always asserted in a space, while a space represents some potential reality. Each space has exactly one primary context. A primary context is defined as a function that maps the truth of a statement in one space onto the satisfaction of a (more complex) statement in another space. For example, the sentence “Bob is a project manager” can be asserted in a space S1 and “Alison believes that [S1]” can be the primary context of S1. This allows mapping the truth of “Bob is a project manager" onto the truth of the sentence “Alison believes that Bob is a project manager", which, in turn, is asserted in some other space. Of course, the semantics of “Bob is a project manager” would be very different in a space S2 whose primary context was, for instance, something like “John had a dream that [S2]". Dinsmore also introduces a notion of secondary context, which allows for lateral mappings. Intuitively, a mapping is a consequence of the semantics of the primary contexts involved. In other words, a secondary context opens a channel of communication between two spaces.
### Compose-and-conquer Approaches
In these approaches, a context is a local theory, namely a (partial, approximate) representation of the world, in a network of relations with other local, domain specific theories. There is no global theory of the world, but only many local theories. Each local theory represents a viewpoint on the world and they are the building blocks of the individual’s knowledge. The totality of his/her knowledge is given by composing such local theories through a collection of rules that connect them into a (still partial) representation of the world. A local theory is not a partition of some global theory, but it represents (partial) knowledge about some portion of the world, from a given perspective. For example, a local theory can be a representation of physical objects in an office.
In the compose-and-conquer approach, there are no a priori relations between contexts. This is a major difference from divide-and-conquer theories. Namely, in divide-and-conquer systems, global knowledge implies how contexts are related to each other. On the other hand, in compose-and-conquer theories, there is no predefined global knowledge and contexts are autonomous theories. This does not mean that there are no relations between contexts, but only that these relations are established on a peer-to-peer basis.
One example of a compose-and-conquer approach to contextual knowledge representation and reasoning is Ghidini and Giunchiglia's Local Models Semantics [Ghidini & Giunchiglia, 2001], together with its proof-theoretical counterpart [Giunchiglia & Serafini, 1994]. It is based on the following two very general principles:
• principle of locality: reasoning always happens in a local theory (a context);
• principle of compatibility: there may be compatibility constraints between the reasoning processes that happen in different contexts.
### Compatibility constraints
Compatibility constraints represent relations between different contexts. These may be formalized in many different ways.
For instance, in the logic based approach to contextual knowledge representation described in [Ghidini & Giunchiglia, 2001], one starts with a family of languages L1;…,Ln;…, where Li is the representation language of a context ci. Each language Li has its set of models Mi. Every subset MTi of Mi satisfies a set of formulae, each corresponding to a different choice of the theory Ti associated with ci. Once the theory Ti associated with ci is fixed, a model belonging to MTi is called a local model of ci. And then, relations between two contexts are represented by compatibility constraints, which state that the truth of a formula F in c1 is related to the truth of the formula F’ in c2. This is achieved by imposing that: sets of local models c1 and c2 of the two contexts c1 and c2 are such that, if the set of local models of c1 satisfies F, then the set of local models of c2 satisfies F’.
Pairs <c1; c2> satisfying the above relation are said to belong to a compatibility relation and define a model for the pair of contexts c1 and c2. Using the notion of compatibility, a wide range of relations between contexts can be formalized. For example, if a context c1 represents Bob's beliefs at day d, and that it contains the statement “I have a board meeting” and if c2 represents Bob’s beliefs at day d+1, then there can be a relationship between the two contexts such that the sentence “Yesterday I had a board meeting" is true in c2. The above relations are based on considering models, while there is also a deductive counterpart to imposing compatibility constraints.
## Other Approaches to Contextual Knowledge Representation
Contextual knowledge representation is used in many different fields, including modeling psychological and cognitive processes. The most important real-world application of contextual knowledge representation at present is in distributed and mobile computing, an area that has to deal with computing devices working in changing contexts. More pertinently to the present book, context based frameworks have been also used in knowledge and data integration – for the integration of information (or knowledge) coming from different sources (see, for instance, [Farquhar et al., 1995]).
Different information sources integrated in a unique system can be thought of as partial views (or contexts) on a world. Available information sources are often distributed, redundant, partial, and autonomous. Consequently, information sources may adopt different conceptual schemata (including domains, relations, naming conventions, etc). The relations between the different domains and between the interpretation domains of the different databases can be established.
For instance the different (but related) meanings of the predicate costs(x; y) in database 1 and database m can be represented by using the following formula:
1 : costs(x, y) ⇒ m : ∃y’ (costs(x,y’) ∧ y’ = 1.07*y)
The meaning of the above relations is that if the models of database 1 satisfy the formula costs(x; y), then any models of the database m must satisfy the formula
∃y’ (costs(x,y’) ∧ y’ = 1.07*y)
which means that item x has price y’ which is obtained by adding taxes to the price y.
One application of these ideas is to traditional relational databases. For instance, in one research project, a context-based knowledge was applied to the problem of specifying redundancy among different databases while maintaining a high degree of autonomy [Mylopoulos & Motschnig-Pitrik, 1995]. In this work, there are mechanisms for change propagation: i.e., mechanisms that establish whether the effects of a change operation performed in one context are visible in other contexts.
In another sort of application, a rich contextual knowledge representation is required for the maintainance of a massive knowledge base such as CYC [Lenat & Sierra, 1999]. The (controversial) philosophy underlying the CYC AI project is that huge, hand-created knowledge bases covering a wide area of (human) knowledge are required in order to create generally intelligent programs (the authors give an example of a logic-based expert system with the rule that amphibians lay eggs in water; however, the system cannot answer the question whether amphibians lay eggs at all, because its rule base is not sufficiently rich to encompass such commonsense facts). However, maintaining and using a huge knowledge base, brings different challenges, one of which is the use of contexts, as the Cyc team has observed:
“… as the CYC common sense knowledge base grew ever larger, it became increasingly diffcult to shoehorn every fact and rule into the same at world. Finally, in 1989, as CYC exceeded 100,000 rules in size, we found it necessary to introduce an explicit context mechanism. That is, we divided the KB up into a lattice of hundreds of contexts, placing each CYC assertion in whichever context(s) it belonged."
Contexts in CYC have a fine internal structure with a dozen mostly-independent dimensions along which contexts vary. Each region of this 12-dimensional space implicitly defines a context. The capability of importing an assertion from one context into another is provided by lifting assertions. They have a general form:
ist(c; p) ⇒ ist(c0; p0)
For instance, the fact that gold is more expensive than silver in the stock market, can be exported to the context of the black market by the assertion:
ist(StockMarket; MoreExpensive(gold,silver))
ist(BlackMarket; MoreExpensive(gold,silver)
## Contextual Knowledge in Probabilistic Logic Networks
PLN has a special link to handle contextual knowledge called ContextLink (or simply Context). For instance one can express in the context of programming a method is a class function by
Context
Programming
Similarity
ClassFunction
Method
The semantics of ContextLink is formally defined by the following equivalence:
Context
C
R A B
is equal to
R (A AND C) (B AND C)
Context can be mixed, as above, or else purely intensional or extensional. The purely intensional version is defined as follows:
InternationalContext
C
R A B
is equal to
R $\displaystyle ($ A AND_Int C$\displaystyle )$ $\displaystyle ($ B AND_Int C$\displaystyle )$
So reasoning with context simply amounts to reasoning with non-contextual standard operators, in certain stereotypical patterns. However since contexts are often used it may be valuable to introduce some inference “macro rules” handling context, like the following, used in an inference example given later on:
Context C A
Implication A B
Context C B
Informally this rule says that if A is true in the context C and A implies B then B is true in the context C.
Also, one may note that the definition involving the Context link assumes that the context applies on a relationship (noted R in the definition given above), nevertheless a truth value over a concept A can be seen as the measure of how much the universe inherits A, that is:
A $\displaystyle $
is equal to
Inheritance <w>
U
A
where U symbolizes the universe. Given this, we can define contextual versions of concepts as well as relationships.
## User Models as Contexts
One area where contextual logic has not frequently been applied, but perhaps should be, is the modeling of the preferences, biases and other properties of individual users and user groups of information systems. (Bry and Jacquenet [Bry and Jacquenet, 2005] have laid out the logic for application of first-order logic to user modeling in the semantic web in some detail, but they do not consider contextual logic in specific.)
### User Modeling in Information Retrieval Systems
A great deal of work has been done to date on user modeling in information retrieval (IR) systems. The relevance of the output of an IR system is not absolute, but relative to specific users; and because of this, the lack of user modeling is one of the main sources of weakness of most contemporary such systems. For example, a tourist and a programmer may use the same word “java” to search for different information, but many retrieval systems would return the same results. It was observed, already in late 70s and early 80s, that information retrieval systems can and should be personalized for users by means of profiles [Rich, 1997; Myaeng et al., 1986]. Since then, a lot of efforts have been invested in the area of individual user modelling and individual user profiles.
The main objective of user modeling is to extract and store information about a user, and to adapt the retrieval tool to the user's needs and interests in order to improve the relevance of the results. Such, personalized search is still considered as one of the major challenges in modern information retrieval. For estimating the user-specific relevance of documents different computing methodologies are used, ranging from genetic algorithms, fuzzy logics, modal logic, to classical first-order logic approaches.
User profiles can model different users' preferences. For example, for some users it is preferable to have the list of all results to the query quickly, even with a number of bad results. On the other hand, some users prefer to have a short list of results, even if producing it takes more time. Some users prefer results only in some specific file formats, etc. These criteria and also criteria that define quality or relevance for a specific user define his/her profile. Generally, a (user) profile consists of a set of preferences with regard to behavior of a search engine as well constraints on the results it presents to the user [Van Gils & Schabell, 2003]. For example, the following criteria may define a particular user-profile:
I prefer a maximum of 25 results per page, and by selecting a relevant resource (clicking on the link) will open a new window. I prefer HTML and PDF formats and refuse the Microsoft DOC format. Furthermore, the size of the resource should not exceed 25Mb.
Profiles can be used for post-processing the results to the query (e.g., some resources can be converted into a prefered format), and most interestingly, can also be used for guiding search.
An important distinction in a user modeling context is between explicit and implicit preferences. Explicit preferences consist of information given by a user explicitly. On the other hand, implicit preferences refer to any context information available to the information retrieval system during a user's session. Relevance feedback (given by the user to the information retrieval system) is one way for providing more context explicitly and can be effective for improving retrieval accuracy. However, it is often unrealistic to motivate user to explicitly rate the results obtained. Therefore, implicit context information and implicit feedback is thus more interesting to exploit and it has been attracting more and more attention (see, for instance, [Kelly & Teevan, 2003] for a bibliography of implicit feedback). In addition, both user’s explicit and implicit preferences may change over time.
There are different sources of implicit user preferences. For instance, the user often need to modify his query in a number of iterations until he is satisfied. In such scenario, the information to be used by the information retrieval system is not just the current query, but also the complete user's search history, information about which documents the user has chosen to view, and even for how long the user has read specific documents, etc. For instance, a simple user’s profile can be based on keywords from his queries performed or documents that he read.
Many approaches use user's actions and navigation through data as implicit feedback. For instance, some web browsers record user actions and navigation, including dwelling times, mouse clicks, mouse movements, scrolling and elapsed times and user explicit ratings of web pages. Some experimental results show that implicit feedback, especially the dwelling time on a page, amount of scrolling on a page, the combination of time and scrolling, how a user exit a result or end a search session have all a strong correlation with explicit relevance ratings [Claypool et al., 2001; Fox et al., 2005]. However, there are also some experimental results showing that there is no general direct relationship between display time and relevance (and that the display time depends on the specific tasks and specific users) [Kelly & Belkin, 2004].
There is a range of techniques for using user’s query history in building his/her profile. Some web browsers use browsing history in past n days for personalized search. Some approaches store user’s interests and can distinguish between long term interests and ad hoc queries. Thank to this, modification of queries that can not appropriately be supported by the user profile are not applied.
User modeling is widely used not only in Internet browsing, but also in recommender systems providing advice to users about items they might wish to purchase or examine. Recommendations made by such systems can help users navigate through large information spaces of product descriptions, news articles, social information, or other items [Burke, 2000].
A general survey covering many aspects of user modeling in information retrieval systems can be found in [Kobsa, 2007]. Another survey of the field, including extensive discussion of different forms of communication with the user (querying, navigation through structures, and visualization) can be found in [Nurnberger & Detyniecki, 2005]. A survey of the field of adaptive user interfaces can be found in [Dieterich et al., 2003].
### User modeling from the cognitive perspective
Beyond the specific domain of information retrieval, there are several types of computer related models considered by cognitive psychology. Apart from the user models – the computer's model of the user (as described above in the context of information retrieval systems), there are also
• mental models – user's model of the system;
• conceptual models – models presented to the user by the system designer.
User models can be splitinto two major categories: empirical quantitative models and analytical-cognitive models [Palermiti & Polity, 1995].
Empirical quantitative models are based on users’ performances when using a given system. Empirically collected performance data are used for constructing abstract formalizations and for defining groups (e.g., skill groups such as "experts”, “beginners"). These user models do not express beliefs, reasoning or cognitive processes, but rather external performances.
Analytical cognitive models attempt to construct a more qualitative behavioral modeling. The models aim to detect the purposes, strategies, plans or beliefs of the user, so that the system may issue predictions and draw conclusions. These models are not static but dynamic, able to evolve according to different tasks or categories of users, and adaptive, by using the cognitive features detected in the user behavior. Their role is to help the system cooperate, to indicate hypotheses, and to single out areas of interest.
User models can also be classified according to three main dimensions:
• models of a single "typical" user versus collections of individual models;
• explicit models defined by the designer versus models inferred by the computer on the basis of the user's behavior;
• long term user characteristics such as areas of interest and expertise versus short term characteristics such as the subject of the last sentence typed.
### Logic based user modeling
Instead of using approximate models and reasoning, logic based user modeling aim at employing formal reasoning for deducing facts about the user. This family of user modeling approaches is relatively small, but has a number of important ideas and techniques.
Some of the logic-based representations used in user modeling are [Kobsa, 1993]:
• PROLOG-based, which offers a comparatively rich representation language with built-in backward reasoning and the possibility of a smooth migration from knowledge representation to programming;
• Predicate logic, which offers more expressiveness than PROLOG, as is needed in many application domains;
• Languages with second-order predicates and modal logic, which allow representing assumptions about beliefs and goals of different agents in the same representation language;
• Connectionist networks, which have been particularly employed for classification tasks.
An overview of different reasoning techniques used in logic-based user modeling systems can be also found in [Kobsa, 1993]. Typically, most logic-based systems use simple forward or backward reasoning (using the knowledge from the knowledge base), or some their combinations.
### Contextual Logic for User Modeling
One potentially promising avenue for applying logical methods to user recommendation involves deploying contextual logic, in such a fashion that each user connotes a distinct context for inference. As this approach seems not to have been touched in the literature yet, we explore it here via a “thought experiment” style example.
Our examples concerns book recommendation systems which, when a user locates one book, recommend to him/her other books of potential interest. In such systems today, models of individual users are not necessarily built at all; instead most recommendations are generic – the same for all users locating a certain book. Typically, relationships between books, as a basis for recommendations, are inferred from purchases of individual users, i.e., from the database of all purchases. For a certain book A, it is calculated how many users that bought the book A also purchased a book B, a book C, etc. Then the books (a fixed number of them – say 10) with the highest score are recommended to the user, as of potential interest. This simple and generic model, commonly used in practice, can be refined in many ways, for instance by applying a specific ordering among the 10 recommended books, derived with respect to the specific user. Let us consider one possible approach for this through a simple example.
Suppose there are the following books available:
(A) Katya Walter The Tao of Chaos
(B) Katya Walter: Dream Mail: Secret Letters for Your Soul
(C) Martin Schonberger: I Ching & the Genetic Code: The Hidden Key to Life
Suppose there is also stored the relevance of these books to two areas: spiritual literature and popular science. These relevance data (given by weight factors from 0 to 1) may then be taken to form two contexts:
Popular science context: relevance(A,0.5) relevance(B,0.2) relevance(C,0.7)
Spiritual literature context: relevance(A,0.6) relevance(B,0.8) relevance(C,0.3)
Let there also be models of two users, Alice and Bob, and their interests, again described by weighting factors. Assume these weighting factors are computed based on their past purchases.
Alice’s model: 1. interested_in(popular science,0.2) 2. interested_in(spiritual literature,0.7)
Bob’s model: 1. interested_in(popular science,0.9) 2. interested_in(spiritual literature, 0.2)
Now, let us assume that Alice searches the book recommendation system and locates the book (A). The engine queries the database and finds that other users that purchased this book most often also purchased books (B) and (C). So, these two books should be recommended to Alice in some suitable ordering and the relevance of these books to Alice should be estimated.
Information from the contexts of certain areas apply to all specific users; so, in our example, all facts from all area contexts are also used in specific user’s contexts by, so-called, bridge rules:
Inference rule in User N’s model:
Context for the area X: relevance(Y,w1) interested_in(X,w2) relevance(Y,w3) relevance(Y ,max(w3,w1*w2))
The meaning of this rule is as follows: if, in the context of the area X, the relevance of the book Y is w1, and if w2 is the weight of the user N’s interest in X, then the relevance of the book Y for the user N is w1*w2. Such relevance factors can be calculated for all available areas, and the final relevance factor for the book Y is the maximum over all areas (the initial relevance factor is 0).
In our example, we should apply, within Alice’s model, the above rule twice for each of the books (B) and (C):
Popular science context: relevance(B,0.2) interested_in(popular science,0.2) relevance(B,0) relevance(B, 0.04)
Spiritual literature context: relevance(B,0.8) interested_in(spiritual literature,0.7) relevance(B, 0.04) relevance(B, 0.56)
Popular science context: relevance(C,0.7) interested_in(popular science,0.2) relevance(C,0) relevance(C, 0.14)
Spiritual literature context: relevance(C,0.3) interested_in(spiritual literature,0.7) relevance(C, 0.14) relevance(C, 0.21)
Hence, the estimated relevance of the book (B) for Alice is 0.56, and the relevance for the book (C) is 0.21 and the ordering in recommendation is (B), (C).
We can also apply, within Bob’s model, the same rules:
Popular science context: relevance(B,0.2) interested_in(popular science,0.9) relevance(B,0) relevance(B, 0.18)
Spiritual literature context: relevance(B,0.8) interested_in(spiritual literature,0.2) relevance(B, 0.16) relevance(B, 0.18)
Popular science context: relevance(C,0.7) interested_in(popular science,0.9) relevance(C,0) relevance(C, 0.63)
Spiritual literature context: relevance(C,0.3) interested_in(spiritual literature,0.2) relevance(C, 0.06) relevance(C, 0.63)
Hence, the estimated relevance of the book (B) for Bob is 0.18, and the relevance for the book (C) is 0.63 and the ordering in recommendation is (C), (B).
It is not hard to see how the same form of contextualization could play a role in other sorts of user modeling besides product purchase investigations. For instance, suppose Bill and Betty are two criminal investigators, investigating a crime involving an individual named Giulio (A). Suppose Giulio has connections to
(B) Elias, a known drug dealer who is suspected to have sold drugs to Giulio's uncle
(C) Jonas, a banker convicted of fraud, who works at the bank where Giulio used to work
Suppose that Bill's expertise is in drug crimes, whereas Betty's expertise is in bank fraud. Then, in the context of Bill's investigation, the association between A and B should be more prominent; whereas in the context of Betty's investigation, the association between A and C should be more prominent. The structure is exactly the same as in the book purchasing example. (And of course there are many subtler examples of how contextual reasoning could be used to aid in user modeling; we have just presented a simple sort of example to make the conceptual connection clear.
## General Considerations Regarding Contextual Inference
Now we turn to issues involving the combination of contextual representation with logical inference systems. The explicit incorporation of context into inference is a subtle matter; a number of approaches have been posited in the research literature, yet none have yet been battle-tested in real-world applications. Here we will review some general considerations regarding contextual inference, then go into more detail regarding PLN-based contextual inference, and give a fairly complex, realistic example of the latter.
Generally speaking, any deductive system (say, Hilbert style first-order logic, or temporal logic) can be used contextually, by restricting its rules to certain domains. But this is different than explicitly contextual reasoning, in which a notion of context is incorporated in both the knowledge representation and the inference rules. In practice, the way explicitly contextual reasoning systems tend to work is that inference rules are supplied for specific contexts, and then special rules are also applied for bridging pairs of contexts.
For example, the relationship “to be superior” can be described by the following (implicitly) universally quantified axioms:
team_leader$\displaystyle (x,z)\Rightarrow superior(x,z)$
superior(x,y) \land superior(y,z) \Rightarrow superior(x,z)[/itex]
The axioms can be associated by an additional parameter describing a context – for instance, a company or a sport club. They can then be exported, by lifting rules, to specific contexts:
team_leader$\displaystyle (x,z,context) \Rightarrow superior(x,z,context)$
superior$\displaystyle (x,y,context) \land superior(y,z,context) \Rightarrow superior(x,z,context)$
or, to the following notational form:
context: team_leader$\displaystyle (x,z) \Rightarrow superior(x,z)$
context: superior$\displaystyle (x,y) \land superior(y,z) \Rightarrow superior(x,z)$
The basic idea here, due originally to McCarthy, is to tackle the problem of generality by using more general axioms if the context is irrelevant, and less general axioms otherwise.
In addition to inference rules local to specific contexts, contextual inference systems also utilize rules that link different contexts. For instance, in the inference system described in [Ghidini & Giunchiglia, 2001], there are bridge rules. Bridge rules are rules whose premises and conclusion belong to different contexts. For instance, the bridge rule corresponding to the compatibility constraint.
if the set of local models of C1 satisfies F, then the set of local models of C2 satisfies F’
would be the following:
$\displaystyle \dfrac{C_1:F}{C_2:F'}$
where C1:F is the premise of the rule and C2:F’ is the conclusion. Obviously, bridge rules are conceptually different from local rules, rules used in individual contexts. Bridge rules can have different forms and can involve more than two contexts. A deduction is a tree of local deductions, obtained by applying only local rules, concatenated with one or more applications of bridge rules. In a special case when there is no relation between the two contexts, there are no constraints on what is true in the two contexts.
Using the machinery of compatibility, a wide range of relations between contexts can be formalized. For example, let c1 represent the prices on the American market at the present moment and assume “Gold gets cheaper" is true in C1. If C2 represents the prices on the Asian market at the present moment, then the sentence “In America, gold gets cheaper" must be true in C2. This conclusion is based on the inference rule of the form:
$\displaystyle \dfrac{C_1:F}{C_2:F \quad \text{is true in} C_1}$
### Uncertain Contextual Inference
Although there is very little literature in this area, there is no reason contextual reasoning can’t be applied in the context of uncertain logics.
For instance, to “uncertainize” the example given above, the fact “Gold gets cheaper” could have a degree of certainty of 0.9 in the context c1 of the American market. Furthermore, the relevant bridge rules could also have degrees of certainty associated with them. So, for instance, there could be a bridge rule
$\displaystyle \frac{C_1 : \text {Gold gets cheaper}} {C_2 : \text {Gold gets cheaper}}$
where c1 is the context of the American market, while c2 is the context of the Asian market. This rule could have a degree of certainty, say, 0.8, indicating that it is likely that if gold gets cheaper on the American market, that it also gets cheaper on the Asian market. So, finally, the degree of certainty that gold gets cheaper on the Asian market could be calculated as the product of the degree of certainty that gold gets cheaper on the American market (0.9) and the degree of certainty of the bridge rule (0.8), giving 0.72.
On the other hand, there could be a bridge rule
$\displaystyle \frac{C_1 : \text {local currency gets stronger}} {C_2 : \text {local currency gets stronger}}$
where c1 is the context of USA, while c2 is the context of Japan. This rule could have a degree of certainty, say, 0.2, indicating that it is unlikely that if US dollar gets stronger, then yen also gets stronger. If the degree of certainty that US dollar gets stronger is, say, 0.95, then the degree of certainty that yen gets stronger is 0.95*0.2=0.19.
## A Detailed Example Requiring Contextual Inference
In this section we present a simple, specific example of contextual inference, and explain how the ideas discussed above may be applied in this case. In working this example, we will use the approach close to [Ghidini & Giunchiglia, 2001] and first order logic as the underlying logic, but, for the sake of simplicity, instead of first-order notation, we will formulate the statements in natural language form. In a following section, we will run through this same example using PLN inference.
The basic assumptions of the inference, expressed in common informal English, are as follows:
Alison is an accountant who is also a musician. Alison is emotional in the context of music, but not in the context of accounting. She frequently mentions Canadian place names in the context of music (maybe she's a Canadian music fan), but not in the context of accounting. Bob is in a similar situation, but he frequently mentions Canadian related stuff in both the music and accounting contexts. Clark is also in a similar situation, but he frequently mentions Canadian related stuff only in the accounting context, not the music context. Trivially, Canadian places are associated with Canadian people. People who have a lot to do with Canadian people, and a lot to do with money, have a chance of being involved in suspicious money-laundering activities. Trivially, accounting has to do with money.
To formalize the above using contextual inference, we will consider the following contexts: music, accounting, to-be-involved, and Canada. We can encode all the above information as axioms in these contexts, while we also have to add some bridge rules (linking different contexts and corresponding to compatibility constraints). Each inference step is made in a specific context. If a person is an accountant, we can encode this information as being that the person is knowledgeable in the context of accounting. It is similar for the context of music.
More explicitly, the contexts under consideration are as follows:
Music context: (1) Alison is knowledgeable. (2) Alison is emotional. (3) Alison frequently mentions Canadian place names. (4) Bob is knowledgeable. (5) Bob is emotional. (6) Bob frequently mentions Canadian place names. (7) Clark is knowledgeable. (8) Clark is emotional. (9) Clark does not frequently mention Canadian place names.
Accounting context: (10) Alison is knowledgeable. (11) Alison is not emotional. (12) Alison does not frequently mention Canadian place names. (13) Bob is knowledgeable. (14) Bob is not emotional. (15) Bob frequently mention Canadian place names. (16) Clark is knowledgeable. (17) Clark is not emotional. (18) Clark frequently mention Canadian place names. (19) Money is important.
To-be-involved context: (20) If someone is knowledgeable and X is important, then he/she is highly involved with X. (21) If someone frequently mentions place names from X, he/she is highly involved with these places. (22) If someone is highly involved with places from X, he/she is highly involved with anything that is associated with these places.
Canada context: (23) Canadian places are associated with Canadian people (24) If someone is highly involved with Canadian people and with money, then he/she has a chance of being involved in money-laundering.
To keep things relatively comprehensible and informal, we won’t explicitly specify the languages of the contexts, and we’ll assume that all relevant information can be expressed in any of the contexts (for instance, we assume that in the music context it can be expressed that someone is involved with Canadian people). Also, we use the following bridge rules:
$\displaystyle \frac{\text {To be involved}: F} {\text {Music}: F}$
$\displaystyle \frac{\text {Canada}: F} {\text {Music}: F}$
$\displaystyle \frac{\text {To be involved}: F} {\text {Accounting}: F}$
$\displaystyle \frac{\text {Canada}: F} {\text {Accounting}: F}$
In the described system, for example, one can infer that Clark has a chance of being involved in money-laundering (in the context of accounting).
The inference in the accounting context goes as follows:
Step Inferred information Justification (C1) If someone is knowledgeable and X is important, then he/she is highly involved with X. bridge from (20) (C2) If someone is knowledgeable and money is important, then he/she is highly involved with money. instance of (C1) (C3) Clark is highly involved with money. from (16), (19), and (C2) (C4) If someone frequently mentions place names from X, he/she is highly involved with these places. bridge from (21) (C5) If someone frequently mentions place names from Canada, he/she is highly involved with these places. instance of (C4) (C6) Clark is highly involved with places from Canada. from (18) and (C5) (C7) If someone is highly involved with places from X, he/she is highly involved with anything that is associated with these places. bridge from (22) (C8) If someone is highly involved with places from Canada, he/she is highly involved with anything that is associated with these places. instance of (C7) (C9) Canadian places are associated with Canadian people. bridge from (23) (C10) Clark is highly involved with Canadian people. from (C6), (C9), and (C8) (C11) If someone is highly involved with Canadian people and with money, then he/she has a chance of being involved in money-laundering. bridge from (24) (C12) Clark has a chance of being involved in money-laundering. from (C3), (C10), and (C11)
Notice that one can prove that Alice is highly involved with Canadian people in the context of music, but cannot prove that Alice is highly involved with Canadian people in the context of accounting. Hence, it cannot be proved that Alison has a chance of being involved in money-laundering.
If one use a bridge rule connecting the two contexts:
$\displaystyle \frac{\text {music}: F} {\text {accounting}: F}$
for all sentences F that can be expressed in the two contexts, then one can infer that Alison is highly involved with Canadian people in the context of accounting, too (if all relevant properties can be expressed in that context). However, such a bridge rule would be unlikely used.
Notice also that, in the same manner as for Clark, one can prove that Bob is highly involved with Canadian people in the context of accounting, and further that Bob has a chance of being involved in money-laundering. One may wonder if there can be made a distinction between Bob and Clark. Namely, Bob is highly involved with Canadian people in both contexts music and accounting, while Clark is highly involved with Canadian people only in the context of accounting. In addressing this there are several issues. First, one might eliminate the axiom (24), and instead add the following bridge rule:
$\displaystyle \frac{ \begin{array}[b]{r} \left( \text {Accounting}: X \text{is involved with Canadian people and with money} \right)\\ \left(\text {Music}: X \text{is not involved with Canadian people} \right) \end{array} }{ \left( \text {Accounting}: \text{has a chance of being involved in money-laundering} \right) }$
Intuitively, this would eliminate the conclusion that Bob has a chance of being involved in money-laundering and keep the conclusion for Clark. However, it is not necessarily the case. Namely, in the context of music, one cannot prove that Clark is involved with Canadian people, but it still does not mean that one can prove that Clark is not involved with Canadian people. Different default logics address this issue and hence can be used for reasoning of the above sort.
Using the variants of the systems described above one can infer crisp conclusions as to whether Alison, Bob, or Clark have a chance of being involved in money-laundering or other information. However, one cannot infer information on how likely it is that Alison/Bob/Clark is involved in money-laundering. For such information, one might use some fuzzy/probabilistic logic as the underlying logic and keep the rest of the inference system. In such a modified system, one could end up with a conclusion that Clark is the most likely to be involved in money-laundering, Bob is second most likely, and Alison is third most likely. Also, if one would replace “accounting” with “marketing” in the above examples, then these degrees of suspicion should decrease, while if “accounting” is replaced with “narcotics”, then the degrees of suspicion should increase. However, these specifics will depend on the inference system in question.
# Causal Reasoning
Many of the inferences one wants to draw about real-world, spatiotemporal, contextual knowledge involve cause and effect. But the relation between causation and logical inference is subtle and storied. As outlined above, deductive reasoning aims at deriving consequences (or effects, outcomes) from premises (or causes). Abductive reasoning aims at deriving possible causes from effects. Finally, inductive reasoning aims at deriving relationships between causes and effects, rules that lead from one to another. Causal reasoning is generally considered a form of inductive reasoning. More concretely, causal reasoning aims at an epistemological problem of establishing precise causal relationships between causes and effects, with focusing on detecting genuine, real causes for some effects, and genuine, real effects of some causes.
Although they share a lot in their background and inferential process, causal inference and statistical inference are different. Statistical inference is concerned with associational inference and used for finding associations between exposures (causes) and outcomes, rather then for inferring causation relationships from observations. Causal relationship implies correlation between two events, but the opposite does not hold.
The first attempts at dealing with causality date back to the old Greek philosophers, including Aristotle, but systematic approaches had to wait centuries to come [Danks2006]. In his Novum Organum (1620), Francis Bacon introduced the notions of:
• The table of presence (tabula praesentiae)
• The table of absence (tabula absentiae)
• The table of degrees (tabula graduum)
According to Bacon, the cause of a phenomenon is the set of properties that explains every case in each of the three tables. In the mid-nineteenth century, John Stuart Mill proposed an improved form of Bacon’s method. But already by then, the limitations of such approaches were understood – since in the eighteenth century, David Hume noticed that causal inference, often intuitive and natural, cannot be formally justified using deduction. This fact is actually a general property of inductive reasoning. It is not possible to justify the pattern “the future will continue the pattern of the past” via deductive reasoning based on observations, without falling into circular reasoning. One can reason “the sun will rise tomorrow morning because it rose for the last 100 mornings”, but this relies on the assumption “the future will continue the pattern of the past”, and it’s circular to say “the future will continue to obey the pattern ’the future will continue the pattern of the past’ because it obeyed it in the past.”
From the twentieth century till the present, philosophers, statisticians, and computer scientists have developed, often building on statistics, various approaches and methods for representing causal structures and solving problems of causal inference. Some of them follow the view of the philosopher Karl Popper in which falsification of a hypothesis is more informative than corroboration of a hypothesis. There could be a number of cases that are consistent with a false hypothesis, but a single counterexample requires modifying the hypothesis. A hypothesis that has survived many attempts to refute it is more likely to be true than one that has been corroborated many times.
Today as in the past, the problem of causal inference is a central challenge for most of the empirical sciences. Causal effect may be the effect of a given drug or therapy for a specific disease, the effect of education on employment and earnings, the effect of training courses on the labor market, etc. If a causal relationship is discovered and established, it may be possible to control future events to some extent by producing (or preventing) occurrence of certain outcomes. There are many real-world applications of causal reasoning, including in economics, law, social sciences, and human-computer interactions, but most important are perhaps those in various branches of medicine and health research.
In the following text, we will briefly discuss some of the central problems in causal reasoning and some of the most significant approaches for modeling it. For a more detailed survey, see, for instance, [Kluve, 2001].
## Correlation does not imply causation
The assumption that correlation and causation are equivalent often leads to significant flaws in reasoning. The phrase “correlation does not imply causation” stresses this fact, meaning that if there is a correlation between two events, it still does not necesserily mean that one causes the other (although it is possible that it is the case). A prototypical example of such flawed reasoning, employed in medicine, is discussed in [Lawlor, 2004]. A number of studies showed that women who were taking combined hormone replacement therapy also had a lower-than-average incidence of coronary heart disease. This correlation led to a widespread idea that hormone therapy was protective against coronary diseases. However, carefully designed experiments demonstrated that this was not true. The explanation was that women taking hormone therapy were mostly from higher socio-economic groups, with better nutritive regimes and, hence, with lower incidence of heart disease. Thus, lower incidence of heart disease was not a consequence of taking hormone therapy, but both were effects of a common cause.
Coming to the conclusion that an event X is caused by an event Y if there is a correlation between the two is generally wrong, although this causation can be indeed present. Other potential explanations for correlation between X and Y are the following:
• Y is caused by X.
• It is both the case that X is caused by Y and Y is caused by X.
• There is a common cause for both X and Y.
• Correlation is present due to a pure chance, or due to reasons that are so complex and deep that they cannot be considered as causation between X and Y.
## Other Challenges in Causal Reasoning
Causal relationships can be very complex and difficult to deal with. For instance, there are situations when there are several hypotheses about a causal relationship, and it is difficult to select the one that is most likely true. Also, one of the problems is recognizing irrelevant events or causes with small impact on the observed outcome. Let us consider the example, given in [Modern Epidemiology, 2008], that illustrates the process of understanding causality with a description of a child learning that moving a light switch causes the light to turn on. However, in wider contexts, the turning-on of the light was sometimes caused also by other subjects:
• The mother who replaced the burned-out light bulb.
• The electrician who replaced a defective circuit breaker.
• The lineman who repaired the transformer that was disabled by lightning.
• The social service agency that arranged to pay the electricity bill.
• The power company, the political authority awarding the franchise, the investment bankers who raised the financing, the Federal Reserve that eased interest rates, the politician who cut taxes, and the health care providers who contributed to the child's safe birth and healthy development.
And there are other deep issues with causal reasoning. Consider an example from [Pearl, 2000], illustrating the problem known as Simpson’s Paradox. Suppose that a teacher discover that a higher proportion of his students who smoke received a final grade of A than students who do not smoke, as shown in Table 8.1a. Confused by the hint that smoking has a positive impact on grades, the teacher tries to find a rational explanation and partition the same data differently, looking at students with low parental income (Table 8.1b) separately from those with high parental income (Table 8.1c). Then, surprisingly, he finds that the situation has been completely reversed: smoking has negative impact on grades in both groups.
A B %A smokers 5 5 50% non-smokers 2 3 40% a: Proportion of Student smokers and non-smokers who received A A B %A smokers 5 4 56% non-smokers 1 0 100% b: Low-income students A B %A smokers 0 1 0% non-smokers 1 3 25% c: Low-income students
Table 8.1: Illustration of Simpson’s Paradox
According to Pearl, Simpson’s Paradox comes from trying to understand causality solely through probability and statistics: “It is an embarrassing yet inescapable fact that probability theory, the official language of many empirical sciences, does not permit us to express sentences such as ’Mud does not cause rain’; all we can say is that the two events are mutually correlated, or dependent—meaning that if we find one, we can expect to encounter the other” [Pearl, 2000].
## Mill's Methods
In his 1843 book A System of Logic: Ratiocinative and Inductive, John Stuart Mill described five methods for drawing conclusions about causal relationships. The account of the Mill’s methods given here and the running example are based on [Kemerling, 2006]. The running example is based on the following scenario: in a college, one day an unusual number of students are suffering from severe indigestion. The college nurse naturally suspects that this symptom results from something the students ate for lunch, and she wants to find evidence that will support a conclusion that eating a certain food caused indigestion. The Mill’s methods are the following.
• Method of Agreement: this method applies to cases in which the effect that occurred reveals only one prior circumstance that all of them shared. It is based on the expectation that similar effects are likely to arise from a similar cause. As an example, suppose that out of the four students with indigestion, one had pizza, coleslaw, orange juice, and a cookie; the second had a hot dog and french fries, coleslaw, and iced tea; the third ate pizza and coleslaw and drank iced tea; and the fourth ate only french fries, coleslaw, and chocolate cake. The nurse, by the method of agreement can conclude that eating coleslaw caused the indigestion.
• ''''Method of Difference: by this method, a comparison of a case in which the effect occurred and a case in which the effect did not occur, reveals that only one prior circumstance was present in the first case but not it the second. In such situations, it is supposed that, other things being equal, different effects are likely to arise from different causes. As an example, suppose that the two students with indigestion ate together, but one became ill while the other did not. The first had eaten a hot dog, french fries, coleslaw, chocolate cake, and iced tea, while the other had eaten a hot dog, french fries, chocolate cake, and iced tea. Again, the nurse concludes that the coleslaw is what made the first student ill.
• Joint Method of Agreement and Difference: This method is a combination of the first two methods, and it assumes that genuine causes are necessary and sufficient conditions for their effects. Consider the following example: eight students come to the nurse: four of them suffered from indigestion, and with each of these four there is another who did not. Each pair of students had exactly the same lunch, except that everyone in the first group ate coleslaw and no one in the second group did. The nurse arrives at the same conclusion as above.
• Method of Concomitant Variation: this method applies when evidences appear to show that there is a correlation between the degree to which the cause occurred and the degree to which the effect occurred. This conforms to our exprectations that effects are typically proportional to their causes. This method does not only notice occurrence or non-occurrence of the causal terms, but also the extent to which each of them takes place. As an example, suppose that there are five students with indigestion: the first ate no coleslaw and feels fine; the second had one bite of coleslaw and felt a little queasy; the third had half a dish of coleslaw and is fairly ill; the fourth ate a whole dish of coleslaw and is violently ill; and the fifth ate two servings of coleslaw and had to be taken to the hospital. The conclusion is again that coleslaw caused the indigestion.
• Method of Residues: many elements of a complex effect are shown to result, by reliable causal beliefs, from several elements of a complex cause; whatever remains of the effect must then have been produced by whatever remains of the cause. As an example, suppose that the nurse, during prior investigations of student illness, has already established that pizza tends to produce a rash and iced tea tends to cause headaches. Today, a student arrives at the nurse's office complaining of headache, indigestion, and a rash; this student reports having eaten pizza, coleslaw, and iced tea for lunch. Since she can recognize most of the student's symptoms as the effects of known causes, the nurse concludes that the additional effect of indigestion must be caused by the additional circumstance of eating coleslaw.
Notice that in all of the above methods, the issue of relevance is crucial. The nurse began with the assumption that what students had eaten for lunch was relevant to their digestive health in the afternoon. That is a reasonable assumption, but the real cause could have been something completely different, something about which the nurse never thought to ask. Therefore, application of Mill’s methods succeeds only if every relevant suspected cause is taken into account, but that is impossible to guarantee in advance. Indeed, the most difficult cases are those in which the real cause was excluded from the analysis as being unobserved or considered irrelevant. Thus, Mill's methods can't help to discover causes unless the list of all potential causes is already known. Problems with using Mill's methods for proving that one event is the cause of another, are even bigger Kemerling, 2006]. Because of their limitations, Mill's methods should rather be considered as a tool for confirming (and not for discovering) hypotheses. If there are several hypotheses about a causal relationship, then Mill’s methods can be helpful, since they will often enable eliminating most of the considered causes and the last remaining hypotheses will likely be valid.
## Hill’s Criteria
Causal reasoning in different forms had been widely used in medicine in twentieth century. In 1965. Austin Bradford Hill made a seminal summary of criteria to be used in causal inference in epidemiology [Hill, 1965]. The basic underlying questions in Hill's criteria for causal inference are: “is the association real or artifactual?” and “is the association secondary to a ‘real’ cause?”. Hill’s criteria are widely recognized as a basis for inferring causality in epidemiology, but not only in epidemiology:
1. Strength of the association – the stronger an association, the less it could reflect the influence of some other factor(s). This criterion includes consideration of the statistical precision and methodological rigor of the existing studies with respect to bias.
2. Consistency – replication of analysis by different investigators, at different times, in different places, in different populations, with different methods leads to identical or similar findings. Namely, identical or similar findings are not likely to be all due to error or artifact. In addition, there should be reasonable and convincing explanation for different findings.
3. Specificity of the association – the more accurately defined the disease and exposure, the stronger the observed relationship should be. But the fact that one agent contributes to multiple diseases is not evidence against its role in other diseases.
4. Temporality – the ability to establish that the putative cause preceded in time the presumed effect.
5. Biological gradient (dose-response) – strength of disease changes with changes in exposure. Still, there could be a “threshold effect”.
6. Plausibility (credibility) –general knowledge and beliefs should be able to explain the observed causal relationship. Still, the observed relationship could be beyond the current knowledge.
7. Coherence – causal interpretation should not conflict with observations and with known facts about the natural history of the disease.
8. Experiment – not a guideline, but a method for testing a specific causal hypothesis. If available, well designed and conducted experimental studies (e.g., with controlled conditions and changing the exposure) provide strong evidence for or against causation.
9. Analogy – use analogies or similarities between the observed causalities and other causalities.
Hill himself did not intend his criteria to be used as a self-contained framework, but rather as guidelines: „Here there are nine different viewpoints from all of which we should study association before we cry causation... None of my nine viewpoints can bring indisputable evidence for or against the case-and-effect hypothesis and none can be required as a sine qua non. What they can do, with greater or lesser strength, is to help us make up our minds on the fundamental question - is there any other way of explaining the set of facts before us, is there any other answer equally, or more, likely than cause and effect“.
## Graphical models
Next, in graphical models of causality, causal relationships are represented by causal graphs [Greenland,1999; Robins, 2001]. A directed acyclical graph (DAG) is causal if every directed edge represents the presence of an effect of the parent (causal) variable on the child (affected) variable. In a causal graph, a directed path represents a causal pathway, and an X-to-Y directed edge represents a direct effect of X on Y. Absence of a directed path from X to Y in the graph corresponds to the causal null hypothesis that no change of the distribution of X can change the distribution of Y. Causal graphs provide simple visual and graph theory methods to check for confounding factors (common cause for two event analyzed for causal relationship) and other relevant properties.
## Potential-outcomes (counterfactual) models
In counterfactual (or potential-outcomes) approaches to causal reasoning, statements about causality are considered in forms of counterfactual statements [Lewis, 2000]. Reasoning focuses on what would have happened if, contrary to fact, the exposure had been something other than what it really was. For instance, the statement that a coleslaw ate by a student caused his indigestion is equivalent to the statement that had not the student eaten the colesaw, he would not have suffered from indigestion. This approach is justified by the fact that there are sistematic ways for dealing with counterfactual statements. For instance, Pearl gave an axiomatic system with clear semantics and effective algorithms for computing counterfactuals [Pearl, 2000]. In his framework, for instance, one can calculate a probability that the student would not have indigestion had he not eaten colesaw if it is the case that the student has eaten colesaw and is suffering from indigestion. The potential outcomes framework has applications in epidemiology and medical research, economics, education, psychology; and social science [Gong, 2008]. Because of Donald Rubin's contributions this is sometimes referred to as the „Rubin Model“.
In potential outcomes models, all possible outcomes, both observable and unobservable, are considered simultaneously, forming outcome vectors. The framework can briefly be describes as follows [Greenland, 2002]. Suppose we have a population of individual units under study (e.g. mice, people, counties) indexed by i = 1,…,N, a treatment or exposure with M + 1 levels (or actions) 0, 1, ..., M, and an outcome variable of interest Y. The standard potential-outcome model assumes that:
• Each individual could have received any of the treatment levels.
• For each individual i and treatment level j, the outcome for the individual i if the individual gets treatment level j is considered even if the individual does not in fact get j; this value is called the potential outcome.
The variable Y represents a generic variable for the actual outcome under the treatment actually given. Then, Yi(j) will be an indicator for the outcome for individual i if that individual is given treatment level j. The vector [Yi(0), Yi(1), …, Yi(M)] is the potential outcome vector for the individual i. Notice that, in practice, for each individual only one of the potential outcomes Yi(j) is observable, since an individual receives only one possible treatment (and the other treatment states and associated potential outcomes are counterfactuals). Outcomes that are not observable can only be estimated. This problem, called missing data problem, is one of the fundamental problem of causal inference. The causal effect is defined as a quantity that contrasts the components of the potential outcomes vector. The choice of treatment is said to have had no effect on Y for individual i if Yi(j) = Yi(k) for every possible pair of treatment levels j and k; otherwise, treatment choice could have had an effect. Treatment choice is said to have had no effect on the population if it had no effect on any individual in the population.
As an illustration, let us consider the simplest case when the treatment is binary, i.e., M = 1 (corresponding, for instance, to situations when there is and there is no treatment). Let Ti be a level of the actual treatment for the individual i. Then, the vector [Yi(0), Yi(1)] is the potential outcome vector for the individual i. The outcome Yi is equal Yi(0) if Ti=0 and Yi(1) if Ti=1. This can be written as:
Yi = Yi(0) + Ti(Yi(1)-Yi(0))
The difference of the outcomes with and without treatment is characterized by Yi(1)-Yi(0), the benefit of treatment. The average treatment effect is equal to (where E denotes expected value):
ATE= E[Y(1)-Y(0)]
The average treatment on the treated individuals is equal to:
ATT= E[Y(1)-Y(0) | T=1]
These quantities cannot be computer because of unobserved potential outcomes. For instance, let the available data are given in the table given below. The observed values Yi(j) are printed in bold. The quantities ATE and ATT can be computed only using estimated values for Yi(j) and this gives ATE=2, ATT=1. Different approaches for estimating these and other unobservable quantities by observable quantities are discussed in [Angrist,1996].
Individual Treatment Yi(0) Yi(1) Yi(1)-Yi(0) 1 0 3 5 2 2 1 2 5 3 3 1 5 4 -1 4 0 2 7 5 5 1 1 2 1
One of the practical results of the potential-outcomes models is the identification of a sufficient set of variables could yield the correct causal effect between variables of interest. That characterization of variables, called "backdoor" criterion, helps in identifying sets of variables worthy of observing.
## Structural-equation models
In structural equation approach, a network of causation is modeled by a system of equations and independence assumptions (see, for instance, [Greenland,2002]. Each equation shows how an individual response (outcome) variable changes as its direct (parent) causal variables change. The ‘individual’ may be any unit of interest, such as a person or aggregate. In the system, a variable may appear in no more than one equation as a response variable, but may appear in any other equation as a causal variable. A variable appearing as a response in the system is said to be endogenous (within the system); otherwise it is exogenous. Relationships between variables can be linear but can also be much more complex. Structural equations can be viewed as formulas for computing potential outcomes under various actions. For instance, equations can assert that one variable will not vary with another variable, if some other variables remain constant. Structural equations differ from ordinary regression equations (that represent only associations of actual outcomes with actual values of the covariates as one moves across individuals). Structural equations with unknown parameters specify the functional form of effects, but do not provide the exact values of effects; thus, they don't not fully quantify causal relations.
## Probabilistic causation
If a causal relationship “A causes B” is interprered deterministically then it states that A must be always followed by B. “Drinking alchohol causes headache” is often true, but still not always, so this causal relationship could be considered invalid. Probabilistic causation tends to overcome simple yes or no causation and in this approach, cause only raises the probability of the effect (rather then> implies effect), all else being equal. In other words, A probabilistically causes B if occurrence of A increases the probability of B [Hitchcock, 2002]. In this approach, causal relationships are explored by using the apparatus of the probability theory. We will not elaborate on this approach here, but it will rise to significance in later chapters of the book. Standard Bayesian approaches to causal inference will be discussed in Chapter 11 in the context of pattern mining; and then the PLN approach to RWR, discussed in Part III of the book, will include a different variant of probabilistic causal inference in an essential role, tightly integrated with probabilistic approaches to other aspects of inference such as deduction, induction and abduction.
Part II: Acquiring, Storing and Mining Logical Knowledge
# Extracting Logical Knowledge from Raw Data
Logical methods are extremely powerful when supplied with appropriate data, but this begs the question of where the data comes from. How does data from the real world get into logical format in the first place?
So far we have discussed simple toy examples, involving a small number of relationships; or else relationships that come from an already-formalized domain such as the Minesweeper game. But the real world is large, messy and not pre-formalized. Our overall goal here is to discuss the application of logic-oriented methods to large, heterogeneous knowledge stores; so we cannot entirely bypass the critical question of how one would construct large, heterogeneous stores of logical information. Essentially the question is how to usefully translate raw data observed in the real world, into sets of logical expressions in appropriate formal languages. This is not an easy question and there is no single answer: the answer is roughly as heterogeneous as the data involved. This is not our main focus here, so in this brief chapter we will merely overview some of the issues involved, giving a few references into some relevant literatures where appropriate.
In some cases, the transformation process is completely obvious and straightforward. For instance, Figure 9.1 shows how Twitter metadata regarding a message can simply and immediately be transformed into formal relationships (which could easily be mapped from diagrammatic into logical form):
For instance, if one has a logical term T1 corresponding to the entity bob_dobbs, and a logical term M1 corresponding to the message “Where r u?”, then one may create from this diagram a logical predicate-argument relationships such as
SentTo(M1,T1)
and interpret this within many of the formal-logic systems discussed in Part I above.
To take a different sort of example, in the project described in [</nowiki>Goertzel & de Garis, 2008; Goertzel 2008] an architecture is described for controlling physical or virtual robots using a logic-based cognitive engine. In this case the logical relationships come directly from the outputs of sensors, and from the commands needed to be issued to actuators. Here there are no major difficulties in representation, though there are significant difficulties in reasoning! For instance, we may represent a certain perceptual relationship by stating that there is a logical relationship of the form
tangentialProperPart(P1, P2)
between polygons P1 and P2, whose coordinates are then indexed in a special data structure (tangentialProperPart is a spatial logical relationship to be discussed a little later on). And we might represent a relationship regarding the movement of an actuator by
moveJoint(7, 1.3, 1.4, 2)
indicating the movement of joint 7 at speed 2 in the direction with θ=1.3 radians and φ=1.4 radians (referring to the standard spherical coordinates). Here, as in the Twitter metadata example, the translation of life into logic is relatively straightforward.
However, in other cases – including many cases relevant to the topic of this book -- the transformation process involves much more difficult choices. In general, transforming raw data into logic is a highly nontrivial matter, which requires the best of current technologies; but it is certainly within the scope of the feasible rather than the impossible. In practice subtle decisions must be made about how much intelligence to put into the transformation process, versus how much to leave to the logical-inference processes acting on the logical knowledge base.
For example, if a software system is given the text “Dogs eat bones,” one simple approach at logicizing this input would be to simply turn it into a sequence of propositions about the individual characters of the text: essentially, propositions of the form “At time so-and-such, I received a text message with ‘g’ as the third character” and so forth. This kind of proposition can be represented easily in formal logic, but this is not necessarily the most useful thing to do.
In the remaining sections of this chapter, we will very briefly consider two cases of the extraction of logical information from nonlogical sources: tabular and relational data, and linguistic data. Other cases also exist, of course: for instance audio, video and so forth; but reviewing the literatures in all these areas would take us too far afield.
## Extracting Logical Knowledge from Tabular and Relational Data
Conceptually, it seems straightforward enough to map tabular or relational data into logical format. Figure 9.2 shows a simple example of a spreadsheet mapped into formal semantic relations:
Figure 9.2: Extracting formal semantic relations from tabular data
In real life, however, this sort of mapping is extremely difficult, because of the problem of figuring out the semantics of the rows and columns of spreadsheets and databases. The field of “table recognition” confronts this issue, and is summarized in [Zanibbi et al, 2003].
## Extracting Logical Knowledge from Graphs, Drawings, Maps and Tables
A yet more difficult issue is the extraction of formal relationships from graphs, engineering drawings, maps and tables that are encoded as bitmap images or vector drawings. The set of techniques in charge of extracting semantics out of graphical information is grouped under the term “Diagram Recognition” and has been given some attention for the past two decades. Several algorithms, techniques and toolkits have been developed and work well in many cases. However in general it remains a hard problem, probably much harder than one who is unfamiliar with the domain may realize at first. This is due to the variety of manners one can choose to convey information graphically. Sometimes to interpret correctly a diagram one even needs contextual information possibly scattered in the rest of the document or relying on common sense knowledge.
Various domain dependent algorithms have been formulated and applied. However recent work has focused on unifying these techniques into a single framework using formal grammars comparable in a way a compiler generates machine code out of a program written in some programming language [Blostein02]. Except that here the program is a 2D image and the machine code is a diagram model. A diagram model encodes the semantics of the image, for instance if the diagram is a graph, the diagram model may be a list of relationships, and the "compiler", or rather called diagram recognizer, would produce an XML file containing the list of nodes and relationships represented by the graph. Following that approach a diagram recognizer may carry out several grammar parsing and data production passes, like:
1. a layout pass, that captures the spatial structure of the image and encodes it into a tree, like inside(circle, left_to_right("B", "o", "b"))
2. a lexical pass, that tries to group symbols into lexical tokens, for instance "B", "o", "b" becomes "Bob"
3. syntactic and semantic passes to finally generate the diagram model, for instance node("Bob") that expresses that the diagram is a graph with a single node called "Bob".
Similarly regarding table recognition there exists several techniques and recent work has been focused on unifying them [Zanibbi et al., 2004; Zanibbi et al., 2006; Blostein et al., 2000].]. Again the process is divided in several passes, where each pass analyze a certain layer to produce more abstract knowledge for the upper layer and so on until the semantic model is built.
## Extracting Logical Knowledge from Natural Language Text
Extracting logical relations from text requires multiple intermediate stages of Natural Language Processing, each of which is subtle and complex in itself.
For instance, using state-of-the-art natural language technology, one can transform a sentence such as “The dog ate the bone” into relationships such as
eat(dog, bone)
which is an abstract relationship embodying the semantic meaning of the sentence. Or, further, one can transform it into
eat_2(dog_1, bone_1)
which indicates that the sense of “eat” used in the sentence is the second one in the system’s reference dictionary, whereas the senses of “dog” and “bone” used are the first ones in the system’s reference dictionary. But performing similar operations on more complex sentences pushes the boundaries of what today’s technology can do.
One can frame this problem of “natural language information extraction” [</nowiki>Cowie and Wilks, 2000] in terms of three stages:
• mapping text into a syntactic representation
• mapping the syntactic representation into a semantic representation
• mapping the semantic representation into a more abstract logical representation
Figure 9.3 shows a relevant example of these stages, produced using the open-source RelEx software system created by Novamente LLC [Goertzel et al, 2006], which incorporates as a major subsystem the link parser [Grinberg et al., 1995] created at Carnegie-Mellon University. In the figure, the input sentence is first transformed into a set of low-level syntactic relations between words. Then these relations are translated into dependency relations such as “subj” and “obj” (representing subject and object relations). Finally these dependency relations are translated into formal relationships that can easily be given logical interpretation. The presence of these multiple stages illustrates the complexity of the natural language information extraction process.
The deepest problem in natural language information extraction has to do with the various sorts of ambiguity that exist in natural language. Words may have multiple meanings; sentences may have multiple parses that all seem syntactically plausible but have varying semantic and pragmatic sensibleness; words may refer back to other words, and so forth. Computational linguistics provides only heuristic and approximative techniques for handling these methods (e.g. [Jurafsky and Martin, 2008]; [Manning and Shutze, 1999]), so, although one may currently make software systems that map natural language text into sets of logical relationships, such systems cannot be expected to work perfectly even for simple sentences, and can provide highly erratic results for complex sentences.
Figure 9.3: Successive transformations of text into syntactic, grammatical and finally formal relationships (that are easily transformable to logical relationships)
# Scalable Spatiotemporal Logical Knowledge Storage
Having dealt with the representation of logical knowledge of various sorts, and briefly discussed the problem of translating nonlogical knowledge into logical knowledge, we now turn to the question of how large amounts of logical knowledge can pragmatically be stored. This chapter presents a brief and relatively nonmathematical interlude before we plunge into the more technical topics of mining patterns in logical knowledge stores, and carrying out inferences regarding changes and other patterns in logical knowledge stores.
Suppose that we represent temporal and spatiotemporal knowledge, appropriately contextualized, in one of the multiple logical formalisms briefly discussed above. If we apply these formalisms to real-world situations we are going to obtain incredibly huge numbers of logical propositions, which presents various potential practical difficulties. The mathematics is the same whether one has a dozen logical propositions or a trillion, but the pragmatics of information-management differs significantly!
In this chapter we review the various available technologies for managing massive amounts of logical terms and relationships.
## Comparison of Available Storage Technologies
The following table summarizes the strengths and weaknesses of available data storage technologies from the perspective of storing and managing large amounts of logical information.
Technology Plus Minuses Leaders Relation DBs Mature, enterprise grade solutions Ease of integration with other systems Bad conceptual fit for graph dataset Poor scalability performance Rigid data model Oracle, IBM, Microsoft, Sybase, MySQL Object Oriented DBS Better conceptual fit than relational DBs Performance and scalability Mature, enterprise grade solutions Single data model Not designed for mining and analysis Objectivity, Versant Graph DBs Flexible data model Performance and scalability Designed with mining and analysis in mind Less mature solutions than relational or OO Neo, Cogito, Open source Hypergraph DBS All the advantages of graph DBs, but with an even more flexible data model Only currently exists at the alpha stage HGDB (open source) RDF Triplestores Semantic web friendly Immature solutions Rigid data model Niche players and open source
Table 10.1: Strenghts and weaknesses of data storage technologies
First of all: representing and querying large graph data stores using traditional relational databases is certainly possible, but it would lead to profound scalability and performance problems. Despite efforts from industry leaders in the RDBMS arena such as Oracle and Sybase, relational databases and graph data are a poor conceptual and implementation fit.
Graph data typically has a flexible structure where connections among similar objects (representing people, entities, time points, spatial locations, and so forth) are numerous. And these connections are the whole point of graph data stores. The natural way to map these connections among similar objects to relational databases requires self-joins, since the connected objects are typically stored in the same table. Contemporary RDBMS technology is unable to perform these self-joins through anything but brute force un-optimized graph traversals, which creates large bottlenecks both for querying at scale and writing to the database [Lightstone et al, 2007].
Object-oriented databases (OODBMS) are mature technologies that provide a better fit for graph data, since object instances are naturally persisted as graphs. Objectivity (objectivity.com) and Versant (versant.com) are the industry leaders in this arena, and they offer enterprise-grade solutions that are highly scalable and distributed, and support multiple programming languages and platforms. Their products can also interoperate with relational databases and support SQL queries for easier integration with legacy systems.
Despite those benefits, there is a major drawback for OODBMS in a graph dataset context, which is the implicit assumption of a single object model. While this is adequate for large-scale object-oriented applications, graph data mining and analysis sometimes requires that the stored data be interpreted in different ways, especially when spatial and temporal dimensions are added. There's always a 1-to-1 mapping between a OODBMS representation and some OO design -- but for analyzing graph data we often want to switch the design without changing the stored data format (which is expensive for a large database), because we care more about the answers we can get than about retrieving the original data in a way that's fully consistent with how it was stored.
Technology that's explicitly oriented towards graph data (as opposed to using graphs to persist object instances) has been developed over the past few years. Neo (neotechnology.com) and Cogito (cogitoinc.com) have introduced commercial graph databases. These new tools emphasize the ability to traverse relations and discover new connections. This explicit focus makes such tools a better fit for data analysis and mining projects, especially ones involving probabilistic logic. Graph databases aren't as mature and enterprise-ready as their OOBDMS counterparts, however.
Aside from these commercial products, there is a fully featured open source version of Neo's DB (neo4j.org); and the Hypergraph DB (HGDB) project (kobrix.com/hgdb.jsp), also open source, was conceived with AI and data mining applications in mind. In fact HGDB goes beyond the graph database paradigm and constitutes a hypergraph DB, involving a basic representation that allows n-ary links and links pointing to links as well as nodes. This is convenient because, as will be elaborated a little later, sets of logical predicates are more naturally represented as hypergraphs than graphs.
Of course, graph DBs may be used as hypergraph DBs via transforming hypergraphs to and from graphs; but this introduces a performance penalty for the translation, and also has the drawback that, after the translation, some simple hypergraph queries become significantly more complex graph queries. But HGDB is still at the alpha stage of the development, and the potential advantages of hypergraph DBs over standard graph DBs are still relatively unexplored.
Finally, the growing interest in the semantic web has led to a number of commercial and open source products for storing knowledge encoded in the RDF data model. These storage solutions, known as Triplestores, can scale up to billions of RDF triples. The triple-based data model in RDF, however, is morelimited than a free-form graph, which impacts scalability (as a more verbose knowledge representation is needed) and analysis (as algorithms have to be tailored to the more rigid RDF format, sometimes with a significant performance penalty).
Overall, our conclusion is that graph databases, whether commercial or open-source, are the current best alternative for storing and analyzing very large graph datasets.
## Transforming Logical Relationship-Sets into Graphs
The discussion in the above section bypasses one issue: the effective representation of sets of logical relationships as graphs. This is not a problematic issue, but bears brief comment because, most literally interpreted, sets of logical relationships would better be represented as mathematical structures called “generalized hypergraphs” than as graphs per se. So one encounters the problem of translating generalized hypergraphs into traditional graphs, using appropriate, hopefully not too complex transformation rules.
Recall that a graph, mathematically, is a set of nodes together with a set of links, where each link is construed as an ordered or unordered pair of nodes. Links and nodes may be labeled and may have various numerical weights attached to them (such as fuzzy or probabilistic truth values). A hypergraph extends this model, in that links may join more than two targets. This is useful for representing logical relationships such as
give(Jim, Bob, ball)
which naturally relate three rather than two entities.
Of course, one can work around the need for hypergraph links via using labeled binary links, for example
subj(give,Jim)
obj(give,Bob)
obj2(give,ball)
which is how the RelEx NLP system (mentioned earlier) analyzes the sentence “Jim gives the ball to Bob” (and other dependency parsers would do it similarly). Similarly, the relation
eat(Ben, steak)
could be represented using a ternary link, or a labeled binary link; or, as RelEx would have it, as a set of labeled binary links
subj(eat, Ben)
obj(eat, steak)
However, attaching a broad variety of semantic labels to links is not always the desired strategy. In general it is desirable to support a broad variety of representational mechanisms, as different approaches to logical formalization of commonsense information are going to choose different ways to set up relationships.
In any case, it is straightforward to eliminate hypergraph links via introducing a “phantom node” corresponding to each hypergraph link, and having the phantom node link binarily to the targets of the hypergraph link. What’s required is that the links emanating from the phantom node be indexed with numbers or some other distinct markers, if the targets of the hypergraph links were so marked. Figure 10.1 illustrates some examples of this, for the above examples.
Figure 10.1: Some hypergraph representations for the “Ben eats steak” example
Next, what we call a “generalized hypergraph” extends the hypergraph model further, via allowing links to point to links, which is the most natural way to represent statements like “Ben believes Bob likes Jim”, e.g.
believes(Ben, likes(Bob, Jim))
Alternately, a RelEx-style representation of the above would be
subj(believes,Ben)
obj(believes, likes)
subj(likes, Bob)
obj(likes, Jim)
Mapping generalized hypergraphs into graphs is also simply accomplished using phantom nodes, as illustrated in Figure 10.2
Figure 10.2: Hypergraph to graph conversions for the “Ben believes Bob likes Jim” example
Just to make sure the point is clear, we next give some examples involving more complex logical constructs such as actually arise in using PLN for carrying out inferences involving changes in complex knowledge bases.
EvaluationLink
believes
Ben
Inheritance(Bob, busy) )
and
IntensionalContext <.9>
Accounting
Evaluation <.01>
Mention
List
Bob
and finally one that is more complex and involves variables
AverageAll $X,$Y
Implication <.9>
Evaluation <.5>
Mention
List
$X$Y
IntensionalInheritance
$X$Y
In summary, using this sort of mapping based on phantom nodes, one can straightforwardly store logical relationships, interpreted as generalized hypergraphs, in graph databases. The transformation required is fairly simple and does not require the same sort of inefficient manipulations as mapping logical propositions into tabular structures as required to store them in standard relational databases.
However, the subtle question in mapping hypergraphs into graphs is: which graph operations will have the expected results when mapped back into hypergraph operations. For instance, if we map a hypergraph into a graph and then find the shortest path P between two nodes N and M in the graph ... is the hypergraph path corresponding to P the shortest path between the hypergraph nodes or links corresponding to N and M? Similarly, does a minimum-cost spanning tree in the graph derived from a hypergraph, correspond to a minimum-cost spanning hypertree in the original hypergraph? Is the set of nodes within radius R of graph node N, closely related to the set of hypergraph nodes/links within radius R of the hypergraph node/link corresponding to N? These issues go beyond the scope of the present book, and are in most cases not extremely difficult to resolve, but do require real care.
# Mining Patterns from Large Spatiotemporal Logical Knowledge Stores
Once one has stored a large knowledge base of logical relationships, then what? One can query the knowledge base -- if one knows what one wants to ask for. One can carry out reasoning toward various goals. And another important question is how to find “unknown unknowns” – patterns in the knowledge base that are surprising and interesting yet unexpected. This quest goes by multiple names – data mining, pattern mining, information exploitation, and so forth. Whatever you call it, it’s a difficult challenge because in any large dataset, the number of possible patterns to search through is mind-boggling.
Many different pattern mining algorithms exist, and a large subset of these are applicable to the case of mining patterns among logical relationships. Here we will review only two classes of algorithms: frequent subgraph mining, and causal network inference. These are important approaches, but are by no means the only approaches of interest.
Furthermore, as will be emphasized in later chapters, pattern mining algorithms in themselves are unlikely to be sufficient for the task of finding relevant and interesting relationships in large logical knowledge bases. The problem is that without significant background knowledge, and the capability to deploy this background knowledge intelligently for analogical inference, it’s very hard to tell interesting patterns from uninteresting ones. So, in order to really do a good job of spotting interesting patterns in large logical knowledge bases, it’s likely to be necessary to combine pattern mining algorithms with uncertain and causal inference algorithms.
That is, one will need to use pattern mining to produce a moderate-sized pool of potentially interesting patterns, and then use inference to filter this down into a smaller set of probably-interesting patterns. As many pattern mining algorithms (including the ones considered here) an be instructed to look for new patterns “in the neighborhood” of a set of target patterns, the patterns identified as interesting by inference may then be used to seed further pattern mining. This kind of hybridized approach has not been explored much if at all in the research literature, but there is little doubt it will be necessary as the sorts of applications envisioned in this book become realities.
To add to the challenges, pattern mining in extremely large bodies of knowledge poses particular difficulties in terms of scalability. For instance, algorithms must cope with the inability to store the whole knowledge base in the memory of any one machine. This is an area computer science is just beginning to explore. For instance, in the following section we will discuss algorithms for identifying “surprisingly frequent subgraphs” in large graph knowledge bases, following [Hsu et al., 2008]. The unique aspect of their approach is a clever mechanism for recursively decomposing a large graph into a large number of smaller subgraphs, recognizing patterns in the subgraphs, and then assembling overall graph patterns from the patterns recognized in the subgraphs. This general sort of idea likely has much more general applicability.
Furthermore, even within the scope of what can be stored within a single machine, there can be sufficient data to render standard pattern mining algorithms inapplicable. So as well as crafting distributed algorithms, one must devise special algorithms capable of handling large bodies of knowledge efficiently within a single machine. We will consider one example of this below: algorithms for finding “partial causal networks” in large bodies of knowledge, which essentially are simplifications and scalings-up of better-known algorithms for finding full causal networks in smaller datasets.
## Mining Frequent Subgraphs of Very Large Graphs
The staple of standard “data mining” in relational databases is a technique called “frequent itemset mining” or FIM [Goethals & Zaki, 2004], which seeks to find the most frequent combinations of data items. There are variants of FIM which seek the most surprising combinations of data items; these are essentially algorithmically identical to FIM, with slightly different underlying mathematics [Chakrabarti et al, 1998; Gallo et al, 2007].
In the graph domain, the analogue of FIM is “frequent subgraph mining,” an area in which there are numerous publications and a handful of open-source software toolkits. An overview of the field is given in [Ivancsy & Vajk, 2005]. These algorithms are directly relevant to our problem of mining patterns in large stores of logical knowledge, because logical predicates may be mapped into graph structures as we discussed in the previous chapter.
Two simple examples of frequent subgraphs that might be found in large graphs in the Twitter domain are as follows:
• Women in East Anglia often send each other private messages about the band Coldplay
• Young Chinese in London often express positive sentiment about the Chinese government on the day after the Chinese soccer team wins a game
Figure 11.1: Stages of distributed pattern mining in large graphs, in the HLW algoritm. Top, left: original graph. Top, right: partition into subgraphs fitting RAM of individual machines. Bottom, left: identification of frequent subgraphs. Bottom, right: merge of subgraphs embodying repeated
Datamining large graph bases is a challenging problem, because most of the highly scalable datamining algorithms available were designed to operate on tabular data, and perform poorly when adapted to graphs. These adaptations often require a fixed graph structure, which isn't practical. Spatiotemporal databases [Yeung and Hall, 2007] make the problem even harder due to their continuously changing nature. A datamining algorithm for a large spatiotemporal graph database must fullfil at least the following requirements:
• Ability to handle data too voluminous to fit in RAM without severe performance degradation.
• Ability to incrementally mine the database, including the ability to consider only new information
• Ability to find patterns that are frequent in space (occur often across different locations), time (occur frequently over time) and both. These patterns can be static or dynamic with regards to time and/or space.
One such algorithm, that we’ve explored in detail, is due to Hsu, Lee and Wang (hence we nickname it HLW) and it has three phases, loosely illustrated in Figure 11.1:
• Partition the graph database into units that fit into RAM.
• Apply a standard graph datamining algorithm to each unit, generating a set of patterns.
• Merge the obtained patterns from each unit, obtaining database-wide patterns.
Alternative algorithms exist and others can be developed. We don't think this particular algorithm is necessarily ideal, but we believe that any software system designed to identify patterns in a large spatiotemporal logic database needs to include an algorithm that fullfils the above requirements.
## Learning Partial Causal Networks
Another important example of data analysis that must be performed on large spatiotemporal logical knowledge bases is the search for causal patterns. Note the key distinction between correlation and causation, as depicted in Figures 11.2 – 11.3: put roughly, causation may be characterized as the combination of correlation with the presence of a plausible causal mechanism ... where the assessment of the plausibility of a causal mechanism always depends upon contextual understanding.
Figure 11.2: Correlation is not causation
Figure 11.3: The basics of Causation
Techniques for inferring networks of causal relationships from databases of events are well known, and are mainly based on the interpretation of a causal network as a Bayesian belief network with causal links [Pearl94].
Figure 11.4: A Causal network
Figure 11.5 shows some causal relationships in the Twitter context that may be represented this way: a causal relation between message contents of a given person, and a causal relation between message content and follower subscription.
Figure 11.5: Example of a causal relation between contents of a given person and a causal relation between message content and follower subscripto
None of the standard Bayes net based methods scale up at all well. On the other hand, there are some modern variations of these methods that do deal with reasonably large datasets, via scaling down their ambition and searching for “partial causal networks” rather than complete ones. Building a partial causal network (still capturing most of the causal relations) is relatively tractable even for processes involving tens of thousands of variables. To apply these methods to really huge datasets, one would then combine them with the same sort of graph-partitioning scheme described in the previous section in the context of frequent itemset mining. However, articulating the details of this combination would go beyond the scope of this book.
## Scalable Techniques for Causal Network Discovery
The leading algorithm for scalably discovering partial causal networks in massive datasets is Local Causal Discovery (LCD), a straightforward technique which has many specialized variants [Silverstein98, Mani01].
The basic idea underlying LCD is simple : testing Conditional Independence for two variables assuming one cause, instead of assuming a conjunction of causes. Recall that the notation X//Y|Z stands for X and Y are independent knowing Z, or more formally P(X|Y,Z)=P(X|Z) and P(Y|X,Z)=P(Y|Z). In these terms, the basic idea of LCD is assuming X//Y|Z instead of X//Y|C where C is a set of variables. This core principle underlies all variations of the algorithm. This sort of algorithm can be reasonably efficient; but has the serious limitation that can only discover causal relations involving one cause at a time.
There are also global approaches that can handle events with multiple causes, via approaches such as
• making special causal assumptions [Cheng97]
• pre-processing dependence over the graph (that is computing a dependency measure for each pair of variables) [Cheng97]
• pre-processing the Markov Boundary (which is the minimal Markov Blanket) [Margaritis99]
The pre-processing approach allows one to reduce the number of conditional independence tests because many configurations of causes are ruled out after the pre-processing (or/and by causal assumptions).
A good example of a more scalable global method is [Margaritis99]. This approach operates by first estimating the "Markov boundary" (i.e. the minimal set of variables that isolates, that is makes independent, a given variable from the rest of the network) of each variables to limit the conditional independence search over the Markov boundary of a given variable. This works excellently when for instance a variable has a small number of direct causes. Another, related approach is [Nielsen08], which uses the algorithm of the previous paragraph but in an incremental way, based on the assumption that the joint probability distribution is changing over time.
Part III: : Probabilistic Logic Networks for Real-World Reasoning
# Probabilistic Logic Networks
The preceding portion of this book has largely constituted “literature review”; this final part is a little different and presents original material, designed to cover important areas that seem omitted by the approaches reviewed above.
Above we have reviewed aspects of the representation of uncertain spatiotemporal knowledge, and also systems for reasoning on real-world knowledge; but there are major gaps between those two sets of ideas as we have presented them so far. Our discussion of representation focused heavily on fuzzy and probabilistic spatiotemporal knowledge, but the logical reasoning systems discussed don’t handle these sorts of uncertainty in a sophisticated or integral way. We suggest that one prerequisite for effective, scalable RWR using a logic-based approach, is to have a logic system that incorporates fuzziness and probability into spatiotemporal, contextual and causal inference in a fundamental way.
In the chapters in this Part of the book, we aim to show how to do this via creating and manipulating special logical relationship types within the Probabilistic Logic Networks (PLN) formalism that we have introduced in prior publications [Goertzel et al., 2008], and developed in the context of our work on the Novamente Cognition Engine [Goertzel et al., 2004] and OpenCog [Goertzel & Hart, 2008] integrative AI architectures, and will use in some of the detailed examples in later chapters. A complete exposition of PLN would be out of place here; our goal will be to explain enough of the elements and the notation to make the examples given in later chapters comprehensible.
## Motivations Underlying PLN
The guiding motivation behind the design of PLN was the desire to create an uncertain inference framework capable of encompassing all the sorts of inference that may confront a general intelligence operating in the everyday human world -- including reasoning based on uncertain knowledge, and/or reasoning leading to uncertain conclusions (whether from certain or uncertain knowledge). Among the general high-level requirements underlying the development of PLN were the following:
• To enable uncertainty-savvy versions of all known varieties of logical reasoning, including for instance higher-order reasoning involving quantifiers, higher-order functions, and so forth
• To reduce to crisp “theorem prover” style behavior in the limiting case where uncertainty tends to zero
• To encompass inductive and abductive as well as deductive reasoning
• To agree with probability theory in those reasoning cases where probability theory, in its current state of development, provides solutions within reasonable calculational effort based on assumptions that are plausible in the context of real-world embodied software systems
• To gracefully incorporate heuristics not explicitly based on probability theory, in cases where probability theory, at its current state of development, does not provide adequate pragmatic solutions
• To provide “scalable” reasoning, in the sense of being able to carry out inferences involving billions of premises. Of course, when the number of premises is fewer, more intensive and accurate reasoning may be carried out.
• To easily accept input from, and send input to, natural language processing software systems
The practical application of PLN is still at an early stage. Based on our evidence so far, however, we have found PLN to fulfill the above requirements adequately well, and our intuition remains that it will be found to do so in general.
The overall structure of PLN theory may be described as follows. First, PLN involves some important choices regarding knowledge representation, which lead to specific “schematic forms” for logical inference rules. The knowledge representation may be thought of as a definition of a set of “logical term types” and “logical relationship types” (some of which we will elaborate below), leading to a novel way of graphically modeling bodies of knowledge. It is this graphical interpretation of PLN knowledge representation that led to the “network” part of the name “Probabilistic Logic Networks.” It is worth noting that the networks used to represent knowledge in PLN are generalized weighted directed hypergraphs [Bollobás, 1998], much more general for example than the binary directed acyclic graphs used in Bayesian network theory. Later on we will review some methods for translating generalized hypergraphs into ordinary graphs, which can be useful for purposes of visualization, analysis and storage.
Next, PLN involves specific mathematical formulas for calculating the probability value of the conclusion of an inference rule based on the probability values of the premises plus (in some cases) appropriate background assumptions. It also involves a particular approach to estimating the confidence values with which these probability values are held (weight of evidence, or second-order uncertainty). Finally, the implementation of PLN in software requires important choices regarding the structural representation of inference rules, and also regarding “inference control” – the strategies required to decide what inferences to do in what order, in each particular practical situation.
Here we will not be concerned at all with PLN’s probability formulas – they are absolutely critical for performing practical inferences and getting useful answers, but here we will only be concerned with exploring the forms of various inferences, and so we will refer the reader to the Probabilistic Logic Networks book [Goertzel et al, 2008] for discussion of quantitative formulas. In our examples here, we will omit quantitative truth values so as to focus on the forms of inferences. In fact, the quantitative truth value associated with an inference may come out differently depending on the particular parameters of the truth value formulas, as clarified in the PLN book.
## Term and Predicate Logic in PLN
One of the distinguishing features of PLN is the way its inference rules combine predicate logic and term logic. As briefly reviewed above, predicate logic and term logic are two different but related forms of logic, each of which can be used both for crisp and uncertain logic.
Predicate logic is the most-familiar kind, where the basic entity under consideration is the “predicate,” a function that maps argument variables (which are are quantified universally or existentially) into truth values.
On the other hand, in term logic, which dates back at least to Aristotle and his notion of the syllogism, the basic element is a subject-predicate statement, denotable in many ways, for instance
A$\displaystyle \longrightarrow$ B
where $\displaystyle \longrightarrow;$ denotes a notion of inheritance or specialization. Logical inferences take the form of “syllogistic rules,” which give patterns for combining statements with matching terms. (We don’t use the $\displaystyle \longrightarrow$ notation much in PLN, because it’s not sufficiently precise for PLN purposes since PLN introduces many varieties of inheritance; but we will use the $\displaystyle \longrightarrow$ notation in this section since here we are speaking about inheritance in term logic in general rather than about PLN in particular).
Example term logic inference rules are the deduction, induction, and abduction rules:
• A$\displaystyle \rightarrow$ B
• B$\displaystyle \rightarrow$ C
• $\displaystyle \vdash$
• A$\displaystyle \rightarrow$ C
• A$\displaystyle \rightarrow$ B
• A$\displaystyle \rightarrow$ C
• $\displaystyle \vdash$
• B$\displaystyle \rightarrow$ C
• A$\displaystyle \rightarrow$ C
• B$\displaystyle \rightarrow$ C
• $\displaystyle \vdash$
• A$\displaystyle \rightarrow$ B
These rules are simple schematically but subtler when one matches them with uncertain truth value formulas. For instance, when one does so, one finds that deduction is infallible, in the case of absolutely certain premises, but uncertain in the case of probabilistic premises; while abduction and induction are always fallible, even given certain premises. In fact, in PLN one derives abduction and induction from the combination of deduction with a simple rule called inversion
A $\displaystyle \longrightarrow$ B
B $\displaystyle \longrightarrow$ A
whose truth value formula derives from Bayes’ rule.
Predicate logic is generally felt to deal more naturally with deduction than with induction, abduction and other uncertain, fallible inference rules. On the other hand, term logic can deal quite elegantly and simply with all forms of inference. Furthermore, as argued in [Goertzel et al., 2008] the predicate logic formulation of deduction proves less amenable to “probabilization” than the term logic formulation. It is for these reasons, among others, that the foundation of PLN is drawn from term logic rather than from predicate logic. PLN begins with a term logic foundation, then adds on elements of probabilistic and combinatory logic, as well as some aspects of predicate logic, to form a complete inference system, tailored for easy integration with software components embodying other (not explicitly logical) aspects of intelligence.
Sommers and Engelbretsen [Englebretsen & Sommers, 2000] have given an excellent defense of the value of term logic for crisp logical inference, demonstrating that many pragmatic inferences are far simpler in term logic formalism than they are in predicate logic formalism. On the other hand, the pioneer in the domain of uncertain term logic is Pei Wang [Wang, 1996], to whose NARS uncertain term logic based reasoning system PLN owes a considerable debt. To frame the issue in terms of our above discussion of PLN’s relation to traditional probabilistic logic approaches, we may say we have found that many things are significantly easier in a term logic rather than predicate logic context, including:
1) the formulation of appropriate heuristics to guide probabilistic inference in cases where adequate dependency information is not available,
2) and the creation of appropriate methods to extend first-order extensional inference rules and formulas to handle other sorts of inference.
In these respects, the use of term logic in PLN is roughly a probabilization of the use of term logic in NARS; but of course, there are many deep conceptual and mathematical differences between PLN and NARS, so that the correspondence between the two theories in the end is more historical and theory-structural, rather than being a precise correspondence on the level of content.
## Knowledge Representation in PLN
PLN knowledge representation is conveniently understood according to two dichotomies: extensional vs. intensional, and first-order vs. higher-order. The former is a conceptual (philosophical/cognitive) distinction, between logical relationships that treat concepts according to their members versus those that treat concepts according to their properties. In PLN extensional knowledge is treated as more basic, and intensional knowledge is defined in terms of extensional knowledge via the addition of a specific mathematics of intension (somewhat related to information theory). This is different from the standard probabilistic approach which contains no specific methods for handling intension, and also different from Wang’s approach in which intension and extension are treated as completely symmetric with neither of them being more basic or derivable from the other.
The first-order versus higher-order distinction, on the other hand, is essentially a mathematical one. First-order, extensional PLN is a variant of standard term logic, as originally introduced by Aristotle in his Logic and more recently elaborated by theorists such as Wang (Wang 1996) and Sommers and Engelbretsen (Sommers 2000). First-order PLN involves logical relationships between terms representing concepts, such as
Inheritance cat animal
ExtensionalInheritance Pixel_444 Contour_7565
(where the notation is used that R A B denotes a logical relationship of type R between arguments A and B). A typical first-order PLN inference rule is the standard term-logic deduction rule
A $\displaystyle \longrightarrow$ B
B $\displaystyle \longrightarrow$ C
A $\displaystyle \longrightarrow$ C
which in PLN looks like
ExtensionalInheritance A B
ExtensionalInheritance B C
ExtensionalInheritance A C
As well as purely logical relationships, first-order PLN also includes a fuzzy set membership relationship, and specifically addresses the relationship between fuzzy set membership and logical inheritance, which is closely tied to the PLN concept of intension.
Higher-order PLN, on the other hand, has to do with functions and their arguments. Much of higher-order PLN is structurally parallel to first-order PLN: for instance, implication between statements is largely parallel to inheritance between terms. However, a key difference is that most of higher-order PLN involves either variables or higher-order functions (functions taking functions as their arguments). So for instance one might have
ExtensionalImplication
Inheritance $X cat Evaluation eat ($X, mice)
(using the notation that
R
A
B
denotes the logical relationship R applied to the arguments A and B). Here Evaluation is a relationship that holds between a predicate and its argument-list; so that e.g.
Evaluation eat (Sylvester, mice)
means that the list (Sylvester, mice) is within the set of ordered pairs characterizing the eat relationship. The parallel of the first-order extensional deduction rule given above would be a rule
ExtensionalImplication A B
ExtensionalImplication B C
ExtensionalImplication A C
where the difference is that in the higher-order inference case, the tokens A, B and C denote either variable-bearing expressions or higher-order functions. Some higher-order inference rules involve universal or existential quantifiers as well.
While first-order PLN adheres closely to the term logic framework, higher-order PLN is better described as a mix of term logic, predicate logic and combinatory logic (though the latter aspect will not be emphasized here). The knowledge representation is kept flexible as this seems to lead to the simplest and most straightforward set of inference rules.
## PLN Truth Values and Formulas
Next, one of the less conventional aspects of PLN – which will not play a major role in this book, but still merits brief mention -- is the quantification of uncertainty using imprecise truth values that contain at least two components, and usually more (in distinction from the typical truth value used in probability theory, which is a single number: a probability). PLN’s indefinite probability approach is related to earlier multi-component truth-value approaches due to Keynes, Wand, and
Walley [Keynes (2007); Wang, 2006; Walley, 1991] and others, but is unique in its particulars.
The simplest kind of PLN truth value, called a SimpleTruthValue, consists of a pair of numbers <s,w> called a strength and a confidence. The strength value is a probability; the confidence value is a measure of the amount of uncertainty attached to the strength value. Confidence values are normalized into [0,1].
For instance <.6,1> means a probability of .6 known with absolute certainty. <.6,.2> means a probability of .6 known with a very low degree of certainty. <.6,0> means a probability of .6 known with a zero degree of certainty, which is equivalent to <x,0> for any other probability value x.
Another type of truth value, more commonly used as the default within PLN, is the IndefiniteTruthValue. We introduce the mathematical and philosophical foundations of IndefiniteTruthValues in Chapter 3. Essentially a hybridization of Walley’s imprecise probabilities and Bayesian credible intervals, indefinite probabilities quantify truth values in terms of four numbers <L,U,b,k>: an interval [L,U], a credibility level b, and an integer k called the lookahead. IndefiniteTruthValues provide a natural and general method for calculating the “weight-of-evidence” underlying the conclusions of uncertain inferences.
Beyond the SimpleTruthValues and IndefiniteTruthValues mentioned above, more advanced types of PLN truth value also exist, principally “distributional truth values” in which the strength value is replaced by a matrix approximation to an entire probability. Note that this then provides for three different granularities of approximations to an entire probability distribution. A distribution can be most simply approximated by a single number, a somewhat better approximation being provided by a probability interval, and an even better approximation given by an entire matrix.
[Goertzel et al 2008] defines the various inference rules of PLN, and also associates with each of them a “strength value formula” with each of them (a formula determining the strength of the conclusion based on the strengths of the premises). For example, the deduction rule mentioned above is associated with two strength formulas, one based on an independence assumption and the other based on a different “concept geometry” based assumption. The independence-assumption-based deduction strength formula looks like
B <sB>
C <sC>
ExtensionalInheritance A B <sAB>
ExtensionalInheritance B C <sBC>
ExtensionalInheritance A C <sAC>
where
sAC = sAB sBC + (1-sAB) ( sC – sB sBC ) / (1- sB )
This particular rule is a straightforward consequence of elementary probability theory. Some of the other formulas are equally straightforward; but some are subtler and (as explained in detail in Goertzel et al, 2008]) require heuristic reasoning beyond standard probabilistic tools like independence assumptions.
Since simple truth values are the simplest and least informative of our truth value types, they provide quick, but less accurate, assessments of the resulting strength and confidence values. A valuable enterprise is extending the simple truth value formulas to IndefiniteTruthValues. A careful consideration of this matter shows that indefinite truth values provide a natural approach to measuring weight-of-evidence. IndefiniteTruthValues can be thought of as approximations to entire distributions, and so provide an intermediate level of accuracy regarding strength and confidence. Finally, PLN inference formulas may also be modified to handle entire distributional truth values. Distributional truth values provide more information than the other truth value types. As a result, they may also be used to yield even more accurate assessments of strength and confidence.
## Some Relevant PLN Relationship Types and Inference Rules
In this section we briefly review the specific PLN relationship types and inference rules that will be used in the inference examples given in later chapters. Contextual, spatial and temporal relationships will not be covered here, as these will be described later on in the appropriately specialized chapters.
As seen above the PLN formalism allows one to express relationships over predicates and relationships. For this purpose it uses higher order operators (comparable in some ways to first order logic connectives and
quantifiers), such as Implication, Equivalence, AverageAll and ThereExists (there is also a ForAll operator but as it is not used in the inference examples we will not elaborate further on it). The semantics of Equivalence and Implication are easily definable using the SatisfyingSet operator that we define below.
### SatisfyingSet and Member
The SatisfyingSet operator allows us to express the concept of a set whose members are all elements that satisfy the predicate. We also recall the Member relationship type that expresses how much an element belongs to a concept (with a truth value that is fuzzy rather than probabilistic).
Let's for instance consider the predicate FriendOfBob, defined by the three elements Jack, John and Jill as follows:
Evaluation <.7>
FriendOfBob
Jack
Evaluation <.6>
FriendOfBob
John
Evaluation <.8>
FriendOfBob
Jill
According to the definition of the SatisfyingSet operator, we would then have:
Member <.7>
Jack
SatisfyingSet(FriendOfBob)
Member <.6>
John
SatisfyingSet(FriendOfBob)
Member <.8>
Jill
SatisfyingSet(FriendOfBob)
### Equivalence and Implication
Now we can define Equivalence and Implication as follows:
Equivalence
A
B
is equal to
Similarity
SatisfyingSet(A)
SatisfyingSet(B)
and
Implication
A
B
is equal to
Inheritance
SatisfyingSet(A)
SatisfyingSet(B)
We have defined the mixed versions of Equivalence and Implication. The extensional and intensional versions are analogously defined, we give only the definition of ExtensionalEquivalence:
ExtensionalEquivalence
A
B
is equal to
ExtensionalSimilarity
SatisfyingSet(A)
SatisfyingSet(B)
### Quantifiers, Average and ThereExists
Quantification in uncertain logic is a somewhat subtle matter. PLN handles it largely via the AverageAll construct, which is a kind of "average quantifier": the truth value of
AverageAll
$X F($X)
can be defined as the weighted average of the truth value of F($X), i.e. as the sum w($X) F($X) / N $\displaystyle \frac{\Sigma_x W(x)F(X)}{\Sigma_x W(X)}$ where w($X) is defined as the truth value of $X in the system and N is the sum of all w($X). (Note that this approach has been elaborated in detail only for finite domains, as the intended application is to the set of knowledge contained explicitly within an AI system. Extension to infinite domains is thought to be possible but would require additional theoretical elaboration.)
There may be other ways to define the truth value of AverageAll, possibly more advantageous in various respects, but the one outlined above has the advantage of being rather tractable, and it is the approach taken in the current PLN software system.
ThereExists is an existential quantification; it is the dual of the quantifier ForAll and we will not recall it in detail here (see the PLN book for more information [Goertzel et al., 2008]). Informally let's just say that the truth value of ThereExists $X F($X) quantifies how much it exists an $X such that F($X) is essentially not zero, i.e. lying within [e,1], where e is a margin of error.
From the detailed treatment of AverageAll and ThereExists given in [Goertzel et al, 2008], it follows that the truth value of ThereExists $X F($X) must be equal to or greater than the truth value of AverageAll $X F($X), at least assuming that e is equal or smaller than the truth value of AverageAll $X F($X). This will be useful in the inference examples given later on.
### Some Inheritance rules
The following rule is also useful in examples given later on; it basically says that if all elements that inherit F, also inherit G, then as a consequence F inherits G:
AverageAll $X ExtensionImplication Inheritance$X F
Inheritance $X G Inheritance F G And the rule below allows to infer an inheritance relation out of a membership relation, i.e. if X is a member of A then the singleton {X} inherits from A. This rule is called M2I for Member to Inheritance Member X A Inheritance {X} A and its truth value formula involves a significantly lower confidence for the conclusion than the premise, reflecting the speculative nature of the inference involved. There also exists the reverse, I2M, though it is not used in our examples. ### Intensional Inheritance As stated earlier in the section, intensional inheritance quantifies how much the properties of a concept inherits from the properties of another. There is more than one way to technically handle intension.The way we have chosen here is by using the relationship ASSOC, where roughly ASSOC(C E) is seen as an approximation of [P(C|E)-P(C|~E)]+, with [x]+ = max(x, 0) is the positive part of x, that is ASSOC(C E) measures how differentially E implies C; and we also assume that ExOut ASSOC C designated the fuzzy set {E|ASSOC(C E)}. Using the above ideas we can define IntensionalImplication_ASSOC as follows: ExtensionalEquivalence IntensionalImplication_ASSOC$X $Y ExtensionalInheritance ExOut ASSOC$X
ExOut ASSOC $Y To more formally define ASSOC in PLN terms, again several ways are possible; here we have chosen ASSOC(C E)=ASSOC_int(C E) OR ASSOC_ext(C E) where ASSOC_int and ASSOC_ext are respectively intensional and extensional versions of ASSOC and OR is the fuzzy disjunction. ASSOC_int is defined below: ExtensionalEquivalence Member$E (ExOut ASSOC_int $C) ExOut Func List IntensionalInheritance$E $C IntensionalInheritance NOT$E
where Func(x, y) = [x-y]+, and $E and$C are conceptNodes. ASSOC_ext is defined similarly where intensional is replaced by extensional.
## Applying PLN
To sum up: the goal underlying the theoretical development of PLN has been the creation of practical software systems carrying out complex, useful inferences based on uncertain knowledge and drawing uncertain conclusions. Toward that end we have implemented most of the PLN theory described in the PLN book, in a “PLN module” incorporated in the Novamente Cognition Engine (NCE), an integrative artificial intelligence software framework [Goertzel, 2006], and the OpenCog engine [Goertzel, 2008; Hart and Goertzel, 2008], an open-source offshoot of the NCE.
By far the most difficult aspect of designing a PLN implementation is inference control – which is really a foundational conceptual issue rather than an implementational matter per se. The PLN framework just tells you what inferences can be drawn, it doesn’t tell you what order to draw them in, in which contexts. The current PLN implementation utilizes the standard modalities of forward-chaining and backward-chaining inference control. However, the vivid presence of uncertainty throughout the PLN system makes these algorithms more challenging to use than in a standard crisp inference context. Put simply, the search trees expand unacceptably fast, so one is almost immediately faced with the need to use clever, experience-based heuristics to perform pruning.
The issue of inference control leads into deep issues related to automated reasoning and cognitive science; we briefly mention some of these issues in these pages, but do not fully explore, because that would lead too far afield from the focus of the book. In the final chapter we visit this theme in the specific context of exploring exactly how commonsense knowledge about spatial and temporal events may help guide PLN inference to perform scalably on large stores of real-world knowledge.
The practical application of PLN is still at an early stage; but so far, we have applied PLN to several areas. We have applied it to process the output of a natural language processing subsystem, using it to combine together premises extracted from different biomedical research abstracts to form conclusions embodying medical knowledge not contained in any of the component abstracts [Goertzel et al, 2006]. We have also used PLN to learn rules controlling the behavior of a humanoid agent in a 3D simulation world: for instance, PLN learns to play “fetch” based on simple reinforcement learning stimuli [Goertzel et al, 2007]. Our current research involves extending PLN’s performance in both of these areas, and bringing the two areas together by using PLN to help the NCE and OpenCog carry out complex simulation-world tasks involving a combination of physical activity and linguistic communication; and, additionally, pursuing the sorts of inferences described in this book, applying PLN to scalable inference on real-world spatiotemporal knowledge stores.
## Deploying PLN in the OpenCog System
With the above comments in mind, we here briefly describe how PLN has been integrated with OpenCog. OpenCog is a complex framework with a complex underlying theory, and here we will only hint at some of its key aspects. OpenCog is an open-source software framework designed to support the construction of multiple AI systems; and the current main thrust of work within OpenCog is the implementation of a specific AGI design called OpenCogPrime (OCP), which is presented in the online wikibook [Goertzel & Hart, 2008]. Much of the OpenCog software code, and many of the ideas in the OCP design, have derived from the open-sourcing of aspects of the proprietary Novamente Cognition Engine, which has been described extensively in previous publications [Goertzel et al., 2004].
The first key entity in the OpenCog software architecture is the AtomTable, which is a repository for weighted, labeled hypergraph nodes and hyperedges. In the OpenCog implementation of PLN, the nodes and links involved in PLN are stored here. OpenCog also contains an object called the CogServer, which wraps up an AtomTable as well as (among other objects) a Scheduler that schedules a set of MindAgent objects that each (when allocated processor time by the Scheduler) carry out cognitive operations involving the AtomTable.
The essence of the OCP design consists of a specific set of MindAgents (including some carrying out various PLN inference operations) designed to work together in a collaborative way in order to create a system that carries out actions oriented toward achieving goals (where goals are represented as specific nodes in the AtomTable, and actions are represented as Procedure objects indexed by Atoms in the AtomTable, and the utility of a procedure for achieving a goal is represented by a certain set of probabilistic logical links in the AtomTable, etc.). OpenCog is still at an experimental stage but has been used for such projects as statistical language analysis, probabilistic inference, and the control of virtual agents in online virtual worlds (see opencog.org). We believe it could also be of significant value in the many other RWR applications.
# Temporal and Contextual Reasoning in PLN
In this chapter we review the temporal and causal PLN relationship types and rules that are used to guide these sorts of inference in PLN, and give some simple examples to illustrate how they are used. In the following chapters we will present more elaborate examples of spatiotemporal reasoning, using these constructs and ideas.
## Temporal relationship types
First, the notation
holdsThroughout
E
T
means that the event E holds during T, where T is a time interval.
So for example:
holdsThroughout
Evaluation
Sick
Bob
[Tuesday:March:2007, Friday:March:2007]
AtTime link is also frequently used instead of holdsThroughout, they are equivalent, one can use one or the other indifferently.
The time format in the examples is arbitrary and matters little, in practice it is an integer corresponding to the number of time units -a time unit could be 10ms for instance- that have passed since a referential beginning date, the zero time.
The relationships intiatedAt and terminatedAt represent respectively when an event starts and stops. So for instance the example above can be similarly expressed by:
AND
initiatedAt
Evaluation Sick Bob
Tuesday:March:2007
terminatedAt
Evaluation Sick Bob
Friday:March:2007
Sometimes, that notation is not enough to characterize the temporal aspect of an event. For instance one may want to express that an event has started within an interval, or similarly ended within an interval. For that the temporal relationships initiatedTroughout and terminatedThroughout are used:
For instance if Bob has gotten progressively sick and healed progressively too:
AND
initiatedThroughout
Evaluation Sick Bob
[TuesdayMorning:March:2007, WednesdayEvening:March:2007]
terminatedThroughout
Evaluation Sick Bob
[ThrusdayMorning:March:2007, FridayEvening:March:2007]
It is possible to express other temporal relationships like OverlapTime which represents how much 2 time intervals overlap:
Or During which represents how much and interval is included in another one:
OverlapTime
[Monday, Wednesday]
[Tuesday, Friday]
Of course these two relationships can be considered as shorthands as they can be expressed using initiatedAt and terminatedAt.
## PLN Temporal Inference in Action
Next we give a concrete example of PLN doing temporal inference, according to the representational mechanisms described above.
Suppose a user has submitted to a logical knowledge based system a query regarding which people were in the same place as Jane last week. Suppose Susie and Jane use the same daycare center, but Jane uses it everyday, whereas Susie only uses it when she has important meetings (otherwise she works at home with her child). Suppose Susie sends a message stating that Tuesday she has a big meeting with a potential funder for her business. Inference is needed to figure out that on Tuesday she’s likely to put her child in daycare, and hence (depending on the time of the meeting!) potentially to be at the same place as Jane sometime on Tuesday. To further estimate the probability of the two women being in the same place, one has to do inference based on the times Jane usually picks up and drops off her child, and the time Susie is likely to do so based on the time of her meeting.
So: how do we use PLN to infer the truth value of the proposition that Susie was at the same Place as Jane last week?
Formally, in PLN notation our target theorem looks like:
ThereExists $Place,$TimeInterval1, $TimeInterval2 AND AtTime(AtPlace(Susie,$Place), $TimeInterval1) AtTime(AtPlace(Jane,$Place), $TimeInterval2) OverlapTime($TimeInterval1, $TimeInterval2) During($TimeInterval1, LastWeek)
During($TimeInterval2, LastWeek) where atPlace is a predicate that indicates if a given person is at a given place. We make the following assumptions for the purpose of the example inference: Axioms Axioms related to Jane: 1) “Jane is at the daycare center everyday of the week between 7am and 7:30am and between 16pm and 16:30pm (when she brings and fetch her child).” 1.a) AverageAll$Day
AND
IsWeekDay($Day) AtTime(AtPlace(Jane, daycare), [$Day:7am, $Day:7:30am]) 1.b) AverageAll$Day
AND
IsWeekDay($Day) AtTime(AtPlace(Jane, daycare), [$Day:16am, $Day:16:30am]) Axioms related to Susie: 2) “When Susie has an important meeting at time interval T, she will be in the daycare center during 30 minutes an hour before the beginning of T and after the end of T” Implication AtTime(ImportantMeeting(Susie), T) AND AtTime AtPlace(Susie, daycare) [beginning(T)-1h, beginning(T)-1:30h] AtTime AtPlace(Susie, daycare) [end(T)+1h, end(T)+1:30h] 3) “Susie had an important meeting last Tuesday between 1:30pm and 3:15pm” AtTime ImportantMeeting(Susie) [LastTuesday:1:30pm, LastTuesday:3:15pm] Inference chain: 1) “Susie was at the daycare center Tuesday between 4:15pm and 4:45pm”. Using axioms #2 and #3: AND AtTime AtPlace(Susie, daycare) [LastTuesday:12:30pm, LastTuesday:1pm] AtTime AtPlace(Susie, daycare) [LastTuesday:4:15pm, LastTuesday:4:45pm] Then using PLN inference rules to deal with AND AtTime AtPlace(Susie, daycare) [LastTuesday:4:15pm, LastTuesday:4:45pm] 2) “Jane was at the daycare center Tuesday between 4:pm and 4:30pm”. Using axioms #1.b AND isWeekDay(Tuesday) AtTime AtPlace(Jane, daycare) [LastTuesday:4pm, LastTuesday:4:45pm] Then using PLN inference rules to deal with AND AtTime AtPlace(Jane, daycare) [LastTuesday:4pm, LastTuesday:4:45pm] 3) Then we can infer an instance of the target theorem using the conclusion of inference step #1 and #3 + other axioms related to temporal relationships AND AtTime AtPlace(Susie, daycare) [LastTuesday:4:15pm, LastTuesday:4:45pm] AtTime AtPlace(Jane, daycare) [LastTuesday:4pm, LastTuesday:4:45pm] OverlapTime [LastTuesday:4:15pm, LastTuesday:4:45pm] [LastTuesday:4pm, LastTuesday:4:45pm] During [LastTuesday:4:15pm, LastTuesday:4:45pm] LastWeek During [LastTuesday:4pm, LastTuesday:4:45pm] LastWeek 4) And the target theorem is reached using step #3 and PLN existential quantifier axioms, by setting $Place=daycare
$TimeInterval1=[LastTuesday:4:15pm, LastTuesday:4:45pm],$TimeInterval2=[LastTuesday:4pm, LastTuesday:4:45pm]
and thus concluding
ThereExists $Place,$TimeInterval1, $TimeInterval2 AND AtTime(AtPlace(Susie,$Place), $TimeInterval1) AtTime(AtPlace(Jane,$Place), $TimeInterval2) OverlapTime($TimeInterval1, $TimeInterval2) During($TimeInterval1, LastWeek)
## PLN Causal Relationship Types
PLN represents the notion of causality with the PredictiveImplication relationship and some variants thereof.
In the inference to be given below, we will use only IntensionalPredictiveImplication so we will only define that one; the other variants are very similar anyway. Formally IntensionalPredictiveImplication is defined as follows:
IntensionalPredictiveImplication <w, T>
A
B
is equal to
IntensionalImplication <w>
A
AND_Seq <T>
A
B
where T is a time interval representing the time between the moment A is initiated and B is initiated, which is what AND_Seq <T> A B formally means.
That is, if SS_initiatedAt(A) is the start time of A and SS_initiatedAt(B) is the start time of B then:
AND_Seq <T>
A
B
is equal to
AND
A
B
SS_initaitedAt(B) - SS_initiatedAt(A) lies within T
## PLN Contextual Inference in Action
Finally, in this section, we run the contextual inference example given in Chapter 7 above using PLN rather than the more traditional contextual inference approach explored earlier.
First we enumerate the assumed axioms, describing each one in natural language and then formally in PLN terms:
Axioms for Music Context
1) In the context of music Alice frequently mentions Canadian place names
IntensionalContext <.9>
Music
Evaluation <.5>
Mention
List
Alice
2) In the context of music Bob frequently mentions Canadian place names
IntensionalContext <.9>
Music
Evaluation <.5>
Mention
List
Bob
3) In the context of music Clark does not frequently mention Canadian place names
IntensionalContext <.9>
Music
Evaluation <.01>
Mention
List
Clark
CandianPlaceNames
Axioms for Accounting Context
4) In the context of accounting Alice does not frequently mention Canadian place names
IntensionalContext <.9>
Accounting
Evaluation <.01>
Mention
List
Alice
5) In the context of accounting Bob frequently mentions Canadian place names
IntensionalContext <.9>
Accounting
Evaluation <.01>
Mention
List
Bob
6) In the context of accounting Clark frequently mentions Canadian place names
IntensionalContext <.9>
Accounting
Evaluation <.5>
Mention
List
Clark
Non-Context-Specific Axioms
7) Accounting is associated with Money
IntensionalInheritance <.7>
Accounting
Money
IntensionalInheritance
10) If someone X frequently mention Y then he/she is highly involved with Y
AverageAll $X,$Y
Implication <.9>
Evaluation <.5>
Mention
List
$X$Y
IntensionalInheritance
$X$Y
IntensionalInheritance <.7>
12) Canadian people sharing properties with foreigner people and money are associated with money-laundering
IntensionalInheritance <.8>
Money-laundering
ExtensionalInheritance <.99>
Clark
14) Probability of being in the context of Accounting
Accounting <.5>
15) Probability of being in the context of Music
Music <.5>
What is the chance of Clark being involved with Money-laundering?
IntensionalInheritance <?>
Clark
Money-laundering
Next we show one possible inference trail via which PLN’s inference rules may estimate the truth value of the target logical relationship, based on the assumption of the above axioms. Of course, there exist many other inference trails as well, and in reality an automated PLN inference system (such as the ones implemented in the NCE or OCP AI systems) will find many of these and produce an overall truth value formed by revising their various conclusions. However, for expositional purposes, it seems sufficient to recount a single inference trail in detail, just show “how such inferences go.” The problem of inference control – i.e. of how an inference engine may be guided to create inferences like this in a reasonable amount of time – will be discussed in a later chapter.
Inference trail
1) Instantiate axiom #10, with $X=Clark and$Y=CanadianPlaceNames
Implication <.9>
Evaluation <.5>
Mention
List
$X$Y
IntensionalInheritance
$X$Y
2) Using the conclusion of step #1 and axiom #6 as premises of the macro PLN rule (Context C A, Implication A B ==> Context C B)
IntensionalContext
Accounting
IntensionalInheritance
Clark
3) Rewrite the conclusion of step #2 under a non-contextual form
IntensionalInheritance
Clark AND_Int Accounting
4) Using axiom #7 (and applying an axiom embodying a basic congruence rule) to replace Accounting by Money in the conclusion of step #3
IntensionalInheritance
Clark AND_Int Money
IntensionalInheritance
Clark AND_Int Money
IntensionalInheritance
Clark AND_Int Money
7) Using axiom #12 and some PLN rules to assess the IntensionalInheritance from the ExtensionalInheritance
IntensionalInheritance
Clark
8) Using the conclusion of step #6 and step #7
IntensionalInheritance
Clark AND_Int AND_Int Money
8) Using axiom #12 with step #6
IntensionalInheritance
Clark AND_Int Money
Money-laundering
9) Then we remove Money from the conclusion of step #7, or that we need to compute the strength of Clark AND_Int Money => Clark, which can be approximated by the extensional case Implication Clark AND_Ext Money => Clark, based on P(A) and P(A AND B).
IntensionalInheritance
Clark
Money-laundering
# Inferring the Causes of Observed Changes
In this chapter we consider the specific question of how the ideas of the previous chapters contribute to carrying out reasoning regarding the potential causes of salient changes in large knowledge stores. The following would be three sorts of examples of change-related inference, in the Twitter domain:
• Looking for changes in a particular person’s patterns of social interaction (a significant new contact, a number of casual acquaintances with similar profiles, etc.), and potential causes of these changes
• Looking for groups with changes in sentiment toward a certain person or organization (say, the Tory party), and potential causes of these changes
• Looking for places with a significant change in their relationship to some specific place (say, East Anglia), and potential causes of these changes
Corresonding examples on other application areas, such as robotics, are obvious to formulate. In this section we give a detailed exposition of an inference regarding the first of these example cases. The others would be handled similarly.
We will make use of the PLN logic framework here, although others could have been utilized as well. In fact no existing logic framework has been fleshed out in great detail in the context of precisely this sort of application, so whatever logical formalism one chooses, in order to approach examples like this, one is going to be carrying out a certain amount of creative improvisation. Due to our prior experience with PLN we felt most comfortable carrying out this invention in this context.
Specifically, we consider the following scenario :
• Before March 2007, Bob never had any Canadian friends except those who were also friends of his wife.
• After March 2007, Bob started acquiring Canadian friends who were not friends of his wife.
• In late 2006, Bob started collecting Pokemon cards. Most of the new Canadian friends Bob made between March 2007 and Late 2007 are associated with Pokemon cards
• In late 2006, Bob started learning French. Most of the new Canadian friends Bob made between March 2007 and Late 2007 are Quebecois.
These are the sorts of patterns that might be identified via the pattern mining algorithms discussed above, for finding surprising relationships in large logical knowledge bases. The question we consider here is: suppose such a pattern has been identified, then how do we figure out what its cause might be? We will consider two cases of the above scenario, one involving temporal reasoning only, and one involving both spatial and temporal reasoning.
The importance of this sort of question should be clear: it is not a matter of doing obscure analytical detective work, it’s a matter of figuring out whether a pattern that arises from a pattern-mining algorithm is actually interesting enough to merit anyone paying attention to it. Pattern mining algorithms tend to find a lot of patterns, and most of them are pretty uninteresting. When a pattern arises from such an algorithm, it is worthwhile to know whether there is an obvious cause for the pattern – and if so, whether the cause is the kind of cause that is interesting to the humans who are receiving the output of the pattern mining algorithm. Thus, causal inference may actually be viewed as an integral part of the pattern mining process. We may in fact posit a repeated process such as:
1. Mine patterns from the knowledge base, biased by a set of concepts and patterns called the “focus”
2. Perform causal inference to find a set P of patterns that are significant but have no known cause, or have causes believed to be interesting to the human users
3. If any of the patterns in P are estimated to be sufficiently interesting to any of the human users, report them to these human users
We have already discussed the pattern-mining portion of this process; now we turn to the causal inference aspect.
## The Case of Bob and His New Friends, with Temporal Inference Only
In this section, we will formalize the above example in PLN, and then use PLN inference to assess the strength of a few possible causal relationships that can explain why Bob has gotten new friends from Canada apart from his wife's (after March 2007).
### Axioms
1) Before March 2007, Bob never had any Canadian friends except those who were also friends of his wife.
Let's first define a predicate equivalent to “Bob's Canadian friends” :
AverageAll $X ExtensionalEquivalence Evaluation CanadianFriendOfBob$X
AND
Evaluation
FriendOf
List
Bob
$X ExtensionalInheritance$X
Then the axiom itself:
holdsThroughout <0.99>
AverageAll $X <0.99> AND_Ext Evaluation CanadianFriendOfBob$X
Evaluation
Friend
List
Bob's wife
$X Before_March_2007 2) After March 2007, Bob started acquiring Canadian friends who were not friends of his wife. First let's define the following predicate equivalent to “Bob's Canadian friends who are not friends of his wife”: AverageAll$X
ExtensionalEquivalence
Evaluation
$X AND Evaluation CanadianFriendOfBob$X
NOT Evaluation
FriendOf
List
$X WifeOfBob Then the axiom itself: initiatedThroughout Evaluation <.99> NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Between_March_2007_And_Late_2007 3) In late 2006, Bob started collecting Pokemon cards. Let's first define the following 0-ary predicate (we are passing over some subtlety here regarding defining the equivalence of a 0-ary term: but, to make a long story short, one can interpret an equivalence between 0-ary terms as an equivalence between 1-ary predicates [Goertzel et al., 2008] via considering it as a shorthand to the same equivalence holding over all contexts$C, where the context $C then becomes the argument of the predicates) ExtensionalEquivalence Evaluation BobCollectingPokemonCards Evaluation Collecting List Bob PokemonCards We may then state the axiom: initiatedAt Evaluation BobCollectingPokemonCards Late_2006 4) The process of collecting$Y shares associations with $Y itself AverageAll$X, $Y ExtensionalImplication Evaluation Collecting List$X
$Y IntensionalInheritance$Y
SatisfyingSet(Collecting)
5) Most of the new Canadian friends Bob made after March 2007 (who are not friends of his wife) are associated with Pokemon cards.
Let's first define a predicate equivalent to “Bob's friends associated with Pokemon cards”:
AverageAll $X <1> ExtensionalEquivalence <1> Evaluation FriendOfBobAssociatedWithPokemonCards$X
AND
Evaluation
FriendOf
List
Bob
$X IntensionalInheritance$X
PokemonCards
And then the axiom is defined below:
AverageAll $X <.7> ExtensionalImplication initiatedThroughout Evaluation CanadianFriendOfBob$X
Between_March_2007_And_Late_2007
Evaluation
FriendOfBobAssociatedWithPokemonCards
$X And then the axiom is defined below: AverageAll$X <.7>
ExtensionalImplication
initiatedThroughout
Evaluation
$X Between_March_2007_And_Late_2007 Evaluation FriendOfBobAssociatedWithPokemonCards$X
6) In late 2006, Bob started learning French.
Let's define the 0-ary predicate (recalling the remark made above concerning the 0-ary predicate BobCollectingPokemonCards)
ExtensionalEquivalence
Evaluation
BobLearningFrench
Evaluation
Learning
List
FrenchLanguage
Bob
So we can now define the axiom:
initiatedAt
Evaluation
BobLearningFrench
Late_2006
7) The process of learning $X shares associations with$X itself
AverageAll $X,$Y
ExtensionalImplication
Evaluation
Learning
List
$X$Y
IntensionalInheritance
$Y SatisfyingSet(Learning) 8) Most of the new Canadian friends Bob made after March 2007 (who are not friends of his wife) are Quebecois. Let's first define a predicate equivalent to “Bob's Quebecois friends” : AverageAll$X <1>
ExtensionalEquivalence <1>
Evaluation
QuebecoisFriendOfBob
$X AND Evaluation FriendOf List Bob$X
ExtensionalInheritance
$X Quebecois Now let's formulate the axiom : AverageAll$X <.7>
ExtensionalImplication
initiatedThroughout
Evaluation
$X Between_March_2007_And_Late_2007 Evaluation QuebecoisFriendOfBob$X
AverageAll $X ExtensionalImplication Evaluation Quebecois$X
Evaluation
$X 10) Quebecois are associated with French language IntensionalInheritance FrenchLanguage SatisfyingSet(Quebecois) ### Inference Trails We now describe three PLN inference trails, aimed at evaluating the validity of the following inference targets: Target theorem 1: “Bob's Pokemon cards interest is the cause of his new Canadian friendships”: IntensionalPredictiveImplication <?, 3_months_to_a_year> Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatifsyingSet(CanadianFriendOfBobButNotHisWife) Target theorem 2: “Bob starting learning French is the cause of his new Canadian friendships”: IntensionalPredictiveImplication <?, 3_months_to_a_year> Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Target theorem 3: “The disjunction of Bob starting learning French and collecting Pokemon cards is the cause of his new Canadian friendships” IntensionalPredictiveImplication <?, 3_months_to_a_year> AND Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) #### An Evaluation of Theorem 1 To illustrate one path for evaluating the truth value of Theorem 1 using PLN, we will begin by presenting four steps that go backward from the target theorem to the axioms. 1) Target 1: “Bob starting collecting Pokemon cards is the cause of his new Canadian friendships” IntensionalPredicativeImplication <?, 3_months_to_a_year> Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) 2) Using step #1 and the definition of IntensionalPredictiveImplication IntensionalImplication <?> Evaluation BobCollectingPokemonCards AND_Seq <3_months_to_a_year> Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) 3) Using step #2 and the definition of AND_Seq <T> IntensionalImplication <?> Evaluation BobCollectingPokemonCards AND AND Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) BobCollectingPokemonCards_BobsNewFriends_Delay is between 3 months to a year time, where BobCollectingPokemonCards_BobsNewFriends_Delay is a shorthand for: SS_InitiatedAt(Evaluation(BobCollectingPokemonCards)) - SS_InitiatedAt(Evaluation(NonEmpty, SatisfyingSet(CanadianFriendOfBobButNotHisWife)) where SS_InitiatedAt(E) is the start time of event E. 4) The next step proceeds using step #3 and the definition of IntensionalImplication (see Section 12.5.5) ExtensionalInheritance ExOut ASSOC Evaluation BobCollectingPokemonCards ExOut ASSOC AND AND Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobCollectingPokemonCards_BobsNewFriends_Delay is between 3 months to a year time For the next steps we create 2 sub-target theorems corresponding to the 2 ExOut ASSOC of the step above in order to assess the ExtensionInheritance thereof. 5) Let's first consider ExOut(ASSOC(Evaluation(BobCollectingPokemonCards))). In principle, at this stage, we should compute ASSOC( Evaluation( BobCollectingPokemonCards ) E) for all terms E in the system. However it is clear that what relates BobCollectingPokemonCards and his new Canadian friends in our example is the concept Pokemon cards (as most of them collect Pokemon cards as well). (In practice this would need to be concluded by an adaptive inference control mechanism, to avoid untoward combinatorial explosions during the chaining process.) Therefore we will only consider ASSOC( Evaluation( BobCollectingPokemonCards ) PokemonCards ). Since Evaluation(BobCollectingPokemonCards) is a relationship we convert it first into its concept representation, SatisfyingSet( BobCollectingPokemonCards ). It is clear that ASSOC_ext(SatisfyingSet(BobCollectingPokemonCards PokemonCards) is null, if we assume (as is most reasonable) there is no knowledge in the system establishing extensional relationship between BobCollectingPokemonCards and PokemonCards. However using an instance of axiom #4 with$X=Bob and $Y=PokemonCards we can calculate IntensionalInheritance PokemonCards SatisfyingSet(BobCollectingPokemonCards) We also need to consider IntensionalInheritance NOT PokemonCards SatisfyingSet(BobCollectingPokemonCards) which is trivially evaluated as null due to the lack of axioms leading to that relationship. Thus, given the above, finally we can compute numerically the strength of ASSOC(SatisfyingSet(BobCollectingPokemonCards) PokemonCards). 6) Let's consider the second relationship now: ExOut ASSOC AND AND Evaluation BobCollectingPokemonCards Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobCollectingPokemonCards_BobsNewFriends_Delay is between 3 months to a year time Since SS_InitiatedAt(BobCollectingPokemonCards) = Late_2006, the above conjunction is equivalent to: ExOut ASSOC AND Evaluation BobCollectionPokemonCards initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 ExOut ASSOC Evaluation(BobCollectionPokemonCards) has already been addressed, so let's now focus on the second term of the conjunction; that is, we are trying to assess the strength of: IntensionalInheritance PokemonCards SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 We set the PLN formula above as new target goal and we will go forward with the inference from that point. 6) Using the definition of FriendOfBobAssociatedWithPokemonCards from axiom #5 and considering that Equivalence(A, B) is Implication(A, B) AND Implication(B, A) we can conclude that: AverageAll$X <.7>
ExtensionalImplication
Evaluation
FriendOfBobAssociatedWithPokemonCards
$X AND Evaluation FriendOf List Bob$X
IntensionalInheritance
$X PokemonCards 7) Using the conclusion of step #6 and standard logic rules: AverageAll$X <.7>
ExtensionalImplication
Evaluation
FriendOfBobAssociatedWithPokemonCards
$X IntensionalInheritance$X
PokemonCards
8) Using the conclusion of step #7 and a PLN rule to represent Evaluation as membership of the satisfying set of the predicate:
AverageAll $X <.7> ExtensionalImplication Member$X
SatisfyingSet
FriendOfBobAssociatedWithPokemonCards
IntensionalInheritance
$X PokemonCards 9) Using the conclusion of step #8 and the M2I PLN rule (that assesses how a member of a set intensionally inherits from that set): AverageAll$X <.7>
ExtensionalImplication
IntensionalInheritance
$X SatisfyingSet FriendOfBobAssociatedWithPokemonCards IntensionalInheritance$X
PokemonCards
10) Using step #9 and the following PLN rule:
AverageAll $X ExtensionImplication Inheritnace$X F
Inheritance $X G Inheritance F G we can conclude: IntensionalInheritance SatisfyingSet(FriendOfBobAssociatedWithPokemonCards) PokemonCards 11) we rewrite axiom #5 using SatisfyingSet to convert the implication into an inheritance: Inheritance <.7> SatisfyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 SatisfyingSet FriendOfBobAssociatedWithPokemonCards 12) Using the conclusions of steps #11 and #10, and the PLN deduction rule we can conclude: Inheritance <?> SatisfyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 PokemonCards 13) Using the conclusion of step #12, and the PLN inversion rule we can conclude: Inheritance <?> PokemonCards SatisfyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 Which allows us to calculate: ASSOC_int SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 PokemonCards #### An Evaluation of Theorem 2 Now we give a similar demonstration of an inference trail leading to the evaluation of the truth value of Theorem 2 above. We will go backward from the target theorem to axioms for the next four steps. 1) Target 2: “Bob starting learning French is the cause of his new Canadian friendships”: IntensionalPredictiveImplication <?, 3_months_to_a_year> Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) 2) using step #1 and definition of IntensionalPredictiveImplication IntensionalImplication <?> Evaluation BobLearningFrench AND_Seq <3_months_to_a_year> Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) 3) using step #2 and the definition of AND_Seq <T> IntensionalImplication <?> Evaluation BobLearningFrench AND AND Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time where BobLearningFrench_BobsNewFriends_Delay is a shorthand for: SS_InitiatedAt(Evaluation(BobLearningFrench)) - SS_InitiatedAt(Evaluation(NonEmpty, SatisfyingSet(CanadianFriendOfBobButNotHisWife)) Where SS_InitiatedAt(E) is the start time of event E. 4) Using step #3 and the definition of IntensionalImplication we can rewrite the conclusion of step #3: ExtensionalInheritance ExOut ASSOC Evaluation BobLearningFrench ExOut ASSOC AND AND Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time For the next steps we create 2 sub-target theorems corresponding to the 2 ExOut ASSOC of the step above in order to assess the ExtensionInheritance thereof. 5) Let's first consider ExOut(ASSOC(Evaluation(BobLearningFrench))). As in the case of Theorem 1, in principle we should compute ASSOC(Evaluation(BobLearningFrench) E) for all terms E in the system. However it is clear that what relates BobLearningFrench and his new Canadian friends in our example is the French language (as it's new Canadian friends are in majority Quebecois whom main language is French). Therefore we will only consider ASSOC(Evaluation(BobLearningFrench) FrenchLanguage). Since Evaluation(BobLearningFrench) is a relationship we convert it first into its concept representation, StatisfyingSet(BobLearningFrench). It is clear that ASSOC_ext( SatisfyingSet(BobLearningFrench FrenchLanguage ) is null because there is no knowledge in the system establishing an extensional relationship between BobLearningFrench and FrenchLanguage. However using an instance of axiom #7 with$X=Bob and $Y=FrenchLanguage we can calculate IntensionalInheritance( FrenchLanguage SatisfyingSet( BobLearningFrench ) ). We also need to consider IntensionalInheritance(NOT FrenchLanguage SatisfyingSet( BobLearningFrench ) ) which is trivially evaluated as null due to the lack of axioms leading to that relationship. Thus finally we can compute numerically the strength of ASSOC( SatisfyingSet( BobLearningFrench ) FrenchLanguage). 6) Let's consider the second one: ExOut ASSOC AND AND Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time Since SS_InitiatedAt(BobLearningFrench) = Late_2006, the above conjunction is equivalent to: ExOut ASSOC AND Evaluation BobLearningFrench initiatedThroughout Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Between_March_2007_And_Late_2007 ExOut ASSOC Evaluation(BobLearningFrench) has already been addressed, so let's now focus on the second term of the conjunction, that is we are trying to assess the strength of: IntensionalInheritance FrenchLanguage SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 We set the PLN formula above as new target goal and we will go forward in the inference this time 7) Using the definition of QuebecoisFriendOfBob in axiom #8 we can directly conclude: ExtensionalInheritance SatisfyingSet(QuebecoisFriendOfBob) SatisfyingSet(Quebecois) 8) Using the conclusion of step #7 and axiom #10, and mixed PLN abduction we can conclude: IntensionalInheritance FrenchLanguage SatisfyingSet(QuebecoisFriendOfBob) 9) We rewrite axiom #8 using SatisfyingSet to convert the implication into an inheritance: Inheritance <.7> SatisfyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 SatisfyingSet QuebecoisFriendOfBob 10) Using axiom #8, the conclusion of step #9, and the abduction PLN rule we can conclude: IntensionalInheritance FrenchLanguage SatisfyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 11) Using the axiom #2 and the definition of the strength of CanadianFriendOfBobButNotHisWife as the average strength over all arguments considered relevant: Implication initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 initiatedThroughout Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Between_march_2007_And_Late_2007 12) Rewriting the conclusion of step #11 into an inheritance relation using SatisfyingSet: Inheritance SatifyingSet initiatedThroughout CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Between_march_2007_And_Late_2007 13) Using the conclusions of steps #10 and #12 and the PLN deduction rule: IntensionInheritance FrenchLanguage SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 This allows us to calculate: ASSOC_int SatisfyingSet initiatedThroughout Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife Between_March_2007_And_Late_2007 FrenchLanguage #### An Evaluation of Theorem 3 Finally, we will not detail the inference trail of target theorem 3 because it exhibits nothing new; instead we will informally explain how to reuse the two previous inferences to derive this conclusion. Recall from above that Target 3: “The disjunction of Bob starting collecting Pokemon cards and learning French is the cause of his new Canadian friendships” IntensionalPredicativeImplication <?, 3_months_to_a_year> OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Analogously to target theorems 1 and 2 we need to assess the following PLN formula (following the same steps backward): ExtensionalInheritance ExOut ASSOC OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench ExOut ASSOC AND AND OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time Assuming we have ExOut ASSOC Evaluation(BobCollectingPokemonCards) which is basically PokemonCards, and ExOut ASSOC Evaluation(BobLearningFrench) which is basically FrenchLanguage, it is no difficulty to conclude that: ExOut ASSOC OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench corresponds to the set {PokemonCards, FrenchLanguage} using standard PLN logical operator rules (it is represented a crisp set but in fact of course it is a fuzzy set). Similarly the same can be concluded for ExOut ASSOC AND AND OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time That is, it corresponds to {PokemonCards, FrenchLanguage} (again one should interpret it a fuzzy set). Of course it is expected that the strength of that last IntensionalInheritance is higher than the strength of the target theorem 1 or 2. This is a matter of the truth value formulas corresponding to the inference rules in the trail, which have not been entered into here as we are concerned with exposition, and the forms of the inferences involved, rather than practical calculation. #### An Evaluation of Theorem 3 Finally, we will not detail the inference trail of target theorem 3 because it exhibits nothing new; instead we will informally explain how to reuse the two previous inferences to derive this conclusion. Recall from above that Target 3: “The disjunction of Bob starting collecting Pokemon cards and learning French is the cause of his new Canadian friendships” IntensionalPredicativeImplication <?, 3_months_to_a_year> OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet(CanadianFriendOfBobButNotHisWife) Analogously to target theorems 1 and 2 we need to assess the following PLN formula (following the same steps backward): ExtensionalInheritance ExOut ASSOC OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench ExOut ASSOC AND AND OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time Assuming we have ExOut ASSOC Evaluation(BobCollectingPokemonCards) which is basically PokemonCards, and ExOut ASSOC Evaluation(BobLearningFrench) which is basically FrenchLanguage, it is no difficulty to conclude that: ExOut ASSOC OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench corresponds to the set {PokemonCards, FrenchLanguage} using standard PLN logical operator rules (it is represented a crisp set but in fact of course it is a fuzzy set). Similarly the same can be concluded for ExOut ASSOC AND AND OR Evaluation BobCollectingPokemonCards Evaluation BobLearningFrench Evaluation NonEmpty SatisfyingSet CanadianFriendOfBobButNotHisWife BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time That is, it corresponds to {PokemonCards, FrenchLanguage} (again one should interpret it a fuzzy set). Of course it is expected that the strength of that last IntensionalInheritance is higher than the strength of the target theorem 1 or 2. This is a matter of the truth value formulas corresponding to the inference rules in the trail, which have not been entered into here as we are concerned with exposition, and the forms of the inferences involved, rather than practical calculation. ## Incorporating Spatial Inference into Analysis of Change In this section we consider a variation of the inference given above, modified to include spatial reasoning involving Canada and its neighbors. This is an illustration of how temporal and spatial inference rules may be combined to carry out commonsense reasoning regarding potential causes of changes in large knowledge bases. See Figure 14.1 for the map used in the example. Figure 14.1: Canada and its neighbors We consider a query similar to one discussed above, but slightly relaxed; instead of the causes of Bob's new Canadian friendships we are interested in the causes of Bob's new friends who are Canadian and associated with Canadian; that is more, formally speaking, who inherit extensionally and/or intensionally from Canadian (instead of only extensionally as in the previous inference). So, the definition of Bob's Canadian friend in axiom #1 of the previous inference is replaced by: AverageAll$X
ExtensionalEquivalence
Evaluation
$X AND Evaluation FriendOf List Bob$X
Inheritance
$X Canadian Then we introduce some spatial knowledge by adding the predicate near(X,Y) and a “curried” version of it, nearCanada(X)=near(X,Canada), represented by a fuzzy grid in Figure 14.2, where the level of opacity of the red color of each cell of the grid corresponds to the degree of its proximity to Canada. And we extract from that grid two regions, Canada in red and Ottawa in green as shown in Figure 14.3, to illustrate the use of the Region Connection Calculus (RCC) as discussed above. And finally we add an axiom that relates geographic proximity and intensional association. Figure 14.2: Fuzzy Grid of the predicate near Canada ### New Axioms We can now formalize the above knowledge in PLN (one should consider it as a set of additional axioms, to be added to the ones from the previous example, which is why they are numbered from 11): 11) Nome Alaska is not near Canada Let's first define the predicate nearCanada: AverageAll$X
ExtensionalEquivalence
Evaluation
$X Evaluation near List$X
Figure 14.3: Canada Region in red, Ottawa Region in green
and the axiom (according to the fuzzy grid of Figure 14.2):
Evaluation <0>
Nome
Actually in practice this knowledge would be deduced from lower-level knowledge such as the following:
Evaluation <0>
occupiesRedCell
List
Nome
Cell(8,8)
where the predicate occupiesRedCell indicates whether (or to what degree) a location occupies a red cell of the grid, and using the axiom:
AverageAll $X,$Y, $Z ExtensionalImplication Evaluation occupiesRedCell List$X
Cell($Y,$Z)
Evaluation
$X But for the sake of expositional simplicity, here we will directly use the predicate nearCanada. 12) How near McCarthy Alaska is to Canada Evaluation <.6> nearCanada McCarthy 13) How near Northern Maine is to Canada (the north of ME as indicated in the map Figure 14.1). Evaluation <.8> nearCanada NorthernMaine 14) Nearby locations have intensional similarity AverageAll$X, $Y ExtensionalImplication <.6> Evaluation near$X
$Y IntensionalSimilarity$X
$Y 15) People of a given location intensionally inherit from that location AverageAll$X, $Y ExtensionalImplication <.7> Evaluation liveIn List$X
$Y IntensionalInheritance$X
$Y 16) Canadians live in Canada Evaluation liveIn List Canadian Canada 17) The disjunction of NTPP (Non-Tangential Proper Part) and TPP (Tangential Proper Part) is transitive (from the Region Connection Calculus defined above) Let's define first PP for Proper Part AverageAll$X, $Y ExtentionalEquivalence Evaluation PP List$X
$Y OR Evaluation NTPP List$X
$Y Evaluation TPP List$X
$Y And then axioms stating the transitivity of PP AverageAll$X, $Y,$Z
ExtensionalImplication
AND
Evaluation
PP
List
$X$Y
Evaluation
PP
List
$Y$Z
Evaluation
PP
List
$X$Z
18) If someone lives somewhere he/she is a Proper Part of that place
AverageAll $X,$Y
ExtensionalImplication
Evaluation
liveIn
List
$X$Y
Evaluation
PP
List
$X$Y
19) Jack lives in McCarthy
Evaluation
liveIn
List
Jack
McCarthy
20) Jim lives in Nome
Evaluation
liveIn
List
Jim
Nome
21) John lives in Ottawa
Evaluation
liveIn
List
John
Ottawa
22) Ottawa is a proper part of Canada (as represented in Figure 14.3)
Evaluation
PP
List
Ottawa
AverageAll $X ExtensionalImplication Evaluation liveIn List$X
ExtensionalInheritance
$X Canadian And that’s it for the axioms. The target theorems are unchanged from the prior example. We will not describe the entire inference here, but only the additional steps dealing with the spatial knowledge introduced and connecting them to the previous inference. What we will show is how to infer that Jack (living in McCarthy) and John (living in Ottawa) inherit from Canadian, but Jim (living further away in Nome) does not. ### Evaluating the Theorems As previously, we now have the following 3 sub-target theorems: Inheritance <?> Jack Canadian Inheritance <?> John Canadian Inheritance <?> Jim Canadian We now discuss each of these, using the new spatial information available. #### Evaluation of Theorem 1 1) Using the definition of near in axiom #1 and axiom #12 Evaluation <.6> near List McCarthy Canada 2) Using the conclusion of step #1 and an instantiation of axiom #14 with$X=McCarthy and $Y=Canada and the PLN deduction rule IntensionalSimilarity <?> McCarthy Canada 3) Using axiom #19, an instantiation of axiom #15 with$X=Jack and $Y=McCarthy and the PLN deduction rule IntensionalInheritance <?> Jack McCarthy 4) Using the definition of IntensionalSimilarity and the conclusion of step #2 AND IntensionalInheritance McCarthy Canada IntensionalInheritance Canada McCarthy 5) Using the conclusion of step #4 (stripping away the second term of the conjunction), the conclusion of step #3 and the PLN deduction rule IntensionalInheritance <?> Jack Canada 6) Using axioms #16, #15 and the PLN deduction rule IntensionalInheritance <?> Canadian Canada 7) Using the conclusions of steps #5, #6 and the PLN abduction rule IntensionalInheritance <?> Jack Canadian 8) Using the definition of Inheritance (disjunction of Extensional and Intensional inheritance) Inheritance <?> Jack Canadian So due to the additional items of spatial knowledge we can conclude that Jack inherits from Canadian, therefore assuming that after March 2007 Jack is a friend of Bob and not a friend of his wife -- knowledge which may have an influence in the calculation of the axiom #2 of the inference of the previous section, even though Jack isn't in fact from Canada strictly speaking. Of course it is clear that Jack inherits from Canadian with a lower strength than a real Canadian individual because the latter extensionally inherits from Canadian, but nevertheless Jack’s (mixed) inheritance from Canadian has a non null strength. A similar inference could be built for someone living in northern Maine. #### Evaluation of Theorem 2 1) Using axiom #21 and an instantiation of axiom #18 with$X=John, $Y=Ottawa Evaluation PP List John Ottawa 2) Using step #1, axiom #22 and the PLB deductive rule Evaluation PP List John Canada 3) Using step #2, axiom #18 with$X=John, $Y=Canada Evaluation liveIn List John Canada 4) Using step #3 and an instantiation of axiom #23 with$X=John
ExtensionalInheritance <?>
John
5) Using the definition of Inheritance (disjunction of extensional and intensional inheritance)
Inheritance <?>
John
So using the Region Connection Calculus within PLN, one can conclude that John is Canadian and assuming that John is a friend of Bob after March 2007 and not a friend of his wife may have a influence over the strength of axiom #2 in the initial example.
#### Evaluation of Theorem 3
This inference is essentially identical to the inference of Theorem 1, so we will not recall it.The only difference is that the strength of (IntensionalInheritance Jim Canadian) is null, and since Nome is not in Canada this ExtensionalInheritance is null as well, therefore
Inheritance <0>
Jim
As the discussion in the above chapters has hopefully made clear, every one of the tasks involved in logic-based RWR has been carefully approached by computer scientists, using various formalisms and software prototypes. We have produced more fully fleshed-out examples using PLN than other formalisms, because PLN is the inference framework we’re most familiar with and because we believe it most adequately integrates rich uncertainty representations with powerful inference mechanisms; but similar explorations could be produced using other inference formalisms, and of course one could also go into far greater detail than has been done above.
But there’s a catch, of course: None of the formalisms or prototypes described in the literature (including PLN) comes along with fully confident theoretical or empirical knowledge telling one how to deal effectively with real-world-scale data stores. This is not just a quantitative problem: it’s a qualitative problem. Addressing this problem adequately requires the implementation and adoption of fundamentally new ideas, complementing existing approaches,
Assuming a logic-based approach, scalability of pattern mining, query processing and analysis boils down to efficient inference control. Which of course, is in general terms a monstrously hard problem. Making inference control generally efficient across all possible large knowledge stores is almost surely impossible.
The need for efficient inference control should be very clear via looking at the detailed PLN inference trails supplied above. At each step in any one of those inference trails, there were many other inferences that could have been done, aside from the ones shown. The question becomes how to choose these exact ones, or others with similar effect. If there are 20 steps in an inference, and 100 possibilities for each step (actually a terrible underestimate of the real situation), then there are 10020 possible inference trails to be theoretically considered. If there are several thousand viable possibilities for each step, which is more likely the case in an inference involving a large knowledge store, then the number of possible inference trails becomes even more astronomical. Fortunately there are also many paths leading to useful conclusions, not just one – but even so, if you view it as a search problem, finding useful inference trails amidst the space of all possible ones is a lot worse than finding a few needles in a planet-sized haystack.
Even without the complications of large and uncertain knowledge stores, the inference trail pruning problem is a very difficult one. This is the essential reason why automated theorem proving has not yet obsoleted human mathematicians. Automated theorem provers, using formal logic, can carry out individual inference steps adeptly (and without making the “stupid mistakes” that even smart humans are sometimes prone to), but in most cases they are vastly inferior to humans at choosing which inference steps to take. Some of the biggest successes in automated theorem proving have occurred in highly restricted contexts (where there aren’t that many possibilities to choose from) or else using a semi-automated methodology, where a human expert intervenes to make choices at places where the AI gets confused at the variety of choices and lacks the experience to make an informed decision.
So, clearly, in its fully general scope, the problem of inference control is practically unsolvable! But, does that mean the whole research programme outlined in this book is infeasible? No, because we don’t need to solve the general problem. To make the programme described here work, the problem needs to be solved only in the context of commonsensical inferences involving the types of knowledge actually stored in the knowledge store. And, to put the point mathematically, it’s clear that real-world temporal and spatiotemporal knowledge stores have special statistical properties, which have mostly not been carefully characterized. So, we suggest that a key focus for research going forward must be the creation of inference control mechanisms that adaptively exploit the statistical structure of real-world temporal and spatiotemporal data, so as to achieve efficient inference control over large real-world knowledge stores.
## Specific Examples Requiring Adaptive Inference Control
The above point may sound like a very abstract one, so to make it concrete, let’s consider a couple inference steps drawn from one of the above inference trails: the proof of Theorem 1 given above.
Quoting directly from there (but leaving out a small amount of explanatory text), we had the following inference step
3) Using step #2 and the definition of AND_Seq <T>
IntensionalImplication <?>
Evaluation
BobCollectingPokemonCards
AND
AND
Evaluation
BobCollectingPokemonCards
Evaluation
NonEmpty
BobCollectingPokemonCards_BobsNewFriends_Delay is between 3 months to a year time,
4) The next step proceeds using step #3 and the definition of IntensionalImplication.
Based on this we can rewrite the conclusion of step #3:
ExtensionalInheritance
ExOut ASSOC
Evaluation
BobCollectingPokemonCards
ExOut ASSOC
AND
AND
Evaluation
BobCollectingPokemonCards
Evaluation
NonEmpty
BobCollectingPokemonCards_BobsNewFriends_Delay is between 3 months to a year time
For the next steps we create 2 sub-target theorems corresponding to the 2 ExOut ASSOC of the step above in order to assess the ExtensionInheritance thereof.
The above is perfectly straightforward once you’ve decided to do it, but where does that decision come from? Basically the decision in moving from Step 3 to Step 4 in this inference is to expand the IntensionalImplication in 3 into an ExtensionalInheritance as shown in 4. However, the utility of this step is obvious only in hindsight. One doesn’t always want to handle an IntensionalImplication by expanding it into an ExtensionalInheritance. Sometimes, for instance, one might want to try to produce it via deduction from other IntensionalImplication -- or via unification by binding values to variables in some other IntensionalImplication with a similar form and some overlapping terms, but a lot more free variables. One could also try to derive it using Bayes rule from some other existing IntensionalImplications. Et cetera.
In many contexts, expanding an IntensionalImplication into an ExtensionalInheritance may be viewed as a “tactic of last resort”, because it’s often going to devolve into a detailed analysis of many special cases. So in many contexts (not necessarily all) it will be an advantageous strategy to first try to evaluate an IntensionalImplication via first-order-inference-style rules from other IntensionalImplications, or via unification from more abstract knowledge, before resorting to the reduction to extensionality. But even if this strategy is followed, there’s the question of how much effort to expend on other methods before resorting to extensionality. And this is bound to be context-dependent. For instance, it may happen that in certain contexts, reasoning purely on the level of intensional relationships has systematically proved difficult, and reductions to extensionality have proved necessary very frequently; whereas in other contexts, reduction to extensionality has rarely proved necessary.
Proceeding to the next step in the above inference, and again quoting directly (with some explanatory text removed), we had:
5) Let's first consider ExOut(ASSOC(Evaluation(BobCollectingPokemonCards))).
In principle, at this stage, we should compute ASSOC( Evaluation( BobCollectingPokemonCards E) ) for all terms E in the system. However it is clear that what relates BobCollectingPokemonCards and his new Canadian friends in our example is the Pokemon cards (as most of them collect Pokemon cards as well). (In practice this would need to be concluded by an adaptive inference control mechanism, to avoid untoward combinatorial explosions during the chaining process.) Therefore we will only consider ASSOC( Evaluation( BobCollectingPokemonCards PokemonCards)).
As noted there, without some kind of mechanism to restrict focus of E to PokemonCards, the inference process will consume an excessive amount of computational resources evaluating irrelevant associations. An example of the kind of mechanism that can be helpful here is activation spreading (see e.g. [Collins et al, 1975; Anderson, 1983; Crestani, 1997; Nilsson, 1998]), which is carried out in various existing AI systems using a variety of mechanisms (for instance, in neural nets it uses activation spreading; in the NCE/OpenCog architecture it uses artificial economics methods (Goertzel, 2007); etc.). Using activation spreading, if all the terms involved in Steps 1-4 were to spread some sort of activation to the terms and relationships related to them, and this activation were to then further propagate to the relatives of these terms and relationships (and so forth) – then, quite likely, PokemonCards would get more activation than nearly any other category in the overall knowledge store, and would be nominated for investigation as indicated in the above text.
Note that in principle one could do the above purely by inference, without introducing an external mechanism such as activation spreading. For instance, one could do inference on links such as (Association A B) denoting generic associative semantics, using probabilistic inference rules designed for such inference. In fact the NCE/OpenCog design contains this possibility. However, unless one adopts an extremely constrained control mechanism for this associational inference, then one is faced once again here with a potential combinatorial explosion problem. And if one does adopt an extremely restrained control mechanism here, then in effect one is basically using uncertain inferential algebra to carry out spreading activation (which may be a fine thing to do).
The inference proceeds:
But, ASSOC(C E) can be approximated as ASSOC_int OR ASSOC_ext ...
It is clear that ASSOC_ext(SatisfyingSet(BobCollectingPokemonCards PokemonCards) is null, if we assume (as is most reasonable) there is no knowledge in the system establishing extensional relationship between BobCollectingPokemonCards and PokemonCards.
However using an instance of axiom #4 with $X=Bob and$Y=PokemonCards we can calculate
IntensionalInheritance
PokemonCards
SatisfyingSet(BobCollectingPokemonCards)
Again, while this is a simple step, there were many other options here besides using Axiom #4. This is another case where some sort of activation spreading mechanism could help a lot, via providing a strong associative link between PokemonCards and BobCollectingPokemonCards. If such an associative link exists, and there is an inference control heuristic biasing PLN toward building inferential links that generally follow the path laid down by associational links, then the above inference step would indeed emerge as obvious.
We could similarly “psychoanalyze” every step in the above inference trails, studying the various alternatives and how they could be explored using intelligent inference control heuristics. However, we will restrict ourselves to giving two more examples, intended to illustrate the way in which commonsense knowledge about time and space is important for inference control.
### Using Commonsense Knowledge about Space in Inference Control
As an example of using commonsense knowledge about spatial relations to control inference, let us first recall the following axiom from the above spatial PLN inference:
18) If someone lives somewhere he/she is a Proper Part of that place
AverageAll $X,$Y
ExtensionalImplication
Evaluation
liveIn
List
$X$Y
Evaluation
PP
List
$X$Y
This axiom was used in the following inference step:
2) Using step #1, axiom #22 and the PLN deductive rule
Evaluation
PP
List
John
3) Using step #2, axiom #18 with $X=John,$Y=Canada
Evaluation
liveIn
List
John
The question here pertaining to inference control is: why would a PLN inference engine choose to make this step rather than some other one? Obviously, not every proper-part relationships is going to be equally promising for hypothetical transformation into a live-in relationship. The answer here however is conceptually obvious: where people and places are concerned, living-in is a common variety of proper-part relationship. This conceptually obvious answer may however be expressed formally in a number of different ways.
It could be expressed explicitly as an implication, e.g.
18’)
AverageAll $X,$Y
ExtensionalImplication
AND
Inheritance $X Human Inheritance$Y Place
Evaluation
PP
List
$X$Y
Evaluation
liveIn
List
$X$Y
In this case, this more specialized axiom 18’ could be used instead of Axiom 18. It would be more likely to be chosen if there were an inference control heuristic in use stating that: The weight assigned to an application of the variable unification rule between a more general Atom A and a more specific Atom B, should be proportional to how surprising it is to find Atoms that bind with A as well as B does. In this case, the point would be that the assignment $X=John,$Y=Canada matches 18’ surprisingly well (relative to most possibly assignments); and that it matches 18’ more surprisingly well than it matches 18.
However, the same effect could be achieved via an activation spreading mechanism. All one needs is for associative linkages such as the following to exist:
• Association John Human
• Association Human liveIn
• Association Place liveIn
Of course many other variants are also possible; for instance one could have Inheritance relations in place of the first two Association relations in the above list; or one could have an association
Association (Human AND Place) liveIn
and so forth. This is a simpler approach than the approach of introducing 18’ in place of 18, but ultimately, in this context, the two achieve the same thing.
One way or another, the key here is that the system has specialized knowledge about the connections between spatial relationships (ProperPart in this case) and specific content (people living in places, in this case), and that it deploys this specialized knowledge to guide it along the right inference trail, thus avoiding the combinatorial explosion of getting dragged down irrelevant trajectories.
### Using Commonsense Knowledge about Time in Inference Control
Time plays a role quite similar to space in inference control, yet even more fundamental. Without inference control heuristics that specifically take into account the habitual connections between temporal relationships and specific content, doing scalable temporal reasoning is not likely to be possible.
Let us consider the following inference step, drawn from the foregoing:
2) using step #1 and definition of IntensionalPredictiveImplication
IntensionalImplication <?>
Evaluation
BobLearningFrench
AND_Seq <3_months_to_a_year>
Evaluation
BobLearningFrench
Evaluation
NonEmpty
3) using step #2 and the definition of AND_Seq <T>
IntensionalImplication <?>
Evaluation
BobLearningFrench
AND
AND
Evaluation
BobLearningFrench
Evaluation
NonEmpty
BobLearningFrench_BobsNewFriends_Delay is between 3 months to a year time
Here, BobLearningFrench_BobsNewFriends_Delay is a shorthand for:
SS_InitiatedAt(Evaluation(BobLearningFrench))
The question here, control-wise, is why the decision would be made to expand the AND_Seq into a regular AND, thus splitting out the time-constraint into an explicit logical relationship rather than leaving it in the metadata attached to the AND_Seq. The judgment call that must be made, to justify attaching a high weight to this step, regards the likelihood that the sequentiality is really important in the given example. Is it the case that the two things just happened to occur in sequence?; or is the fact that they occur in sequence a critical aspect of their interrelationship? Of course, this won’t generally be known except in hindsight after more of the inference has been completed, but one has to make a reasonable guess. If the guess is that sequentiality is not critical, then the inference step taken in 3 above may likely be the right one, and should properly be assigned a relatively high weight in a bandit problem based inference control mechanism.
It seems the simplest heuristic that would do the trick here is the assumption that, if there is no special knowledge that A and B really need to be sequential, then the pathway of expanding (AND_Seq A B) into (AND A B) is potentially well worth exploring. For instance, if we have
AND_Seq
Evaluation TurnOn (Ben, BensCar)
Evaluation Drive (Ben, BensCar)
then there may be prior knowledge that sequence is important here, for instance because it may be that in general
AND_Seq
Evaluation TurnOn ($X, Car) Evaluation Drive ($X, Car)
is much more frequent than
AND_Seq
Evaluation Drive ($X, Car) Evaluation TurnOn ($X, Car)
So, in this case, morphing
AND_Seq <T>
Evaluation TurnOn (Ben, BensCar)
Evaluation Drive (Ben, BensCar)
into
AND
AND
Evaluation TurnOn (Ben, BensCar)
Evaluation Drive (Ben, BensCar)
TurnOnDriveDelay lies in T
would likely not be productive, unlike in the case discussed above with BobLearningFrench_BobsNewFriends_Delay.
So, here we have a case where some fairly sophisticated investigation of the knowledge base may be useful for adjusting the weights attached to various optional paths in the inference tree. And, analogously to the spatial reasoning case, the key here is knowledge connecting temporal relationships with specific content (language, friends, turning-on, driving, etc.). This kind of specialized knowledge embodies the particular statistics of real-world knowledge stores, which is fortunately not the same as the statistics of a random knowledge store (or else real-world inference would not be feasible).
## General Issues Raised by the Above Examples
A key point we wish to emphasize is that the full, realistic problem of inference control is not encountered when one runs inference engines as “toy systems” or prototypes. If one adds a relatively small number of knowledge into an inference engine, it’s not terribly hard to supply it with simple, general inference control heuristics allowing it to do inferences like the examples given in previous chapters. The really hard part comes when you couple an inference engine with a massive knowledge store, because then the problem of drawing inferences becomes inextricably tangled up with the problem of figuring out which data items are relevant enough to be sensibly used in a given inference.
To handle the “scalable inference control problem,” we have hypothesized, will require a combination of specialized heuristics (such as the ones described in the immediately preceding sections) and the integration of inference with non-inferential methods such as activation spreading. But whether or not this hypothesis is correct, what is plain is that any logic-based approach that is going to handle large knowledge stores (especially large, complex spatiotemporal knowledge stores) is going to need to deal with this problem somehow – and there is not much current research explicitly addressing this area.
As noted above, the idea of “exploiting the statistics implicit in real-world knowledge stores” arises here implicitly. For instance, if drawing inferences that follow the lines of associations works effectively, this is an example of “special statistics” – in a general, mathematically random knowledge store, this kind of heuristic would provide little if any value. And if there are certain contexts in which, systematically, reducing intensional relations to extensional ones is differentially more useful than in the average context – again, this is an example of “special statistics” that would not likely occur in a random knowledge store. To the extent that appropriate heuristics and design principles can be created, implicitly or explicitly representing systematic biases that tend to be present in real-world data, the inference control problem will be solvable.
### Inference Control and Cognitive Architectures
Work in the area of cognitive architecture also has some relevance here; architectures like SOAR [Wray & Jones, 2005] and ACT-R [Stewart & West, 2006] are specifically aimed at processing information in a way that avoids combinatorial explosions via restricting cognition’s attention to information items that are directly relevant to the tasks at hand. Underlying such architectures is the idea of implicitly embedding assumptions about the statistical regularities of the real world in the cognitive architecture itself. However, these architectures have not yet been applied in a really scalable way either, and nor have they been tested in combination with highly powerful formal inference systems. SOAR has been used in conjunction with real-world spatiotemporal data in some contexts, e.g. the TacAir flight simulator project [Jones et al., 1993], which is promising.
So it seems that, to some extent, the final solution to the all-important inference control problem may involve a fusion of ideas from the temporal, spatial and uncertain logic literatures with ideas from the cognitive architecture literature. This is one way of describing aspects of the direction we’ve taken in our own work with the NCE and OpenCog; but of course it’s an idea that can be fleshed out in many different ways. However, the elaboration of the ways that various cognitive architectures might enable effective scalable inference control would lead us too far beyond the scope of this book.
## Inference Control in the OpenCog Cognitive Architecture
We end this chapter with some brief remarks about the specific flavor that adaptive inference control takes in the OpenCog cognitive architecture, within which PLN is currently embedded.
### Activation Spreading and Inference Control in OpenCog
A number of our remarks on inference control above mentioned “activation spreading.” This is a general notion with a long history in cognitive science and AI, which may be implemented in many different ways. In this section we will briefly describe one such way, ECAN, which is implemented in the OpenCog framework and is currently being integrated with the PLN inference engine.
ECAN, or Economic Attention Networks, constitutes a novel method for simultaneously storing memories and allocating resources in AI systems. It bears some resemblance to the spread of activation in attractor neural networks, but differs via explicitly differentiating two kinds of “activation” (Short Term Importance, related to processor allocation; and Long Term Importance, related to memory allocation), and in using equations that are based on ideas from economics rather than approximative neural modeling.
An ECAN is a graph, consisting of untyped nodes and links, and also links that may be typed either HebbianLink or InverseHebbianLink. It is also useful sometimes to consider ECANs that extend the traditional graph formalism and involve links that point to links as well as to nodes. The term Atom will be used to refer to nodes and links collectively. Each Atom in an ECAN is weighted with two numbers, called STI (short-term importance) and LTI (long-term importance). Each Hebbian or InverseHebbian link is weighted with a probability value.
The equations of an ECAN explain how the STI, LTI and Hebbian probability values get updated over time. The metaphor underlying these equations is the interpretation of STI and LTI values as (separate) artificial currencies. The motivation for this metaphor has been elaborated somewhat in (Goertzel, 2007) and will not be recapitulated here. The fact that STI (for instance) is a currency means that the total amount of STI in the system is conserved (except in unusual instances where the ECAN controller decides to introduce inflation or deflation and explicitly manipulate the amount of currency in circulation), a fact that makes the dynamics of an ECAN dramatically different than that of, say, an attractor neural network (in which there is no law of conservation of activation).
Conceptually, the STI value of an Atom is interpreted to indicate the immediate urgency of the Atom to the ECAN at a certain point in time; whereas the LTI value of an Atom indicates the amount of value the ECAN perceives in the retention of the Atom in memory (RAM). An ECAN will often be coupled with a Forgetting process that removes low-LTI Atoms from memory according to certain heuristics.
STI and LTI values will generally vary continuously, but the ECAN equations we introduce below contain the notion of an AttentionalFocus (AF), consisting of those Atoms in the ECAN with the highest STI value. The AF is given its meaning by the existence of equations that treat Atoms with STI above a certain threshold differently.
Conceptually, the probability value of a HebbianLink from A to B is the odds that if A is in the AF, so is B; and correspondingly, the InverseHebbianLink from A to B is weighted with the odds that if A is in the AF, then B is not. A critical aspect of the ECAN equations is that Atoms periodically spread their STI and LTI to other Atoms that connect to them via Hebbian and InverseHebbianLinks; this is the ECAN analogue of activation spreading in neural networks.
In an OpenCog context, ECAN consists of a set of Atom types, and then a set of MindAgents carrying out ECAN operations such as HebbianLinkUpdating and ImportanceUpdating. OCP also requires many other MindAgents carrying out other cognitive processes such as probabilistic logical inference according to the PLN system (Goertzel et al, 2008) and evolutionary procedure learning according to the MOSES system (Looks, 2006). The interoperation of the ECAN MindAgents with these other MindAgents is a subtle issue that will be briefly discussed in the final section of the chapter, but the crux is simple to understand.
The CogServer is understood to maintain a kind of central bank of STI and LTI funds. When a non-ECAN MindAgent finds an Atom valuable, it sends that Atom a certain amount of Stimulus, which results in that Atom’s STI and LTI values being increased (via equations to be presented below, that transfer STI and LTI funds from the CogServer to the Atoms in question). Then, the ECAN ImportanceUpdating MindAgent carries out multiple operations, including some that transfer STI and LTI funds from some Atoms back to the CogServer – keeping the flow of money going.
All this represents one among many possible ways of implementing the general notion of activation spreading mentioned in the above inference control examples. A key point is that making activation spreading work well for inference control will likely require extremely tight integration between the inference engine and the association-spreading engine; and the OpenCog approach presents one way of providing such integration.
### Working around the Frame Problem via Integrative AGI
We noted in Chapter 2 that nonmonotonic logic is not the only route for circumventing the frame problem; and in fact in our own work with PLN in the NCE and OpenCog, we have taken a significantly different approach. Our own approach involves two key ingredients:
• Heavy use of activation spreading to guide inference, as illustrated in several of the examples given in the previous section. This mitigates against a reasoning system actually spending its time doing inferences that are useless because they pertain to assumptions that are implicit and considered obvious. The idea is that these background assumptions just don’t get stimulated with much “juice” (which may formally be an activation level, an importance value, or take some other form depending on the AI system in question; in NCE or OCP it is a ShortTermImportance currency value, as briefly discussed above).
• Use of a hierarchical ontology within inference control, to provide a form of “default logic” that is contained in the inference control mechanism rather than, in nonmonotonic logic, in the logical formalism itself. Effective use of the hierarchical ontology relies on the existence of a robust activation spreading mechanism, in a way that will be described below.
To exemplify the notion of default inheritance, consider again the case of penguins, which do not fly, although they are a subclass of birds, which do fly. When one discovers a new type of penguin, say an Emperor penguin, one reasons initially that they do not fly – i.e., one reasons by reference to the new type’s immediate parent in the ontological hierarchy, rather than its grandparent. In standard default logic frameworks, the notion of hierarchy is primary and default inheritance is wired in at the inference rule level. But this is not the case with PLN – in PLN, correct treatment of default inheritance must come indirectly out of other mechanisms, this can be achieved in a fairly simple and natural way.
Consider the two inferences
A)
Implication penguin fly <0>
Implication bird penguin <.02>
Implication bird fly
B)
Implication penguin bird <1>
Implication bird fly <.9>
Implication penguin fly
The correct behavior in these cases, according to the default inheritance idea is that, in a system that already knows at least a moderate amount about the flight behavior of birds and penguins, inference A should be accepted but inference B should not. That is, evidence about penguins should be included in determining whether birds can fly – even if there is already some general knowledge about the flight behavior of birds in the system. But evidence about birds in general should not be included in estimating whether penguins can fly, if there is already at least a moderate level of knowledge that in fact penguins are atypical birds in regard to flight.
But how can the choice of A over B be motivated in terms of PLN theory? The essence of the answer is simple: in case B the independence assumption at the heart of the deduction rule is a bad one. Within the scope of birds, being a penguin and being a flier are not at all independent. On the other hand, looking at A, we see that within the scope of penguins, being a bird and being a flier are independent. So the reason B is ruled out is that if there is even a moderate amount of knowledge about the truth-value of (Inheritance penguin fly), this gives a hint that applying the deduction rule’s independence assumption in this case is badly wrong.
On the other hand, what if a mistake is made and the inference B is done anyway? In this case the outcome could be that the system erroneously increases its estimate of the strength of the statement that penguins can fly. On the other hand, the revision rule may come to the rescue here. If the prior strength of (Inheritance penguin fly) is 0, and inference B yields a strength of .9 for the same proposition, then the special case of the revision rule that handles wildly different truth-value estimates may be triggered. If the 0 strength has much more confidence attached to it than the .9, then they won’t be merged together, because it will be assumed that the .9 is an observational or inference error. Either the .9 will be thrown out, or it will be provisionally held as an alternate, non-merged, low-confidence hypothesis, awaiting further validation or refutation.
What is more interesting, however, is to consider the implications of the default inference notion for inference control. It seems that the following may be a valuable inference control heuristic:
1. Arrange terms in a hierarchy; e.g., by finding a spanning DAG of the terms in a knowledge base, satisfying certain criteria (e.g., maximizing total strength*confidence within a fixed limitation on the number of links).
2. When reasoning about a term, first do deductive reasoning involving the term’s immediate parents in the hierarchy, and then ascend the hierarchy, looking at each hierarchical level only at terms that were not visited at lower hierarchical levels.
This is precisely the “default reasoning” idea – but the key point is that in PLN it lives at the level of inference control, not inference rules or formulas. In PLN, default reasoning is a timesaving heuristic, not an elementary aspect of the logic itself. Rather, the practical viability of the default-reasoning inference-control heuristic is a consequence of various other elementary aspects of the logic, such as the ability to detect dependencies rendering the deduction rule inapplicable, and the way the revision rule deals with wildly disparate estimates.
One way to embody the above ideas in concrete AI system design is to define a notion of OntologicalInheritance, based on the spanning DAG mentioned above. One can build a default logic based on the spanning DAG G, as follows. Suppose X lies below A in the DAG G. And, suppose that, for predicate F, we have
F(A) <t>
meaning that F applies to A with truth value t. Then, we may say that “If ~F(X) is not known, then F(X)”. In other words, we may assume by default that X possesses the properties of the terms above it in the hierarchy D. More formally, we might propose a “default inference rule” such as:
Implication
AND
OntologicalInheritance X A
Evaluation F A
NOT
Known( NOT (Evaluation F X ) )
Evaluation F X
There is no reason that a rule like this can’t be implemented within PLN. Note that implementing this rule within PLN gives you something nice, which is that all the relationships involved (OntologicalInheritance, Known, NOT, etc.) may be probabilistically quantified, so that the outcome of the inference rule may be probabilistically quantified. The “Known” predicate is basically the K predicate from standard epistemic logic (commonly denoted Ka, where a is the reasoning system itself in this case) [Fagin et al, 2003; Rescher, 2005].
The hierarchy G needs to be periodically rebuilt as it is based on abstracting a DAG from a graph of probabilistic logical relations. And, the results of the above default inference rule may be probabilistic and may be merged with the results of other inference rules.
The results from using this rule, in principle, should not be so different from if the rule were not present. However, the use of an ontological hierarchy in this way leads to a completely different dynamics of inference control, which in practice will lead to different results.
And, note finally that for this sort of ontological approach can also be used in the context of activation spreading. In this case, it may take the form (for example) of ontology-based pruning of Assocation relationships. The Association between pigeon and flyer may be pruned because it is essentially redundant with a set of two Association relationships that are implicit in the hierarchy:
Association pigeon bird
Association pigeon flyer
Using an ontology-pruned set of Association links to guide inference that is partially based on ontology-driven default inference rules, provides an alternative approach to solving the frame problem, that doesn’t require introducing a notion of hierarchy into the underlying logic.
In this section we have already, perhaps, ventured too far from the focus of this book, which is logic, into the domain of integrative cognitive architecture. But we have done so with a purpose – to illustrate our belief that, in fact, the most likely route to effective control of RWR lies in integrative cognitive architecture, i.e. in mixing up reasoning with other cognitive processes that are not explicitly inferential. This may be done in different ways within different cognitive architectures, and is in our view an extremely important subject of research in the quest for effective RWR for both AGI and near-term application purposes.
# Conclusion
As emphasized from the start, the main purpose of this book has been to serve as a sort of “thought experiment.” We began with a very general problem of “querying, mining and analyzing huge real-world knowledge stores” – i.e., Real-World Reasoning. We then made the choice to characterize this problem in terms of formal logic, and in this context considered its various aspects:
• How to represent real-world knowledge in logical form?
• How to pragmatically translate knowledge from other formats into logical form?
• How to store, retrieve and manipulate large amounts of logical knowledge?
• How should software applications query large logical knowledge stores?
• How to best manage the uncertainty that is rampant in real-world knowledge and conclusions?
• How to represent and reason about time, space and context within an uncertain logic based system?
• How to mine patterns from large logical knowledge stores, and how and why to interface pattern-mining algorithms with inference algorithms
• How to control and direct inference algorithms so as to make them scalable to large knowledge stores?
For each of these questions, we have surveyed the available literature, and in some cases introduced our own ideas when the literature appeared insufficient.
As we have not (yet) actually constructed a scalable, uncertain logic-based system for querying, mining and analyzing large stores of real-world knowledge, we cannot claim any definitiveness for the original ideas we’ve introduced here, following our extensive literature review. But we do believe we have demonstrated, at least, that this is an extremely promising area for future investigation. Having pursued the RWR thought-experiment as far as we have done in these pages, we have little doubt that the construction of scalable uncertain logic based systems would constitute a viable approach to the problem of querying, mining and analyzing changes and other relevant patterns in large real-world knowledge stores.
# References
Aehlig, K., & Beckmann, A. (2007). Propositional Logic for Circuit Classes. Proceedings of the 21st International Workshop, CSL 2007, 16th Annual Conference of the EACSL, Lausanne, Switzerland, September 11-15, 2007.
Aiello, M. & Pratt-Hartmann, I. & van Benthem, J. editors (2007). Handbook of Spatial Logics, pp. 497-564. Springer.
Akman, V., & Surav, M. (1996). Steps toward formalizing context. AI Magazine 17 (3), 55-72, 1996.
Allen, J. F. (1983). Maintaining Knowledge about Temporal Intervals. Communications of the ACM 26, 11, 832-843, November 1983.
Allen, J. F. (1984). Towards a general theory of action and time, Artificial Intelligence, volume 23, pages 123-154, 1984.
Anderson, J. (1983). "A spreading activation theory of memory." Journal of Verbal Learning and Verbal Behavior
Angrist, J.D. & Imbens G.W. & Rubin, D.B. (1996). Identification of Causal Effects Using Instrumental Variables: Rejoinder. Journal of the American Statistical Association, Vol. 91, No. 434., pp. 468-472, 1996.
Åqvist, L. (2005). Combinations of tense and deontic modality: On the Rt approach to temporal logic with historical necessity and conditional obligation. J. Applied Logic 3(3-4): 421-460 (2005)
Bacchus, F., Tenenberg, J.D., Koomen, J.A.G.M. (1991). A Non-Reified Temporal Logic. Artif. Intell. 52(1): 87-108 (1991)
Barrett, C. & Sebastiani, R. & Seshia, S.A. & Tinelli C. (2009). Satisfiability Modulo Theories. In: Handbook of Satisfiability 2009.
Batson, B., Lamport, L. (2003). High-Level Specifications: Lessons from Industry, Formal Methods for Components and Objects, Lecture Notes in Computer Science number 2852, 242-262, Springer, 2003.
Bettini, C. & Brdiczka, O. & Henricksen, K. & Indulska, J. & Nicklas, D. & Ranganathan, A. & Riboni, D. (2010). A survey of context modelling and reasoning techniques. Pervasive and Mobile Computing 6(2): 161-180 (2010)
Bloch, I., & Saffiotti, A. (2003). On the Representation of Fuzzy Spatial Relations in Robot Maps. In: B. Bouchon-Meunier, L. Foulloy and R.R. Yager (eds) Intelligent Systems for Information Processing, pp. 47-57. Elsevier, NL, 2003.
Blostein, D., E. Lank, R. Zanibbi, "Treatment of Diagrams in Document Image Analysis", Lecture Notes in Computer Science, vol. 1889, pages 330-344, 2000 Blostein, D., J.R. Cordy, R. Zanibbi, "Applying Compiler Techniques to Diagram Recognition", Proc. 16th IAPR International Conference on Pattern Recognition, 2002
Bollobás, B. (1998). Modern Graph Theory. Springer, 1998
Bennacer, N. (2006). Formalizing Mappings for OWL Spatiotemporal Ontologies. Lecture Notes in Computer Science, Springer Berlin / Heidelberg.
Bouquet, P., Ghidini, C., Giunchiglia, F., & Blanzieri, E. (2001). Theories and uses of context in knowledge representation and reasoning. Technical Report DIT-02-010, Informatica e Telecomunicazioni, University of Trento, 2001.
Bresina J.L., Jonsson A.K., Morris P.H., Rajan K. (2005). Activity Planning for the Mars Exploration Rovers. Proceedings of the International Conference on Automated Planning and Scheduling/Artificial Intelligence Planning Systems - ICAPS(AIPS), 2005.
Brezillon, P. (1999). Context in problem solving: A survey. The Knowledge Engineering Review 14 (1), 1-34, 1999.
Bry, F. & Jacquenet, F. (2005). First Order Logic for Learning User Models in the Semantic Web: Why Do We Need It? Workshop on Machine Learning for User Modeling: Challenges, Edinburgh, Scotland, 2005, http://www-connex.lip6.fr/~artieres/UM2005/
Broersen J., Dignum F., Dignum, V., & Meyer, J-J. Ch. (2004). Designing a Deontic Logic of Deadlines. 7th International Workshop on Deontic Logic in Computer Science, DEON 2004, Madeira, Portugal, May 26-28, 2004.
Burke, R. (2000). Knowledge-based recommender systems, Encyclopedia of Library and Information Science, Vol. 69(32), 2000.
Buvac, S., & Mason, I. (1993). Propositional logic of context. Proc. of the 11th American National Conference on Artificial Intelligence (AAAI-93), 412-419, 1993.
Chakrabarti, S., S. Sarawagi, and B. Dom (1998) Mining surprising patterns using temporal description length. Proceedings of the 24th International Conference on Very Large Data Bases .
Cheng, J. (2006). Temporal Deontic Relevant Logic as the Logical Basis for Decision Making Based on Anticipatory Reasoning. IEEE International Conference on Systems, Man and Cybernetics, 2006. SMC '06.
Claypool, M., Le, P., Waseda, M., & Brown, D. (2001). Implicit interest indicators. In Proceedings of Intelligent User Interfaces 2001, pages 33–40, 2001.
Cohn, A.G., Bennett, B., Gooday, J.M., Gotts, N. (1997). RCC: a calculus for Region based Qualitative Spatial Reasoning. GeoInformatica 1(3):275-316 (1997)
Collins, A. Loftus, E. (1975). "A spreading-activation theory of semantic processing", Psychological Review. 1975 Nov Vol 82(6) 407-428
Copi, I., & Cohen, C. (1998). Introduction to Logic. Prentice Hall College Div; 10th edition (August 1998).
Cowie J. & Wilks Y. (2000). Information Extraction. Handbook of natural language processing, edited by Robert Dale, Hermann Moisl, H. L. Somers, 2000. Marcel Dekker.
Craven, M., DiPasquo, D., Freitag, D., McCallum, A., Mitchell, T., Nigam, K., & Slattery, S. (1998). Learning to Extract Symbolic Knowledge from the World Wide Web, CMU-CS-98-122, Carnegie Mellon University, September 1, 1998
Crestani, F. (1997). "Application of Spreading Activation Techniques in Information Retrieval". Artificial Intelligence Review
Danks, D. (2006). Causal Statistical Inference, Carnegie Mellon Summer School, 2006.
Delgrande, J., & Schaub, T. (2003). A Consistency-Based Approach for Belief Change, Artificial Intelligence Journal 151, 1-2, 2003, pp. 1-41.
Dieterich, H., Malinowski, U., Kühme, T., & Schneider-Hufschmidt M. (1993). State of the art in Adaptive User Interfaces, in Schneider-Hufschmidt M., Kuehme T., Malinowski U. (eds), Adaptive User Interfaces, Elsevier, 1993, pp. 13-48.
Dignum, F., & Kuiper, R. (1997). Combining dynamic deontic logic and temporal logic for the specification of deadlines. Proceedings of the Thirtieth Hawaii International Conference on System Sciences, 1997.
Dimopoulos, Y., Kakas, A. (1996). Abduction and Induction: an AI Perspective. Proceedings of the ECAI ’96 Workshop on Abductive and Inductive Reasoning.
Dinsmore, J. (1991). Partitioned Representations. Kluwer Academic Publishers, 1991.
Emerson, E.A. (1991). Temporal and modal logic, Handbook of theoretical computer science (vol. B): formal models and semantics, MIT Press, 1991.
Englebretsen, G., & Sommers, F.T. (2000). An Invitation to Formal Reasoning - The Logic of Terms. Ashgate Pub Ltd, 2000.
Fagin, R., J. Halpern, Y. Moses, M, Vardi (2003). Reasoning about Knowledge. Cambridge: MIT Press.
Farquhar, A., Dappert, A., Fikes, R., & Pratt, W. (1995). Integrating Information Sources Using Context Logic. In: Proceedings of AAAI Spring Symposium on Information Gathering from Distributed Heterogeneous Environments, 1995.
Fox, S., Karnawat, K., Mydland, M., Dumais S.T., & White T. (2005). Evaluating implicit measures to improve web search. ACM Transaction of Information System, 23(2):147–168, 2005.
Frank, A. (1991). Qualitative spatial reasoning with cardinal directions. In Proceedings of the Seventh Austrian Conference on Artificial Intelligence.
Gallo, A., T. De Bie, N. Cristianini (2007). MINI: Mining Informative Non-redundant Itemsets. Proceedings of the 11th conference on Principles and Practice of Knowledge Discovery in Databases (PKDD07), Warsaw, September 2007.
Galton, A. (1991). Reified temporal theories and how to unrelfy them, In Proc IJCAI-91, pp 1177-1182, (1991)
Ghidini, C., Giunchiglia, F. (2001). Local Models Semantics, or Contextual Reasoning = Locality + Compatibility. Artificial Intelligence 127 (4), 221-259, 2001.
Giunchiglia, F., & Serafini, L. (1994). Multilanguage hierarchical logics. Artificial Intelligence 65, 29-70, 1994.
Goertzel, B. (2008). Achieving Advanced Machine Consciousness through Integrative, Virtually Embodied Artificial General Intelligence. Nokia Workshop on Machine Consciousness 2008.
Goertzel B., & Bugaj, S.V. (2008). Stages of Ethical Development in Artificial General Intelligence Systems. Proceedings of the First Conference on Artificial General Intelligence, 2008.
Goertzel, B. & de Garis, H. (2008). XIA-MAN: An Extensible, Integrative Architecture for Intelligent Humanoid Robotics. Biologically Inspired Cognitive Architectures, AAAI 2008 Fall Symposium Series: November 7-9, Arlington, Virginia, USA.
Goertzel, B., Iklé, M., Goertzel, I.F., Heljakka A. (2008). Probabilistic Logic Networks - A Comprehensive Framework for Uncertain Inference. Springer, 2008.
Goertzel, Ben, Ari Heljakka, CassProbabilistic Logic Based Reinforcement Learning of Simple Embodied Behaviors in a 3D Simulation World, Welter Silva, Cassio Pennachin, Andre Senna, Izabela Goertzel, Teemu Keinonen, Matthew Ikle’, Sanjay Padmane, Proceedings of International Symposium on Intelligence Computation and Applications (ISICA) 2007
Goertzel, Ben (2007). Virtual Easter Egg Hunting: A Thought-Experiment in Embodied Social Learning, Cognitive Process Integration, and the Dynamic Emergence of the Self, in Advances in Artificial General Intelligence, IOS Press.
Goertzel, Ben, Hugo Pinto, Ari Heljakka, Michael Ross, Izabela Goertzel, Cassio Pennachin. Using Dependency Parsing and Probabilistic Inference to Extract Gene/Protein Interactions Implicit in the Combination of Multiple Biomedical Research Abstracts, Proceedings of BioNLP-2006 Workshop at ACL-2006, New York
Goertzel, B., Iklé, M., & Goertzel, I. (2006c). Indefinite Probabilities for General Intelligenge. AGIRI Workshop 2006.
Goertzel B., Looks M., Heljakka A. & Pennachin C. (2006a). Toward a Pragmatic Understanding of the Cognitive Underpinnings of Symbol Grounding. In Gudwin R and Queiroz J (eds), Semiotics and Intelligent Systems Development, Idea Group Publishing.
Goertzel, Ben, Ari Heljakka, Stephan Vladimir Bugaj,
Cassio Pennachin, Moshe Looks, Exploring Android Developmental Psychology in a Simulation World, Symposium “Toward Social Mechanisms of Android Science”, Proceedings of ICCS/CogSci 2006, Vancouver
Goertzel, Ben (2006). Patterns, Hypergraphs and General Intelligence. Proceedings of WCCI 2006, Vancouver
Goertzel, B., Looks, M., & Pennachin, C. (2004). Novamente: An Integrative Architecture for General Intelligence. AAAI Fall Symposium on Achieving Human-Level Intelligence.
Goertzel, B., & Pennachin, C. (2006). Artificial General Intelligence. Springer.
Goertzel, B., Pennachin, C., de Souza Coelho, L., Gurbaxani, B., Maloney, E.M. & Jones, J.F. (2006b). Combinations of single nucleotide polymorphisms in neuroendocrine effector and receptor genes predict chronic fatigue syndrome. Pharmacogenomics. 2006 Apr;7(3):475-83.
Goertzel B. & Wang P (2006). Introduction: Aspects of Artificial General Intelligence. Proceedings of the AGIRI Workshop 2006.
Goethals, B. and Zaki, M. J. (2004). Advances in frequent itemset mining implementations: report on FIMI'03. SIGKDD Explor. Newsl. 6, 1 (Jun. 2004), 109-117.
Gong, Z. (2008). Parametric Potential-Outcome Survival Models for Causal Inference, PhD Thesis, University of Canterbury, 2008.
Gounder R. S., Esterline, A. C. (1998). Fuzzy Versions of Epistemic and Deontic Logic, NASA-URC98, Huntsville, AL.
Greenland, S. & Brumback B. (2002). An overview of relations among causal modelling methods. International Journal of Epidemiology 31:1030–1037, 2002.
Greenland, S. & Pearl, J. & Robins, J.M. (1999). Causal diagrams for epidemiologic research. Epidemiology, 10:37–48, 1999.
Grinberg, D., Lafferty J., & Sleator, D. (1995). A robust parsing algorithm for link grammars. Carnegie Mellon University Computer Science technical report CMU-CS-95-125, and Proceedings of the Fourth International Workshop on Parsing Technologies, Prague, September, 1995.
Grindrod, P. & Kibble, M. (2004). Expert Review of Proteomics, August 2004, Vol. 1, No. 2, Pages 229-238 , DOI 10.1586/14789450.1.2.229 (doi:10.1586/14789450.1.2.229)
Grindrod, P. & Kibble, M. (2004). Review of the uses of network and graph theory concepts within proteomics, Expert Rev Proteomics, 1(2), 2004
Hart, D. and Goertzel, B. (2008). OpenCog: A Software Framework for Integrative Artificial General Intelligence. Proceedings of the First Conference on Artificial General Intelligence, 2008.
Hayes, B. (1997). Computing Science: Can't Get No Satisfaction. American Scientist, Vol. 85, No. 2, March-April 1997, pages 108-112.
Held, A., Buchholz, S., & Schill, A. (2002). Modeling of context information for pervasive computing applications. In Proceedings of SCI 2002/ISAS 2002, 2002.
Hill A. B. (1965). The environment and disease: association or causation? Proc Royal Soc Medicine 1965, 58:295-300.
Hitchcock, C. (2002). Probabilistic Causation, Stanford Encyclopedia of Philosophy, 2002, on-line at: http://plato.stanford.edu/entries/causation-probabilistic/
Hsu, W., Lee, M. L., & Wang, J. (2008). Temporal and Spatio-Temporal Data Mining. Igi Global, 2006.
Huang, B., & Claramunt, C. (2002). STOQL: an ODMG-based spatio-temporal object model and query language. In: D. Richardson, P. Oosterom (Eds.), Advances in Spatial Data Handling, Springer-Verlag, Berlin, 2002.
Ivancsy, R., & Vajk, I. (2005). A Survey of Discovering Frequent Patterns in Graph Data. Proceedings of Databases and Applications (DBA 2005).
Jones, R. Tambe, M., Laird, J., Rosenbloom, P. (1993). Intelligent automated agents for flight training simulators. In Proceedings of the Third Conference on Computer Generated Forces and Behavioral Representation. University of Central Florida. IST-TR-93-07.
Jurafsky, D. and J. Martin (2008). Speech and Language Processing, Prentice Hall
Kelly, D., & Belkin, N.J. (2004). Display time as implicit feedback: Understanding task effects. In Proceedings of SIGIR 2004, 2004.
Kelly, D., & Teevan, J. (2003). Implicit feedback for inferring user preference: A bibliography. SIGIR Forum, 37(2):18–28, 2003.
Kemerling, G. (2006). Philosophy Pages, on-line at http://www.philosophypages.com/, 2006.
Kluve, J. (2001). On the Role of Counterfactuals in Inferring Causal Effects of Treatments. University of Heidelberg and IZA, Bonn, Discussion Paper No. 354, 2001.
Kobsa, A. (1993). User Modeling: Recent Work, Prospects and Hazards, In Adaptive User Interfaces: Principles and Practise. Elsevier, 1993.
Kobsa A. (2007). Generic User Modeling Systems, In: P. Brusilovsky, A. Kobsa, W. Nejdl, eds.: The Adaptive Web: Methods and Strategies of Web Personalization. Berlin, Heidelberg, New York: Springer Verlag, 136-154, 2007.
Kontchakov, R. & Kurucz, A. & Wolter, F. & Zakharyaschev, M. (2007). Spatial logic + temporal logic = ? In M. Aiello, I. Pratt-Hartmann and J. van Benthem, editors, Handbook of Spatial Logics, pp. 497-564. Springer, 2007.
Kontchakov, R. & Pratt-Hartmann, I. & Wolter, F. & Zakharyaschev, M. (2010). Spatial logics with connectedness predicates. Logical Methods in Computer Science 6(3), 2010.
Kontchakov, R. & Pratt-Hartmann, F. & Zakharyaschev, M. (2010). Interpreting Topological Logics over Euclidean Spaces. In F. Lin, U. Sattler and M. Truszczynski, editors, Proceedings of KR (Toronto, Canada, May 9-13), AAAI Press, 2010.
Kozen, D. (1983). Results on the Propositional μ-Calculus. Theoretical ComputerScience, 27 (1983).
Krötzsch, M., Hitzler, P., Vrandecic, D., & Sintek, M. (2006). How to reason with OWL in a logic programming system. In Thomas Eiter, Enrico Franconi, Ralph Hodgson, Susie Stephens, eds.: Proceedings of the Second International Conference on Rules and Rule Markup Languages for the Semantic Web (RuleML-06), pp. 17–26. IEEE Computer Society 2006.
Lamport, L. (1994). The temporal logic of actions. ACM Transactions on Programming Languages and Systems, 16(3):872-923, 1994.
Laschi C., Dario P., Carrozza M.C., Guglielmelli E., Teti G., Taddeucci D., Leoni F., Massa B., Zecca M. & Lazzarini R. (2000). Grasping and Manipulation in Humanoid Robotics. S7: General Control & Simulation, First IEEE-RAS Workshop on Humanoids - Humanoids 2000, Boston, MA, September 7-8, 2000.
Lawlor, D.A. & Smith, G.D. & Ebrahim. S. (2004). The hormone replacement - coronary heart disease conundrum: is this the death of observational epidemiology? International Journal of Epidemiology, 33:464-467, 2004.
Lenat, D., & Sierra, C. (1999). The Dimensions of Context Space. Tech. rep. CYCorp, http://www.cyc.com/context-space.rtf,doc,txt., 1999.
Lewis, D. (2000). Causation as Influence, The Journal of Philosophy 97: 182-197, 2000.
Lightstone, S., T. Teorey and T. Nadeau (2007). Physical Database Design. Morgan Kaufmann
Lynce, I., & Marques-Silva, J. P. (2002). Building State-of-the-Art SAT solvers. In: Proceedings of the European Conference on Artificial Intelligence, July 2002.
Ma, J., & Knight, B. (2001). Reified Temporal Logics: An Overview, Artificial Intelligence Review, Volume 15, Number 3, pp189--217, 2001.
Manning, C. and H. Schuetze (1999). Foundations of Statistical Natural Language Processing, The MIT Press
Martin, C., Moravec, H. P. (1996). Robot Evidence Grids. Technical rept., Robotics Institute, Carnegie-Mellon University.
McCarthy, J., (1958). Programs with common sense, Symposium on Mechanization of Thought Processes. National Physical Laboratory, Teddington, England, 1958.
McCarthy, J., & Hayes, P.J. (1969). Some philosophical problems from the standpoint of artificial intelligence. Machine Intelligence, 4:463-502.
McCarthy, J. (1987). Generality in Artificial Intelligence. Communications of ACM, 30(12):1030-1035, 1987.
McNamara, P. & Prakken, H. (1999). Norms, Logics and Information Systems: New Studies in Deontic Logic and Computer Science. Amsterdam: IOS Press.
Mendelson, E. (1997). Introduction to Mathematical Logic. Fourth Edition. International Thomson Publishing, 1997, 440 pp.
Meulen, A. (2001), Logic and Natural Language, in Goble, Lou, ed., The Blackwell Guide to Philosophical Logic. Blackwell.
Mitra, D. (2004). Modeling and Reasoning with Star Calculus. Eighth International Symposium on Artificial Intelligence and Mathematics (AMAI 2004).
Moszkowski, B. (1994). Some very compositional temporal properties. In E.-R. Olderog, editor, Programming Concepts, Methods and Calculi, volume A-56 of IFIP Transactions, pages 307–326. IFIP, Elsevier Science B.V. (North–Holland), 1994.
Muller, M. (2001). Inducing Content Based User Models with Inductive Logic Programming Techniques, UM2001 Workshop on Machine Learning for User Modeling, 2001.
Myaeng, S. H., & Korfhage, R. R. (1986). Towards an intelligent and personalized retrieval system. In Proceedings of the ACM SIGART international symposium on Methodologies for intelligent systems, pages 121–129, Knoxville, Tennessee, United States. ACM Press, 1986.
Mylopoulos, J., & Motschnig-Pitrik, R. (1995). Partitioning Information Bases with Contexts. In: Third International Conference on Cooperative Information Systems. Vienna, 1995.
Nilsson, N. (1998). "Artificial Intelligence: A New Synthesis". Morgan Kaufmann Publishers, Inc., San Francisco, California, 1998, pages 121-122
Nurnberger, A., & Detyniecki, M. (2005). Adaptive Multimedia Retrieval: From Data to User Interaction, In Do Smart Adaptive Systems Exist? Studies in Fuzziness and Soft Computing, Volume 173, pp341-370, Springer, 2005.
Palermiti R., & Polity, Y. (1995). Desperately seeking user models in information retrieval systems : benefits and limits of cognitivist and marketing approaches, The new review of information and library research, vol 1, 1995, p.57-65.
Pani A. K., & Bhattacharjee, G. P. (2001). Temporal representation and reasoning in artificial intelligence : A review, Mathematical and computer modelling, 2001, vol. 34, number 1-2, pp. 55-80.
Pearl, J. & Russel, S. (1998). "Bayesian Networks", 1998.
Pearl, J. (2000). Causality: Models, Reasoning and Inference. Cambridge University Press, 2000.
Pnueli, A. (1977). The temporal logic of programs. Proceedings of the 18th IEEE Symposium on Foundations of Computer Science, pages 46-67, 1977.
Pohl, W., & Hohle, J. (1997). Mechanisms for Flexible Representation and Use of Knowledge in User Modeling Shell Systems. In Anthony Jameson, Cécile Paris, and Carlo Tasso (Eds.), User Modeling: Proceedings of the Sixth International Conference, UM97, CISM, 1997.
Pohl, P. (1999). Logic-Based Representation and Reasoning for User Modeling Shell Systems, User Modeling and User-Adapted Interaction 9: 217–282, 1999.
Przulj, N. (2005). Graph Theory Analysis of Protein-Protein Interactions, a chapter in "Knowledge Discovery in Proteomics", edited by Igor Jurisica and Dennis Wigle, CRC Press, 2005.
Radlinski, F., & Joachims, T. (2005). Query chains: Learning to rank from implicit feedback. In Proceedings of SIGKDD 2005.
Reiter R. (1980). A logic for default reasoning. Artificial Intelligence, 13:81-132.
Rescher, Nicolas (2005). Epistemic Logic: A Survey Of the Logic Of Knowledge. University of Pittsburgh Press.
Rich, E. (1979). Building and Exploiting User Models. PhD. Thesis, Department of Computer Science. Pittsburgh, PA: Carnegie-Mellon University, 1979.
Richardson, M. & Domingos, P. (2006). Markov logic networks. Machine Learning, Volume 62, Numbers 1-2, February 2006 , pp. 107-136(30)
Robins, J.M. (2001). Data, design, and background knowledge in etiologic inference. Epidemiology 12:313–20, 2001.
Rothman, K.J. (1976). Causes. Am J Epidemiol. 104:587–92, 1976.
Rothman, K.J. & S. Greenland & C. Poole, T. & Lash (2008). Causation and Causal Inference. In: Modern Epidemiology (3rd edition). Wolter Kluwer, 2008.
Samulowitz, M., Michahelles, F., Linnhoff-Popien, C. (2001). CAPEUS: An Architecture for Context-Aware Selection and Execution of Services. Conference on New Developments in Distributed Applications and Interoperable Systems, 2001.
Schockaert Steven, De Cock Martine, Cornelis Chris and Kerre Etienne E. (2008). Fuzzy region connection calculus: Representing vague topological information. Approx. Reasoning, Volume 48, Number 1, 2008, pages 314-331, Elsevier Science Inc.
Scott, D., & De Bakker, J. (1969). A theory of programs. Unpublished manuscript, IBM, Vienna, 1969.
Shanahan, M.P., & Baars, B.J. (2005). Applying Global Workspace Theory to the Frame Problem, Cognition, vol. 98, no. 2 (2005), pages 157-176
Sheng L., Ozsoyoglu, Z.M. & Ozsoyoglu, G. (1999). A graph query language and its query processing. Proceedings of the 15th International Conference on Data Engineering, 1999.
Shoham, Y. (1987). Temporal logics in AI: Semantical and ontological considerations. Artificial Intelligence, 33(1):89-104, 1987.
Smigrodzki R., Goertzel B., Pennachin C., Coelho L., Prosdocimi F. & Parker W.D. Jr. (2005). Genetic algorithm for analysis of mutations in Parkinson's disease. Artif Intell Med. 2005 Nov;35(3):227-41. Epub 2005 Oct 3.
Stewart, T. C. & West, R. L. (2006). Deconstructing ACT-R. In Proceedings of the Seventh International Conference on Cognitive Modeling (pp. 298-303). Trieste, Italy.
Strang T., & Linnhoff-Popien, C. (2004). A Context Modeling Survey. Workshop on Advanced Context Modelling, Reasoning and Management, 2004.
Sugiyama K., Hatano K., & Yoshikawa, M. (2004). Adaptive web search based on user profile constructed without any effort from users. In Proceedings of WWW 2004, pages 675–684, 2004.
Sutcliffe G., Gao Y. & Colton S. A Grand Challenge of Theorem Discovery. In Proceedings of the CADE-19 workshop on Grand Challenges and Novel Applications of Automated Reasoning, 2003.
Tarski, A. (1994). Introduction to Logic and to the Methodology of the Deductive Sciences. (New edition of book originally published in Polish in 1936), New York, Oxford University Press.
Thiele, H. (1993). On the Definition of Modal Operators in Fuzzy Logic. In Proceedings of the 23rd IEEE International Symposium on Multiple-Valued Logic, 1993: 62-67.
Lewis, D. (2000). Causation as Influence, The Journal of Philosophy 97: 182-197, 2000.
Van Gils, B., & Schabell, E. D. (2003). User-profiles for Information Retrieval, In Proceedings of the 15th Belgian-Dutch Conference on Artificial Intelligence.
Vilain, M., Kautz, H., & van Beek, P. (1989). Constraint propagation algorithms for temporal reasoning: a revised report. Readings in Qualitative Reasoning about Physical Systems.
Walley, P. (1991). Statistical reasoning with imprecise probabilities. London ; New York : Chapman and Hall, 1991. Monographs on statistics and applied probability : 42.
Wang, P. (1996). Heuristics and normative models of judgment under uncertainty. International Journal of Approximate Reasoning, Vol.14, No.4, Pages 221-235, 1996.
Wang, P. (2006a). From NARS to a Thinking Machine. Proceedings of the Artificial General Intelligence Research Institute Workshop, Washington DC, May 2006.
Wang, P. (2006b). Rigid Flexibility - The Logic of Intelligence. Springer, 2006.
Weischedel R.M. (2006). Challenges in Natural Language Processing. Cambridge University Press, 2006.
Westervelt E.R., Grizzle J.W., Chevallereau C., Choi J.H. & Morris B. (2007). Feedback control of dynamic bipedal robot locomotion (Control & automation, Vol. 1). Taylor & Francis/CRC, 2007.
Weyhrauch, R., 1980. Prolegomena to a Theory of Mechanized Formal Reasoning. Artificial Intelligence 13 (1), 133-176, 1980.
Wheeler R. & Aitken S. (2000). Multiple algorithms for fraud detection. Knowledge-Based Systems, Volume 13, Issues 2-3, April 2000, Pages 93-99.
Wieringa, R. J. & Meyer, J-J. Ch. (1993). Applications of deontic logic in computer science: A concise overview. in Deontic Logic in Computer Science: Normative System Specification (1994).
Wood, W.G. (1989). Temporal Logic Case Study, Technical Report CMU/SEI-89-TR-024, Carnegie Mellon University, 1989.
Wray, R.E., & Jones, R.M (2005). An introduction to Soar as an agent architecture. In, R. Sun (ed), Cognition and Multi-agent Interaction: From Cognitive Modeling to Social Simulation, Cambridge University Press, pp 53-78, 2005.
Yeung, Albert K.W. & Hall, G. Brent (2007). Spatial Database Systems: Design, Implementation and Project Management, edited by Albert K.W. Yeung, Springer, 2007.
Ying, M.S. (1988). On a standard models of fuzzy modal logics. Fuzzy Sets and Systems.
Zadeh, L. A. (1965). Fuzzy sets. Inf. Control 8, 338-353, 1965.
Zadeh, L. A. (1978). Fuzzy sets as a basis for a theory of possibility, Fuzzy Sets and Systems 1, 3-28, 1978.
Zanibbi, R. & Blostein, D. & Cordy, J.R. (2003). A Survey of Table Recognition: Models, Observations, Transformations, and Inferences International Journal of Document Analysis and Recognition, 2003.
Zanibbi, R. & Blostein, D. & Cordy, J.R. (2004). A survey of table recognition Models, observations, transformations, and inferences, International Journal on Document Analysis and Recognition, Volume 7, Number 1, 2004.
Zanibbi, R. & Blostein, D. & Cordy, J.R. (2006). Decision-Based Specification and Comparison of Table Recognition Algorithms, Western New York Image Processing Workshop, 2006.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 150, "math_score": 0.7243205308914185, "perplexity": 1621.7070582767606}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-51/segments/1544376829542.89/warc/CC-MAIN-20181218164121-20181218190121-00350.warc.gz"}
|
https://hsm.stackexchange.com/questions/2580/english-translation-of-omar-khayy%C3%A1ms-mathematical-work
|
# English translation of Omar Khayyám's mathematical work
Is there any current English translation of the mathematical works of Omar Khayyám?
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9558864235877991, "perplexity": 6323.055174282717}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178389472.95/warc/CC-MAIN-20210309061538-20210309091538-00109.warc.gz"}
|
https://kimberlywynne.com/calculus/ode/
|
This is the 1st HTML5 slant canvas
# k i w y
### Need a light for your site?
This is the 2nd HTML5 slant canvas
## 1st order ODE
\frac{dy}{dt} + p(t) y = q(t)
The general solution is
y = \frac{\int \mu(t) q(t) dt + C}{\mu(t)}
where the integration factor μ is
\mu(t) = e^{\int p(t) dt}
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8789364099502563, "perplexity": 15626.114892741072}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057202.68/warc/CC-MAIN-20210921101319-20210921131319-00462.warc.gz"}
|
https://bitbucket.org/evzijst/django-recaptcha
|
# Overview
Django reCAPTCHA form field/widget integration app.
Contents
django-recaptcha uses a modified version of the Python reCAPTCHA client which is included in the package as client.py.
## Installation
RECAPTCHA_PUBLIC_KEY = '76wtgdfsjhsydt7r5FFGFhgsdfytd656sad75fgh'
This can be seperately specified at runtime by passing a public_key parameter when constructing the ReCaptchaField, see field usage below.
RECAPTCHA_PRIVATE_KEY = '98dfg6df7g56df6gdfgdfg65JHJH656565GFGFGs'
This can be seperately specified at runtime by passing a private_key parameter when constructing the ReCaptchaField, see field usage below.
RECAPTCHA_USE_SSL = True
If you don't add this setting the default behaviour is to NOT use SSL. This can be seperately specified at runtime by passing a use_ssl parameter when constructing the ReCaptchaField, see field usage below.
## Usage
### Field
The quickest way to add reCAPTHCA to a form is to use the included ReCaptchaField field type. A ReCaptcha widget will be rendered with the field validating itself without any further action required from you. For example:
from django import forms
To allow for runtime specification of keys and SSL usage you can optionally pass private_key, public_key or use_ssl parameters to the constructor, i.e.:
captcha = ReCaptchaField(
private_key='98dfg6df7g56df6gdfgdfg65JHJH656565GFGFGs',
use_ssl=True
)
If specified these parameters will be used instead of your reCAPCTHA project settings.
The reCAPTCHA widget supports several Javascript options variables customizing the behaviour of the widget, such as theme and lang. You can forward these options to the widget by passing an attr parameter containing a dictionary of options to ReCaptchaField, i.e.:
captcha = ReCaptchaField(attrs={'theme' : 'clean'})
The captcha client takes the key/value pairs and writes out the RecaptchaOptions value in JavaScript.
### Unit Testing
django-recaptcha detects for DEBUG = True in the settings.py to facilitate unit tests. When DEBUG = TRUE, using PASSED as the recaptcha_response_field value.
Example:
form_params = {'recaptcha_response_field': 'PASSED'} form = RegistrationForm(form_params) # assuming only one ReCaptchaField form.is_valid() # True
Passing any other values will cause django-recaptcha to continue normal processing and return a form error.
### django-registration
django-recaptcha ships with a django-registration backend extending the default backend to include a reCAPTCHA field. This is included mostly as an example of how you could intergrate a reCAPTCHA field with django-registration. I suggest you familiarize yourself with the django-registration docs for more comprehensive documentation.
Note
This backend will only work with django-registration versions 0.8-alpha-1 and up.
To use the reCAPTHCA backend complete these steps:
(r'^accounts/', include('captcha.backends.default.urls')),
3. Add an ACCOUNT_ACTIVATION_DAYS setting to the project's settings.py file. This is the number of days users will have to activate their accounts after registering, as required by django-registration, i.e.:
ACCOUNT_ACTIVATION_DAYS = 7
4. Implement the various templates as required by django-registration.
Once done you should be able to access /accounts/register/ and see the reCAPTCHA field in action as part of the registration process.
## Credits
Inspired Marco Fucci's blogpost titled Integrating reCAPTCHA with Django
client.py taken from recaptcha-client licenced MIT/X11 by Mike Crawford.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.15852439403533936, "perplexity": 9962.769409217271}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187825227.80/warc/CC-MAIN-20171022113105-20171022133105-00519.warc.gz"}
|
https://docs.asciidoctor.org/asciidoctor/latest/convert/contexts-ref/
|
# Convertible Contexts
When Asciidoctor converters a document, it delegates to the converter to convert each node (block or inline element) in the document as it visits each node in document order (though it’s the #content method on a block that continues the traversal of its child nodes). Asciidoctor then combines the result of all those calls to produce the output document.
To convert a node, Asciidoctor calls the convert method on the converter and passes in the node. In some cases, it also passes in an extra transform value (e.g., embedded) to differentiate between different uses cases for a node of the same context. When Asciidoctor is using a converter that extends Asciidoctor::Converter::Base, the converter will delegate the call to a method whose name starts with convert_ and is followed by the context of the node (e.g., convert_paragraph).
The following table lists the contexts for all the built-in nodes that Asciidoctor converts, along when the name of the method the base converter will look for. Extensions can introduce additional contexts.
The built-in contexts of nodes that are converted by the converter
Context Convert method on base converter
:audio
convert_audio
:colist
convert_colist
:dlist
convert_dlist
:document
convert_document
:embedded
convert_embedded
:example
convert_example
:floating_title
convert_floating_title
:image
convert_image
:inline_anchor
convert_inline_anchor
:inline_break
convert_inline_break
:inline_button
convert_inline_button
:inline_callout
convert_inline_callout
:inline_footnote
convert_inline_footnote
:inline_image
convert_inline_image
:inline_indexterm
convert_inline_indexterm
:inline_kbd
convert_inline_kbd
:inline_quoted
convert_inline_quoted
:listing
convert_listing
:literal
convert_literal
:olist
convert_olist
:open
convert_open
:outline
convert_outline
:page_break
convert_page_break
:paragraph
convert_paragraph
:preamble
convert_preamble
:quote
convert_quote
:section
convert_section
:sidebar
convert_sidebar
:stem
convert_stem
:table
convert_table
:thematic_break
convert_thematic_break
:toc
convert_toc
:ulist
convert_ulist
:verse
convert_verse
:video
convert_video
The converter is not called to handle nodes with the :list_item or :table_cell contexts. Instead, it’s up to the converter to access these nodes from the parent node and convert them directly.
When a method to convert a block is called, the inline markup has not yet been parsed. That parsing and subsequent conversion happens when the #content method is called on the node. This is also what triggers the processor to visit the child nodes of that block.
Some methods for converting blocks have to handle the specialized behavior as indicated by the style. For example, the convert_listing method also handles source blocks (listing blocks with the source style). And convert_dlist handles qanda lists (dlist blocks with the qanda style).
:embedded is not a true context, but rather a transform of the :document context. It’s convert method is called instead of the one for :document when the document is loaded in embedded mode (i.e., the :standalone option on the processor is false).
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4366081953048706, "perplexity": 2081.3943818525686}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335276.85/warc/CC-MAIN-20220928180732-20220928210732-00076.warc.gz"}
|
https://math.meta.stackexchange.com/questions/12582/when-do-we-include-axiom-of-choice-as-a-tag
|
# When do we include "Axiom of Choice" as a tag?
My question is the following:
Suppose that the question is about a functional equation, where assuming the function we seek is continuous, we can show the existence of one particular type solution, whereas if continuity is not assumed, then the fact that $\mathbb R$ possesses a Hamel basis (as a linear space over $\mathbb Q$), guarantees the existence of discontinuous solutions of the functional equation.
And my question is the following: It is non-appropriate to include the "Axiom of Choice" among the tags?
It is important to say that many, if not most, of the subscribers who try to solve such functional equations are not familiar the Axiom of Choice/Zorn's Lemma, as such problems are popular in the IMOs and similar competitions.
• – Asaf Karagila Mod
Jan 25, 2014 at 20:30
I think the prevailing view is that since most mathematicians consider the Axiom of Choice a "fundamental truth" (or at least use it without much consideration), the tag should only in included when Choice is somehow of more central interest to the question itself.
Examples of such question templates would be:
• How do you prove that $\Phi$ is equivalent to the Axiom of Choice?
• Can $\Psi$ be proven without the Axiom of Choice?
• How much Choice is needed to prove $\Xi$?
• Does $\Theta$ imply Choice?
Just as I would not include the tag for most questions dealing with cardinal arithmetic, even in cases where $\mathsf{AC}$ and $\neg \mathsf{AC}$ may yield different answers (such as this recent example), I would not include it in questions from other mathematical areas where whether or not Choice holds could possibly change the answer (and the OP has not indicated interest in the connection between Choice and the particular problem).
Of course, this doesn't preclude users from submitting answers which point out the connection between Choice and the given question. Such answers may be quite enlightening (though admittedly possibly not to the OP).
• How about: "This functional equation has a discontinuous solution, but this requires Zorn's Lemma" - What happens then? Jan 25, 2014 at 21:03
• @Yiorgos: How about you just post the link to the question?
– Asaf Karagila Mod
Jan 25, 2014 at 21:10
• See for example: math.stackexchange.com/questions/641416/… Jan 25, 2014 at 21:13
• @Yiorgos: If your question was "Do discontinuous such functionals exist without assuming the Axiom of Choice," then, yes, I would include the axiom-of-choice tag. But I would rephrase your question to explicitly state that this is your central interest. As it stands, your Update seems to say that Choice yields a solution to your question, but this does not make your question about Choice. Jan 25, 2014 at 21:23
• Remark: much of the advice of this answer is independent of the axiom-of-choice tag. The phrase "when [blank] is somehow of more central interest to the question itself" is the general guideline for tagging. Another example: don't tag a linear PDE question with semigroups just because solutions can be expressed in terms of a continuous semigroup of mappings; use it when somehow the structure of a semigroup is critical to the question. Jan 27, 2014 at 8:35
I think an "axiom of choice equivalents" might be a reasonable tag. This includes gobs of important stuff.
• I'm against this. I wrote on this in the thread linked on the comments of the question.
– Asaf Karagila Mod
Jan 27, 2014 at 1:16
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8542274832725525, "perplexity": 585.8116999429294}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030336674.94/warc/CC-MAIN-20221001132802-20221001162802-00564.warc.gz"}
|
http://mathhelpforum.com/advanced-algebra/144237-integral-domain.html
|
Math Help - Integral Domain
1. Integral Domain
Hello I am just trying to work out some problems from herstein. Can you please tell me if this one is right.
If D is an integral domian and if na=0 for some a not 0 in D and for some integer n not 0 , prove that D is of finite characterstic.
so we have for some a not 0, na = 0
if b not 0 is another element in D then nab = 0b (ie) (na)b = 0
na is an element of D, an integral domain which has zero divisors, so(na)b=0 where na = 0 is justified.
However nab=0 => (nb)a=0
Here nb is an element of D and a is nonzero,
So nb = 0.
Hence the proof.
Am i right?
2. Originally Posted by poorna
Hello I am just trying to work out some problems from herstein. Can you please tell me if this one is right.
If D is an integral domian and if na=0 for some a not 0 in D and for some integer n not 0 , prove that D is of finite characterstic.
so we have for some a not 0, na = 0
if b not 0 is another element in D then nab = 0b (ie) (na)b = 0
na is an element of D, an integral domain which has zero divisors, so(na)b=0 where na = 0 is justified.
However nab=0 => (nb)a=0
Here nb is an element of D and a is nonzero,
So nb = 0.
Hence the proof.
Am i right?
That looks good to me.
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.933155357837677, "perplexity": 1232.2448355359695}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1412037663754.5/warc/CC-MAIN-20140930004103-00382-ip-10-234-18-248.ec2.internal.warc.gz"}
|
https://isr-publications.com/jmcs/articles-9689-on-unified-gould-hopper-based-apostol-type-polynomials
|
# On unified Gould-Hopper based Apostol-type polynomials
Volume 24, Issue 4, pp 287--298
Publication Date: March 25, 2021 Submission Date: January 08, 2021 Revision Date: January 28, 2021 Accteptance Date: February 27, 2021
• 205 Views
### Authors
Waseem Ahmad Khan - Department of Mathematics and Natural Sciences, Prince Mohammad Bin Fahd University, P. O. Box 1664, Al Khobar 31952, Saudi Arabia. Kottakkaran Sooppy Nisar - Department of Mathematics, College of Arts and Sciences, Wadi Aldawaser, Prince Sattam bin Abdulaziz University, Saudi Arabia. Mehmet Acikgoz - Department of Mathematics, Faculty of Arts and Sciences, University of Gaziantep, TR-27310 Gaziantep, Turkey. Ugur Duran - Department of the Basic Concepts of Engineering, Faculty of Engineering and Natural Sciences, Iskenderun Technical University, TR-31200 Hatay, Turkey. Abdallah Hassan Abusufian - Department of Mathematics, College of Arts and Sciences, Wadi Aldawaser, Prince Sattam bin Abdulaziz University, Saudi Arabia.
### Abstract
In this paper, we consider unified Gould-Hopper based Apostol-type polynomials and investigate some of their formulas including several implicit summation formulae and some symmetric identities by the series manipulation method. Moreover, we acquire several new results for unified Gould-Hopper based Apostol-type polynomials using appropriate operational rules.
### Share and Cite
##### ISRP Style
Waseem Ahmad Khan, Kottakkaran Sooppy Nisar, Mehmet Acikgoz, Ugur Duran, Abdallah Hassan Abusufian, On unified Gould-Hopper based Apostol-type polynomials, Journal of Mathematics and Computer Science, 24 (2022), no. 4, 287--298
##### AMA Style
Khan Waseem Ahmad, Nisar Kottakkaran Sooppy, Acikgoz Mehmet, Duran Ugur, Abusufian Abdallah Hassan, On unified Gould-Hopper based Apostol-type polynomials. J Math Comput SCI-JM. (2022); 24(4):287--298
##### Chicago/Turabian Style
Khan, Waseem Ahmad, Nisar, Kottakkaran Sooppy, Acikgoz, Mehmet, Duran, Ugur, Abusufian, Abdallah Hassan. "On unified Gould-Hopper based Apostol-type polynomials." Journal of Mathematics and Computer Science, 24, no. 4 (2022): 287--298
### Keywords
• Gould-Hopper polynomials
• monomiality principle
• unified Apostol-type polynomials
• summation formula
• symmetric identity
• 11B68
• 33C45
• 05A10
### References
• [1] L. C. Andrews, Special functions for engineers and mathematicians, Macmillan Co., New York (1985)
• [2] T. M. Apostol, On the Lerch zeta function, Pacific J. Math., 1 (1951), 161--167
• [3] P. Appell, Sur une classe de polynomes, Ann. Sci. Ecole Norm. Sup., 9 (1880), 119--144
• [4] G. Bretti, C. Cesarno, P. E. Ricci, Laguerre-type exponentials and generalized Appell polynomials, Comput. Math. Appl., 48 (2004), 833--839
• [5] G. Dattoli, Hermite-Bessel and Laguerre-Bessel functions: a by-product of the monomiality principle, Proc. Melfi Sch. Adv. Top. Math. Phys., 1 (2000), 147--164
• [6] H. W. Gould, A. T. Hopper, Operational formulas connected with two generalization of Hermite polynomials, Duke Math. J., 29 (1952), 51--63
• [7] W. A. Khan, Some properties of the generalized Apostol type Hermite-based polynomials, Kyungpook Math. J., 55 (2015), 597--614
• [8] S. Khan, M. W. Al-Saad, R. Khan, Laguerre-based Appell polynomials: properties and applications, Math. Comput. Modelling, 52 (2010), 247--259
• [9] S. Khan, N. Raza, General-Appell polynomials within the context of monomiality principle, Int. J. Anal., 2013 (2013), 11 pages
• [10] S. Khan, G. Yasmin, R. Khan, N. A. M. Hassan, Hermite-based Appell polynomials, properties and applications, J. Math. Anal. Appl, 351 (2009), 756--764
• [11] V. Kurt, Some symmetry identities for the unified Apostol-type polynomials and multiple power sums, Filomat, 30 (2016), 583--592
• [12] V. Kurt, On the unified family of generalized Apostol-type polinomials of higher order and multiple power sums, Filomat, 30 (2016), 929--935
• [13] B. Kurt, Notes on unified q-Apostol-type polynomials, Filomat, 30 (2016), 921--927
• [14] B. Kurt, S. Bilgic, Symmetry identities for the 2-variable unified Apostol-type polynomials, J. Inequal. Spec. Funct., 8 (2017), 96--103
• [15] Q.-M. Luo, Fourier expansions and integral representations for the Apostol-Bernoulli and Apostol-Euler polynomials, Math. Comp., 78 (2009), 2193--2208
• [16] Q.-M. Luo, The multiplication formulas for the Apostol-Bernoulli and Apostol-Euler polynomials of higher order, Integral Transforms Spec. Funct., 20 (2009), 377--391
• [17] Q.-M. Luo, Some formulas for Apostol-Euler polynomials associated with Hurwitz zeta function at rational arguments, Appl. Anal. Discrete Math., 3 (2009), 336--346
• [18] Q.-M. Luo, Fourier expansions and integral representations for the Genocchi polynomials, J. Integer Seq., 12 (2009), 1--9
• [19] Q.-M. Luo, q-extensions for the Apostol-Genocchi polynomials, Gen. Math., 17 (2009), 113--125
• [20] Q.-M. Luo, Extensions of the Genocchi polynomials and their Fourier expansions and integral representations, Osaka J. Math., 48 (2011), 291--309
• [21] Q.-M. Luo, H. M. Srivastava, Some generalizations of the Apostol-Bernoulli and Apostol-Euler polynomials, J. Math. Anal. Appl., 308 (2005), 290--302
• [22] Q.-M. Luo, H. M. Srivastava, Some relationships between the Apostol-Bernoulli and Apostol-Euler polynomials, Comput. Math. Appl., 51 (2006), 631--642
• [23] Q.-M. Luo, H. M. Srivastava, Some generalizations of the Apostol-Genocchi polynomials and the Stirling numbers of the second kind, Appl. Math. Comput., 217 (2011), 5702--5728
• [24] M. A. Ozarslan, Hermite-based unified Apostol-Bernoulli, Euler and Genocchi polynomials, Adv. Difference Equ., 2013 (2013), 13 pages
• [25] M. A. Ozarslan, Unified Apostol-Bernoulli, Euler and Genocchi polynomials, Comput. Math. Appl., 62 (2011), 2452--2462
• [26] H. Ozden, Generating function of the unified representation of the Bernoulli, Euler and Genocchi polynomials of higher order, AIP Conf. Proc., 1389 (2011), 349--352
• [27] H. Ozden, Y. Simsek, H. M. Srivastava, A unified presentation of the generating functions of the generalized Bernoulli, Euler and Genocchi polynomials, Comput. Math. Appl., 60 (2010), 2779--2287
• [28] M. A. Pathan, W. A. Khan, Some implicit summation formulas and symmetric identities for the generalized Hermite-based polynomials, Acta Univ. Apulensis Math. Inform., 39 (2014), 113--136
• [29] M. A. Pathan, W. A. Khan, Some implicit summation formulas and symmetric identities for the generalized Hermite-Bernoulli polynomials, Mediterr. J. Math., 12 (2015), 679--695
• [30] M. A. Pathan, W. A. Khan, A new class of generalized polynomials associated with Hermite and Euler polynomials, Mediterr. J. Math., 13 (2016), 913--928
• [31] J. Sandor, B. Crstici, Handbook of number theory. II, Kluwer Academic Publishers, Dordrecht (2004)
• [32] Y. Smirnov, A. Turbiner, , Errata: "Lie algebraic discretization of differential equations'', Modern Phys. Lett. A, 10 (1995), 1795--1802
• [33] H. M. Srivastava, S. Araci, W. A. Khan, M. Acikgoz, A Note on the Truncated-Exponential Based Apostol-Type Polynomials, Symmetry, 11 (2019), 1--20
• [34] H. M. Srivastava, H. L. Manocha, A treatise on generating functions, Ellis Horwood Limited. Co., New York (1984)
• [35] J. F. Steffensen, The poweroid, an extension of the mathematical notion of power, Acta Math., 73 (1941), 333--366
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8419355750083923, "perplexity": 12641.50683224033}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-17/segments/1618039554437.90/warc/CC-MAIN-20210421222632-20210422012632-00344.warc.gz"}
|
https://aimsciences.org/article/doi/10.3934/dcds.2005.12.283
|
# American Institute of Mathematical Sciences
February 2005, 12(2): 283-302. doi: 10.3934/dcds.2005.12.283
## Equivariant versal unfoldings for linear retarded functional differential equations
1 Faculty of Science, University of Ontario Institute of Technology, 2000 Simcoe St. North, Oshawa, ON L1H 7K4, Canada 2 Department of Mathematics and Statistics, University of Ottawa, Ottawa, ON K1N 6N5, Canada
Received August 2003 Revised August 2004 Published December 2004
We continue our investigation of versality for parametrized families of linear retarded functional differential equations (RFDEs) projected onto finite-dimensional invariant manifolds. In this paper, we consider RFDEs equivariant with respect to the action of a compact Lie group. In a previous paper (Buono and LeBlanc, J. Diff. Eqs., 193 , 307-342 (2003)), we have studied this question in the general case (i.e. no a priori restrictions on the RFDE). When studying the question of versality in the equivariant context, it is natural to want to restrict the range of possible unfoldings to include only those which share the same symmetries as the original RFDE, and so our previous results do not immediately apply. In this paper, we show that with appropriate projections, our previous results on versal unfoldings of linear RFDEs can be adapted to the case of linear equivariant RFDEs. We illustrate our theory by studying the linear equivariant unfoldings at double Hopf bifurcation points in a $\mathbb D_3$-equivariant network of coupled identical neurons modeled by delay-differential equations due to delays in the internal dynamics and coupling.
Citation: Pietro-Luciano Buono, V.G. LeBlanc. Equivariant versal unfoldings for linear retarded functional differential equations. Discrete & Continuous Dynamical Systems - A, 2005, 12 (2) : 283-302. doi: 10.3934/dcds.2005.12.283
[1] Ovide Arino, Eva Sánchez. A saddle point theorem for functional state-dependent delay differential equations. Discrete & Continuous Dynamical Systems - A, 2005, 12 (4) : 687-722. doi: 10.3934/dcds.2005.12.687 [2] Ismael Maroto, Carmen NÚÑez, Rafael Obaya. Dynamical properties of nonautonomous functional differential equations with state-dependent delay. Discrete & Continuous Dynamical Systems - A, 2017, 37 (7) : 3939-3961. doi: 10.3934/dcds.2017167 [3] Ismael Maroto, Carmen Núñez, Rafael Obaya. Exponential stability for nonautonomous functional differential equations with state-dependent delay. Discrete & Continuous Dynamical Systems - B, 2017, 22 (8) : 3167-3197. doi: 10.3934/dcdsb.2017169 [4] Víctor Guíñez, Eduardo Sáez. Versal Unfoldings for rank--2 singularities of positive quadratic differential forms: The remaining case. Discrete & Continuous Dynamical Systems - A, 2005, 12 (5) : 887-904. doi: 10.3934/dcds.2005.12.887 [5] Tomás Caraballo, Gábor Kiss. Attractivity for neutral functional differential equations. Discrete & Continuous Dynamical Systems - B, 2013, 18 (7) : 1793-1804. doi: 10.3934/dcdsb.2013.18.1793 [6] Michael Dellnitz, Mirko Hessel-Von Molo, Adrian Ziessler. On the computation of attractors for delay differential equations. Journal of Computational Dynamics, 2016, 3 (1) : 93-112. doi: 10.3934/jcd.2016005 [7] Hermann Brunner, Stefano Maset. Time transformations for delay differential equations. Discrete & Continuous Dynamical Systems - A, 2009, 25 (3) : 751-775. doi: 10.3934/dcds.2009.25.751 [8] Klaudiusz Wójcik, Piotr Zgliczyński. Topological horseshoes and delay differential equations. Discrete & Continuous Dynamical Systems - A, 2005, 12 (5) : 827-852. doi: 10.3934/dcds.2005.12.827 [9] Xiuli Sun, Rong Yuan, Yunfei Lv. Global Hopf bifurcations of neutral functional differential equations with state-dependent delay. Discrete & Continuous Dynamical Systems - B, 2018, 23 (2) : 667-700. doi: 10.3934/dcdsb.2018038 [10] Sylvia Novo, Carmen Núñez, Rafael Obaya, Ana M. Sanz. Skew-product semiflows for non-autonomous partial functional differential equations with delay. Discrete & Continuous Dynamical Systems - A, 2014, 34 (10) : 4291-4321. doi: 10.3934/dcds.2014.34.4291 [11] Abdelhai Elazzouzi, Aziz Ouhinou. Optimal regularity and stability analysis in the $\alpha-$Norm for a class of partial functional differential equations with infinite delay. Discrete & Continuous Dynamical Systems - A, 2011, 30 (1) : 115-135. doi: 10.3934/dcds.2011.30.115 [12] Fuke Wu, Shigeng Hu. The LaSalle-type theorem for neutral stochastic functional differential equations with infinite delay. Discrete & Continuous Dynamical Systems - A, 2012, 32 (3) : 1065-1094. doi: 10.3934/dcds.2012.32.1065 [13] Rafael Obaya, Ana M. Sanz. Persistence in non-autonomous quasimonotone parabolic partial functional differential equations with delay. Discrete & Continuous Dynamical Systems - B, 2019, 24 (8) : 3947-3970. doi: 10.3934/dcdsb.2018338 [14] Ya Wang, Fuke Wu, Xuerong Mao, Enwen Zhu. Advances in the LaSalle-type theorems for stochastic functional differential equations with infinite delay. Discrete & Continuous Dynamical Systems - B, 2017, 22 (11) : 1-14. doi: 10.3934/dcdsb.2019182 [15] Vitalii G. Kurbatov, Valentina I. Kuznetsova. On stability of functional differential equations with rapidly oscillating coefficients. Communications on Pure & Applied Analysis, 2018, 17 (1) : 267-283. doi: 10.3934/cpaa.2018016 [16] Olesya V. Solonukha. On nonlinear and quasiliniear elliptic functional differential equations. Discrete & Continuous Dynamical Systems - S, 2016, 9 (3) : 869-893. doi: 10.3934/dcdss.2016033 [17] Pierluigi Benevieri, Alessandro Calamai, Massimo Furi, Maria Patrizia Pera. On general properties of retarded functional differential equations on manifolds. Discrete & Continuous Dynamical Systems - A, 2013, 33 (1) : 27-46. doi: 10.3934/dcds.2013.33.27 [18] John A. D. Appleby, Denis D. Patterson. Subexponential growth rates in functional differential equations. Conference Publications, 2015, 2015 (special) : 56-65. doi: 10.3934/proc.2015.0056 [19] Nguyen Thieu Huy, Ngo Quy Dang. Dichotomy and periodic solutions to partial functional differential equations. Discrete & Continuous Dynamical Systems - B, 2017, 22 (8) : 3127-3144. doi: 10.3934/dcdsb.2017167 [20] Serhiy Yanchuk, Leonhard Lücken, Matthias Wolfrum, Alexander Mielke. Spectrum and amplitude equations for scalar delay-differential equations with large delay. Discrete & Continuous Dynamical Systems - A, 2015, 35 (1) : 537-553. doi: 10.3934/dcds.2015.35.537
2018 Impact Factor: 1.143
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3812699019908905, "perplexity": 4456.963139673348}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514574588.96/warc/CC-MAIN-20190921170434-20190921192434-00419.warc.gz"}
|
https://www.physicsforums.com/threads/correspond-momentum-of-a-rigid-body-to-rotational-forces.553119/
|
# Correspond Momentum of a rigid body to rotational forces
1. Nov 23, 2011
### _chris_198
1. The problem statement, all variables and given/known data:
I have the momentum acting on the center of mass of a rigid body consisting of 3 bodies.
I would like to have the corresponding rotational forces (of this momentum) on the 3 bodies.
2.
I derive the general form:
Frot_i= Inertia_i*(RixM)/Total_Inertia*(Ri)^2
where:
Inertia_i=mass_i*(RixM)^2
Total_Inertia=SUM(Inertia_i)
3.
First calculate the vectors
Rix = xi-xc_of_mass
Riy = yi-yc_of_mass
Riz = zi-zc_of_mass
Then calculate the Ri^2:
=Rix^2 + Riy^2 + Riz^2
Then calculate the vectors Vi=(RixM):
Vix=Riy*Mz-Riz*My
Viy=Riz*Mx-Rix*Mz
Viz=Rix*My-Riy*Mx
Then calculate the scalar Si=(RixM)^2
Si=Vix^2 +Viy^2+Viz^2
Then I subtitute to the above equation and calculate the rotational forces of each of the 3 beads i in x y z component:
Frot_ix
Frot_iy
Frot_iz
When I sum up the x component of the force on the three beads I come out with a non-zero number. Shouldn't it be zero? Same is happening with y and z also.
Frot_1x+Frot_2x+Frot_3x !=0
Is the equation for calculating the rotational force correct?
I would be grateful for any reply.
cheers,
Chris
Can you offer guidance or do you also need help?
Similar Discussions: Correspond Momentum of a rigid body to rotational forces
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9326179623603821, "perplexity": 1153.2276600538992}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934805687.20/warc/CC-MAIN-20171119153219-20171119173219-00387.warc.gz"}
|
https://math.stackexchange.com/questions/1260038/when-does-one-use-succeeds-and-when-does-one-use-greater-than
|
# When does one use 'succeeds' and when does one use 'greater than'?
I am reading a text on convex optimisation, and there is a line:
$f_i(\tilde{x})\leq0$ and $h_i(\tilde{x})=0$, and $\lambda \succeq 0$
and I was just wondering why for one term, $\leq$ is used and for the other, $\succeq$ is used.
I have a computer science background and for some reason we never were taught much formal mathematical notation.
$\succeq$ is used typically in the context of matrices and vectors.
• If used in the context of vectors, it typically means that all elements of the vectors are non-negative, i.e., $\vec{\lambda} \succeq 0$, if $\lambda_i \geq 0$ for all $i$.
• If used in the context of matrices, it typically means that the matrix is non-negative definite, i.e., $A \succeq 0$, if $x^TAx \geq 0$ for all $x \in \mathbb{R}^{n}$, where $A \in \mathbb{R}^{n \times n}$. However, on extremely rare occasions, this symbol could also mean that all entries in a matrix are non-negative, i.e., $A \succeq 0$, if $A(i,j) \geq 0$ for all $i,j$.
In a different context, succeed might be used when there is "discrete" ordering, e.g. for the natural numbers, as opposed to "continuous" ordering, e.g. the reals. For example, $5$ succeeds $4$ (as there are no intermediate natural numbers between $4$ and $5$) it is also said that $5$ is the immediate successor of $4$. On the other hand there is not such thing as an immediate successor for real numbers, e.g. one could say that $4.6>4.2$ but we could always insert more numbers in between, $4.6>4.56>4.48374883748>4.2$. So here one would only say that $4.6$ is greater than $4.2$
• @guskenny83 Thank you. In yet another context one may use the same notation $\prec$ with a different word and different meaning, that is for (usually open) covers $\mathcal U$ and $\mathcal V$ of a topological space, one says that $\mathcal U$ refines $\mathcal V$, written $\mathcal U\prec \mathcal V$ if if for every $U\in\mathcal U$ there is some $V\in\mathcal V$ such that $U\subseteq V$. More generally $\prec$ is sometimes used for pre-orders: When $x\prec y$ and $y\prec x$ does not necessarily imply that $x=y$. – Mirko May 1 '15 at 2:31
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9556264281272888, "perplexity": 121.70951147520923}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439739182.35/warc/CC-MAIN-20200814070558-20200814100558-00248.warc.gz"}
|
https://simple.m.wikipedia.org/wiki/Chloroform
|
# Chloroform
chemical compound
Structure of a molecule of chloroform
Chloroform (also called trichloromethane) is a chemical substance. It is an organic compound. Chloroform is one of the intermediate substances that occur in the production of Polytetrafluoroethylene, better known as Teflon. Chloroform is used as a solvent. In the 19th century, it was a widely used anaesthetic.[1]
## History
The first who reported to have produced Chloroform was the French chemist Eugène Soubeiran, in 1831. Soubeiran took acetone and ethanol, and used bleach powder for the reaction.[2] The American physician Samuel Guthrie prepared gallons of the material and described its "deliciousness of flavor."[3] Independently, Justus von Liebig also described it.[4] Chloroform was named and chemically characterised in 1834 by Jean-Baptiste Dumas.[5]
## Production
Chloroform is produced by mixing methane and chlorine at a temperature of 400 to 500°C. In this reaction, the result is a mix of chloromethane, dichloromethane, trichloromethane and tetrachloromethane.
${\displaystyle \mathrm {CH_{4}+Cl_{2}\longrightarrow CH_{3}Cl+HCl} }$
${\displaystyle \mathrm {CH_{3}Cl+Cl_{2}\longrightarrow CH_{2}Cl_{2}+HCl} }$
${\displaystyle \mathrm {CH_{2}Cl_{2}+Cl_{2}\longrightarrow CHCl_{3}+HCl} }$
${\displaystyle \mathrm {CHCl_{3}+Cl_{2}\longrightarrow CCl_{4}+HCl} }$
The four compounds get then separated with a distillation.
In laboratories, chloroform can be produced by giving sodium hypochlorite and acetone together.
## Uses
The main uses of Chloroform are:
## Problems
An illustration showing the effects of Chloroform,from the 1840s
Chloroform was used as an anaesthetic during childbirth and surgery, from about 1847. It replaced ether which was used before. Chloroform is very poisonous, and can cause breathing problems, and problems with the heart. Death from chloroform can come from cardiac arrest. When Chloroform is stored for a longer time period, it can decay into Phosgene. Phosgene was a chemical weapon (poison gas) used during the First World War. Chloroform can cause birth defects and lead to miscarriages. It may cause cancer.
## References
1. "Chloroform - used, first, anesthetic, blood, body, Anesthetic Chloroform, Simpson Discovers Chloroforms Potency". discoveriesinmedicine.com. 2012 [last update]. Retrieved 17 July 2012. Check date values in: |year= (help)
2. Eugène Soubeiran (1831). ".". Ann. Chim. 48: 131.
3. Samuel Guthrie (1832). "New mode of preparing a spirituous solution of Chloric Ether". Am. J. Sci. and Arts 21: 64.
4. Justus Liebig (1832). "Ueber die Verbindungen, welche durch die Einwirkung des Chlors auf Alkohol, Aether, ölbildendes Gas und Essiggeist entstehen". Annalen der Pharmacie 1 (2): 182–230. doi:10.1002/jlac.18320010203.
5. Jean-Baptiste Dumas (1834). "Untersuchung über die Wirkung des Chlors auf den Alkohol". Annalen der Pharmacie 107 (41): 650–656. doi:10.1002/andp.18341074103.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 4, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6756191849708557, "perplexity": 23072.837263057325}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496665985.40/warc/CC-MAIN-20191113035916-20191113063916-00393.warc.gz"}
|
http://wbsjosaka.com/expedition/onosato08.htm
|
j
NO. @@ 7/26 8/23 9/27 10/25 11/22 12/27
17 JE 8 2 7 7 2 1
29 _CTM 1 3 @ 2 @ @
31 RTM 3 7 6 10 3 @
34 AITM 3 5 12 5 2 4
47 }K @ @ @ 2 @ 4
49 RK @ @ 10 8 12 17
53 qhK @ @ @ 26 48 72
72 ~TS @ 1 1 5 @ 4
74 gr @ 1 4 7 5 2
87 nuT @ @ @ 1 @ @
91 `EQ{E @ @ 1 @ @ @
108 CJ`h 7 4 11 12 11 1
109 V`h 8 1 1 @ @ @
115 P @ @ 5 @ @ @
126 n}VM 2 @ @ @ @ @
146 LAVVM 10 3 3 @ @ @
147 C\VM @ 2 5 5 3 3
148 \nVVM 1 @ @ @ @ @
153 `EVNVM 15 2 @ @ @ @
169 ZOJ @ @ 1 2 1 6
170 IIZOJ @ @ @ @ @ 1
174 E~lR 11 25 30 16 1 1
188 LWog 3 10 8 12 9 7
207 JZ~ @ @ 1 1 @ @
216 qo @ 1 1 6 5 5
218 co 20 15 5 @ @ @
224 nNZLC 5 11 4 3 5 3
225 ZOZLC 1 3 3 3 2 2
227 rYC @ @ @ @ @ 2
231 qh 5 2 7 12 22 20
233 Y @ @ 4 6 4 2
248 WEr^L @ @ @ @ 2 @
249 mr^L @ @ @ 2 @ @
251 C\qh @ 2 2 2 2 2
259 cO~ @ @ @ @ 2 12
261 EOCX @ @ @ @ 1 1
275 ZbJ 2 2 @ @ 1 @
294 zIW @ @ @ 2 2 2
301 AIW @ @ @ @ 3 2
304 IIW @ @ @ @ @ 3
307 Jq 2 2 1 6 2 21
320 XY 40 23 39 50 260
324 Nh 13 10 18 19 55 40
331 nV{\KX 1 4 2 5 3 6
332 nVugKX 5 3 3 4 4 1
@ XEv 22 25 28 29 28 30
1227
@ƂɁARTMA_CTMAJƂFp͂ȂBB̃Z_iLn[ɁANhAcO~AqhQ~HׂĂBqhKtAN̓JJYɒ[ȂBROJ]AN߂ƂẮAT~TJCBNjɂzFl肪Ƃ܂BN܂B
1122
1025
C݂rTMR̍XnŃCJ`hAn̑ŃzIWAmr^LQoAK̗ǂX^[gBACɓAAg̊FAC̐|BCb^Cpɐ|ĂƂ̂ƁB̂gAp͏ȂBłÃ𑽂~TSĂA|H˂ԃnuT܂pȂώ@łAQꂽF͖lqB
927
oƁASnJ[lj߂B̃VꂷɃ`EQ{EbzoOcJB܂QmAŌAgljыBCł̓E~lRJьAYqrnARKĂBgECHB
823
WVESEɂJAQ͂ȂBX^btRnQ`Bnƃ~TS_CrO܂BdȂAvĂAVME`hpȂBeg|bgɃ`EVNVMLAVVMpBr1030vJcȂAr~ƂBleeJQȂAЂ肵ĂAC\qhۗĂB
726
@AHn^n܂Ă̂A`EVNVMALAVVMȂǂꂽBɐBĂACJ`hړĂĂBiJł̃S~EAVM`hpijقƂ悤CɂȂB
NO. @@ 1/26 2/23 3/22 4/26 5/24 6/28
4 JCcu @ 1 @ @ @ @
17 JE 1 2 3 1 5 1
29 _CTM @ 2 2 2 1 @
31 RTM 1 2 4 6 5 2
34 AITM 5 7 3 1 3 5
47 }K 11 13 4 @ @ @
48 JK @ @ 2 2 @ @
49 RK 32 32 15 17 @ @
53 qhK 66 68 72 25 @ @
54 AJqh @ 1 1 @ @ @
70 E~ACT @ @ 1 @ @ @
72 ~TS 1 @ 1 @ @ @
74 gr 2 1 1 1 1 1
76 II^J @ @ @ 2 @ 1
79 nC^J @ @ @ 1 @ @
81 mX 1 @ @ @ @ @
87 nuT @ 1 @ @ @ @
100 o 1 1 @ @ @ @
107 R`h @ @ 3 1 @ 1
108 CJ`h 3 7 @ 2 @ @
109 V`h 16 1 11 1 1 @
110 _C`h @ @ 1 6 1 @
114 _C[ 2 2 @ @ @ @
115 P @ 1 2 2 2 2
126 n}VM 1 @ 7 1 @ @
146 LAVVM @ @ @ 11 6 @
147 C\VM 3 4 5 3 @ 2
148 \nVVM @ @ @ 1 3 @
150 II\nVVM @ @ @ 3 @ @
153 `EVNVM @ @ @ 23 4 6
168 J 3 6 3 2 @ @
169 ZOJ 3 28 2 1 @ @
170 IIZOJ 2 18 @ @ @ @
173 J @ 3 @ @ @ @
174 E~lR 2 8 2 1 @ 2
184 RAWTV @ @ @ @ 8 @
188 LWog 3 2 7 8 11 5
207 JZ~ 1 @ @ @ @ @
216 qo 20 5 6 3 2 11
218 co @ @ 4 23 11 30
224 nNZLC 4 7 3 2 @ 3
225 ZOZLC 2 1 1 @ @ @
229 ^qo @ 3 @ @ @ @
231 qh 14 2 6 5 3 @
233 Y 1 1 2 @ @ @
251 C\qh @ @ 1 @ 1 1
259 cO~ 16 10 15 4 @ @
261 EOCX 2 1 1 @ @ @
267 IIVL @ @ @ 1 @ @
275 ZbJ @ @ @ 1 2 3
293 W 6 16 3 @ @ @
294 zIW 3 @ 1 @ @ @
301 AIW 4 5 2 @ @ @
304 IIW 5 @ @ @ @ @
307 Jq 18 3 4 3 6 3
318 V @ @ 1 @ @ @
320 XY 43 12 12 20 20 19
324 Nh 8 18 20 16 21 38
331 nV{\KX 3 5 4 3 2 2
332 nVugKX 3 @ 5 1 3 5
@ vSEPC 37 37 40 37 23 21
628
ȂJjVBO@B㔼Jj̊ώ@Beg̏ɂUH̃`EVNVMC\qh̎pBcoNh挎葽BJj̓PtTC\KjAAJeKjAAVnKjANxPCKjAn}KjAJNxPCKjAn}KjA}gITKjAVI}lLAnNZVI~}lL̂XAjł͐ł̂ł́HƂgrn[̎pB
grn[
nNZVI}lL
524
J͍~ĂȂE~肻ȋ͗lBQ҂͂ȂvAÂQQnBC݂ɓt|c|cJ~oBCɃJEARAWTVAJVMXNьe͏BPԒJ{u~ƂȂAVEEĂPPIB
426
̎lœ키ALł̉e̎p͑SȂBŊPUFQVƒ̏]ǂȂÃJjTĂ`EiJVNVM̌Q̒II\nVVMAۂŃLAVVM̌QJjTĂpgB̉âlqVME`h̎ʂ̃|CgǂWbNgƊώ@B^XNVM`h̐͏Ȃ`JA^̃VM^Ő݂AQF͖ꂽlqB
322
Wꏊ߂̌ŃV̏o}BjɌr̋TMRՒn̍XnɃPQHBPH͒nʂɂ܂ĂBЂƂƕHł̓coBj͌ɓƁASn悢̍肪YBJjBoĂBJ̐ȂȂĂBVM`̎ށA͂܂ȂA悢j̋GߓƂBy݂ByސlĂB
223
@͕BC݂ɒƍXɕȂ萁ꂻBQH̃_C[̎pmFB}ɍȂJ~oB}篋Ɍꎞҋ@B̂ނ̎p͏ȂBVH̃CJ`h]̒ɂ܂ĂBJނʂ悤Ƃr[AĂɔїBƕӂnƏɃnuT̎pBJ̌Q̒ɃAJqh̎pB쒹̎pȂ悤ɊA荇킹ł͐挎ƓRVmFB
126
C݂ŃqoQȂĔщĂBȌi͒BIIWV̌sɕtĂ鍩HׂĂlqJZ~AmX̔ĂȂǂ܋߂ŃWbNƊώ@oB_C[V`hAn}VM̎pȂAOlA̒ł̒TłAQꂽF͊YꖞꂽlqłB
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.898631751537323, "perplexity": 1657.3719301414526}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370524043.56/warc/CC-MAIN-20200404134723-20200404164723-00482.warc.gz"}
|
https://www.nature.com/articles/s41598-021-86394-w?error=cookies_not_supported&code=57039e71-4202-4edf-85b8-ffc55141d2ba
|
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
# Smooth transportation of liquid metal droplets in a microchannel as detected by a serially arranged capacitive device
## Abstract
Gallium alloy liquid metals (Galinstan) possessing fluidity, electric conductivity, and low toxicity are attractive for use in flexible devices and microfluidic devices. However, the oxide skin of Galinstan in the atmosphere adheres to the microchannel surface, preventing the transportation of Galinstan in the channel. To tackle the problem of the adhesion of Galinstan to microchannel, we introduced liquid with Galinstan into a channel with a diameter of 1000 μm. Then, we found that the cylindrical shape of the channel enabled smooth transportation of Galinstan independently of both the liquid and the channel material. The liquid introduced with Galinstan not only prevents adhesion but also improves the spatial controllability of Galinstan in the channel. We can control the position of Galinstan with 100 μm resolution using highly viscous (> 10 cSt) liquid. In addition, we combined the microchannel with patterned electrodes, fabricating a serially arranged capacitive device. The local capacitance detected by the patterned electrodes changed by more than 6% via the smooth transportation of Galinstan. The analysis results based on an equivalent circuit quantitatively agree with our experimental results. We can modulate the serially arranged capacitors using the smooth transportation of Galinstan in the channel.
## Introduction
Liquid metal (LM) has both fluidity and high electrical conductivity and is an appropriate material for electrical wiring and sensors in flexible devices1,2,3. Since LM is soft and has a smooth shape, it is easily introduced into microchannels, indicating that it has promising microfluidic device applications, such as in switches and valves4. A well-known LM is mercury, which shows fluidity and high electric conductivity at room temperature. Hence, mercury has been used broadly in electric switches, diodes, and pressure gauges. However, because of the large burden on the environment and the high toxicity of mercury, its usage has recently been avoided5,6. Gallium alloy-LM (Galinstan) is a fascinating alternative to mercury because Galinstan shows low toxicity as well as fluidity and electrical conductivity7,8. Galinstan is applicable to the flexible devices as well as mercury. In addition, the fluidity of Galinstan is useful in the photocatalysis application9 and the conductive patterning application10.
The application of Galinstan to microdevices presents the following challenges: in the atmosphere, gallium at the surface is oxidized, and as a result, a thin (1–3 nm) oxide skin is spontaneously formed2,11. The oxide skin is adhesive to surfaces, which inhibits Galinstan from being transported smoothly in the microchannel and causes disconnection and deterioration of flexible devices12. In addition, the gallium oxide skin is an electric insulator (~ a few MΩ)13. Thus, the application of Galinstan based on its electrical conductivity is difficult without nontrivial treatment to remove the oxide skin.
By chemical surface treatment with a strong acid or base, the oxide skin can be removed, and thus both the fluidity and the electric conductivity can be maintained14,15. However, Galinstan in the atmosphere continues to be oxidized spontaneously. To take advantage of the promising properties of Galinstan with chemical treatment, continuous chemical treatment is necessary; that is, microdevices must contain a strong acid or base with Galinstan. Therefore, the chemical treatment requirement is a disadvantage in terms of practical device applications.
Recently, it was reported that liquid, such as water, introduced with Galinstan into microchannels can prevent the oxide skin from adhering to the surfaces of microfluidic devices16. The fluidity of Galinstan dispersed in liquid can be maintained without harsh chemicals, which is an advantage over chemical treatment. However, the mechanism for preventing gallium oxide from adhering is unclear. Because of the fluidity, Galinstan is appropriate material for the microfluidic application. Thus, the condition to transport Galinstan through the microchannel must be revealed. In principle, with this method, it is impossible to remove the oxide skin, making applications based on electric resistance difficult without additional treatment. Therefore, it is important to use electrical properties that can be detected without direct contact17,18. For example, capacitance is promising candidate because the position of Galinstan in the microchannel should couple with the local capacitance which can be measured indirectly.
In this paper, we solved the problem of the adhesion of Galinstan to microchannel, focusing on the geometry of and materials in the microfluidic devices. After introducing Galinstan into the microchannel, we applied pressure to the Galinstan. To control the position of the Galinstan in the microchannel, we prevented Galinstan from adhering to the surface of the channel and stably transported Galinstan while keeping its shape intact. In particular, we quantitatively estimated the reproductivity, which is an important parameter for the spatial control of Galinstan. Then, we fabricated a serially arranged capacitive device in which Galinstan are transported smoothly. We modulated the local capacitance using the smooth transportation of Galinstan. We discussed the local capacitive modulation and the spatial resolution of Galinstan droplet in the microchannel.
## Results
Figure 1a shows a microscopic image of a Galinstan and water in the cylindrical microchannel (diameter: ϕ = 600 μm). The Galinstan did not adhere to the surface of the channel and maintained a smooth, intact shape. While pressure was applied to the inlet, the moving Galinstan never adhered to the channel surface. Then, to investigate the effect of the channel material of the microfluidic device, we introduced Galinstan into cylindrical channels composed of Teflon (ϕ = 1 mm and 0.8 mm), vinyl chloride (ϕ = 1 mm), and glass (ϕ = 0.6 mm). As a result, we confirmed that the Galinstan is transported smoothly without adhesion to the surface of the channels. In addition, we observed the transportation of Galinstan with silicone oil, Fluorinert, or vegetable oil, whose chemical properties differ from each other. As a result, we confirmed that the oxide skin of Galinstan does not adhere to the channel surface independently of the liquid used. These results suggest that the materials that make up the channel and the liquid do not affect the transportation of Galinstan.
Figure 1b shows a microscope image of the Galinstan and water in a square pole-shaped channel. Since the Galinstan adhered to the channel surface just its injection with water, the Galinstan does not show fluidity in the square pole-shaped channel. Next, we fabricated a hybrid microchannel composed of glass and polydimethylsiloxane (PDMS), which is a typical combination in fluid microelectromechanical systems (MEMSs)19,20. However, the Galinstan in the square pole-shaped hybrid microchannel adhered to the channel surface. Therefore, the cylindrical shape of the microchannel is important for the smooth transportation of Galinstan.
Since the spatial resolution of the Galinstan in the microchannel is directly related to the detection resolution of devices, we focused on the spatial control of Galinstan. We sealed the outlet and repeatedly applied positive (50 kPa) and negative pressure on Galinstan in the microchannel. After 50 compression processes, we measured the spatial deviation from the initial position. As a result, the deviation of Galinstan in water was larger than 2 mm. This is caused by the time lag between the switching of the pressure direction and that of the transporting direction of Galinstan due to the inertia of Galinstan in low viscosity liquid. Then, we used highly viscous liquids, that is, vegetable or silicone oil, instead of water. Figure 2(a1–a3) shows optical microscopy images of Galinstan and silicone oil under positive pressure. Although the kinetic viscosity of silicone oil is much higher than that of water, the Galinstan is transported smoothly. In contrast, the Galinstan is transported in an inverse direction under negative pressure, as shown in Fig. 2(b1–b3). We summarized the liquids used and the kinetic viscosity and deviation of Galinstan from the initial position in Table 1. When Galinstan disperse in a highly viscous liquid, the deviation becomes smaller than the observation resolution of the microscope (< 100 μm).
To estimate the adhesion of Galinstan to the microchannel composed of PDMS, we measured the contact angle on the flat PDMS-film. We put the flat PDMS-film on the bottom of a container and filled the container with water. Then, we dropped 4 μl of Galinstan on into water. Sideview of Galinstan droplet was obtained by the microscope. Figure 3a shows the microscope image of the Galinstan droplet under water environment. Contact angle was estimated at 142°, which indicates that Galinstan droplet dewets PDMS. We removed water and observed Galinstan droplets under air environment. The contact angle decreases to 90°, which indicates that Galinstan adhere to PDMS-film. Therefore, we concluded that liquid prevents the Galinstan droplet from adhering to the microchannel.
To fabricate a serially arranged capacitive device using the smooth transportation of Galinstan droplets, we investigated the relation between capacitance and the position of Galinstan in a microchannel. Then, we sandwiched the microchannel between patterned electrodes and investigated the local capacitance of Galinstan introduced in the channel. Figure 4a shows a schematic cross section of the serially arranged capacitive device. The diameter of the channel was 1.0 mm. We defined the thickness of the channel composed of PDMS as d and used a channel with d = 2.3 mm. The Galinstan in the microchannel is enlarged in Fig. 4b. We connected the syringe pump to the inlet to control the position of Galinstan by pressure. Figure 4c shows a photograph of the serially arranged capacitive device. The microchannel was sandwiched between two glass substrates, on which we patterned three pairs of aluminum electrodes. Via capacitance measurements, a pair of electrodes is connected to an inductance–capacitance–resistance (LCR) meter. We measured C0 and C(x), where C0 and C(x) are the capacitance without Galinstan, i.e., the channel is occupied by pure silicone oil, and the capacitance as a function of the position of Galinstan, x (we took the origin at the center of our device), respectively. C0 was found to be 3–4 pF. Since the dielectric constant of silicone oil is almost the same as that of PDMS, and C0 = εS/d, substituting the experimental values of ε ~ 3 × 10−11 F/m, S ~ 6 × 10−5 m2, and d ~ 2 × 10−3 m, we obtain C0 ~ 1 pF, which agrees with the experimental value.
The red, green, and blue markers in Fig. 4c indicate the capacitance measured at the left, center, and right electrodes, respectively. Here, we normalized C(x) as C(x)/C0 to neglect the difference between the shapes of those electrodes. When Galinstan reach a position between the electrodes, the normalized capacitance C(x)/C0 takes the maximum value. In contrast, when the Galinstan are far from the electrodes, C(x)/C0 takes a constant minimum value. The change in C(x)/C0 is larger than 2%, which indicates that we can modulate the serially arranged capacitors using the smooth transportation of Galinstan in the microchannel.
One of the important factors that determined the change in the capacitance is the volume ratio of Galinstan to PDMS. We fabricated microfluidic devices composed of PDMS with a thickness of 1.4–5.5 mm. The thickness of PDMS must be greater than 1 mm because the diameter of the channel is 1 mm. Then, we sandwiched the microfluidic device between uniform electrode substrates and measured C0 and C, where C is the capacitance under the condition that the channel is occupied by Galinstan. Figure 5 shows the PDMS thickness dependence of the normalized capacitance C/C0. Red markers indicate the C/C0 measured by uniform electrodes. The blue markers indicate the maximum change in C(x)/C0 estimated from the results shown in Fig. 4c. When the thickness of a device is 1.4 mm, the C/C0 is larger than 6%. In addition, C/C0 monotonically decreases with increasing thickness.
## Discussion
### Smooth transportation of Galinstan
In the cylindrical channel, the Galinstan is subjected to a uniform force from the channel. As a result, the liquid layer is uniformly inserted between Galinstan and the channel, preventing Galinstan from adhering to the channel. In contrast, Galinstan in the square pole-shaped channel are under heterogeneous forces from the channel. For example, the force from the flat part of the channel is different from that of the edge. Since Galinstan are forced against the surface of the channel by heterogeneous forces, the oxide skin adheres to the channel. Therefore, to smoothly transport Galinstan through the microchannel, the shape of the cross section must be a circle, independent of the liquid and channel material.
### Change in the capacitance with and without Galinstan
Since silicone oil in the channel is replaced by Galinstan via smooth transportation, the local capacitance increases. To quantitatively analyze the change in the capacitance, we constructed an equivalent circuit of our device with two approximations: (i) the distribution of electric charge in the parallel plane to the electrodes is uniform, and (ii) the capacitance out of the device is negligible because the capacitances of PDMS and the channel are larger than that of air.
Figure 6 shows a schematic cross section of the channel in PDMS with thickness d and a width of 6 mm. We superimposed the equivalent circuit on the cross section, as shown in Fig. 6. The z-axis is perpendicular to the electrodes. We regarded PDMS in the ranges of (d + 1)/2 < z < d and 0 < z < (d − 1)/2 as capacitors with capacitance,
$$C_{p1} = C_{p3} = \varepsilon_{p} S\frac{2}{d - 1}$$
(1)
where εp and S are the dielectric constant of PDMS and the electrode area, respectively. We regarded the channel and PDMS in the range of (d − 1)/2 < z < (d + 1)/2 as two parallel capacitors because we neglected the heterogeneity of electric charges in the perpendicular plane to the z-axis. The capacitance of PDMS in the range of (d − 1)/2 < z < (d + 1)/2 is given as
$$C_{p2} = \varepsilon_{p} \left( {1 - \frac{\lambda }{6}} \right)S$$
(2)
where λ is the geometric parameter that corresponds to the area ratio of the channel to PDMS. For geometrical reasons, λ < 1. Since silicone oil is mainly composed of PDMS, we regarded the dielectric constant of silicone oil as εp. Then, the capacitance of the channel occupied by silicone oil, CSO, and that occupied by Galinstan, CLM, are denoted as
$$C_{f} = \left\{ {\begin{array}{*{20}c} {C_{SO} = \varepsilon_{p} \frac{\lambda }{6} S \,({\text{silicon oil}})} \\ {C_{LM} = \alpha C_{SO} \,({\text{Galinstan}})} \\ \end{array} } \right.$$
(3)
where α is the ratio of the capacitance with Galinstan to that with silicone oil. Equations (13) indicate that the total capacitance of the equivalent circuit is given as
$$\frac{1}{C} = \frac{1}{{C_{p1} }} + \frac{1}{{C_{f} + C_{p2} }} + \frac{1}{{C_{p3} }}$$
(4)
The normalized capacitance C/C0 is given as
$$\frac{C}{{C_{0} }} = \frac{d}{d - \delta (\alpha , \lambda )}$$
(5a)
and
$$\delta (\alpha , \lambda ) = \frac{\lambda (\alpha - 1)}{{6 +\lambda (\alpha - 1)}}$$
(5b)
Equation (5a) indicates that the electrode gap d is modified as d − δ by replacing silicone oil with Galinstan in the channel. We fitted the experimental results with Eq. (5a), where δ is the only fitting parameter, and showed the best fit curve in Fig. 5 as the green line. Our model based on the equivalent circuit agrees approximately with the experimental results. A slight quantitative disagreement is thought to be caused by our approximations, where we neglected the electric heterogeneity in the parallel plane to the electrodes. The fitting result indicates that δ ~ 70 μm, and α = 1.5—6. This result suggests that the capacitance increases 1.5–6 times by introducing Galinstan into the channel.
### Motion and position detection of transported Galinstan droplets in a microchannel
We succeeded in changing the capacitance of serially arranged capacitors using the smooth transportation of Galinstan. In the serially arranged capacitive device, the normalized capacitance C(x)/C0 depends on the relative positions of Galinstan and the electrodes. When the Galinstan is far from the electrodes, C(x)/C0 takes a constant minimum value. After intruding into the channel sandwiched by the electrodes, C(x)/C0 starts increasing. C(x)/C0 eventually reaches a maximum value at the position where the center of the Galinstan corresponds with that of the electrodes. These results indicate that C(x)/C0 monotonously depends on the volume ratio of Galinstan in the channel between the electrodes.
For microfabrication, (1) the length of the Galinstan droplets and the width of the electrodes should correspond with each other to improve the spatial resolution of the Galinstan near the center of the electrodes. Under the condition that the length of Galinstan is not equal to the width of electrodes, the spatial deviation of Galinstan from the center of the electrodes is not accompanied by a change in C(x)/C0 because the volume fraction of Galinstan is invariant near the center of the electrodes. (2) The length of Galinstan should be longer than the interval between pairs of electrodes. Unless a part of Galinstan overlaps with electrodes, the spatial change in Galinstan does not cause a change in C(x)/C0. Since the position of Galinstan droplet in the channel whose diameter is larger than the size of the Galinstan droplet is unstable along the depth direction, Galinstan must be larger than the diameter of the channel. Thus, we found that the limitation of the microfabrication is ~ 1 pattern/mm. Since the occupation volume of Galinstan in the microchannel is much smaller than that of the liquid, the viscous dissipation in Galinstan is negligible. Here, we assumed that the flow field in the microchannel is Poiseuille flow. The volumetric flow rate is given as
$$Q \sim \frac{\pi }{8\gamma } a^{4} \frac{P}{l},$$
(6)
where P, a, l, and γ are the pressure difference between inlet and outlet, the radius of the microchannel, the length of the microchannel, and viscous coefficient, respectively. Signal frequency is determined as v/d, where v and d are the velocity of Galinstan droplet and electrode interval, respectively. Substituting the experimental values of P ~ 30 kPa, a ~ 500 μm, l ~ 500 mm, and γ ~ 0.1 Pa/s, we obtained Q ~ 15 mm3/s and the average transportation velocity of Galinstan droplet ~ 10–20 mm/s. Because of the limitation of microfabrication (1/d ~ 1 pattern/mm), the signal frequency was estimated as ~ 10 Hz.
We modulated serially arranged capacitors using the smooth transportation of Galinstan droplets, which conversely indicates that we can sense the position of Galinstan in the channel from the change in capacitance. The change in the capacitance is 10−2 pF/mm. Taking into account the resolution of the LCR meter (10−2 ~ 10−3 pF), the position of Galinstan can be detected with a 100 μm resolution. The deviation after the transport process was repeated was found to be approximately 100 μm. Thus, we can sense the position of Galinstan based on serially arranged capacitors with a comparable resolution with spatial controllability.
## Method
### Observation of Galinstan in a microchannel
We introduced Galinstan (eutectic gallium indium stannum, Zairyo-ya.com) and liquid into a microchannel composed of PDMS (DuPont Toray Specialty Materials K.K.). We used water, vegetable oil (Kadoya Sesame Mill inc.), Fluorinert (3 M), and silicone oil (MOMENTIVE) as the liquids. One side of the through-hole of the PDMS microchannel was an inlet, and the other side was an outlet. We fabricated two kinds of channels, that is, a cylindrical channel and a square pole-shaped channel. The diameter of the cylindrical channel ϕ was 600 μm–1 mm, and the dimensions of the square pole channel were 500 μm × 500 μm. For observation of Galinstan in the microchannel, we used a microscope (DIGITAL MICROSCOPE VHX-500F, KEYENCE) and a transmission optical microscope (MX9430, MEJI TECHNO). All observation and experiments in this paper were performed at room temperature.
### Position control of Galinstan in microchannel
We introduced Galinstan and liquid from the inlet. Then, we sealed the outlet with epoxy glue, where a small amount of air between the liquid and outlet was maintained. To transport Galinstan from the inlet to the outlet, we applied 30 kPa pressure to the inlet for 5 s. For transportation in the backward direction, we removed the pressure for 5 s.
### Position of Galinstan and serially arranged capacitors
A PDMS microchannel with ϕ = 1 mm and diameter d was sandwiched by glass plates whose surface was patterned with aluminum electrodes. Electrodes were connected with an LCR meter (LCR METER, KEYSIGHT). We injected Galinstan and silicone oil into the microchannel with a syringe pump. To control the position of Galinstan in the microchannel, we applied pressure to the channel. For capacitance measurements, aluminum electrodes connected to an LCR meter were applied to an AC electric field at 2 V and 500 kHz. We measured the capacitance and the position of Galinstan simultaneously.
The total capacitance of PDMS and the microchannel is approximately εS/d, where ε, S, and d are the dielectric constant of PDMS, the electrode area, and the electrode gap, respectively. The resolution of capacitance increases with decreasing electrode gap. The minimum electrode gap is the diameter of the microchannel. For this reason, we designed an electrode gap on a similar order to the diameter of the microchannel (~ 1 mm).
We used two kinds of electrode substrates to evaluate the capacitance.
1. (1)
Patterned electrodes (Fig. 7a1–2)
We patterned the upper glass substrate with three aluminum electrodes whose dimensions were 6 mm × 10 mm. The intervals of electrodes were 16 mm. The lower glass substrate was patterned with three aluminum electrodes whose width was 10 mm with a 16 mm spacing. We connected a pair of electrodes to an LCR meter to measure the local capacitance. Based on the balance between spatial and measurement resolutions, the electrode area S was set to 60 mm2.
1. (2)
Uniform electrodes (Fig. 7 (b1-2))
We deposited uniform aluminum electrodes on glass substrates to accurately measure the capacitance. PDMS microchannels with a width of 6 mm and diameter d were sandwiched by uniform electrode substrates whose electrode area S was ~400 mm2. Upper and lower electrodes were connected to an LCR meter to measure the total capacitance of the fluidic device.
## Conclusion
In this paper, we overcame the problem of the adhesion of Galinstan to microchannel. We showed that Galinstan introduced into a cylindrical channel with liquid are transported smoothly independently of the liquid and the channel material because the Galinstan in the cylindrical channel is subjected to a uniform force from the channel. The position of Galinstan can be controlled with a resolution of 100 μm using highly viscous (~ 10 cSt) liquid. These results indicate that the smooth transportation of Galinstan droplets is applicable for microfluidic device with high spatial resolution. Then, we fabricated serially arranged capacitive device and measured the local capacitance indirectly. The local capacitance of the patterned electrode device can be modulated by the position of Galinstan in the channel. The prediction of our equivalent circuit model agrees with the experimental behavior, which indicates that Galinstan droplet locally increases the capacitance. These results suggest that the smooth transportation of Galinstan through a channel is applicable for the modulation of serially arranged capacitors. Moreover, we showed the possibility of applying a serially arranged capacitive device as a spatial sensor of Galinstan droplets even in a microfabrication system.
## Data availability
All data are included in this published article.
## References
1. 1.
Otake, S., Mori, F. & Konishi, S. Common Flexible Channel Structure for Integration of Three-Axis Force Sensor and Soft Micro Actuator. in 2019 IEEE 32nd International Conference on Micro Electro Mechanical Systems (MEMS) 652–655 (2019). doi:https://doi.org/10.1109/MEMSYS.2019.8870621.
2. 2.
Zheng, Y., He, Z., Gao, Y. & Liu, J. Direct desktop printed-circuits-on-paper flexible electronics. Sci. Rep. 3, 1786 (2013).
3. 3.
Kazem, N., Hellebrekers, T. & Majidi, C. Soft multifunctional composites and emulsions with liquid metals. Adv. Mater. 29, 1605985 (2017).
4. 4.
Tang, S.-Y., Lin, Y., Joshipura, I. D., Khoshmanesh, K. & Dickey, M. D. Steering liquid metal flow in microchannels using low voltages. Lab. Chip 15, 3905–3911 (2015).
5. 5.
Boening, D. W. Ecological effects, transport, and fate of mercury: a general review. Chemosphere 40, 1335–1351 (2000).
6. 6.
Clarkson, T. W., Magos, L. & Myers, G. J. The toxicology of mercury–current exposures and clinical manifestations. N. Engl. J. Med. 349, 1731–1737 (2003).
7. 7.
Liu, T., Sen, P. & Kim, C. Characterization of nontoxic liquid-metal alloy Galinstan for applications in microdevices. J. Microelectromechanical Syst. 21, 443–450 (2012).
8. 8.
Dickey, M. D. Stretchable and soft electronics using liquid metals. Adv. Mater. 29, 1606425 (2017).
9. 9.
Ghasemian, M. B. et al. Self-limiting galvanic growth of MnO2 monolayers on a liquid metal—applied to photocatalysis. Adv. Funct. Mater. 29, 1901649 (2019).
10. 10.
Rahim, M. A. et al. Polyphenol-Induced Adhesive Liquid Metal Inks for Substrate-Independent Direct Pen Writing. Adv. Funct. Mater. n/a, 2007336.
11. 11.
Dickey, M. D. Emerging Applications of Liquid Metals Featuring Surface Oxides. ACS Appl. Mater. Interfaces 6, 18369–18379 (2014).
12. 12.
Boley, J. W., White, E. L., Chiu, G.T.-C. & Kramer, R. K. Direct writing of gallium-indium alloy for stretchable electronics. Adv. Funct. Mater. 24, 3501–3507 (2014).
13. 13.
Mori, F., Hirata, A., Kakehi, Y. & Konishi, S. An Effect of a Needle Electrode for the Ohmic Contact Between an Electrode and Liquid Metal Droplet in Electrical Switching. in 2020 IEEE 33rd International Conference on Micro Electro Mechanical Systems (MEMS) 957–960 (2020). doi:https://doi.org/10.1109/MEMS46641.2020.9056437.
14. 14.
Khoshmanesh, K. et al. Liquid metal enabled microfluidics. Lab. Chip 17, 974–993 (2017).
15. 15.
Kim, D. et al. Recovery of nonwetting characteristics by surface modification of gallium-based liquid metal droplets using hydrochloric acid vapor. ACS Appl. Mater. Interfaces 5, 179–185 (2013).
16. 16.
Khan, M. R., Trlica, C., So, J.-H., Valeri, M. & Dickey, M. D. Influence of Water on the Interfacial Behavior of Gallium Liquid Metal Alloys. ACS Appl. Mater. Interfaces 6, 22467–22473 (2014).
17. 17.
So, J.-H. et al. Reversibly deformable and mechanically tunable fluidic antennas. Adv. Funct. Mater. 19, 3632–3637 (2009).
18. 18.
Chen, C., Whalen, J. & Peroulis, D. Non-Toxic Liquid-Metal 2–100 GHz MEMS Switch. in 2007 IEEE/MTT-S International Microwave Symposium 363–366 (2007). doi:https://doi.org/10.1109/MWSYM.2007.380446.
19. 19.
Fujii, T. PDMS-based microfluidic devices for biomedical applications. Microelectron. Eng. 61–62, 907–914 (2002).
20. 20.
Zhou, J., Ellis, A. V. & Voelcker, N. H. Recent developments in PDMS surface modification for microfluidic devices. Electrophoresis 31, 2–16 (2010).
## Acknowledgements
This work was partially supported by Ritsumeikan Global Innovation Research Organization.
## Author information
Authors
### Contributions
S.K. conceived the direction of research and experiment. S.K., Y.K., F.M., and S.B. performed experiments. S.K. and S.B. wrote the manuscript. All author discussed the results and reviewed the manuscript.
### Corresponding author
Correspondence to Satoshi Konishi.
## Ethics declarations
### Competing interests
The authors declare no competing interests.
### Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
## Rights and permissions
Reprints and Permissions
Konishi, S., Kakehi, Y., Mori, F. et al. Smooth transportation of liquid metal droplets in a microchannel as detected by a serially arranged capacitive device. Sci Rep 11, 7048 (2021). https://doi.org/10.1038/s41598-021-86394-w
• Accepted:
• Published:
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7715956568717957, "perplexity": 3060.042011994951}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243989756.81/warc/CC-MAIN-20210518063944-20210518093944-00066.warc.gz"}
|
https://www.jobilize.com/course/section/parametric-surfaces-surface-integrals-by-openstax?qcr=www.quizover.com
|
# 6.6 Surface integrals
Page 1 / 27
• Find the parametric representations of a cylinder, a cone, and a sphere.
• Describe the surface integral of a scalar-valued function over a parametric surface.
• Use a surface integral to calculate the area of a given surface.
• Explain the meaning of an oriented surface, giving an example.
• Describe the surface integral of a vector field.
• Use surface integrals to solve applied problems.
We have seen that a line integral is an integral over a path in a plane or in space. However, if we wish to integrate over a surface (a two-dimensional object) rather than a path (a one-dimensional object) in space, then we need a new kind of integral that can handle integration over objects in higher dimensions. We can extend the concept of a line integral to a surface integral to allow us to perform this integration.
Surface integrals are important for the same reasons that line integrals are important. They have many applications to physics and engineering, and they allow us to develop higher dimensional versions of the Fundamental Theorem of Calculus. In particular, surface integrals allow us to generalize Green’s theorem to higher dimensions, and they appear in some important theorems we discuss in later sections.
## Parametric surfaces
A surface integral is similar to a line integral, except the integration is done over a surface rather than a path. In this sense, surface integrals expand on our study of line integrals. Just as with line integrals, there are two kinds of surface integrals: a surface integral of a scalar-valued function and a surface integral of a vector field.
However, before we can integrate over a surface, we need to consider the surface itself. Recall that to calculate a scalar or vector line integral over curve C , we first need to parameterize C . In a similar way, to calculate a surface integral over surface S , we need to parameterize S . That is, we need a working concept of a parameterized surface (or a parametric surface ), in the same way that we already have a concept of a parameterized curve.
A parameterized surface is given by a description of the form
$\text{r}\left(u,v\right)=⟨x\left(u,v\right),y\left(u,v\right),z\left(u,v\right)⟩.$
Notice that this parameterization involves two parameters, u and v , because a surface is two-dimensional, and therefore two variables are needed to trace out the surface. The parameters u and v vary over a region called the parameter domain, or parameter space —the set of points in the uv -plane that can be substituted into r . Each choice of u and v in the parameter domain gives a point on the surface, just as each choice of a parameter t gives a point on a parameterized curve. The entire surface is created by making all possible choices of u and v over the parameter domain.
## Definition
Given a parameterization of surface $\text{r}\left(u,v\right)=⟨x\left(u,v\right),y\left(u,v\right),z\left(u,v\right)⟩,$ the parameter domain of the parameterization is the set of points in the uv -plane that can be substituted into r .
## Parameterizing a cylinder
Describe surface S parameterized by
$\text{r}\left(u,v\right)=⟨\text{cos}\phantom{\rule{0.2em}{0ex}}u,\text{sin}\phantom{\rule{0.2em}{0ex}}u,v⟩,\text{−}\infty
To get an idea of the shape of the surface, we first plot some points. Since the parameter domain is all of ${ℝ}^{2},$ we can choose any value for u and v and plot the corresponding point. If $u=v=0,$ then $\text{r}\left(0,0\right)=⟨1,0,0⟩,$ so point (1, 0, 0) is on S . Similarly, points $\text{r}\left(\pi ,2\right)=\left(-1,0,2\right)$ and $\text{r}\left(\frac{\pi }{2},4\right)=\left(0,1,4\right)$ are on S .
Although plotting points may give us an idea of the shape of the surface, we usually need quite a few points to see the shape. Since it is time-consuming to plot dozens or hundreds of points, we use another strategy. To visualize S , we visualize two families of curves that lie on S. In the first family of curves we hold u constant; in the second family of curves we hold v constant. This allows us to build a “skeleton” of the surface, thereby getting an idea of its shape.
First, suppose that u is a constant K . Then the curve traced out by the parameterization is $⟨\text{cos}\phantom{\rule{0.2em}{0ex}}K,\text{sin}\phantom{\rule{0.2em}{0ex}}K,v⟩,$ which gives a vertical line that goes through point $\left(\text{cos}\phantom{\rule{0.2em}{0ex}}K,\text{sin}\phantom{\rule{0.2em}{0ex}}K,v\right)$ in the xy -plane.
Now suppose that v is a constant K. Then the curve traced out by the parameterization is $⟨\text{cos}\phantom{\rule{0.2em}{0ex}}u,\text{sin}\phantom{\rule{0.2em}{0ex}}u,K⟩,$ which gives a circle in plane $z=K$ with radius 1 and center (0, 0, K ).
If u is held constant, then we get vertical lines; if v is held constant, then we get circles of radius 1 centered around the vertical line that goes through the origin. Therefore the surface traced out by the parameterization is cylinder ${x}^{2}+{y}^{2}=1$ ( [link] ).
Notice that if $x=\text{cos}\phantom{\rule{0.2em}{0ex}}u$ and $y=\text{sin}\phantom{\rule{0.2em}{0ex}}u,$ then ${x}^{2}+{y}^{2}=1,$ so points from S do indeed lie on the cylinder. Conversely, each point on the cylinder is contained in some circle $⟨\text{cos}\phantom{\rule{0.2em}{0ex}}u,\text{sin}\phantom{\rule{0.2em}{0ex}}u,k⟩$ for some k , and therefore each point on the cylinder is contained in the parameterized surface ( [link] ).
Is there any normative that regulates the use of silver nanoparticles?
what king of growth are you checking .?
Renato
What fields keep nano created devices from performing or assimulating ? Magnetic fields ? Are do they assimilate ?
why we need to study biomolecules, molecular biology in nanotechnology?
?
Kyle
yes I'm doing my masters in nanotechnology, we are being studying all these domains as well..
why?
what school?
Kyle
biomolecules are e building blocks of every organics and inorganic materials.
Joe
anyone know any internet site where one can find nanotechnology papers?
research.net
kanaga
sciencedirect big data base
Ernesto
Introduction about quantum dots in nanotechnology
what does nano mean?
nano basically means 10^(-9). nanometer is a unit to measure length.
Bharti
do you think it's worthwhile in the long term to study the effects and possibilities of nanotechnology on viral treatment?
absolutely yes
Daniel
how to know photocatalytic properties of tio2 nanoparticles...what to do now
it is a goid question and i want to know the answer as well
Maciej
Abigail
for teaching engĺish at school how nano technology help us
Anassong
Do somebody tell me a best nano engineering book for beginners?
there is no specific books for beginners but there is book called principle of nanotechnology
NANO
what is fullerene does it is used to make bukky balls
are you nano engineer ?
s.
fullerene is a bucky ball aka Carbon 60 molecule. It was name by the architect Fuller. He design the geodesic dome. it resembles a soccer ball.
Tarell
what is the actual application of fullerenes nowadays?
Damian
That is a great question Damian. best way to answer that question is to Google it. there are hundreds of applications for buck minister fullerenes, from medical to aerospace. you can also find plenty of research papers that will give you great detail on the potential applications of fullerenes.
Tarell
what is the Synthesis, properties,and applications of carbon nano chemistry
Mostly, they use nano carbon for electronics and for materials to be strengthened.
Virgil
is Bucky paper clear?
CYNTHIA
carbon nanotubes has various application in fuel cells membrane, current research on cancer drug,and in electronics MEMS and NEMS etc
NANO
so some one know about replacing silicon atom with phosphorous in semiconductors device?
Yeah, it is a pain to say the least. You basically have to heat the substarte up to around 1000 degrees celcius then pass phosphene gas over top of it, which is explosive and toxic by the way, under very low pressure.
Harper
Do you know which machine is used to that process?
s.
how to fabricate graphene ink ?
for screen printed electrodes ?
SUYASH
What is lattice structure?
of graphene you mean?
Ebrahim
or in general
Ebrahim
in general
s.
Graphene has a hexagonal structure
tahir
On having this app for quite a bit time, Haven't realised there's a chat room in it.
Cied
what is biological synthesis of nanoparticles
how did you get the value of 2000N.What calculations are needed to arrive at it
Privacy Information Security Software Version 1.1a
Good
Got questions? Join the online conversation and get instant answers!
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 24, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8970869779586792, "perplexity": 953.0127175960356}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627999817.30/warc/CC-MAIN-20190625092324-20190625114324-00388.warc.gz"}
|
http://www.solved-problems.com/circuits/electrical-circuits-problems/641/nodal-analysis-dependent-current-source/
|
Deploy nodal analysis method to solve the circuit and find the power of the dependent source.
Solution
I. Identify all nodes in the circuit. Call the number of nodes $N$.
The circuit has 4 nodes:
Therefore, $N=4$.
II. Select a reference node. Label it with reference (ground)
symbol.
All nodes have the same number of elements. We prefer to select one of the nodes connected to the voltage source to avoid having to use a supernode.
III. Assign a variable for each node whose voltage is unknown.
We label the remaining three nodes as shown above.
IV. If there are dependent sources in the circuit, write down equations that express their values in terms of other node voltages.
There is one dependent source, which is a current controlled current source. We need to write $-2I_1$ in terms of node voltages. $I_1$ is the current passing through the $2\Omega$ - resistor. Applying the Ohm's law,
$I_1 = \frac{V_2-V_3}{2\Omega}$.
Hence, $-2I_1 = V_3-V_2$.
V. Write down a KCL equation for each node.
Node $V_1$: $\frac{V_1}{3 \Omega}+\frac{V_1-V_2}{1\Omega}-2I_1=0$.
$\to 4V_1-6V_2+3V_3=0$ (Eq. 1).
Node $V_2$: $2I_1+\frac{V_2-V_1}{1\Omega}-2A+\frac{V_2-V_3}{2\Omega}=0$.
Please note that we avoid using all unknowns except node voltages. Using $I_1$ in this KCL equation introduces an unnecessary unknown to the equations set. Substituting $2I_1 = V_2-V_3$ and rearranging results in:
$-2V_1+5V_2-3V_3=4$ (Eq. 2).
Node $V_3$ has a voltage source connected to. Therefore, $V_3=10V$. Substituting this in Eq. 1 and Eq.2 leads to
$\left\{\begin{array}{l} 4V_1-6V_2=-30 \\ -2V_1+5V_2=34 \end{array}\right.$.
By solving the system of equations,
$V_1=6.75V$ and $V_2=9.5$.
Now, we need to find the voltage across the dependent current source and the current passing through it. Lets start with $I_1$.
$I_1 = \frac{V_2-V_3}{2\Omega}=-0.25 A$.
Assuming positive terminal placed on the node of $V_1$, the voltage across the dependent current source is $V_1-V_2=-2.75V$. The current flowing through the dependent current source is $-2I_1=0.5A$. Therefore the power of the dependent current source is $-2.75 \times 0.5 = -1.375W$. Because the current direction and the voltage polarity is in accordance with the passive sign convention and the power is negative, the dependent current source is supplying power.
Hi! Yaz is here. I am passionate about learning and teaching. I try to explain every detail simultaneously with examples to ensure that students will remember them later too.
## Join the Conversation
1. khans sahil says:
Thank you sir
2. Ram Kaushik says:
thanks a lot !! im passing my test coz of u .... cheers mate 🙂 🙂
3. Juan dela Cruz says:
Thanks, but correct me if i'm wrong but all nodes doesn't have the same number of elements:
*Node 4 has 3 elements = 2 ohms, 3 ohms and 10V
*Node 2 has 4 elements = -2I1 dependet source, 2A current source, 1 ohms and 2 ohms.
*Node 3 has 4 elements = 2 ohms, 10V, 2A current source and 2 ohms.
*Node 1 has 3 elements = 1 ohms, 3 ohms and -2I1 dependet source.
Why choose node 4 with 3 elements as reference node if node 2 and node 3 have 4 elements or the most number of elements?
|
{"extraction_info": {"found_math": true, "script_math_tex": 26, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 26, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7270705103874207, "perplexity": 1015.2775676252679}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250607407.48/warc/CC-MAIN-20200122191620-20200122220620-00156.warc.gz"}
|
http://mathhelpforum.com/pre-calculus/83387-word-problems-help.html
|
1. ## Word Problems help
1. The displacement of a spring vibrating in damped harmonic motion is given by the following equation. Find the times when the spring is at its equilibrium position, y = 0.
2. As the moon revolves around the earth, the side that faces the earth is usually just partially illuminated by the sun. The phases of the moon describe how much of the surface appears to be in sunlight. An astronomical measure of phase is given by the fraction F of the lunar disc that is lit. When the angle between the sun, earth, and moon is θ (0 ≤ θ < 360°), then the fraction F is given by the formula below. Determine the angles θ that correspond to the following phases.
3. Consider the equation.
(a) Use an addition or subtraction formula to simplify the equation.
(b) Find all solutions in the interval [0, 2π). (
2. did you check the last post you made with the same questions?
http://www.mathhelpforum.com/math-he...solutions.html
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9155239462852478, "perplexity": 374.98154603786367}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560280730.27/warc/CC-MAIN-20170116095120-00324-ip-10-171-10-70.ec2.internal.warc.gz"}
|
https://stacks.math.columbia.edu/tag/04GZ
|
Lemma 7.30.1. Let $\mathcal{C}$ be a site. Let $\mathcal{F}$ be a sheaf on $\mathcal{C}$. Then the category $\mathop{\mathit{Sh}}\nolimits (\mathcal{C})/\mathcal{F}$ is a topos. There is a canonical morphism of topoi
$j_\mathcal {F} : \mathop{\mathit{Sh}}\nolimits (\mathcal{C})/\mathcal{F} \longrightarrow \mathop{\mathit{Sh}}\nolimits (\mathcal{C})$
which is a localization as in Section 7.25 such that
1. the functor $j_\mathcal {F}^{-1}$ is the functor $\mathcal{H} \mapsto \mathcal{H} \times \mathcal{F}/\mathcal{F}$, and
2. the functor $j_{\mathcal{F}!}$ is the forgetful functor $\mathcal{G}/\mathcal{F} \mapsto \mathcal{G}$.
Proof. Apply Lemma 7.29.5. This means we may assume $\mathcal{C}$ is a site with subcanonical topology, and $\mathcal{F} = h_ U = h_ U^\#$ for some $U \in \mathop{\mathrm{Ob}}\nolimits (\mathcal{C})$. Hence the material of Section 7.25 applies. In particular, there is an equivalence $\mathop{\mathit{Sh}}\nolimits (\mathcal{C}/U) = \mathop{\mathit{Sh}}\nolimits (\mathcal{C})/h_ U^\#$ such that the composition
$\mathop{\mathit{Sh}}\nolimits (\mathcal{C}/U) \to \mathop{\mathit{Sh}}\nolimits (\mathcal{C})/h_ U^\# \to \mathop{\mathit{Sh}}\nolimits (\mathcal{C})$
is equal to $j_{U!}$, see Lemma 7.25.4. Denote $a : \mathop{\mathit{Sh}}\nolimits (\mathcal{C})/h_ U^\# \to \mathop{\mathit{Sh}}\nolimits (\mathcal{C}/U)$ the inverse functor, so $j_{\mathcal{F}!} = j_{U!} \circ a$, $j_\mathcal {F}^{-1} = a^{-1} \circ j_ U^{-1}$, and $j_{\mathcal{F}, *} = j_{U, *} \circ a$. The description of $j_{\mathcal{F}!}$ follows from the above. The description of $j_\mathcal {F}^{-1}$ follows from Lemma 7.25.7. $\square$
In your comment you can use Markdown and LaTeX style mathematics (enclose it like $\pi$). A preview option is available if you wish to see how it works out (just click on the eye in the toolbar).
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 2, "x-ck12": 0, "texerror": 0, "math_score": 0.9994243383407593, "perplexity": 191.31850267179033}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-18/segments/1555578727587.83/warc/CC-MAIN-20190425154024-20190425175111-00046.warc.gz"}
|
http://www.physicspages.com/2015/07/01/ideal-gas-relation-of-average-speed-of-molecules-to-temperature/
|
# Ideal gas: relation of average speed of molecules to temperature
Reference: Daniel V. Schroeder, An Introduction to Thermal Physics, (Addison-Wesley, 2000) – Problems 1.18 – 1.20.
By using the ideal gas law and an analysis on the molecular scale, we can derive a relation between the speed of the molecules in an ideal gas and the temperature. Schroeder does this in detail in his section 1.2 so I won’t go through the whole derivation again; I’ll just summarize the main ideas.
The model places a single molecule inside a cylinder with a movable piston at one end. The molecule is moving at some speed ${v}$ and whenever it hits a wall or the piston it bounces off elastically. If the axis of the cylinder is the ${x}$ axis, then whenever the molecule hits the piston, the ${x}$ component of its velocity, ${v_{x}}$, is reversed, which means that its ${x}$ momentum changes by ${\Delta p_{x}=-2mv_{x}}$. The rate of change of momentum is the force exerted on the piston, and the force per unit area is the pressure, so we can relate ${v_{x}}$ to the pressure exerted on the piston. Since we’re considering only one molecule, the pressure is felt only at the moments when the molecule collides with the piston, so what we’re really interested in is the time average of the pressure. It turns out that this is
$\displaystyle \bar{P}=\frac{mv_{x}^{2}}{V} \ \ \ \ \ (1)$
where ${m}$ is the mass of the molecule and ${V}$ is the volume of the cylinder.
We can now extend the argument by putting a large number ${N}$ of molecules in the cylinder. In this case, the ${v_{x}^{2}}$ factor is replaced by the average of ${v_{x}^{2}}$ over all the molecules, so we get
$\displaystyle \bar{P}V=Nm\overline{v_{x}^{2}} \ \ \ \ \ (2)$
Comparing with the ideal gas law
$\displaystyle PV=NkT \ \ \ \ \ (3)$
we see that
$\displaystyle m\overline{v_{x}^{2}}=kT \ \ \ \ \ (4)$
or, since ${\frac{1}{2}m\overline{v_{x}^{2}}}$ is the kinetic energy from the ${x}$ motion of the molecules,
$\displaystyle \frac{1}{2}m\overline{v_{x}^{2}}=\frac{1}{2}kT \ \ \ \ \ (5)$
However, since the molecules are moving at random, there’s nothing special about the ${x}$ direction, so we’d expect the same contribution to the kinetic energy from the ${y}$ and ${z}$ directions, giving the relation
$\displaystyle \bar{K}_{trans}=\frac{1}{2}m\overline{v^{2}}=\frac{3}{2}kT \ \ \ \ \ (6)$
where ${\bar{K}_{trans}}$ is the average kinetic energy due to the translational motion of the molecules (if the molecules contain two or more atoms, then we can also have rotational and vibrational kinetic energy, so the total kinetic energy is greater than ${\frac{3}{2}kT}$).
A reasonable estimate of the average speed of molecules is the root mean square speed, defined as
$\displaystyle v_{rms}\equiv\sqrt{\overline{v^{2}}}=\sqrt{\frac{3kT}{m}} \ \ \ \ \ (7)$
Example 1 The molecules in a gas at room temperature are actually moving pretty fast. For example, for a nitrogen molecule (nitrogen makes up around 4/5 of the air) at room temperature (293 K), we have
$\displaystyle m$ $\displaystyle =$ $\displaystyle 28.0134\mbox{ amu}=4.65\times10^{-26}\mbox{ kg}\ \ \ \ \ (8)$ $\displaystyle k$ $\displaystyle =$ $\displaystyle 1.38\times10^{-23}\mbox{ m}^{2}\mbox{kg s}^{-2}\mbox{K}^{-1}\ \ \ \ \ (9)$ $\displaystyle v_{rms}$ $\displaystyle =$ $\displaystyle 510.7\mbox{ m s}^{-1} \ \ \ \ \ (10)$
Example 2 Consider a gas containing hydrogen and oxygen molecules in thermal equilibrium. The ratio of their rms speeds is
$\displaystyle \frac{v_{H}}{v_{O}}=\sqrt{\frac{m_{O}}{m_{H}}}=\sqrt{\frac{31.9988}{2.016}}=3.984 \ \ \ \ \ (11)$
where the masses are in amu. The hydrogen molecules are moving about 4 times faster than the oxygen molecules.
Example 3 To separate the two naturally occurring isotopes of uranium ${^{235}\mbox{U}}$ and ${^{238}\mbox{U}}$, the uranium is combined with fluorine to make uranium hexafluoride gas ${\mbox{UF}_{6}}$. The two isotopes will result in different rms speeds for the two types of molecules. We have
$\displaystyle m_{235}$ $\displaystyle =$ $\displaystyle 235.04+6\times18.998\mbox{ amu}=5.794\times10^{-25}\mbox{ kg}\ \ \ \ \ (12)$ $\displaystyle m_{238}$ $\displaystyle =$ $\displaystyle 238.02891+6\times18.998\mbox{ amu}=5.843\times10^{-25}\mbox{ kg}\ \ \ \ \ (13)$ $\displaystyle v_{235}$ $\displaystyle =$ $\displaystyle \sqrt{\frac{3k\times293}{m_{235}}}=144.692\mbox{ m s}^{-1}\ \ \ \ \ (14)$ $\displaystyle v_{238}$ $\displaystyle =$ $\displaystyle \sqrt{\frac{3k\times293}{m_{238}}}=144.084\mbox{ m s}^{-1} \ \ \ \ \ (15)$
Thus the speed difference is quite small.
## 8 thoughts on “Ideal gas: relation of average speed of molecules to temperature”
1. Pingback: Compression work | Physics pages
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 51, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9430686235427856, "perplexity": 167.12871658357884}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917125532.90/warc/CC-MAIN-20170423031205-00328-ip-10-145-167-34.ec2.internal.warc.gz"}
|
http://hamiltonianfunction.blogspot.com/2015/08/a-box-of-rocks.html?showComment=1441006623350
|
“Be kind, for everyone you meet is fighting a hard battle” - Often attributed to Plato but likely from Ian McLaren (pseudonym of Reverend John Watson)
## Sunday, August 30, 2015
### A box of rocks?
I've blogged on a few occasions regarding energy storage, most recently a "LightSail Energy redux" (though I'll be updating that in the near future to revise an inaccuracy with respect to LightSail's ability to produce commercially in their existing Berkeley facility). But there are lots of ideas for storage out there, from molten salt to bags of air beneath the sea and many others. But a new one caught my eye not too long ago that seemed out of cloud cuckoo land.
The firm touting this technology is Heindl Energy and their concept is to cut an annulus around rock and space beneath the rock, thereby creating a rock mass piston in a rock mass cylinder. Water would be pumped in to lift the rock when excess energy (or cheap energy if arbitrage is the name of the game) is available and let the rock descend, pumping the water through turbines when energy is needed or expensive.
The idea is that the piston diameter is equal to its length, and it would be raised and lowered a length equal to its radius, so half or more of the piston would always be beneath the ground surface.
One claimed advantage is that, unlike pumped hydro or underground compressed air energy storage, the geological feature necessary for the storage is more easily found (though, obviously, they can't be built just any old place).
Another is that the density of rock is greater than that of water and so a smaller volume of rock is needed for a given potential energy availability (though the factor is only about 2.5 or so).
It's easy to show that the available energy, ignoring efficiency losses of various kinds, is $E=(2\rho_{r}+\frac{3}{2}\rho_w)\pi gr^{4}$ where $\rho_r$ is rock density, $\rho_w$ is water density, $\pi$ is, well, $\pi$, $g$ is the acceleration of gravity, and $r$ is the radius of the rock piston. Note that the length of the piston is $2r$ and the height to which it's raised is $r$. Thus, the storage available scales with the fourth power of the radius. But, since construction is really all about the surface of the piston, construction time and cost scales approximately with the square of the radius. So, in this case, size truly matters! Doubling the radius gives approximately 16 times the capacity for "only" four times the construction cost and difficulty.
A bit of an issue is that the Heidl site "Idea & Function" page gives the energy as $(2\rho_r\frac{3}{2}\rho_w)\pi gr^{4}$. I'm sure it's a typo and I've emailed them to mention it but still, it doesn't lend confidence. Nevertheless, I used Wolfram Mathematica to check on the validity of the table shown on that page and it's actually conservative. They claim efficiency of 85% but I have to reduce efficiency to 57% or so to hit their numbers. Update: I received an email from Dr. Eduard Heindl recognizing the typo and stating that it has now been fixed. Dr. Heindl agreed that such an error on the technical page should not have taken place.
Unlike many storage scheme sites and descriptions, Heidl limits their discussion to quantity (gigawatt hours) and doesn't, as far as I could find, discuss power (the rate at which energy can be delivered by such a system). Both metrics are, of course, crucial and this site has the opposite of my typical frustration. They also provide no discussion of any load following capabilities.
We're talking here about a very large project. Using their numbers, 8 gigawatt hours of storage requires a 125 meter radius piston (250 meters or over 2 1/2 football fields across). Such a piston would weigh about 35.2 million (short) tons! To lift it would require water pressure of about 64 bar (though their table shows 52 bar).
Such pressures would demand a lot from the seals between the piston and cylinder walls, otherwise, the system would act as a 250 meter diameter circular fountain! Heidl discusses the seal system here and it appears to be quite innovative (a rolling seal against, apparently, a steel cylinder sleeve) but I see no indication that a pilot plant has confirmed that it will work. The devil, as always, is in the details.
Finally, how many? If a single 250 meter diameter rock piston can store 8 gigawatt hours, what is required to provide stable delivery from renewable sources so that the need for fossil fuel burning base load, spinning reserve, and peaker plants can be minimized with increasing penetration of intermittent renewable sources? According to Heidl's site, Germany currently needs 1,600 gigawatt hours of storage, of which only 40 have so far been provided. Therefore, Germany currently needs 195 such storage facilities!
In the Q & A, Heidl estimates that the system is of comparable cost to pumped hydro storage at a radius of 100 meters, and less expensive above that due to the scaling factors mentioned above. They also estimate a minimum of 2 years of planning and 3 to 4 years of construction per facility. And my experience is that the grander the scale of the project, the less likely it is to meet budget and schedule estimates. And this has never been tried.
While I think that it's a long shot that this type of system will ever see widespread use, the bigger picture is the scale of the undertaking needed to provide sufficient storage for deep grid scale intermittent penetration regardless of the system used. I think that distributed generation and local storage are key to providing a sustainable energy future through renewable sources.
Note: R.I.P. BB King.
#### 1 comment:
MrGantri said...
Man! I never understand what you're talking about BUT I always come away happily ignorant - lol
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6817173361778259, "perplexity": 1069.0919934379738}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463609817.29/warc/CC-MAIN-20170528120617-20170528140617-00321.warc.gz"}
|
http://www.quantwolf.com/doc/simplestrategies/appa.html
|
# Appendix AReview of Discrete Probability
This appendix provides a review of the essentials of discrete probability. It is concerned with observations, experiments or actions that have a finite number of unpredictable outcomes. The set of all possible outcomes is called the sample space (standard terminology) and is denoted by the symbol $$\Omega$$. An element of $$\Omega$$ (an individual outcome) will be denoted by $$\omega$$. A coin toss for example, has two possible outcomes: heads (H) or tails (T). The sample space is $$\Omega=\{H,T\}$$ and $$\omega=H$$ is one of the possible outcomes. Another example is the roll of a dice which has 6 outcomes so that $$\Omega=\{1,2,3,4,5,6\}$$. A subset of the sample space is called an event and is denoted by a capital letter such as $$A$$ or $$B$$. In the dice example, let $$A$$ be the event that an even number is rolled, then $$A=\{2,4,6\}$$.
Each outcome, $$\omega$$, will have a probability assigned to it, denoted $$P(\omega)$$. The probability is a real number ranging from $$0$$ to $$1$$ that signifies the likelihood that an outcome will occur. If $$P(\omega)=0$$ then $$\omega$$ will never occur and if $$P(\omega)=1$$ then $$\omega$$ will always occur. An intermediate value such as $$P(\omega)=1/2$$ means that $$\omega$$ will occur roughly half the time if the experiment is repeated many times. In general, if you perform the experiment a large number of times, $$N$$, and the number of times that $$\omega$$ occurs is $$n(\omega)$$, then the ratio $$n(\omega)/N$$ should approximately equal the probability of $$\omega$$. It is possible to define $$P(\omega)$$ as the limit of this ratio.
$$\tag{A.1} P(\omega) = \lim_{N \to \infty} \frac{n(\omega)}{N}$$
The function $$P(\omega)$$, which assigns probabilities to outcomes, is called a probability distribution. We will now look at some of its defining properties. To begin with, if the probabilities are defined as in equation A.1, then clearly the sum of all the probabilities must equal 1.
$$\tag{A.2} \sum_{\omega \in \Omega} P(\omega) = 1$$
It is often necessary to determine the probability that one of a subset of all the possible outcomes will occur. If $$A$$ is a subset of $$\Omega$$ then $$P(A)$$ is the probability that one of the outcomes contained in $$A$$ will occur. Using the definition in equation A.1 it should be obvious that:
$$\tag{A.3} P(A)=\sum_{\omega \in A} P(\omega)$$
Many other properties can be derived from the algebra of sets. Let $$A + B$$ be the set of all elements in either $$A$$ or $$B$$ (no duplicates) and let $$AB$$ be the the set of all elements in both $$A$$ and $$B$$, then:
$$\tag{A.4} P(A + B) = P(A) + P(B) - P(AB)$$
If $$A$$ and $$B$$ have no elements in common then they are exclusive events, i.e. they can not both occur simultaneously. In this case equation A.4 reduces to $$P(A + B) = P(A) + P(B)$$. In general, the probability that any one of a number of exclusive events will occur is just equal to the sum of their individual probabilities.
Conditional probabilities and the closely related concept of independence are very important and useful in probability calculations. Let $$P(A|B)$$ be the probability that $$A$$ has occurred given that we know $$B$$ has occurred. In short, we will refer to this as the probability of $$A$$ given $$B$$ or the probability of $$A$$ conditioned on $$B$$. What $$P(A|B)$$ really represents is the probability of $$A$$ using $$B$$ as the sample space instead of $$\Omega$$. If $$A$$ and $$B$$ have no elements in common then $$P(A|B)=0$$. If they have all elements in common so that $$A=B$$ then obviously $$P(A|B)=1$$. In general we have
$$\tag{A.5} P(A|B) = \frac{P(AB)}{P(B)}$$
Using a single fair dice roll as an example, let $$A=\{1,3\}$$ and $$B=\{3,5\}$$ then $$AB=\{3\}$$, $$P(AB)=1/6$$, $$P(B)=1/3$$ and
$$\tag{A.6} P(A|B) = \frac{1/6}{1/3} = \frac{1}{2}$$
Knowledge that $$B$$ has occurred has increased the probability of $$A$$ from $$P(A)=1/3$$ to $$P(A|B)=1/2$$. The result can also be deduced by simple logic. We know that $$B$$ has occurred therefore the roll was either a 3 or a 5. Half of the $$B$$ events are caused by a 3 and half by a 5 but only the 3 also counts as an $$A$$ event also, therefore $$P(A|B)=1/2$$.
Conditional probabilities are not necessarily symmetric. $$P(B|A)$$ need not be equal to $$P(A|B)$$. Using the definition in equation A.5, you can show that
$$\tag{A.7} P(A|B) P(B) = P(B|A) P(A)$$
so the two conditional probabilities are only equal if $$P(A)=P(B)$$. Another useful thing to keep in mind is that conditional probabilities obey the same properties as non-conditional probabilities. This means for example that if $$A$$ and $$B$$ are exclusive events then $$P(A+B|C) = P(A|C) + P(B|C)$$.
The concept of independence is naturally related to conditional probability. Two events are independent if the occurrence of one has no effect on the probability of the other. In terms of conditional probabilities this means that $$P(A|B)=P(A)$$. Independence is always symmetric, if $$A$$ is independent of $$B$$ then $$B$$ is independent of $$A$$. Using the definition in equation A.5 you can see that independence also implies that
$$\tag{A.8} P(AB) = P(A) P(B)$$
This is often taken as the defining relation for independence.
Another important concept in probability is the law of total probability. Let the sample space $$\Omega$$ be partitioned by the sets $$B_1$$ and $$B_2$$ so that every element in $$\Omega$$ is in one and only one of the two sets and we can write $$\Omega = B_1 + B_2$$. This means that the occurrence of $$A$$ coincides with the occurrence of $$B_1$$ or $$B_2$$ but not both and we can write
$$\tag{A.9} A = A B_1 + A B_2 = A (B_1 + B_2) = A \Omega$$
The probability of $$A$$ is then
$$\tag{A.10} P(A) = P(A B_1) + P(A B_2)$$
This can be extended to any number of sets that partition $$\Omega$$.
To carry out any kind of probabilistic analysis we need random variables. A random variable is a bit like the probability distributions discussed above in that it assigns a number to each of the elements in the sample space. It is therefore really more like a function that maps elements in the sample space to real numbers. A random variable is usually denoted with an upper case letter such as $$X$$ and the values it can assume are given subscripted lower case letters such as $$x_i$$ for $$i=1,2,\ldots,n$$ where $$n$$ is the number of possible values. The mapping from an element $$\omega$$ to a value $$x_i$$ is denoted as $$X(\omega)=x_i$$. Note that it is not necessary that every element be assigned a unique value and the particular value assigned will depend on what you want to analyze.
A simple example is a coin toss betting game. You guess what the result of the toss will be. If your guess is correct you win $1 otherwise you loose$1. The sample space consists of only two elements, a correct guess and an incorrect guess $$\Omega=\{\mathrm{correct},\mathrm{incorrect}\}$$. If you are interested in analyzing the amounts won and lost by playing several such games then the obvious choice for the random variable is $$X(\mathrm{correct})=1$$, $$X(\mathrm{incorrect})=-1$$. If you are just interested in the number of games won or lost then the random variable $$Y(\mathrm{correct})=1$$, $$Y(\mathrm{incorrect})=0$$ would be better. Often an analysis in terms of one variable can be converted into another variable by finding a relation between them. In the above example $$X = 2Y - 1$$ could be used to convert between the variables.
As another example consider tossing a coin three times. The sample space consists of 8 elements $$\Omega=\{TTT,TTH,THT,THH,HTT,HTH,HHT,HHH\}$$ where $$T$$ indicates the toss was a tail and $$H$$ a head. This time we let $$X$$ be the random variable that counts the number of heads in the three tosses. It can have values 0, 1, 2, or 3 and not every element in the sample space has a unique value. The values are $$X(TTT)=0$$, $$X(TTH)=X(THT)=X(HTT)=1$$, $$X(THH)=X(HTH)=X(HHT)=2$$, $$X(HHH)=3$$.
Probability distributions are most often expressed in terms of the values that a random variable can take. The usual notation is
$$\tag{A.11} P(X=x_i) = p(x_i)$$
The function $$p(x_i)$$ is the probability distribution for the random variable $$X$$. It is often also called the probability mass function. Note that it is not necessarily the same as the probability distribution for the individual elements of the sample space since multiple elements may be mapped to the same value by the random variable. In the three coin toss example, each element in the sample space has a probability of $$1/8$$, assuming a fair coin. The probability distribution for $$X$$ however is $$p(0)=1/8$$, $$p(1)=3/8$$, $$p(2)=3/8$$, $$p(3)=1/8$$. It will always be true that the sum over all the probabilities must equal 1.
$$\tag{A.12} \sum_i p(x_i) = 1$$
The two most important properties of a random variable are its expectation and variance. The expectation is simply the average value of the random variable. In the coin toss betting game, $$X$$ can have a value of +1 or -1 corresponding to winning or losing. In $$N$$ flips of the coin let $$k$$ be the number of wins and $$N-k$$ the number of losses. The total amount won is then
$$\tag{A.13} W = k - (N-k)$$
and the average amount won per flip is
$$\tag{A.14} \frac{W}{N} = \frac{k}{N} - (1-\frac{k}{N})$$
As the number of flips becomes very large the ratio $$k/N$$ will equal $$p(1)$$, the probability of winning, and the equation then becomes equal to expectation of the random variable.
$$\tag{A.15} E[X] = p(1) - p(-1)$$
Where $$p(-1)=1-p(1)$$ is the probability of losing and $$E[X]$$ is the usual notation for the expectation of $$X$$. In this case the expectation is the average amount that you can expect to win per flip if you play the game for a very long time.
In general if $$X$$ can take on $$n$$ values, $$x_i$$, $$i=1,2,\ldots,n$$ with corresponding probabilities $$p(x_i)$$ then the expectation is
$$\tag{A.16} E[X] = \sum_{i=1}^n p(x_i) x_i$$
The expectation gives you the average but in reality large deviations from the average may be possible. The variance of a random variable gives a sense for how large those deviations can be. It measures the average of the squares of the deviations. The equation for the variance is:
$$\tag{A.17} \mathrm{Var}[X] = \sum_{i=1}^n p(x_i) (x_i - E[X])^2$$
The equation simplifies somewhat to
$$\tag{A.18} \mathrm{Var}[X] = E[X^2] - E[X]^2$$
where
$$\tag{A.19} E[X^2] = \sum_{i=1}^n p(x_i) x_i^2$$
is the expectation for the square of the random variable. In general the expectation for any function $$g(X)$$ is:
$$\tag{A.20} E[g(X)] = \sum_{i=1}^n p(x_i) g(x_i)$$
Another useful measure of deviation from the average is called the standard deviation, $$\sigma$$. It is found by taking the square root of the variance.
$$\tag{A.21} \sigma = \sqrt{\mathrm{Var}[X]}$$
As we saw above, a sample space can have more than one random variable defined on it. If we have two variables $$X$$ and $$Y$$ then we can define the probability that $$X=x_i$$ at the same time that $$Y=y_j$$. This is called the joint probability distribution for $$X$$ and $$Y$$.
$$\tag{A.22} P(X=x_i,Y=y_j) = p(x_i,y_j)$$
The individual distributions, $$p(x_i)$$ and $$p(y_j)$$, are recovered by summing the joint distribution over one of the variables. To get $$p(x_i)$$ you sum $$p(x_i,y_j)$$ over all the possible values of $$Y$$.
$$\tag{A.23} p(x_i) = \sum_j p(x_i,y_j)$$
and likewise for $$p(y_j)$$
$$\tag{A.24} p(y_j) = \sum_i p(x_i,y_j)$$
From these last two equations it is obvious that if you sum over both variables of the distribution, the result should equal 1.
$$\tag{A.25} \sum_i \sum_j p(x_i,y_j) = 1$$
It is possible to construct a joint distribution for any number of random variables, not just 2. For example $$p(x_i,y_j,z_k)$$ would be a joint distribution for the variables $$X$$, $$Y$$, and $$Z$$.
With a joint distribution you can calculate the expectation and variance for functions of variables. The expectation for the sum $$X + Y$$ is:
\begin{eqnarray}\tag{A.26} E[X+Y] & = & \sum_i \sum_j p(x_i,y_j)(x_i + y_j)\\ & = & \sum_i x_i \sum_j p(x_i,y_j) + \sum_j y_j \sum_i p(x_i,y_j)\nonumber\\ & = & \sum_i x_i p(x_i) + \sum_j y_j p(y_j)\nonumber\\ & = & E[X] + E[Y]\nonumber \end{eqnarray}
The property that the expectation for a sum of variables is equal to the sum of their expectations is called linearity and it is true for the sum of any number of variables. For three variables for example $$E[X+Y+Z]=E[X]+E[Y]+E[Z]$$. Another easily verifiable consequence of linearity is that for any constants $$a$$ and $$b$$
$$\tag{A.27} E[aX+bY] = aE[X] + bE[Y]$$
In the example of the coin toss game we had two random variables that were related by $$X = 2Y -1$$. The linearity property of the expectation means that $$E[X] = 2E[Y] - 1$$, where we used the fact that the expectation of a constant is just the constant.
The expectation for the product $$XY$$ is
$$\tag{A.28} E[XY] = \sum_i \sum_j p(x_i,y_j) x_i y_j$$
If the variables $$X$$ and $$Y$$ are independent then the joint distribution can be factored into a product of the individual distributions, $$p(x_i,y_j) = p(x_i) p(y_j)$$. In this case you can show that the expectation of the product is the product of the expectations, $$E[XY] = E[X] E[Y]$$.
For the variance of a sum we have
\begin{eqnarray}\tag{A.29} \mathrm{Var}[X+Y] & & =\\ & & E[(X - E[X] + Y - E[Y])^2] \end{eqnarray}
after expanding and simplifying this becomes
\begin{eqnarray}\tag{A.30} \mathrm{Var}[X+Y] & & =\\ & & \mathrm{Var}[X] + \mathrm{Var}[Y] +\\ & & 2\mathrm{Cov}[X,Y] \end{eqnarray}
where $$\mathrm{Cov}[X,Y]$$ is called the covariance of $$X$$ and $$Y$$. The covariance is defined as:
$$\tag{A.31} \mathrm{Cov}[X,Y] = E[XY] - E[X]E[Y]$$
For independent variables the covariance is zero. The variance of the sum is then just the sum of the variances.
This completes the review of discrete probability. You do not need to understand everything in this review in order to understand the contents of this book. The more you do understand, the more likely you will be able to extend the concepts in this book to build even more powerful trading strategies.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 28, "x-ck12": 0, "texerror": 0, "math_score": 0.9723559021949768, "perplexity": 85.90405928005389}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267867364.94/warc/CC-MAIN-20180625014226-20180625034226-00467.warc.gz"}
|
https://eprints.utas.edu.au/14422/
|
# Observations on some Tasmanian fishes. Part V
Scott, EOG (1953) Observations on some Tasmanian fishes. Part V. Papers and Proceedings of the Royal Society of Tasmania, 87. pp. 141-166. ISSN 0080-4703
Preview
PDF
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9421254396438599, "perplexity": 24134.057782847867}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948520042.35/warc/CC-MAIN-20171212231544-20171213011544-00599.warc.gz"}
|
http://mathhelpforum.com/pre-calculus/42409-solve-equation-round-nearest-tenth-if-necessary.html
|
# Thread: Solve the equation. Round to the nearest tenth, if necessary.
1. ## Solve the equation. Round to the nearest tenth, if necessary.
If an object is projected upward with an initial velocity of 64 feet per second from a height of 336 feet, then its height t seconds after it is projected is defined by the expression h(t) = -16t2 + 64t + 336. How long after it is projected will it hit the ground?
THANK YOU!!!
2. Originally Posted by cechmanek32
If an object is projected upward with an initial velocity of 64 feet per second from a height of 336 feet, then its height t seconds after it is projected is defined by the expression h(t) = -16t2 + 64t + 336. How long after it is projected will it hit the ground?
THANK YOU!!!
Since h(t) is the height of the object we want h(t)=0 that is
$-16t^2+64t+336=0 \iff -16(t^2-4t-21)=0 \iff -16(t-7)(t+3)=0$
So t=7 or t=-3 since we are concerned with only future events we don't use t=-3. So the object hits the ground after 7 seconds.
Good luck.
3. Hello,
Originally Posted by cechmanek32
If an object is projected upward with an initial velocity of 64 feet per second from a height of 336 feet, then its height t seconds after it is projected is defined by the expression h(t) = -16t2 + 64t + 336. How long after it is projected will it hit the ground?
THANK YOU!!!
Find t such that $h(t)=0$.
$h(t)=-16t^2+64t+336=16(-t^2+4t+21)=-16(t-7)(t+3)$
$t=7$ or $t=-3$.
Because it's a nonsense to have a negative duration, the solution is 7seconds...
Edit :
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 5, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8480011224746704, "perplexity": 509.441149812993}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128320057.96/warc/CC-MAIN-20170623114917-20170623134917-00545.warc.gz"}
|
http://stackoverflow.com/questions/14780858/escape-in-script-tag-contents?answertab=oldest
|
# Escape </ in script tag contents
In HTML, tags and entities aren't parsed within <script> tags, and </ immediately ends the tag. Thus,
<script><b>fun & things</
will give you a script tag with the exact contents <b>fun & things.
If you're including JSON and you want to include the characters </ in your script, then you can replace it with <\/ because the only place for those characters to appear is in a string, and \/ is an escape sequence that turns into a single forward slash.
However, if you're not using JavaScript, then this trick doesn't work. In my case specifically I'm trying to insert a <script type="math/tex"> into the source so that MathJax will process it. Is there a way to escape </ in the original HTML source? (I don't have a particular need for </ but I'm writing a generic tool and want to make it possible to use any text.)
(It's possible to create the script tag in JavaScript and populate its innerText, but I'm working with the raw HTML so I can't do that.)
-
Can't you use <? It is not clear to me what you are trying to do here. – Oded Feb 8 at 20:33
So, basically, you're trying to encode any HTML between <script> and </script> tags without encoding the actual script tags themselves? – sbeliv01 Feb 8 at 20:37
@Oded: No, due to how script tags work, if you use < you'll actually get the four characters &, l, t, ;. – Ben Alpert Feb 8 at 21:28
Can you clarify the question and what you are trying to achieve? It is really difficult to understand what you want to do here and why it is important to have the sequence </. – Oded Feb 8 at 21:30
@Oded: Sorry, perhaps it's a bit clearer now. – Ben Alpert Feb 8 at 21:47
add comment
## 2 Answers
In HTML, as opposite to XHTML, the content of a script element is processed as plain text except for the occurrence of an end tag, so that </ ends processing and must, in conforming documents, start the end tag </script>. There is no general mechanism to avoid this. Any methods that circumvent this feature are unavoidably dependent on the “language” used inside the element. The word “language” is in quotes here, because the content can be just about anything, as long as your code can parse and process it.
So: no general mechanism, but for content other than JavaScript or some of the few other client-side scripting languages recognized by some browsers, you can make your own rules.
-
add comment
More HTML encoding might help? < for the <.
Difficult to know quite what you are doing with it. If you are not sure of what the content between the script tags might be (looks like you might be trying to use it as a template holder of some sort?) then you could/should use a CDATA section:
<script><![CDATA[<b>fun & things</b>]]></script>
That should do it. More description could help give a better answer too :)
-
add comment
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.31043604016304016, "perplexity": 1983.1535798798811}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-48/segments/1386164016462/warc/CC-MAIN-20131204133336-00032-ip-10-33-133-15.ec2.internal.warc.gz"}
|
https://www.mathdoubts.com/quadratic-equations/
|
An equation with one variable, which contains two as its highest power of variable, is called as quadratic equation.
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8887853026390076, "perplexity": 323.61110636529327}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-13/segments/1552912202704.58/warc/CC-MAIN-20190323000443-20190323022443-00531.warc.gz"}
|
http://nbviewer.jupyter.org/github/myuuuuun/MirandaFackler/blob/master/ddp_ex_MF_9_7_8_jl.ipynb
|
# Livestock Feeding¶
Akira Matsushita
Department of Economics, University of Tokyo
From Miranda and Fackler, Applied Computational Economics and Finance, 2002, Section 8.4.8 and 9.7.8
## 8.4.8 Model Formulation and Solution by hand¶
• A livestock producer feeds their stock for $T$ periods and sells it at the beginning of period $T+1$
• Each period $t$, the producer determine the amount of feed: $x_t \geq 0$. The fixed unit cost of grain is $\kappa \geq 0$
• $s_t$, the weight of livestock at time $t$, follows the deterministic process: $s_{t+1} = g(s_t, x_t), s_0 = \bar{s}$. Assume $s_t \geq 0$ for all $t$
• The selling price of the livestock at time $T+1$ is given by $s_{T+1} \times p$
Then the profit maximization problem for producer is formulated as:
\begin{align*} \underset{ \{x_t \geq 0 \}_{t=1}^T}{\max} \delta^T p s_{T+1} - \sum_{t=1}^{T} \delta^{t-1} \kappa x_t \hspace{1em} \text{s.t.} \hspace{1em} \begin{cases} s_{t+1} = g(s_t, x_t) \\ s_0 = \bar{s} \end{cases} \end{align*}
where $\delta \geq 0$ is the discount factor
The corresponding Bellman equation is:
\begin{align*} \begin{cases} V_t(s_t) = \underset{x_t \geq 0}{\max} \{ -\kappa x_t + \delta V_{t+1}(g(s_t, x_t)) \} \hspace{1em} t=1, \ldots, T \\ V_{T+1}(s_{T+1}) = ps_{T+1} \end{cases} \end{align*}
### Euler Conditions¶
Let's derive the Euler Equilibrium Conditions from the Bellman equation
Assume $g: \mathbb{R}_+ \times \mathbb{R}_+ \rightarrow \mathbb{R}_+$ is differentiable on the entire domain (so $g$ is partially differentiable by both $s$ and $x$)
Define $g_x = \frac{\partial g}{\partial x}$ and $g_s = \frac{\partial g}{\partial s}$
Assume $g_x(s, 0)$ is large enough so that the nonnegative constraint of $x$ will not be binded at an optimum (ensures an interior point solution)
Let $X_t^*$ be the optimal solution correspondence, i.e.,
\begin{align*} X_t^*(s) = \left\{ x_t^* \in [0, \infty) \mid V_t(s) = -\kappa x_t^* + \delta V_{t+1}(g(s_t, x_t^*)) \right\} \end{align*}
and $x_t^*(s) \in X_t^*(s)$ be a selection of it. Assume $x_t^*(s)$ is differentiable for all $t$
Note that by using $x_t^*(s)$ the optimal value function at time 1 is derived as
\begin{align*} V_1(\bar{s}) &= -\kappa x_1^*(\bar{s}) + \delta V_2(g(\bar{s}, x_1^*(\bar{s}))) \\ &= -\kappa \left[ x_1^*(\bar{s}) + \delta x_2^*(g(\bar{s}, x_1^*(\bar{s}))) \right] + \delta^2 V_3(g(g(\bar{s}, x_1^*(\bar{s})), x_2^*(g(\bar{s}, x_1^*(\bar{s}))))) \\ &= \ldots \end{align*}
Here we show that $V_t$ is differentiable by $s_t$ for all $t=1, 2, \ldots, T+1$
• At $t = T+1$, $V_{T+1}(s_{T+1}) = ps_{T+1}$ so this is differentiable by $s_{T+1}$
• As an induction hypothesis, assume $V_{t+1}$ is differentiable by $s_t$ ($1 \leq t \leq T$). Since
\begin{align*} V_t(s_t) = -\kappa x_t^*(s_t) + \delta V_{t+1}(g(s_t, x_t^*(s_t))) \end{align*}
and $x_t^*(s_t)$ and $V_{t+1}(s_t)$ are differentiable by $s_t$ and $g(s_t, x)$ is differentiable by both $s_t$ and $x$, $V_t(s_t)$ is also differentiable. The derivative is
\begin{align*} \frac{\partial}{\partial s_t} V_t(s_t) = -\kappa \frac{\partial}{\partial s_t} x_t^*(s) + \delta \frac{\partial}{\partial s_t}V_{t+1}(g(s_t, x_t^*(s_t))) \left\{ g_s(s_t, x_t^*(s_t)) + g_x(s_t, x_t^*(s_t)) \frac{\partial}{\partial s_t} x_t^*(s_t) \right\} \end{align*}
Hence by induction, $V_t$ is differentiable by $s_t$ for all $t$
Next consider the optimality condition of the maximization problem in the Bellman equation: $\underset{x_t \geq 0}{\max} \{ -\kappa x_t + \delta V_{t+1}(g(s_t, x_t)) \}$
Define $\lambda_t(s_t) \equiv V_{t}'(s_t) = \frac{\partial V_t}{\partial s_t}$
Using the chain rule, the FOCs are
\begin{align*} \delta \lambda_{t+1}(g(s_t, x_t^*))g_x(s_t, x_t^*) = \kappa \hspace{1em} t=1, \ldots, T \\ \end{align*}
Then the drivatives of the value functions by $s_t$: $\lambda_t(s_t) = \frac{\partial}{\partial s_t} V_t(s_t)$ become
\begin{align*} \lambda_t(s) &= \underbrace{ \bigl\{ -\kappa + \delta \lambda_{t+1}(g(s_t, x_t^*(s_t))) g_x(s_t, x_t^*(s_t)) \bigr\}}_{=0 \text{ by FOC}} \frac{\partial}{\partial s_t} x_t^*(s_t) + \delta \lambda_{t+1}(g(s_t, x_t^*(s_t))) g_s(s_t, x_t^*(s_t)) \\ &= \delta \lambda_{t+1}(g(s_t, x_t^*(s_t))) g_s(s_t, x_t^*(s_t)) \end{align*}
for $t=1, \ldots, T$ and
\begin{align*} \lambda_{T+1}(s_{T+1}) = p \end{align*}
In summary, the optimal path follows the following equations:
\begin{align*} \begin{cases} \delta \lambda_{t+1}(g(s_t, x_t))g_x(s_t, x_t) = \kappa \hspace{1em} t=1, \ldots, T \\ \lambda_t(s_t) = \delta \lambda_{t+1}(g(s_t, x_t)) g_s(s_t, x_t) \hspace{1em} t=1, \ldots, T \\ \lambda_{T+1}(s_{T+1}) = p \end{cases} \end{align*}
### Solve the equations by hand¶
How to solve the above equations and get the optimal polity $\{x_t\}_{t=1}^{T}$?
• At $t=T$, since $\lambda_{T+1}(s) = p$ regardless of the value of $s$, the 1st equation becomes $$g_x(s_T, x_T) = \frac{\kappa}{\delta p}$$
So we can get the optimal policy at $t=T$ by solving this equation for $x_T$. Then by the 2nd equation, we get $\lambda_{T}(s_T)$
• At $1 \leq t < T$, we can get $x_t$ and $\lambda_t(s_t)$ in the similar way
• Finally we set $s_1 = \bar{s}$ and then we get the concrete values of $\lambda_1(s_1^*), x_1^*, \lambda_2(s_2^*), x_2^*, \ldots$
### Concrete Example¶
Let
• $T = 6$
• $g(x, s) = \alpha s + \sqrt{x}$
• $p = 1$
Then
• $g_x(x, s) = \cfrac{0.5}{\sqrt{x}}$
• $g_s(x, s) = \alpha$
At $t=6$,
• $g_x(s_6, x_6) = \cfrac{0.5}{\sqrt{x_6}} = \cfrac{\kappa}{\delta} \hspace{2em} \therefore x_6(s_6) = \cfrac{\delta^2}{4 \kappa^2}$
• $\lambda_6(s_6) = \delta \alpha$
At $1 \leq t \leq 5$,
• $g_x(s_t, x_t) = \cfrac{0.5}{\sqrt{x_t}} = \cfrac{\kappa}{\delta^{7-t} \alpha^{6-t}} \hspace{2em} \therefore x_t(s_t) = \cfrac{(\delta^{7-t} \alpha^{6-t})^2}{4 \kappa^2}$
• $\lambda_t(s_t) = \delta^{7-t} \alpha^{7-t}$
So the optimal feeding policy is
\begin{align*} x_t^*(s_t) = \cfrac{(\delta^{7-t} \alpha^{6-t})^2}{4 \kappa^2} \hspace{1em} t=1, \ldots, 6 \end{align*}
Note that in this special case the optimal policy $x_1, \ldots, x_6$ does not depend on the initial weight $\bar{s}$
## 9.7.8 Solution by computation¶
In addition to the settings of the example above, assume
• $\alpha = 0.9$
• $\kappa = 0.4$
• $\delta = 0.9$
In order to calculate the value function $V_t(s_t) = \underset{x_t \geq 0}{\max} \{ -\kappa x_t + \delta V_{t+1}(g(s_t, x_t)) \}$ given $s_t$ computationally, we approximate $V_t(s_t)$ as
\begin{align*} V_t(s_t) \simeq \tilde{V_t}(s_t) = \sum_{i=1}^n c_{it} \phi_i(s_t) \end{align*}
where $\{ \phi_i \}_{i=1}^n$ are the predetermined basis functions and $\{ c_{it} \}_{i=1}^n$ are coefficients of them
Also we choose $n$ sample points $\xi_1 < \xi_2 < \ldots < \xi_n$ in the state space
The procedure of calculating optimal policy is as follows:
• At $t=T$
• Using optimizer, solve $V_T(\xi_j) = \underset{x_T \geq 0}{\max} \{ -\kappa x_T + \delta p g(\xi_j, x_T) \}$ for each $\xi_1, \ldots, \xi_n$ and get optimal $V_T(\xi_j)$ and $x_T(\xi_j)$
• Solve $n$ simultaneous linear equations
\begin{align*} \sum_{i=1}^n c_{iT} \phi_i(\xi_j) = V_T(\xi_j) \hspace{1em} j=1, \ldots, n \end{align*}
to determine the $n$ coefficients $\{ c_{iT} \}_{i=1}^n$ and get $\tilde{V_T}(s_T)$
• At $1 \leq t < T$, given $\tilde{V_{t+1}}(\xi_j)$,
• Using optimizer, solve $V_t(\xi_j) = \underset{x_t \geq 0}{\max} \{ -\kappa x_t + \delta \tilde{V_{t+1}}(g(\xi_j, x_t)) \}$ for each $\xi_1, \ldots, \xi_n$ and get optimal $V_t(\xi_j)$ and $x_t(\xi_j)$
• Solve $n$ simultaneous linear equations
\begin{align*} \sum_{i=1}^n c_{it} \phi_i(\xi_j) = V_t(\xi_j) \hspace{1em} j=1, \ldots, n \end{align*}
to determine the $n$ coefficients $\{ c_{it} \}_{i=1}^n$ and get $\tilde{V_t}(s_t)$
• Finally set $s_1 = \bar{s}$ and then we can get all optimal policies
In [1]:
using QuantEcon
using Optim
using BasisMatrices
using Plots
plotlyjs()
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9997541308403015, "perplexity": 3004.0905054536393}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257647584.56/warc/CC-MAIN-20180321063114-20180321083114-00523.warc.gz"}
|
https://stats.stackexchange.com/questions/315836/does-statistical-power-matter-if-we-are-not-interested-in-nhst
|
# Does statistical power matter if we are not interested in NHST?
If we don't care about the probability of finding a real effect if there is one, AKA a statistically significant finding, why care about statistical power and power calculations?
I mean there are obviously benefits to having larger samples such as decreasing our standard errors and increasing precision etc, but would power analyses matter if we don't care about the p-values? What if we were to just focus on effect sizes and other descriptive statistics?
• Power is the probability of rejecting the null hypothesis given some particular alternative holds. If you're not testing, you're can't be rejecting hypotheses, so what do you then mean by power at all? Can you clarify what you're asking (and you should indicate where your premise comes from -- why you think anyone does care. Maybe they do, but what leads you to think they do?) – Glen_b Nov 27 '17 at 9:00
• AKA a Statisticcally significant finding - please elaborate. just to understand what is meant by p- value ? – Subhash C. Davar Nov 27 '17 at 11:03
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8399702310562134, "perplexity": 726.1101275408582}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243991258.68/warc/CC-MAIN-20210517150020-20210517180020-00056.warc.gz"}
|
http://www.objectivebooks.com/2017/12/network-theorems-electrical-engineering.html
|
# Practice Test: Question Set - 02
1. Between the branch voltages of a loop the Kirchhoff s voltage law imposes
(A) Nonlinear constraints
(B) Linear constraints
(C) No constraints
(D) None of the above
2. An ideal voltage source has
(A) Zero internal resistance
(B) Open circuit voltage equal to the voltage on full load
(C) Terminal voltage in proportion to current
(D) Terminal voltage in proportion to load
3. A closed path made by several branches of the network is known as
(A) Branch
(B) Loop
(C) Circuit
(D) Junction
4. The superposition theorem requires as many circuits to be solved as there are
(A) Sources, nodes and meshes
(B) Sources and nodes
(C) Sources
(D) Nodes
5. Kirchhoff s current law states that
(A) Net current flow at the junction is positive
(B) Hebraic sum of the currents meeting at the junction is zero
(C) No current can leave the junction without some current entering it
(D) Total sum of currents meeting at the junction is zero
6. An ideal voltage source should have
(A) Large value of e.m.f.
(B) Small value of e.m.f.
(C) Zero source resistance
(D) Infinite source resistance
7. “In any linear bilateral network, if a source of e.m.f. E in any branch produces a current I in any other branch, then same e.m.f. acting in the second branch would produce the same current / in the first branch”. The above statement is associated with
(A) Compensation theorem
(B) Superposition theorem
(C) Reciprocity theorem
(D) None of the above
8. The resistance LM will be
(A) 6.66 Q
(B) 12 Q
(C) 18 Q
(D) 20 Q
9. A nonlinear network does not satisfy
(A) Superposition condition
(B) Homogeneity condition
(C) Both homogeneity as well as superposition condition
(D) Homogeneity, superposition and associative condition
10. A terminal where three on more branches meet is known as
(A) Node
(B) Terminus
(C) Combination
(D) Anode
Show and hide multiple DIV using JavaScript View All Answers
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8445118069648743, "perplexity": 1925.9203544112258}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084891105.83/warc/CC-MAIN-20180122054202-20180122074202-00410.warc.gz"}
|
https://learnzillion.com/lesson_plans/8274-add-positive-and-negative-numbers-using-a-number-line
|
Instructional video
Add positive and negative numbers using a number line
teaches Common Core State Standards CCSS.Math.Content.7.NS.A.1b http://corestandards.org/Math/Content/7/NS/A/1/b
You have saved this instructional video!
Here's where you can access your saved items.
Dismiss
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8911616206169128, "perplexity": 26685.527881722406}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039741660.40/warc/CC-MAIN-20181114062005-20181114084005-00085.warc.gz"}
|
https://tex.stackexchange.com/questions/332962/rounded-box-around-heading
|
I was wondering if anyone would be willing to provide a code to replicate the attached image in TeX. It is simply a blue text box with rounded edges and a dotted (rather than dashed) line between the heading (I don't mind which font this is in) and a paragraph of extra information. The width of the box should be the same as the width of each line of text and currently I'm working in a document set on A4 Paper with left and right margins both at -0.5 cm, the top and bottom margins being 2 cm.
• Welcome to TeX SX! You should take a look at the tcolorbox package documentation. – Bernard Oct 6 '16 at 20:32
\documentclass{book}
\usepackage{xcolor}
\usepackage{tcolorbox}
\begin{document}
\begin{tcolorbox}[colback=white, colframe=blue]
bla
\tcblower
blub
\end{tcolorbox}
\end{document}
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9738517999649048, "perplexity": 487.8420314500397}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627999876.81/warc/CC-MAIN-20190625172832-20190625194832-00486.warc.gz"}
|
http://stackoverflow.com/questions/3592851/executing-a-command-stored-in-a-variable-from-powershell
|
Executing a Command stored in a Variable from Powershell
I have a command that i have build and stored in a variable in power shell. This command works if i do a write-host and copy and past into a standard cmd.exe window.
How do i execute this command from inside my script?
I have tried several combination of invoke-command or invoke-expression with no luck.
This is how i built the variable:
``````\$cmd1= \$arcprg + \$arcdir + "\" + \$site1+"-"+\$hst+"-"+\$yesterday+".zip "+\$logpath1+"u_ex"+\$yesterday+".log"
``````
This is what the variable looks like if it is printed to the screen:
``````7z.exe a -tzip c:\arc_logs\site-host-at-web1-100827.zip c:\inetpub\logs\logfiles\w3svc1\u_ex100827.log
``````
-
Here is yet another way without `Invoke-Expression` but with two variables (command:string and parameters:array). It works fine for me. Assume `7z.exe` is in the system path.
``````\$cmd = '7z.exe'
\$prm = 'a', '-tzip', 'c:\temp\with space\test1.zip', 'C:\TEMP\with space\changelog'
& \$cmd \$prm
``````
If the command is known (7z.exe) and only parameters are variable then this will do:
``````\$prm = 'a', '-tzip', 'c:\temp\with space\test1.zip', 'C:\TEMP\with space\changelog'
& 7z.exe \$prm
``````
BTW, `Invoke-Expression` with one parameter works for me, too, e.g. this works:
``````\$cmd = '& 7z.exe a -tzip "c:\temp\with space\test2.zip" "C:\TEMP\with space\changelog"'
Invoke-Expression \$cmd
``````
P.S. I usually prefer the way with a parameter array because it is easier to compose programmatically than to build an expression for `Invoke-Expression`.
-
That is great. It is starting to come together now. Instead of 'c:\temp\with space\test1.zip' can i just use a \$variable ? Do i need to place it is "" or '' ? – Travis Aug 30 '10 at 13:50
Try invoking your command with `invoke-expression`.
``````invoke-expression \$cmd1
``````
Here is a working example on my machine:
``````\$cmd = "& 'C:\Program Files\7-zip\7z.exe' a -tzip c:\temp\test.zip c:\temp\test.txt"
invoke-expression \$cmd
``````
-
Tells me The term '7z.exe a -tzip c:\arc_logs\site-host-at-web1-100827.zip c:\inetpub\logs\logfiles\w3svc1\u_ex100827.log' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At :line:14 char:1 + & <<<< \$cmd1 – Travis Aug 28 '10 at 23:29
@Travis: Oops. the ampersand is works if there are no args. I updated the post with a solution for your command. – kbrimington Aug 28 '10 at 23:32
I have tried invoke-expression before and it has not worked. It spits out the error: Bad numeric constant: 7. At :line:1 char:2 + 7z <<<< .exe a -tzip c:\arc_logs\site-host-at-web1-100827.zip c:\inetpub\logs\logfiles\w3svc1\u_ex100827.log It almost seems like it is trying to evaluate it instead of executing it. – Travis Aug 28 '10 at 23:47
@Travis: It may be that 7z.exe is not on your path. Verify that it is on the path and/or try giving the full path name to the executable in your expression. – kbrimington Aug 28 '10 at 23:59
@Travis: I just confirmed on my own system that 7z.exe gave the error you described, just typing from the command prompt, but with the full path to 7z.exe (for me, it was `'C:\Program Files\7-zip\7z.exe'`, I could execute 7z.exe. – kbrimington Aug 29 '10 at 0:04
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8332616686820984, "perplexity": 5413.966610902445}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-15/segments/1398223203841.5/warc/CC-MAIN-20140423032003-00355-ip-10-147-4-33.ec2.internal.warc.gz"}
|
https://gmatclub.com/forum/what-is-the-greatest-prime-factor-of-70126.html
|
GMAT Question of the Day - Daily to your Mailbox; hard ones only
It is currently 15 Nov 2018, 23:41
### GMAT Club Daily Prep
#### Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email.
Customized
for You
we will pick new questions that match your level based on your Timer History
Track
every week, we’ll send you an estimated GMAT score based on your performance
Practice
Pays
we will pick new questions that match your level based on your Timer History
## Events & Promotions
###### Events & Promotions in November
PrevNext
SuMoTuWeThFrSa
28293031123
45678910
11121314151617
18192021222324
2526272829301
Open Detailed Calendar
• ### Free GMAT Strategy Webinar
November 17, 2018
November 17, 2018
07:00 AM PST
09:00 AM PST
Nov. 17, 7 AM PST. Aiming to score 760+? Attend this FREE session to learn how to Define your GMAT Strategy, Create your Study Plan and Master the Core Skills to excel on the GMAT.
• ### GMATbuster's Weekly GMAT Quant Quiz # 9
November 17, 2018
November 17, 2018
09:00 AM PST
11:00 AM PST
Join the Quiz Saturday November 17th, 9 AM PST. The Quiz will last approximately 2 hours. Make sure you are on time or you will be at a disadvantage.
# What is the greatest prime factor of 2^100 - 2^96?
Author Message
TAGS:
### Hide Tags
Manager
Joined: 28 Aug 2008
Posts: 93
What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
11 Sep 2008, 13:11
3
8
00:00
Difficulty:
15% (low)
Question Stats:
72% (00:27) correct 28% (00:35) wrong based on 620 sessions
### HideShow timer Statistics
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
SVP
Joined: 07 Nov 2007
Posts: 1682
Location: New York
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
11 Sep 2008, 13:28
1
2
IgnitedMind wrote:
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
2^100 - 2^96 = 2^4*2^96 -2^96
= 2^96 (16-1) = 2^96 *5*3
5 is the greatest factor
C
_________________
Smiling wins more friends than frowning
Manager
Joined: 09 Jul 2007
Posts: 215
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
11 Sep 2008, 16:25
1
2^100 - 2^96=2^96 ( 2^4-1)=2^96*3*5
so greatest prime factor =5
VP
Joined: 17 Jun 2008
Posts: 1225
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
11 Sep 2008, 18:10
IgnitedMind wrote:
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
IMO C
2^96(16-1)=2^96 * 3*5 => 5 is the greatest prime factor
_________________
cheers
Its Now Or Never
Senior Manager
Joined: 21 Apr 2008
Posts: 261
Location: Motortown
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
11 Sep 2008, 18:14
C
Same explanation as others :
2^96(2^4 -1) = 2^96 * 3*5
Senior Manager
Joined: 05 May 2011
Posts: 354
Location: United States (WI)
GMAT 1: 780 Q49 V50
WE: Research (Other)
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
04 Aug 2011, 05:51
C as well. Factor out 2^96
Director
Joined: 01 Feb 2011
Posts: 664
Re: PS Greatest Prime Factors [#permalink]
### Show Tags
04 Aug 2011, 13:05
=2^96(15) = 2^96*3*5
Greatest prime factor = 5.
Intern
Joined: 14 Jul 2012
Posts: 1
Re: What is the greatest prime factor of 2^100 - 2^96? A. 2 B. [#permalink]
### Show Tags
21 Jan 2013, 09:20
Can someone please explain how we can go from 2^4 * 2^96 - 2^96 to 2^96 (16-1) ?
Thanks!
Math Expert
Joined: 02 Sep 2009
Posts: 50615
Re: What is the greatest prime factor of 2^100 - 2^96? A. 2 B. [#permalink]
### Show Tags
21 Jan 2013, 09:47
2
1
ebliss wrote:
Can someone please explain how we can go from 2^4 * 2^96 - 2^96 to 2^96 (16-1) ?
Thanks!
Sure.
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
$$2^{100} - 2^{96}=2^4*2^{96}-2^{96}$$.
Now, factor out 2^{96}: $$2^4*2^{96}-2^{96}=2^{96}(2^4-1)=2^{96}*15=2^{96}*3*5$$. The greatest prime factor is 5.
Hope it's clear.
_________________
Manager
Joined: 04 Jan 2013
Posts: 71
Re: What is the greatest prime factor of 2^100 - 2^96? A. 2 B. [#permalink]
### Show Tags
21 Jan 2013, 10:33
Bunuel wrote:
ebliss wrote:
Can someone please explain how we can go from 2^4 * 2^96 - 2^96 to 2^96 (16-1) ?
Thanks!
Sure.
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
$$2^{100} - 2^{96}=2^4*2^{96}-2^{96}$$.
Now, factor out 2^{96}: $$2^4*2^{96}-2^{96}=2^{96}(2^4-1)=2^{96}*15=2^{96}*3*5$$. The greatest prime factor is 5.
Hope it's clear.
@bunuel..2^96-2^96=2^1 according to index laws..in our factorised equation 2^4*2^96-2^96 the answer is 16*2=32..can we do prime factorisation on 32 where by k/2 +k/3 +k/5--->32/2+32/3+32/5?would we just take the 5 as our largest factor?but we know 5 is not a factor of 32..where am i missing the details please tell me..thanks in advance
Posted from my mobile device
Math Expert
Joined: 02 Sep 2009
Posts: 50615
Re: What is the greatest prime factor of 2^100 - 2^96? A. 2 B. [#permalink]
### Show Tags
21 Jan 2013, 10:40
1
chiccufrazer1 wrote:
Bunuel wrote:
ebliss wrote:
Can someone please explain how we can go from 2^4 * 2^96 - 2^96 to 2^96 (16-1) ?
Thanks!
Sure.
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
$$2^{100} - 2^{96}=2^4*2^{96}-2^{96}$$.
Now, factor out 2^{96}: $$2^4*2^{96}-2^{96}=2^{96}(2^4-1)=2^{96}*15=2^{96}*3*5$$. The greatest prime factor is 5.
Hope it's clear.
@bunuel..2^96-2^96=2^1 according to index laws..in our factorised equation 2^4*2^96-2^96 the answer is 16*2=32..can we do prime factorisation on 32 where by k/2 +k/3 +k/5--->32/2+32/3+32/5?would we just take the 5 as our largest factor?but we know 5 is not a factor of 32..where am i missing the details please tell me..thanks in advance
Posted from my mobile device
Not sure I understand your post.
First of all, 2^96-2^96=0 not 2.
Next, $$2^4*2^{96}-2^{96}$$ equals to $$2^{96}*3*5$$ not 32.
_________________
Current Student
Joined: 12 Aug 2015
Posts: 2633
Schools: Boston U '20 (M)
GRE 1: Q169 V154
Re: What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
14 Mar 2016, 01:16
Manager
Joined: 09 Jun 2015
Posts: 94
Re: What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
14 Mar 2016, 04:05
IgnitedMind wrote:
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
It can be written as 2^96(2^4-1), which has prime factors as 2, 3, and 5.
Current Student
Joined: 12 Aug 2015
Posts: 2633
Schools: Boston U '20 (M)
GRE 1: Q169 V154
Re: What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
16 Mar 2016, 01:37
Manager
Joined: 09 Jun 2015
Posts: 94
Re: What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
16 Mar 2016, 01:44
IgnitedMind wrote:
What is the greatest prime factor of 2^100 - 2^96?
A. 2
B. 3
C. 5
D. 7
E. 11
The given expression can be written as 2^96(2^4 -1) = 2^96 * 3*5
Non-Human User
Joined: 09 Sep 2013
Posts: 8779
Re: What is the greatest prime factor of 2^100 - 2^96? [#permalink]
### Show Tags
23 Oct 2018, 07:02
Hello from the GMAT Club BumpBot!
Thanks to another GMAT Club member, I have just discovered this valuable topic, yet it had no discussion for over a year. I am now bumping it up - doing my job. I think you may find it valuable (esp those replies with Kudos).
Want to see all other topics I dig out? Follow me (click follow button on profile). You will receive a summary of all topics I bump in your profile area as well as via email.
_________________
Re: What is the greatest prime factor of 2^100 - 2^96? &nbs [#permalink] 23 Oct 2018, 07:02
Display posts from previous: Sort by
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6373715996742249, "perplexity": 10587.258674486806}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039742981.53/warc/CC-MAIN-20181116070420-20181116092420-00374.warc.gz"}
|
https://www.physicsforums.com/threads/eigenvalues-of-two-matrices-are-equal.711169/
|
# Eigenvalues of two matrices are equal
1. Sep 18, 2013
### gopi9
Hi everyone,
I have two matrices A and B,
A=[0 0 1 0; 0 0 0 1; a b a b; c d c d] and B=[0 0 0 0; 0 0 0 0; 0 0 a b; 0 0 c d].
I have to proves theoretically that two of the eigenvalues of A and B are equal and remaining two eigenvalues of A are 1,1.
I tried it by calculating the determinant of A and B and I got close to the result but I am not able to prove it completely.
I got result like this,
sum of roots of determinant of A and B as
p+q+r+s=p1+q1 (p,q,r,s are roots of det of A, p1,q1 are roots of det of B)
Product of roots
p*q*r*s=p1*q1
pqr+qrs+prs+pqs=-2(p1*q1).
Thanks.
2. Sep 18, 2013
### AlephZero
Find the characteristic polynomials of A and B.
Then factorize A - the question tells you two of the factors.
3. Sep 18, 2013
### gopi9
The characteristic equations that I got for A is
and
for B
I cant factorize A polynomial equation, since it does not have simple 1 or -1 as roots.
4. Sep 18, 2013
### AlephZero
The question says two roots of the A polynomial are equal to 1. So if (p-1)^2 isn't a factor, either you made a mistake somewhere, or the question is wrong.
I agree with you that p-1 us not a factor of the A polynomial, so I think the question in your OP is wrong. Are you missing some minus signs in the A matrix?
5. Sep 18, 2013
### gopi9
Matlab gives -1 as an eigenvalue but theoretically i cant prove it. There is no mistake in the theoretical proof, i checked it many times. The signs in A matrix are also correct.
6. Sep 18, 2013
### gopi9
This is an example of A matrix that I have
0 0 1 0
0 0 0 1
-400000 200000 -400000 200000
66666.67 -133333.33 66666.67 -133333.33
I took a=-400000, b=200000, c=66666.67, d= -133333.33
7. Sep 19, 2013
### AlephZero
Take the simpler example of a = b = c = d = 0.
There is obviously something wrong with the question here.
8. Oct 5, 2013
### brmath
Switching around the rows you have A' = $\begin{pmatrix} a & b & a & b\\ c & d & c & d\\ 0 & 0 & 1 & 0\\ 0 & 0 & 0 & 1\\ \end{pmatrix}$ and B = $\begin{pmatrix} 0 & 0 & 0 & 0\\ 0 &0 & 0 & 0\\ 0 & 0 & a & b\\ 0 & 0 & c & d\\ \end{pmatrix}$
Clearly A' has two eigenvalues of 1 -- they are sitting right there on the diagonal; Since A' was obtained by switching each row and even number of times, the eigenvalues of A' are those of A. Clearly also B has two eigenvalues of 0. What are the other eigenvalues of B? If b = c = 0 then they are a and d. If b and c are 0 a and d will also be the eigenvalues of A. You will want to show this, but the computation should be easy with all those zeros in it.
However, b and c don't have to be 0. My guess would be that the 2 eigenvalues of $\begin{pmatrix} a & b\\ c & d\\ \end{pmatrix}$ will also be the other two eigenvalues of A.
Can you show that?
Last edited: Oct 5, 2013
9. Dec 20, 2013
### gopi9
In my case a,b,c,d are not zeros alephzero. Thanks for the reply
10. Dec 20, 2013
### gopi9
Thanks brmath. That helps
11. Jan 13, 2014
### gopi9
Can we obtain relation between eigenvectors of A and B matrices
12. Jan 13, 2014
### brmath
Will try to get back to you later today.
13. Jan 13, 2014
### Office_Shredder
Staff Emeritus
Take brmath's re-arranged matrix A' and consider vectors of the form (x,y,0,0).
14. Jan 13, 2014
### gopi9
Thanks for the reply.I did not understand what u meant by consider vectors of the form (x,y,0,0). I already tried using A' matrix to solve it but could not go any further.
15. Jan 13, 2014
### gopi9
For the example that I took A has eigenvectors
[2.206e-6 6.008e-6 0.6912 -0.3835;
-4.749e-7 9.304e-6 -0.1487 -0.5940;
-0.9776 -0.5424 -0.6912 0.3835;
0.2104 -0.84007 0.1487 0.5940]
and B has
[0 0 1 0;
0 0 0 1;
-0.9776 -0.5424 0 0;
0.2104 -0.84007 0 0].
Eigenvectors of [a b; c d] is
[-0.9776 -0.5424;
0.21043 -0.84007]
16. Jan 13, 2014
### brmath
The eigenvectors present a different kind of problem, and there are a number of different possibilities.
I suggest you start by finding the eigenvectors of B which match the 0 eigenvalues: i.e. Bx = 0. You will get either one or two different x's. I suspect just one.
With the A there are numerous possibilities, which depend on the a, b, c, and d. For example if a = c = 1 and b = d = 0, A will probably have four independent eigenvectors all corresponding to the single eigenvalue 1. If you have a = b = d = 1 and c = 0, A will probably have 3 eigenvectors corresponding to the eigenvalue 1. You will have to work this out.
In either of these cases, B will also have eigenvectors corresponding to 1 - -either two independent ones, or just one.
Whether any of these eigenvectors match up between A and B is something you will have to compute. That is, find the eigenvectors of A which correspond to 1 under the two a,b,c,d scenarios I suggested, and find the eigenvectors of B for those same 1's.
Offhand I see no particular reason to believe they are the same or different -- you'll have to see.
Now the x's that match with the zero eigenvalues of B might or might not be eigenvalues of A. It could be that they are for some values of a,b,c,d and likely not for others. But you should check by multiplying the x's by A.
Once you've gotten through all that, you may have a clue as to whether anything matches up for other values of a,b,c,d.
17. Jan 13, 2014
### Office_Shredder
Staff Emeritus
Calculate A'(x,y,0,0)t and you should notice that it looks a lot like a 2x2 matrix operating on a two dimensional vector.
18. Jan 14, 2014
### brmath
It does, but I think the a,b,c,d create a lot of complications.
19. Jan 20, 2014
### wangchong
A and B are block matrices. So are xI-A and xI-B. Does that give you a way to find their characteristic polynomials? If I have to guess:
charpoly(B,x) = x^2((x-a)(x-d) -bc) and charpoly(A,x)=(x-1)(x-1)((x-a)(x-d) -bc)
so both charpoly(A) and charpoly(B) have common (quadratic) factors ((x-a)(x-d) -bc) .
Similar Discussions: Eigenvalues of two matrices are equal
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8636042475700378, "perplexity": 847.279228080153}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886109157.57/warc/CC-MAIN-20170821152953-20170821172953-00180.warc.gz"}
|
https://ajayshahblog.blogspot.com/2013/06/author-suyash-rai.html
|
## Monday, June 10, 2013
### Author: Suyash Rai
Suyash Rai is a researcher at the National Institute for Public Finance and Policy.
Please note: LaTeX mathematics works. This means that if you want to say $10 you have to say \$10.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2765446603298187, "perplexity": 5739.723047282631}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-04/segments/1484560282935.68/warc/CC-MAIN-20170116095122-00196-ip-10-171-10-70.ec2.internal.warc.gz"}
|
https://worldwidescience.org/topicpages/b/banana+bunchy+top.html
|
Sample records for banana bunchy top
1. Certain growth related attributes of bunchy top virus infected banana ...
African Journals Online (AJOL)
Effect of banana bunchy top virus (BBTV) on morpho-physiological characteristics of banana (Musa sp.) cv., Basrai plants was assessed. Healthy and BBTV infected samples of banana were collected from its open fields and micro-propagated aseptically. These plantlets were established in wire-house for three months.
2. Relative susceptibility of Musa genotypes to banana bunchy top disease in Cameroon and implication for disease management
Science.gov (United States)
Banana bunchy top disease (BBTD) is a serious threat to banana and plantain (Musa spp.) production. BBTD is caused by the Banana bunchy top virus (BBTV, genus Babuvirus) which is spread through infected plant propagules and banana aphid, Pentalonia nigronervosa. A high level of resistance to BBTD in...
3. Molecular Characterization of Geographically Different Banana bunchy top virus Isolates in India.
Science.gov (United States)
Selvarajan, R; Mary Sheeba, M; Balasubramanian, V; Rajmohan, R; Dhevi, N Lakshmi; Sasireka, T
2010-10-01
Banana bunchy top disease (BBTD) caused by Banana bunchy top virus (BBTV) is one of the most devastating diseases of banana and poses a serious threat for cultivars like Hill Banana (Syn: Virupakshi) and Grand Naine in India. In this study, we have cloned and sequenced the complete genome comprised of six DNA components of BBTV infecting Hill Banana grown in lower Pulney hills, Tamil Nadu State, India. The complete genome sequence of this hill banana isolate showed high degree of similarity with the corresponding sequences of BBTV isolates originating from Lucknow, Uttar Pradesh State, India, and from Fiji, Egypt, Pakistan, and Australia. In addition, sixteen coat protein (CP) and thirteen replicase genes (Rep) sequences of BBTV isolates collected from different banana growing states of India were cloned and sequenced. The replicase sequences of 13 isolates showed high degree of similarity with that of South Pacific group of BBTV isolates. However, the CP gene of BBTV isolates from Shervroy and Kodaikanal hills of Tamil Nadu showed higher amino acid sequence variability compared to other isolates. Another hill banana isolate from Meghalaya state had 23 nucleotide substitutions in the CP gene but the amino acid sequence was conserved. This is the first report of the characterization of a complete genome of BBTV occurring in the high altitudes of India. Our study revealed that the Indian BBTV isolates with distinct geographical origins belongs to the South Pacific group, except Shervroy and Kodaikanal hill isolates which neither belong to the South Pacific nor the Asian group.
4. The antagonistic effect of Banana bunchy top virus multifunctional protein B4 against Fusarium oxysporum.
Science.gov (United States)
Zhuang, Jun; Coates, Christopher J; Mao, Qianzhuo; Wu, Zujian; Xie, Lianhui
2016-06-01
The viral-induced banana bunchy top disease and the fungal-induced banana blight are two major causes of concern for industrial scale production of bananas. Banana blight is particularly troublesome, affecting ∼80% of crops worldwide. Strict guidelines and protocols are in place in order to ameliorate the effects of this devastating disease, yet little success has been achieved. From the data presented here, we have found that Banana bunchy top virus (BBTV)-infected bananas are more resistant to Fusarium oxysporum f. sp. cubense (Foc). BBTV appears to be antagonistic towards Foc, thus improving the survivability of plants against blight. The BBTV suppressor of RNA silencing, namely protein B4, displays fungicidal properties in vitro. Furthermore, transgenic tomatoes expressing green fluorescent protein (GFP)-tagged protein B4 demonstrate enhanced resistance to F. oxysporum f. sp. lycopersici (Fol). Differential gene expression analysis indicates that increased numbers of photogenesis-related gene transcripts are present in dark-green leaves of B4-GFP-modified tomato plants relative to those found in WT plants. Conversely, the transcript abundance of immunity-related genes is substantially lower in transgenic tomatoes compared with WT plants, suggesting that plant defences may be influenced by protein B4. This viral-fungal interaction provides new insights into microbial community dynamics within a single host and has potential commercial value for the breeding of transgenic resistance to Fusarium-related blight/wilt. © 2016 BSPP AND JOHN WILEY & SONS LTD.
5. Elucidating the resistance response of irradiated banana cv. Lakatan to banana bunchy top virus (BBTV) infection transmitted by the banana aphid Pentalonia nigronervosa Coquerol
International Nuclear Information System (INIS)
Dela Cueva, F.M.; Sison, M.L.J.; Dinglasan, E.G.; Damasco, O.P.
2014-01-01
Development of banana bunchy top virus (BBTV)-resistant banana variety Lakatan through gamma-irradiation had been successfully done as part of integrated management strategies against the disease. Ten irradiated Lakatan lines exhibited resistance to BBTV. Resistance of these lines was evaluated based on symptomatology and host-virus relationship. Insect colony development on Lakatan banana irradiated lines was monitored by artificially inoculating viruliferous banana aphids, Pentalonia nigronervosa, and counting the resulting number of aphids per plant at weekly intervals. Resistance to virus multiplication of Lakatan irradiated lines was characterized by quantifying the virus titer through Enzyme-Linked Immunosorbent Assay (ELISA). Results showed that not all lines were suitable as hosts in establishing aphid population. The reaction of the mutant lines to the vector and the pathogen varied to some extent. Disease incidence in some cases was correlated with aphid preference. Disease incidence was significantly higher (50%) on lines that were preferred by aphids and lower (50%) in those that were not colonized. Some mutant lines with very low aphid colony count, however showed high BBTV incidence. Variability in the results could be affected by other factors such as the developmental stage of the plant and prevailing environmental conditions during the conduct of the experiment. Virus titer was also reduced on these mutant lines, thus reduced virus multiplication. Non-irradiated (control) Lakatan banana had comparably high population of aphids, high disease incidence, and high virus titer
6. Development of bunchy top virus resistant banana cv lakatan in vitro culture and radiation technology
International Nuclear Information System (INIS)
Estrella, J.D.; Caymo, L.S.; Dizon, T.O.; Dela Cruz, F. Jr; Damasco, O.P.
2002-01-01
Bunchy to virus (BTV) is the most destructive virus disease of banana in the Philippines. Incorporation of resistance to this virus disease by conventional hybridization is not possible due to male and female sterility of most commercial banana cultivars. In vitro culture coupled with radiation technology can be used to develop BTV resistance in banana cv. Lakatan. The sensitivity of banana shot tip explants to gamma irradiation was determined by subjecting the shoot tips to varying doses (5, 10, 20, 25, 30, 40, 60, 80 and 100 Gy) of irradiation. The LD sub 50 for banana shoot tips determined by 50% reduction in growth and shoot proliferation, was observed to around 20-25 Gy. Bulk irradiation of shoot tip explants was conducted using 20-25 Gy. Irradiated cultures were multiplied for 3-5 cycles and plants regenerated were potted out and screened for BTV resistance. A total of 3,447 irradiated plants regenerated from the radiosensitivity experiment (1,847 plants) and bulk irradiation of 20/25 Gy (1,600 plants) were screened for BTV resistance in the greenhouse using artificial BTV inoculation using the aphid vector Pentalonia nigronervosa. One hundred eighteen plants or 3.4% (118/3,447) of the artificially irradiated plants showed seedling resistance after 4-7 months of evaluation. These plants were planted in the field and were subjected to natural BTV infection. To date, 85 (out of the 118) putative seedling resistant plants continuously expressed BTV resistance in the field after more than 10 months of evaluation. The absence of BTV infection in 39 putative resistant plants was confirmed by ELISA test. Suckers from selected putative resistance plants will be collected, propagated and evaluated for the second cycle stability of BTV resistance and detailed characterization of important horticultural traits
7. Localization, Concentration, and Transmission Efficiency of Banana bunchy top virus in Four Asexual Lineages of Pentalonia aphids
Directory of Open Access Journals (Sweden)
Alberto Bressan
2013-02-01
Full Text Available Banana bunchy top virus (BBTV is the most destructive pathogenic virus of banana plants worldwide. The virus is transmitted in a circulative non-propagative manner by the banana aphid, Pentalonia nigronervosa Coquerel. In this work, we examined the localization, accumulation, and transmission efficiency of BBTV in four laboratory-established lineages of Pentalonia aphids derived from four different host plants: taro (Colocasia esculenta, heliconia (Heliconia spp., red ginger (Alpinia purpurata, and banana (Musa sp.. Mitochondrial sequencing identified three and one lineages as Pentalonia caladii van der Goot, a recently proposed species, and P. nigronervosa, respectively. Microsatellite analysis separated the aphid lineages into four distinct genotypes. The transmission of BBTV was tested using leaf disk and whole-plant assays, both of which showed that all four lineages are competent vectors of BBTV, although the P. caladii from heliconia transmitted BBTV to the leaf disks at a significantly lower rate than did P. nigronervosa. The concentration of BBTV in dissected guts, haemolymph, and salivary glands was quantified by real-time PCR. The BBTV titer reached similar concentrations in the guts, haemolymph, and salivary glands of aphids from all four lineages tested. Furthermore, immunofluorescence assays showed that BBTV antigens localized to the anterior midguts and the principal salivary glands, demonstrating a similar pattern of translocations across the four lineages. The results reported in this study showed for the first time that P. caladii is a competent vector of BBTV.
8. Localization, concentration, and transmission efficiency of Banana bunchy top virus in four asexual lineages of Pentalonia aphids.
Science.gov (United States)
Watanabe, Shizu; Greenwell, April M; Bressan, Alberto
2013-02-22
Banana bunchy top virus (BBTV) is the most destructive pathogenic virus of banana plants worldwide. The virus is transmitted in a circulative non-propagative manner by the banana aphid, Pentalonia nigronervosa Coquerel. In this work, we examined the localization, accumulation, and transmission efficiency of BBTV in four laboratory-established lineages of Pentalonia aphids derived from four different host plants: taro (Colocasia esculenta), heliconia (Heliconia spp.), red ginger (Alpinia purpurata), and banana (Musa sp.). Mitochondrial sequencing identified three and one lineages as Pentalonia caladii van der Goot, a recently proposed species, and P. nigronervosa, respectively. Microsatellite analysis separated the aphid lineages into four distinct genotypes. The transmission of BBTV was tested using leaf disk and whole-plant assays, both of which showed that all four lineages are competent vectors of BBTV, although the P. caladii from heliconia transmitted BBTV to the leaf disks at a significantly lower rate than did P. nigronervosa. The concentration of BBTV in dissected guts, haemolymph, and salivary glands was quantified by real-time PCR. The BBTV titer reached similar concentrations in the guts, haemolymph, and salivary glands of aphids from all four lineages tested. Furthermore, immunofluorescence assays showed that BBTV antigens localized to the anterior midguts and the principal salivary glands, demonstrating a similar pattern of translocations across the four lineages. The results reported in this study showed for the first time that P. caladii is a competent vector of BBTV.
9. Transgenic approaches for development of disease resistance in banana
International Nuclear Information System (INIS)
Shekhawat, Upendra K.S.; Ghag, Siddhesh B.; Ganapathi, Thumballi R.
2014-01-01
Banana (Musa spp.) is an important food and cash crop worldwide. Diseases and pests pose the most serious constraint to banana cultivation. Among the diseases, Fusarium wilt and Banana Bunchy Top Virus (BBTV) are the most important economically. We have explored different transgenic approaches for development of efficient resistance in banana against these two diseases. For countering Fusarium wilt, we have over expressed Petunia floral defensins using a strong constitutive promoter in transgenic banana plants. We have also tested a host induced gene silencing strategy targeting two vital fungal genes to obtain Fusarium resistant banana plants. For development of BBTV resistant banana plants also, we have used a host-induced gene silencing approach utilizing the full and partial coding sequence of the viral replication initiation protein. Successful bioassays performed in controlled greenhouse conditions have shown the efficacy of using these strategies to develop disease resistant banana plants. (author)
10. Biology, etiology, and control of virus diseases of banana and plantain.
Science.gov (United States)
Kumar, P Lava; Selvarajan, Ramasamy; Iskra-Caruana, Marie-Line; Chabannes, Matthieu; Hanna, Rachid
2015-01-01
Banana and plantain (Musa spp.), produced in 10.3 million ha in the tropics, are among the world's top 10 food crops. They are vegetatively propagated using suckers or tissue culture plants and grown almost as perennial plantations. These are prone to the accumulation of pests and pathogens, especially viruses which contribute to yield reduction and are also barriers to the international exchange of germplasm. The most economically important viruses of banana and plantain are Banana bunchy top virus (BBTV), a complex of banana streak viruses (BSVs) and Banana bract mosaic virus (BBrMV). BBTV is known to cause the most serious economic losses in the "Old World," contributing to a yield reduction of up to 100% and responsible for a dramatic reduction in cropping area. The BSVs exist as episomal and endogenous forms are known to be worldwide in distribution. In India and the Philippines, BBrMV is known to be economically important but recently the virus was discovered in Colombia and Costa Rica, thus signaling its spread into the "New World." Banana and plantain are also known to be susceptible to five other viruses of minor significance, such as Abaca mosaic virus, Abaca bunchy top virus, Banana mild mosaic virus, Banana virus X, and Cucumber mosaic virus. Studies over the past 100 years have contributed to important knowledge on disease biology, distribution, and spread. Research during the last 25 years have led to a better understanding of the virus-vector-host interactions, virus diversity, disease etiology, and epidemiology. In addition, new diagnostic tools were developed which were used for surveillance and the certification of planting material. Due to a lack of durable host resistance in the Musa spp., phytosanitary measures and the use of virus-free planting material are the major methods of virus control. The state of knowledge on BBTV, BBrMV, and BSVs, and other minor viruses, disease spread, and control are summarized in this review. © 2015 Elsevier Inc
11. Better bananas
International Nuclear Information System (INIS)
1986-01-01
This is a public relations film describing problems associated with the genetic improvement of bananas and plantains. These fruit and food crops have a large economic and nutritional value for tropical regions. The vulnerability of bananas to disease epidemics urgently requires breeding for resistance to black Sigatoka (leaf spot disease). The joint FAO/IAEA division has initiated a programme and developed a biotechnological strategy for genetic improvement of bananas and plantains
12. Modifying Bananas: From Transgenics to Organics?
Directory of Open Access Journals (Sweden)
James Dale
2017-02-01
Full Text Available Bananas are one of the top ten world food crops. Unlike most other major food crops, bananas are difficult to genetically improve. The challenge is that nearly all banana cultivars and landraces are triploids, with high levels of male and female infertility. There are a number of international conventional breeding programs and many of these are developing new cultivars. However, it is virtually impossible to backcross bananas, thus excluding the possibility of introgressing new traits into a current cultivar. The alternative strategy is to “modify” the cultivar itself. We have been developing the capacity to modify Cavendish bananas and other cultivars for both disease resistance and enhanced fruit quality. Initially, we were using transgenes; genes that were derived from species outside of the Musa or banana genus. However, we have recently incorporated two banana genes (cisgenes into Cavendish; one to enhance the level of pro-vitamin A and the other to increase the resistance to Panama disease. Modified Cavendish with these cisgenes have been employed in a field trial. Almost certainly, the next advance will be to edit the Cavendish genome, to generate the desired traits. As these banana cultivars are essentially sterile, transgene flow and the outcrossing of modified genes into wild Musa species. are highly unlikely and virtually impossible in other triploid cultivars. Therefore, genetic changes in bananas may be compatible with organic farming.
13. Response of banana cultivars to banana weevil attack | Kiggundu ...
African Journals Online (AJOL)
East African Highland Bananas (EAHB) (Musa AAA, 'Matooke' group) are a major staple food in the East African region. However, banana weevil (Cosmopolites sorllidus) is a major production constraint to bananas and may cause damage levels of up to 100%. Pesticides can effectively control banana weevil but these are ...
14. Micropropagation of banana.
Science.gov (United States)
Kaçar, Yıldız Aka; Faber, Ben
2012-01-01
Banana (Musa spp. AAA) is propagated vegetatively and can be rapidly and efficiently propagated by micropropagation. Conventional micropropagation techniques, however, may be too costly for commercial purposes. Our laboratory has found that depending on the combination of culture vessel and gelling agent more economic methods can be chosen for successfully micropropagating banana.
15. Bananas go paraelectric
International Nuclear Information System (INIS)
Loidl, A; Krohns, S; Hemberger, J; Lunkenheimer, P
2008-01-01
Using a banana as an example, we demonstrate how the ferroelectric-like hysteresis loops measured in inhomogeneous, conducting materials can easily be identified as non-intrinsic. With simple experiments, the response of a banana to electric fields is revealed as characteristic for an inhomogeneous paraelectric ion conductor. Not even absolute beginners in dielectrics should identify this biological matter as ferroelectric. (viewpoint)
16. Bananas and plantains
International Nuclear Information System (INIS)
1986-01-01
The film shows the germplasm diversity within the Genus Musa and the evolution of cultivated forms of bananas and plantains. Cultivation history and geographical distribution are depicted; features of plant morphology and the floral biology are demonstrated. Economic and nutritional impact and importance of bananas and plantains for developing countries are briefly discussed. The second part of the film surveys problems in the propagation and genetic improvement of bananas and plantains: fruits of these vegetatively propagated plants are usually seedless which complicate the application of conventional plant breeding methods. In-vitro techniques are shown to be useful for plant propagation and germplasm conservation. Cross breeding with some semi-sterile clones of bananas has not led so far to lines which are resistant to the most harmful diseases, e.g. panama disease, black sigatoka. The Joint FAO/IAEA division has initiated an in-vitro mutation breeding programme to improve disease resistance in bananas
17. The Banana Genome Hub
Science.gov (United States)
Droc, Gaëtan; Larivière, Delphine; Guignon, Valentin; Yahiaoui, Nabila; This, Dominique; Garsmeur, Olivier; Dereeper, Alexis; Hamelin, Chantal; Argout, Xavier; Dufayard, Jean-François; Lengelle, Juliette; Baurens, Franc-Christophe; Cenci, Alberto; Pitollat, Bertrand; D’Hont, Angélique; Ruiz, Manuel; Rouard, Mathieu; Bocs, Stéphanie
2013-01-01
Banana is one of the world’s favorite fruits and one of the most important crops for developing countries. The banana reference genome sequence (Musa acuminata) was recently released. Given the taxonomic position of Musa, the completed genomic sequence has particular comparative value to provide fresh insights about the evolution of the monocotyledons. The study of the banana genome has been enhanced by a number of tools and resources that allows harnessing its sequence. First, we set up essential tools such as a Community Annotation System, phylogenomics resources and metabolic pathways. Then, to support post-genomic efforts, we improved banana existing systems (e.g. web front end, query builder), we integrated available Musa data into generic systems (e.g. markers and genetic maps, synteny blocks), we have made interoperable with the banana hub, other existing systems containing Musa data (e.g. transcriptomics, rice reference genome, workflow manager) and finally, we generated new results from sequence analyses (e.g. SNP and polymorphism analysis). Several uses cases illustrate how the Banana Genome Hub can be used to study gene families. Overall, with this collaborative effort, we discuss the importance of the interoperability toward data integration between existing information systems. Database URL: http://banana-genome.cirad.fr/ PMID:23707967
18. Alcohol from bananas
Energy Technology Data Exchange (ETDEWEB)
Hammond, J.B.; Egg, Richard; Coble, C.G. [Texas A and M Univ., College Station, TX (United States). Dept. of Agricultural Engineering; Diggins, Drew
1996-04-01
Laboratory studies were conducted to assess the ethanol production potential from waste bananas. Over a 10-day ripening period, there was a 9% loss of fresh weight by day 6 and a 15% loss by day 10. Ethanol yields from normal ripe bananas were: whole fruit - 0.091, pulp -0.082, and peel -0.006 l/kg of whole fruit. Ripeness effects on ethanol yield were measured as green - 0.090, normal ripe - 0.082, and overripe - 0.069 l/kg of green whole bananas. Enzymatic hydrolysis was necessary for maximum yields. Dilution water was not essential for effective fermentation. Waste parameters of the banana stillage were measured. (Author)
Science.gov (United States)
Zagatto, V. A. B.; Medina, N. H.; Okuno, E.; Umisedo, N. K.
2008-08-01
The content of 40K natural radionuclide in bananas (Musa sapientum) from the Vale do Ribeira region, São Paulo, Brazil, has been measured. We have collected several samples of bananas prata and nanica, its peels, leaves, and also different soils where the banana tree was planted, such as soil with a standard amount of fertilizer, the fertilizer itself and also soil without fertilizer for comparison. We have used the gamma-ray spectroscopy technique with a NaI(T1) crystal inside a 12 cm thick lead shield to detect the gamma-radiation. The results indicate that only part of the available potassium is absorbed by the plant, which is mainly concentrated in the banana peel.
International Nuclear Information System (INIS)
Zagatto, V. A. B.; Medina, N. H.; Okuno, E.; Umisedo, N. K.
2008-01-01
The content of 40 K natural radionuclide in bananas (Musa sapientum) from the Vale do Ribeira region, Sao Paulo, Brazil, has been measured. We have collected several samples of bananas prata and nanica, its peels, leaves, and also different soils where the banana tree was planted, such as soil with a standard amount of fertilizer, the fertilizer itself and also soil without fertilizer for comparison. We have used the gamma-ray spectroscopy technique with a NaI(T1) crystal inside a 12 cm thick lead shield to detect the gamma-radiation. The results indicate that only part of the available potassium is absorbed by the plant, which is mainly concentrated in the banana peel
International Nuclear Information System (INIS)
Huyzers, C.J.; Basson, R.
1985-01-01
Early studies on the radurisation of bananas indicated that this commodity did not benefit substantially from the treatment. This work, which was carried out at Pelindaba, indicated a low threshold dose for radiation damage and little shelf-life extension at this dose. In a second study carried out at Tzaneen more promising results were obtained. The reason for the differences seemed to be due to the time between harvesting and treatment which was much shorter in the Tzaneen study. Consequently it was decided to undertake a third and much larger trial in which the bananas would be treated in Tzaneen and then dispatched to Pretoria for storage and evaluation by a joint Nucor/Banana Board team. Parameters investigated included colour, firmness (by penatrometer testing) and sensory qualities. The bananas were stored for various periods at 15 o C, ripened by exposure to ethylene gas under commerical conditons and then stored at ambient temperature for the remainder of the experiment. Bananas we re irradiated at various doses between 0,3 kGy and 1,5kGy and were compared with control batches which were stored under the same conditions
2. The "Blue Banana" Revisited
NARCIS (Netherlands)
Faludi, A.K.F.
2015-01-01
This essay is about the “Blue Banana”. Banana is the name given subsequently by others to a Dorsale européenne (European backbone) identified empirically by Roger Brunet. In a background study to the Communication of the European Commission ‘Europe 2000’, Klaus Kunzmann and Michael Wegener put
3. Ply tensile properties of banana stem and banana bunch fibres
African Journals Online (AJOL)
2012-03-01
Mar 1, 2012 ... BANANA BUNCH FIBRES REINFORCED NATURAL RUBBER. COMPOSITE ... National Institute for Interdisciplinary Science and Technology, (NIIST) CSIR Trivandrum, India. ..... Handbook of Ceramics and Composites,. Vol.
4. Banana Algebra: Compositional syntactic language extension
DEFF Research Database (Denmark)
Andersen, Jacob; Brabrand, Claus; Christiansen, David Raymond
2013-01-01
We propose an algebra of languages and transformations as a means of compositional syntactic language extension. The algebra provides a layer of high-level abstractions built on top of languages (captured by context-free grammars) and transformations (captured by constructive catamorphisms...... algebra as presented in the paper is implemented as the Banana Algebra Tool which may be used to syntactically extend languages in an incremental and modular fashion via algebraic composition of previously defined languages and transformations. We demonstrate and evaluate the tool via several kinds...
5. Response of banana hybrids to the banana weevil (Cosmopolites ...
African Journals Online (AJOL)
ACSS
Response of banana hybrids to the banana weevil (Cosmopolites sordidus Germar) .... A number of physical and chemical factors are .... The total number of weevils trapped were then counted and recorded. Agronomic characteristics. Bunch weight, girth and height. In addition to corm damage assesment, data was also ...
6. Banana Gold: Problem or Solution?
Science.gov (United States)
Joseph, Garnet
1992-01-01
Since 1955, the British banana industry has dominated the lives of the Caribs and other peoples in Dominica. Banana growing supplants other economic activities, including local food production; toxic chemicals and fertilizers pollute the land; community is dwindling; suicide is common; and child labor diminishes school attendance. (SV)
7. Combating the Sigatoka disease complex on banana
Science.gov (United States)
Banana is the fourth most important staple food in the world behind rice, wheat and maize, with more than 100 million tons produced annually. Although the majority of bananas produced are consumed locally, banana export is a multi-billion dollar business. Bananas are grown in more than 100 countri...
8. Beyond the double banana
DEFF Research Database (Denmark)
Rosenzweig, Ivana; Fogarasi, András; Johnsen, Birger
2014-01-01
PURPOSE: To investigate whether extending the 10-20 array with 6 electrodes in the inferior temporal chain and constructing computed montages increases the diagnostic value of ictal EEG activity originating in the temporal lobe. In addition, the accuracy of computer-assisted spectral source......). Spectral source analysis used source montage to calculate density spectral array, defining the earliest oscillatory onset. From this, phase maps were calculated for localization. The reference standard was the decision of the multidisciplinary epilepsy surgery team on the seizure onset zone. Clinical...... performance was compared with the double banana (longitudinal bipolar montage, 10-20 array). RESULTS: Adding the inferior temporal electrode chain, computed montages (reference free, common average, and source derivation), and voltage maps significantly increased the sensitivity. Phase maps had the highest...
9. Fabrication of Biomembrane from Banana Stem for Lead Removal
Directory of Open Access Journals (Sweden)
Afianti Sulastri
2016-04-01
Full Text Available Heavy metal (i.e. lead (Pb is one of the environmental issues recently due to its danger for human health. Therefore, strategy for removing Pb from waste water treatment is important. One of the prospective methods to remove Pb is membrane biofilter. Here, the purpose of this study was to prepare the membrane biofilter for Pb removal process. In this study, membrane biofilter was produced from banana stem. Banana stem was selected because of its abundant availability in Indonesia. And, for somewhat, this banana stem can be environmental problems (become waste since Indonesia is one of the top producers in the world. In short of the experimental procedure, we conducted three steps of experiments: (1 Preparation of microbial cellulose using Acetobacter xylinum using banana stem for a main source; (2 Synthesis of cellulose acetate; and (3 Preparation of biomembrane from obtained cellulose acetate. To produce membrane biofilter, the cellulose acetate was dissolved into dichloromethane to form a dope solution. Then, the doped solution was printed in Petri dish. Some biomembrane properties were characterized for identification, i.e. infrared spectra, electron microscope, and elemental analysis. Experimental results showed that we succeeded to prepare biomembrane with a pore size of 5 μm. The filtration efficiency of our prepared membrane was 93.7% of Pb when using Pb with a concentration of 10 ppm.
10. In vitro digestibility of banana starch cookies.
Science.gov (United States)
Bello-Pérez, Luis A; Sáyago-Ayerdi, Sonia G; Méndez-Montealvo, Guadalupe; Tovar, Juscelino
2004-01-01
Banana starch was isolated and used for preparation of two types of cookies. Chemical composition and digestibility tests were carried out on banana starch and the food products, and these results were compared with corn starch. Ash, protein, and fat levels in banana starch were higher than in corn starch. The high ash amount in banana starch could be due to the potassium content present in this fruit. Proximal analysis was similar between products prepared with banana starch and those based on corn starch. The available starch content of the banana starch preparation was 60% (dmb). The cookies had lower available starch than the starches while banana starch had lower susceptibility to the in vitro alpha-amylolysis reaction. Banana starch and its products had higher resistant starch levels than those made with corn starch.
11. Study on Banana Cooperatives in Hainan Province
OpenAIRE
Huang, Huide; Zhang, Wanzhen; Liu, Enping; Zhang, Xizhu
2013-01-01
This paper gives an overview of the distribution, member scale, production and operation of banana cooperatives in Hainan Province, and points out the market risk and natural risk faced by the production of banana cooperatives in Hainan Province. In order to promote the banana cooperatives to form new agricultural management system integrating organization and intensification, this paper puts forth the production and operation recommendations, such as joint production of banana cooperatives, ...
Index Scriptorium Estoniae
2008-01-01
Ekspedeerimisfirmade TOP 57. Vt. samas: Tanel Raig. Majandus kukutab ekspedeerimisturgu. Diagramm: Väliskaubanduse statistika; Katrin Raie. Ekspedeerijad hakkavad rohkem koostööle rõhuma. Kommenteerib Jaan Lepp; Müügitulu TOP 10; Müügitulu kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Eestis registreeritud Vene hiiglane; Ekspedeerimisturu kasumiliider kaotas 20 miljonit; Küsimustele vastab OÜ Contimer juht Dmitri Redkin
13. Antioxidant activity of banana flavonoids.
Science.gov (United States)
Vijayakumar, S; Presannakumar, G; Vijayalakshmi, N R
2008-06-01
The antioxidant activity of flavonoids from banana (Musa paradisiaca) was studied in rats fed normal as well as high fat diets. Concentrations of peroxidation products namely malondialdehyde, hydroperoxides and conjugated diens were significantly decreased whereas the activities of catalase and superoxide dismutase were enhanced significantly. Concentrations of glutathione were also elevated in the treated animals.
14. Effects of covering highland banana stumps with soil on banana weevil Cosmopolites sordidus (Coleoptera: Curculionidae) oviposition
NARCIS (Netherlands)
Masanza, M.; Gold, C.S.; Huis, van A.; Ragama, P.E.
2005-01-01
The effect of covering post-harvest banana stumps with soil on banana weevil Cosmopolites sordidus (Germar) oviposition levels was investigated at three locations, Sendusu, Kawanda Agricultural Research Institute (KARI) and Ntungamo district of southwestern Uganda. In the first experiment
15. Marketing of banana and banana products in Uganda: Results of a rapid rural appraisal
OpenAIRE
Digges, Philip
1994-01-01
This report concerns a survey undertaken by NRI in Uganda during September and December 1993, which sought to characterise the banana and banana beer marketing systems. The study follows on from the recommendations of the Banana Based Cropping System Rapid Rural Appraisal (1991), and focuses upon the Kampala market.
16. Production of ethyl alcohol from bananas
Energy Technology Data Exchange (ETDEWEB)
Jones, R.L.; Towns, T.
1983-12-01
The production of ethyl alcohol from waste bananas presents many special problems. During cooking, matting of the latex fibers from the banana peel recongeal when cooled and left untreated. This problem has been addressed by Alfaro by the use of CaC1/sub 2/. Separation of solids prior to distillation of the mashes in an economical fashion and use of the by product are also of concern to banana processors.
17. Enhancing banana weevil ( Cosmopolites sordidus ) resistance by ...
African Journals Online (AJOL)
Enhancing banana weevil (Cosmopolites sordidus) resistance by plant genetic modification: A perspective. Andrew Kiggundu, Michael Pillay, Altus Viljoen, Clifford Gold, Wilberforce Tushemereirwe, Karl Kunert ...
18. Social Interactions in Growing Bananas
DEFF Research Database (Denmark)
Van Den Broeck, Katleen; Dercon, Stefan
This paper analyses whether agricultural information flows give rise to social learning effects in banana cultivation in Nyakatoke, a small Tanzanian village. Based on a village census, full information is available on socio-economic characteristics and banana production of farmer kinship members......, neighbours and informal insurance group members. This allows a test for social learning within these groups and the identification of different types of social effects. Controlling for exogenous group characteristics, the effect of group behaviour on individual farmer output is studied. The results show...... that social effects are strongly dependent on the definition of the reference group. It emerges that no social effects are found in distance based groups, exogenous social effects linked to group education exist in informal insurance groups, and only kinship related groups generate the endogenous social...
19. Direct Effects Of Chronic Gamma Radiation On Musa Acuminata Var. Berangan, A Local Malaysia Banana Cultivar
International Nuclear Information System (INIS)
Maimum Tahir; Azhar Mohamad; Rozeita Laboh; Umikalsum Mohd Bahari
2014-01-01
Musa acuminata var. Berangan, is a popular variety of our local banana known as Pisang Berangan. The variety is a triploid banana, use mainly for dessert and has a great value for commodity fruit crops. However, production of PisangBerangan has been threatened by diseases such as Fusarium wilt, black sigatoka, Fusarium wilt, burrowing nematodes and viral diseases like Banana streak virus, Banana bunch top virus and Banana bract mosaic virus. The scenario becoming worst as Musa has a narrow genetic background for breeding and/or selection program. The banana breeding program of edible bananas is hampered by high sterility, and very limited amounts of seeds. Mutation induction via chronic gamma radiation is an alternative ways in creating more variants for selections towards a better quality and disease tolerance. A total number of 75 samples at nursery stage (1 month) were exposed to chronic gamma radiation in Gamma Greenhouse at Malaysian Nuclear Agency for 28 weeks. The samples were accordingly arranged with distance ranging from 1 m to 15 m from gamma source (Cesium-137). Plant height and new buds were used as measurement parameters in evaluating the direct effects of the chronic gamma radiation. Results showed effective dose of chronic gamma radiation in Pisang Berangan was 20 Gy. Number of new emerging sucker was ranging from 1-3 pieces with the highest at ring-4 and ring-5. Plant height was observed ranging from 22.1 to 110.5 cm. Effects of chronic radiation were observed after 3-4 months in the GGH. The samples revealed as striking leaves, short inter node and new emergence of suckers. The objective of this work is to get a dose response for chronic gamma radiation in Pisang Berangan. As for selection of potential mutant variants, new emerging suckers were tissue cultured in segregating chimeras and to get required numbers of samples for further field evaluation. (author)
20. Diseases threatening banana biodiversity in Uganda ...
African Journals Online (AJOL)
Recent on station and on-farm studies suggest the major diseases threatening banana biodiversity in Uganda include: 1)Black sigatoka which severely affects all East African Highland (EA-AAA) banana cultivars and a range of introduced genotypes; 2) Fusarium wilt which affects several introduced genotypes though all EA ...
1. Banana (Musa. spp.) strain HD-1 appraisal
International Nuclear Information System (INIS)
Longyan, G.; Xinguo, L.; Lingxia, W.; Xuefei, J.
2016-01-01
Being one of the important tropical and subtropical fruit trees, banana (Musa spp.) belongs to the family Musaceae and the order Scitaminae with two genera, Musa and Ensete. In a field survey, research team has discovered a potential banana mutant strain HD-1 with a sound economic value. The results of the finding are as follows: based on Simmonds classification, the pseudostem of banana strain HD-1 is relatively short and purplish red; its upright outward petiole groove has red edges and wraps its pseudostem loosely. Its ploidy is 3, AAA type. Karyotype analysis shows that the number of chromosomes is 33, the karyotype formula is 2n=3x=33=2L + 3 M2 + 4 M1 + 2 S, HD-1 is classified as 1B type. With the help of ISSR molecular markers, we find thatbanana HD-1 has the closest relationship with Pubei and Tianbao dwarf banana; the similarity coefficient is 0.81. In an artificial simulation tests of cold, drought and salt resistance environment changes of physiological and biochemical indexes indicate that HD-1 exhibits stronger defense capability than Brazil banana. By way of inoculation with injury of root dipping method, we respectively treat two kinds of banana seedlings inoculated Banana Fusarium wilt race 4 small species. The results show that their resistance evaluation scores are 3 and 4, disease levels are susceptible and high sensitivity respectively. We conclude that HD-1 has stronger resistance ability to Fusarium wilt than Brazil banana. (author)
2. 33 CFR 117.263 - Banana River.
Science.gov (United States)
2010-07-01
... 33 Navigation and Navigable Waters 1 2010-07-01 2010-07-01 false Banana River. 117.263 Section 117.263 Navigation and Navigable Waters COAST GUARD, DEPARTMENT OF HOMELAND SECURITY BRIDGES DRAWBRIDGE OPERATION REGULATIONS Specific Requirements Florida § 117.263 Banana River. (a) The draw of the Mathers (SR...
3. Market opportunities for Ugandan banana products: National ...
African Journals Online (AJOL)
Mo
the decade between 1991 and 2001, banana consumption has increased by three ... obtain only 17 % of the retail price per bunch of banana, whereas the rest ..... The Middle and Far East will experience the biggest growth rates in terms of ...
4. Banana orchard inventory using IRS LISS sensors
Science.gov (United States)
Nishant, Nilay; Upadhayay, Gargi; Vyas, S. P.; Manjunath, K. R.
2016-04-01
Banana is one of the major crops of India with increasing export potential. It is important to estimate the production and acreage of the crop. Thus, the present study was carried out to evolve a suitable methodology for estimating banana acreage. Area estimation methodology was devised around the fact that unlike other crops, the time of plantation of banana is different for different farmers as per their local practices or conditions. Thus in order to capture the peak signatures, biowindow of 6 months was considered, its NDVI pattern studied and the optimum two months were considered when banana could be distinguished from other competing crops. The final area of banana for the particular growing cycle was computed by integrating the areas of these two months using LISS III data with spatial resolution of 23m. Estimated banana acreage in the three districts were 11857Ha, 15202ha and 11373Ha for Bharuch, Anand and Vadodara respectively with corresponding accuracy of 91.8%, 90% and 88.16%. Study further compared the use of LISS IV data of 5.8m spatial resolution for estimation of banana using object based as well as per-pixel classification and the results were compared with statistical reports for both the approaches. In the current paper we depict the various methodologies to accurately estimate the banana acreage.
5. A screening method for banana weevil ( Cosmopolites sordidus ...
African Journals Online (AJOL)
The banana weevil (Cosmopolites sordidus Germar) is a serious pest in most banana-growing areas of the world. Host-plant resistance is considered to be the most feasible and sustainable method for its control. However, a quick and effective method for screening banana genotypes for resistance against the banana ...
6. Radiation enhances shelf life of Nendra bananas without changing the lectin content of raw and steamed Nendra banana
International Nuclear Information System (INIS)
Coelho, Neil Renault; Nivas, Shashikiran; D'Souza, L.
2016-01-01
Our study shows that the shelf life of bananas is increased with low doses of radiation (300 Gy, 400 Gy, 500 Gy). However, there is no decrease in the lectin content. This improves the keeping quality of nendra bananas without affecting their lectin content. Hence, radiation can be used safely for the bananas distributed to HIV children. The present study was also to compare the lectin content of raw and steamed Nendra bananas with the gamma irradiated bananas
Index Scriptorium Estoniae
2002-01-01
TOP 90. Kinnisvara valdkondade TOP 5. Käibe TOP 30. Käibe kasvu TOP 30. Rentaabluse TOP 30. Kasumi TOP 30. Kasumi kasvu TOP 30. Varade tootlikkuse TOP 30. Kinnisvarafirmade üldandmed. Kinnisvarafirmade finantsandmed
8. Phenylphenalenones Accumulate in Plant Tissues of Two Banana Cultivars in Response to Herbivory by the Banana Weevil and Banana Stem Weevil
OpenAIRE
H?lscher, Dirk; Buerkert, Andreas; Schneider, Bernd
2016-01-01
Phenylphenalenone-type compounds accumulated in the tissues of two banana cultivars—Musa acuminata cv. “Grande Naine” (AAA) and Musa acuminata × balbisiana Colla cv. “Bluggoe” (ABB)—when these were fed on by the banana weevil (Cosmopolites sordidus (Germ.) (Coleoptera: Curculionidae)) and the banana stem weevil (Odoiporus longicollis (Oliver) (Coleoptera: Curculionidae)). The chemical constituents of the banana material were separated by means of chromatographic techniques and identified by N...
9. PROPOLIS EXTRACT IN POSTHARVEST CONSERVATION BANANA ' PRATA'
Directory of Open Access Journals (Sweden)
FLÁVIA REGINA PASSOS
2016-01-01
Full Text Available ABSTRACT In the present work were evaluated the effects of propolis coatings of various botanical sources on quality traits of bananas cv. Prata (Musa sapientum L. stored at room temperature. ´Prata´ bananas were selected and submitted to five postharvest treatments: four coatings applied by immersion in propolis extracts at a concentration of 2.5% (w/v and a control (without coating. Propolis extracts were applied as 1 a wild type aqueous propolis extract, 2 a wild type hydroalcoholic propolis extract, 3 a rosemary green type hydroalcoholic propolis extract and 4 a red type hydroalcoholic propolis extract. The bananas were evaluated at three-day intervals along 12 days for fresh weight losses, flesh firmness, soluble solids (SS, titratable acidity (TA, the ratio SS/TA and pH. Sensory analyses were performed after three and six days of storage by 55 not trained panelists designed for acceptability. At the end of the twelve-day storage period, bananas coated either with the rosemary green hydroalcoholic extract or with the aqueous extract presented lower fresh weight losses in comparison to the bananas of the control treatment. No differences were determined in relation to flesh firmness and along the storage period TA values decreased and pH values increased in bananas of all treatments. SS contents increased towards the end of the storage period that, consequently, contributed to increases in the SS/TA ratio. The most significant increase in SS/TA ratio was determined in bananas coated with the red type hydroalcoholic extract. Taste panelists did not detect significant differences amongst coated and not coated cv. Prata bananas up to six days of storage.
10. In vivo fertilization of banana
Directory of Open Access Journals (Sweden)
Taliane Leila Soares
2014-01-01
Full Text Available The aim of this research was to study the in vivo fertilization process of banana cultivars. The diploid hybrid (AA 091087-01 was the male progenitor. Flower samples were checked for fertilization from the first to the twentieth day after pollination. The size of the diploid ovules increased gradually at the beginning of the seed formation process. On the other hand, in the AAA triploids (Cavendish subgroup, the not fertilized ovules were aborted. In the AAB triploids (Prata subgroup some ovules were fertilized. The flowers of Grand Naine, Nanicão and 'Pacovan' cultivars presented necrosis in the distal part of the ovary on the first day after pollination. Necrosis can hinder pollen tube growth towards the ovule, which might be related to the low seed yield in 'Pacovan' cultivars and to the absence of seeds in the Cavendish subgroup cultivars.
Index Scriptorium Estoniae
2003-01-01
Kinnisvarafirmade TOP 90. Rentaabluse TOP 30. Käibe TOP 30. Käibe kasvu TOP 30. Kasumi TOP 30. Kasumi kasvu TOP 30. Varade tootlikkuse TOP 30. Kinnisvarafirmade üldandmed. Kinnisvarafirmade finantsandmed
Index Scriptorium Estoniae
2006-01-01
Teedeehitusfirmade TOP. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi Kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Teedeehitusfirmade üld- ja finantsandmed
Index Scriptorium Estoniae
2006-01-01
Koolitusfirmade TOP. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Majandustegevuse kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Koolitusfirmade üld- ja finantsandmed
14. Telekommunikatsiooni TOP aastal 2003
Index Scriptorium Estoniae
2004-01-01
Telekommunikatsiooni TOP aastal 2003. Käibe TOP 10. Käibe kasvu TOP 10. Rentaabluse TOP 10. Kasumi TOP 10. Kasumi kasvu TOP 10. Omakapitali tootlikkuse TOP 10. Telekommunikatsioonifirmade üldandmed. Telekommunikatsioonifirmade finantsandmed
15. Põllumajandustootjate TOP 50
Index Scriptorium Estoniae
2006-01-01
Põllumajandustootjate TOP. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Majandustegevuse kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Põllumajandustootjate üld- ja finantsandmed
16. Ehitusmaterjalitootjate TOP 95
Index Scriptorium Estoniae
2006-01-01
Ehitusmaterjalitootjate TOP. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Ehitusmaterjalitootjate üld- ja finantsandmed
17. Molecular Characterization of Cocoa, Mango, Banana and Yam ...
African Journals Online (AJOL)
Molecular Characterization of Cocoa, Mango, Banana and Yam Isolates of Botryodiplodia theobromae in Ghana. ... A total of 25 fungal isolates were sampled from cocoa, mango, banana and yam within four ... HOW TO USE AJOL.
18. Detection of banana streak virus (BSV) Tamil Nadu isolate (India ...
African Journals Online (AJOL)
Yomi
2012-10-09
641 003, Tamil Nadu, India. 2Department of Fruit Crops, ... Hence, attempts were made for diagnosis of BSV and to study the serological relationship with ... Among the five virus diseases of banana, disease caused by banana ...
19. Value-adding post harvest processing of cooking bananas (Musa ...
African Journals Online (AJOL)
Yomi
2010-12-29
Dec 29, 2010 ... It is estimated that more than 30% of the banana production are lost after harvest. The losses .... nutritional qualities are important factors in the production of banana flour and ..... Agriculture (IITA), Ibadan, Nigeria, VII, p. 166.
20. Effect of fermented Banana peel on Broiler Carcass
OpenAIRE
Koni TNI
2013-01-01
This experiment was conducted to examine effect of inclusion of fermented banana peel by Rhyzopus oligosporus in diets on slaughter weight, carcass weight and carcass percentage, weight and percentage abdominal fat of broiler. The experiment was done based on Completely Randomized Design with four treatments and four replications and each replication consisted of six chickens. The treatment were R0 = without banana peel fermented, R1 = 5% banana peel fermented, R2 = 10% banana peel fermented...
1. Research on Risks and Forecasting Countermeasures of Hainan Banana Industry
OpenAIRE
Liu, Yan-qun; Zeng, Xiao-hong; Fang, Jia
2011-01-01
Based on the overviews of the current conditions of Hainan banana industry, the research makes an analysis of the risks faced by Hainan banana industry. They are respectively marketing risks, natural risks, information risks and production risks. In order to promote a sustainable and rapid development of Hainan banana industry, Countermeasures are proposed in the research. The first is to strengthen the leading organization of forecasting mechanisms on banana industry. The second is to establ...
2. 7 CFR 318.13-22 - Bananas from Hawaii.
Science.gov (United States)
2010-01-01
... 7 Agriculture 5 2010-01-01 2010-01-01 false Bananas from Hawaii. 318.13-22 Section 318.13-22... SERVICE, DEPARTMENT OF AGRICULTURE STATE OF HAWAII AND TERRITORIES QUARANTINE NOTICES Regulated Articles From Hawaii and the Territories § 318.13-22 Bananas from Hawaii. (a) Green bananas (Musa spp.) of the...
3. Weed management in banana production: The use of Nelsonia ...
African Journals Online (AJOL)
During a survey of weeds in the Tiko banana plantations, the plant Nelsonia canescens (Lam.) Spreng was found to have invaded large areas of the plantation with no visible adverse effects on the banana crop. The effects of this Acanthaceae on banana yield parameters, snails' population and weed species diversity and ...
4. I Have a Banana Tree in My Classroom
Science.gov (United States)
Williams, Patricia A.
2007-01-01
When the banana is growing, the broadest part of the banana is located at the bottom, while the tapered end points upward. It appears upside down, however, from the banana tree's perspective, it is growing right side up. The author observes that the students in her classroom labeled by society as "at risk," are also, in a sense, "upside down."…
5. Analysis of genetic variation in different banana ( Musa species ...
African Journals Online (AJOL)
The banana (Musa acuminata Colla) is considered as an important crop plant due to its high economic value as good dietary source. Here, we analyze the genetic relationship of four different banana varieties that are cultivated in south India. Random amplified polymorphic DNAs (RAPDs) fingerprinting of these banana ...
6. Farmer evaluation of dried banana based products | Pekke | African ...
African Journals Online (AJOL)
A farmer participatory evaluation of dried banana based products was conducted in various districts of Uganda. Bananas were dried using a tunnel solar dryer developed by Post Harvest Handling and Storage project (PHHS) of Kawanda Post-harvest Programme and improved by the National Banana Research ...
7. Compositional changes in banana ( Musa ssp. ) fruits during ripening
African Journals Online (AJOL)
overripe banana fruits, respectively. The results showed that the nutritional composition of banana pulp was diversely affected by ripening. Changes in mineral composition varied and were not consistent with the stages of ripeness. Bananas are considered a good source of Mg in the diet, and the data obtained herein ...
8. Remote quality monitoring in the banana chain.
Science.gov (United States)
Jedermann, Reiner; Praeger, Ulrike; Geyer, Martin; Lang, Walter
2014-06-13
Quality problems occurring during or after sea transportation of bananas in refrigerated containers are mainly caused by insufficient cooling and non-optimal atmospheric conditions, but also by the heat generated by respiration activity. Tools to measure and evaluate these effects can largely help to reduce losses along the banana supply chain. The presented green life model provides a tool to predict the effect of deviating temperature, relative humidity, and CO2 and O2 gas concentrations on the storage stability of bananas. A second thermal model allows evaluation of the cooling efficiency, the effect of changes in packaging and stowage and the amount of respiration heat from the measured temperature curves. Spontaneous ripening causes higher respiration heat and CO2 production rate. The resulting risk for creation of hot spots increases in positions in which the respiration heat exceeds the available cooling capacity. In case studies on the transport of bananas from Costa Rica to Europe, we validated the models and showed how they can be applied to generate automated warning messages for containers with reduced banana green life or with temperature problems and also for remote monitoring of the ripening process inside the container.
9. Olfactory responses of banana weevil predators to volatiles from banana pseudostem tissue and synthetic pheromone.
Science.gov (United States)
Tinzaara, W; Gold, C S; Dicke, M; van Huis, A
2005-07-01
As a response to attack by herbivores, plants can emit a variety of volatile substances that attract natural enemies of these insect pests. Predators of the banana weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) such as Dactylosternum abdominale (Coleoptera: Hydrophilidae) and Pheidole megacephala (Hymenoptera: Formicidae), are normally found in association with weevil-infested rotten pseudostems and harvested stumps. We investigated whether these predators are attracted to such environments in response to volatiles produced by the host plant, by the weevil, or by the weevil plant complex. We evaluated predator responses towards volatiles from banana pseudostem tissue (synomones) and the synthetic banana weevil aggregation pheromone Cosmolure+ in a two-choice olfactometer. The beetle D. abdominale was attracted to fermenting banana pseudostem tissue and Cosmolure+, whereas the ant P. megacephala was attracted only to fermented pseudostem tissue. Both predators were attracted to banana pseudostem tissue that had been damaged by weevil larvae irrespective of weevil presence. Adding pheromone did not enhance predator response to volatiles from pseudostem tissue fed on by weevils. The numbers of both predators recovered with pseudostem traps in the field from banana mats with a pheromone trap were similar to those in pseudostem traps at different distance ranges from the pheromone. Our study shows that the generalist predators D. abdominale and P. megacephala use volatiles from fermented banana pseudostem tissue as the major chemical cue when searching for prey.
10. Infestation of the banana root borer among different banana plant genotypes
Directory of Open Access Journals (Sweden)
Fernando Teixeira de Oliveira
Full Text Available ABSTRACT: In this study, we aimed to investigate Cosmopolites sordidus (Coleoptera: Dryophthoridae infestation among different banana genotypes in a commercial banana orchard over the course of 30 months. Banana root borer infestation was compared in 20 banana genotypes, including five varieties and 15 hybrids. Overall, we observed that 94.17% of pest infestation cases occurred in the cortex region, and only 5.83% occurred in the central cylinder. Genotypes least sensitive to infestation were the Prata Anã (AAB and Pacovan (AAB varieties, where no damage was recorded. Among the hybrid genotypes, PV 9401 and BRS Fhia 18 showed intermediate levels of sensitivity, while BRS Tropical hybrids (AAAB, PA 9401 (AAAB, BRS Vitoria (AAAB, YB 4203 (AAAB, and Bucaneiro (AAAA were the most sensitive to attack by banana root borer. This study demonstrated that the infestation of the banana root borer varies according banana plant genotype, and the utilization of less susceptible genotypes could reduce infestation rates of C. sordidus.
11. Expression Study of Banana Pathogenic Resistance Genes
Directory of Open Access Journals (Sweden)
Fenny M. Dwivany
2016-10-01
Full Text Available Banana is one of the world's most important trade commodities. However, infection of banana pathogenic fungi (Fusarium oxysporum race 4 is one of the major causes of decreasing production in Indonesia. Genetic engineering has become an alternative way to control this problem by isolating genes that involved in plant defense mechanism against pathogens. Two of the important genes are API5 and ChiI1, each gene encodes apoptosis inhibitory protein and chitinase enzymes. The purpose of this study was to study the expression of API5 and ChiI1 genes as candidate pathogenic resistance genes. The amplified fragments were then cloned, sequenced, and confirmed with in silico studies. Based on sequence analysis, it is showed that partial API5 gene has putative transactivation domain and ChiI1 has 9 chitinase family GH19 protein motifs. Data obtained from this study will contribute in banana genetic improvement.
12. Physicochemical characterization of purple banana fiber
International Nuclear Information System (INIS)
Goncalves, A.P.B.; Guimaraes, D.H.; Miranda, C.S.; Oliveira, J.C.; Cruz, A.M.F.; Luporini, S.; Jose, N.M.
2014-01-01
Due to the environmental appeal that has grown in recent years, researches involving the use of renewable sources raw materials reaffirm this need. The vegetable fibers has excelled as promising materials with possibilities in different applications. The objective of this work is the evaluation of the physicochemical properties of banana fiber. These fibers were extracted from the banana pseudostem of a species not yet reported in the literature, Musa velutina, known as purple banana. For this experiment were used in natura fibers and processed fibers with NaOH 5% which were characterized by TGA, DSC, DRX and FTIR analysis. In the thermal analysis, both tested fibers showed good thermal properties. In DRX analysis, the processed fibers showed higher crystallinity. The use of these materials implies adding value to an agricultural waste in addition to being a more ecologically correct proposal. (author)
13. Unfolding energetics and stability of banana lectin.
Science.gov (United States)
Gupta, Garima; Sinha, Sharmistha; Surolia, Avadhesha
2008-08-01
The unfolding pathway of banana lectin from Musa paradisiaca was determined by isothermal denaturation induced by the chaotrope GdnCl. The unfolding was found to be a reversible process. The data obtained by isothermal denaturation provided information on conformational stability of banana lectin. The high values of DeltaG of unfolding at various temperatures indicated the strength of intersubunit interactions. It was found that banana lectin is a very stable and denatures at high chaotrope concentrations only. The basis of the stability may be attributed to strong hydrogen bonds of the order 2.5-3.1 A at the dimeric interface along with the presence of water bridges. This is perhaps very unique example in proteins where subunit association is not a consequence of the predominance of hydrophobic interactions. (c) 2008 Wiley-Liss, Inc.
14. Hydrolysis of alkaline pretreated banana peel
Science.gov (United States)
Fatmawati, A.; Gunawan, K. Y.; Hadiwijaya, F. A.
2017-11-01
Banana peel is one of food wastes that are rich in carbohydrate. This shows its potential as fermentation substrate including bio-ethanol. This paper presented banana peel alkaline pretreatment and enzymatic hydrolysis. The pretreatment was intended to prepare banana peel in order to increase hydrolysis performance. The alkaline pretreatment used 10, 20, and 30% w/v NaOH solution and was done at 60, 70 and 80°C for 1 hour. The hydrolysis reaction was conducted using two commercial cellulose enzymes. The reaction time was varied for 3, 5, and 7 days. The best condition for pretreatment process was one conducted using 30% NaOH solution and at 80°C. This condition resulted in cellulose content of 90.27% and acid insoluble lignin content of 2.88%. Seven-day hydrolysis time had exhibited the highest reducing sugar concentration, which was7.2869 g/L.
15. Biomethanation of banana peel and pineapple waste
Energy Technology Data Exchange (ETDEWEB)
Bardiya, N.; Somayaji, D.; Khanna, S. [Tata Energy Research Inst., New Delhi (India)
1997-10-01
Biomethanation of banana peel and pineapple wastes studied at various HRTs showed a higher rate of gas production at lower retention time. The lowest possible HRT for banana peel was 25 days, resulting in a maximum rate of gas production of 0.76 vol/vol/day with 36% substrate utilization, while pineapple-processing waste digesters could be operated at 10 days HRT, with a maximum rate of gas production of 0.93 vol/vol/day and 58% substrate utilization. For pineapple-processing waste lowering of retention time did not affect the methane content significantly; however, with banana peel an HRT below 25 days showed a drastic reduction in methane content. (author)
16. Top Earners
DEFF Research Database (Denmark)
Badel, Alejandro; Daly, Moira; Huggett, Mark
of the earnings distribution becomes thicker with age, and (3) the growth rate of earnings over the working lifetime is larger for groups with higher lifetime earnings. Models of top earners should account for these qualitative patterns and, importantly, for how they quantitatively differ across countries....
17. Effect of fermented Banana peel on Broiler Carcass
Directory of Open Access Journals (Sweden)
Koni TNI
2013-06-01
Full Text Available This experiment was conducted to examine effect of inclusion of fermented banana peel by Rhyzopus oligosporus in diets on slaughter weight, carcass weight and carcass percentage, weight and percentage abdominal fat of broiler. The experiment was done based on Completely Randomized Design with four treatments and four replications and each replication consisted of six chickens. The treatment were R0 = without banana peel fermented, R1 = 5% banana peel fermented, R2 = 10% banana peel fermented, R3 = 15% banana peel fermented. Data of the experiment were analyzed, using ANOVA and then continued with Duncan's Multiple Range Test. Result showed that level of fermented banana peel affected slaughter weight and carcass weight. However carcass persentage, weight and percentage of abdominal fat was not affected by treatment. Banana peel fermented by Rhizopus oligosporus could can be used maximally 10% in broiler ration.
18. banana juice as an alternative energy source for banana in vitro
African Journals Online (AJOL)
ACSS
2015-02-23
Feb 23, 2015 ... Corresponding author: [email protected], [email protected] ... However, the cost of tissue culture grade energy sources is high, thus making tissue ..... Treatment (banana juice from different Cvs at 50 ml l-1).
19. Olfactory responses of banana weevil predators to volatiles from banana pseudostem tissue and synthetic pheromone
NARCIS (Netherlands)
Tinzaara, W.; Gold, C.S.; Dicke, M.; Huis, van A.
2005-01-01
As a response to attack by herbivores, plants can emit a variety of volatile substances that attract natural enemies of these insect pests. Predators of the banana weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) such as Dactylosternum abdominale (Coleoptera: Hydrophilidae) and
20. Substituting Wheat Flour with Banana Skin Flour from Mixture Various Skin Types of Banana on Making Donuts
OpenAIRE
Renny Futeri; Pharmayeni Pharmayeni
2014-01-01
Tropical forest plants is a very rich source of chemical compounds or bioactive efficacious . Many of the compounds potential as a source of raw materials in food processing . One is the banana plant , West Sumatra Padang and Bukittinggi is one area in Indonesia with banana . Generally people in West Sumatra just consume or eat the fruit and throw banana skin just because it is considered as waste ( waste banana peel ) . When the banana peel waste is left alone so do not rule out the possibil...
1. Phenylphenalenones Accumulate in Plant Tissues of Two Banana Cultivars in Response to Herbivory by the Banana Weevil and Banana Stem Weevil.
Science.gov (United States)
Hölscher, Dirk; Buerkert, Andreas; Schneider, Bernd
2016-08-25
Phenylphenalenone-type compounds accumulated in the tissues of two banana cultivars-Musa acuminata cv. "Grande Naine" (AAA) and Musa acuminata × balbisiana Colla cv. "Bluggoe" (ABB)-when these were fed on by the banana weevil (Cosmopolites sordidus (Germ.) (Coleoptera: Curculionidae)) and the banana stem weevil (Odoiporus longicollis (Oliver) (Coleoptera: Curculionidae)). The chemical constituents of the banana material were separated by means of chromatographic techniques and identified by NMR spectroscopy. One new compound, 2-methoxy-4-phenylphenalen-1-one, was found exclusively in the corm material of "Bluggoe" that had been fed on by the weevils.
2. Phenylphenalenones Accumulate in Plant Tissues of Two Banana Cultivars in Response to Herbivory by the Banana Weevil and Banana Stem Weevil
Directory of Open Access Journals (Sweden)
Dirk Hölscher
2016-08-01
Full Text Available Phenylphenalenone-type compounds accumulated in the tissues of two banana cultivars—Musa acuminata cv. “Grande Naine” (AAA and Musa acuminata × balbisiana Colla cv. “Bluggoe” (ABB—when these were fed on by the banana weevil (Cosmopolites sordidus (Germ. (Coleoptera: Curculionidae and the banana stem weevil (Odoiporus longicollis (Oliver (Coleoptera: Curculionidae. The chemical constituents of the banana material were separated by means of chromatographic techniques and identified by NMR spectroscopy. One new compound, 2-methoxy-4-phenylphenalen-1-one, was found exclusively in the corm material of “Bluggoe” that had been fed on by the weevils.
3. Review on postharvest technology of banana fruit
African Journals Online (AJOL)
mu
2013-02-13
Feb 13, 2013 ... The aim of this review is in threefold: First, to explore the effect of different preharvest treatments on postharvest ... biochemical changes in banana during development, maturation, ripening and storage were reviewed. Third, postharvest ..... at full mature stage for local domestic market (Gowen,. 1995).
4. introduction and evaluation of improved banana cultivars
African Journals Online (AJOL)
jen
important parameters in banana marketing thus the reason they were considered in this study. The data were analysed using Statistics Analysis. System (SAS) for analysis of variance (ANOVA) and means were separated by the Student-. Newman-Keuls test. RESULTS. The differences in growth parameters of the 10.
5. Relative susceptibility of banana cultivars to Xanthomonas ...
African Journals Online (AJOL)
STORAGESEVER
2009-10-19
Oct 19, 2009 ... African Journal of Biotechnology Vol. 8 (20), pp. 5343-5350, 19 ... and Central Africa. The disease was first reported about. 40 years ago in Ethiopia on Ensete, which is closely related to banana (Yirgou et al., 1968). Outside Ethiopia,. BXW was ... Economic impact of the disease is manifested as result of ...
6. Towards improving highland ban.anas
African Journals Online (AJOL)
The most fertile land races belonged to 'Nakabululu' and 'Nfuuka' clone sets. Viable seeds were obtained from several land races indicating that genetic improvement ofthese highland bananas through cross breeding is possible. The fertile Iandraces should be cross-pollinated with improved diploids to produce resistant ...
7. Love Is Like a Squished Banana
Science.gov (United States)
Brown, Stephen
1976-01-01
An unemployed poet obtained a CETA public service job as a teacher's aide in Marin County, California, where he has guided elementary children's imaginative projects. His experiences are described. He has published a volume of the children's verse under the title "Love Is Like a Squished Banana." (AJ)
8. Ecuadorian Banana Farms Should Consider Organic Banana with Low Price Risks in Their Land-Use Portfolios
Science.gov (United States)
Castro, Luz Maria; Calvas, Baltazar; Knoke, Thomas
2015-01-01
Organic farming is a more environmentally friendly form of land use than conventional agriculture. However, recent studies point out production tradeoffs that often prevent the adoption of such practices by farmers. Our study shows with the example of organic banana production in Ecuador that economic tradeoffs depend much on the approach of the analysis. We test, if organic banana should be included in economic land-use portfolios, which indicate how much of the land is provided for which type of land-use. We use time series data for productivity and prices over 30 years to compute the economic return (as annualized net present value) and its volatility (with standard deviation as risk measure) for eight crops to derive land-use portfolios for different levels of risk, which maximize economic return. We find that organic banana is included in land-use portfolios for almost every level of accepted risk with proportions from 1% to maximally 32%, even if the same high uncertainty as for conventional banana is simulated for organic banana. A more realistic, lower simulated price risk increased the proportion of organic banana substantially to up to 57% and increased annual economic returns by up to US187 per ha. Under an assumed integration of both markets, for organic and conventional banana, simulated by an increased coefficient of correlation of economic return from organic and conventional banana (ρ up to +0.7), organic banana holds significant portions in the land-use portfolios tested only, if a low price risk of organic banana is considered. We conclude that uncertainty is a key issue for the adoption of organic banana. As historic data support a low price risk for organic banana compared to conventional banana, Ecuadorian farmers should consider organic banana as an advantageous land-use option in their land-use portfolios. PMID:25799506 9. Autotranspordifirmade TOP 100 Index Scriptorium Estoniae 2006-01-01 Ilmunud ka: Delovõje Vedomosti : Transport i Logistika 29. nov. lk. 10-11. Autofirmade TOP 100. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Majandustegevuse kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Autotranspordifirmade üld- ja finantsandmed. Delovõje Vedomosti : Transport i Logistika sisaldab tabelit Autofirmade TOP 100 10. Physicochemical evaluation of cooking and dessert bananas (Musa sp.) varieties OpenAIRE Rosales-Reynoso, O. Lidia; Agama-Acevedo, Edith; Aguirre-Cruz, Andres; Bello-Perez, Luis A.; Dufour, Dominique; Gibert, Olivier 2014-01-01 In México, banana (Musa sp.) varieties ate used for human consumption as well as for traditional medicine, but the literature lacks information on local diversity and functional justification for their use. Three varieties of dessert bananas (Valery, Morado, and Enano) and one cooking banana (Macho) were collected in a commercial farm in Tuxtepec, Oaxaca, México, at the agronomic maturity stage, and they were physically and chemically evaluated. A random sampling, ANOVA, and Tukey tests were ... 11. Autotranspordi TOP aastal 2007 Index Scriptorium Estoniae 2008-01-01 TOP 50. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Marika Roomere. Täisteenuse pakkumine kergitas tulemusi; Jupiter Plus otsib järjest uusi kasvuvõimalusi; EST-Trans Kaubaveod teenib kasumit toiduvedamisega 12. Koolitusfirmade TOP aastal 2007 Index Scriptorium Estoniae 2008-01-01 Koolitusfirmade TOP. Vt. samas: Käibe TOP 10; Käibekasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Signe Sillasoo. Invicta tahab lähiaastail laieneda Eestis ja mujalgi; Ketlin Priilinn. Addenda kasutas ära majanduse soodsa seisu. Kommenteerib Heli Sõmer. Juhtide hoiakute muutmisega tõus esikolmikusse 13. Audiitorfirmade TOP aastal 2007 Index Scriptorium Estoniae 2008-01-01 Audiitorfirmade TOP 51. Vt. samas: Urve Vilk. Audiitoriteni pole majanduslangus jõudnud; Intervjuu I.S. Audiitorteenuste OÜ omaniku Irina Somovaga; Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10 14. Koolitusfirmade TOP aastal 2006 Index Scriptorium Estoniae 2007-01-01 Koolitusfirmade TOP. Vt. samas: Käibe TOP10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Varade tootlikkuse TOP 10; Pille Rõivas. Võtmetegur meeskond; Vain & Partnerid: uudsus peitub sisus; Kristo Kiviorg. Teel Baltimaade koolitajate tippu 15. Top reconstruction and boosted top experimental overview CERN Document Server Skinnari, Louise 2015-01-01 An overview of techniques used to reconstruct resolved and boosted top quarks is presented. Techniques for resolved top quark reconstruction include kinematic likelihood fitters and pseudo- top reconstruction. Many tools and methods are available for the reconstruction of boosted top quarks, such as jet grooming techniques, jet substructure variables, and dedicated top taggers. Different techniques as used by ATLAS and CMS analyses are described and the performance of different variables and top taggers are shown. 16. Generalized ripple-banana transport in a tokamak International Nuclear Information System (INIS) Yushmanov, P.N. 1983-01-01 The paper considers the transport of banana particles in a rippled magnetic field over the entire energy range. It is shown that all familiar regimes of ripple transport - ripple-plateau, banana-drift and stochastic - can be described in a unified manner. The general expression obtained for the rippled fluxes of banana particles describes, apart from the already familiar regimes, also the as yet unstudied energy region between the drift and stochastic regimes. A generalized ripple-banana thermal conductivity coefficient, chisub(i)sup(RB), is calculated. (author) 17. Suitability of banana peels for biogas production Energy Technology Data Exchange (ETDEWEB) Meseguer, C.M.; Silesky, F.; Chacon, G. 1983-01-01 Banana (Musa cavendishii) peel in the ripe state (yellow with sufficient spots) has the potential to produce by anaerobic fermentation 0.22 plus or minus 0.03 cubic m biogas/kg dry material. Inhibition of the process can be prevented if the peel is pretreated by oxidation or if the process is carried out at approximately 35 degrees. The inoculate used must be acclimated to the medium. 18. Physicochemical and sensorial quality of banana genotypes Directory of Open Access Journals (Sweden) Ronielli Cardoso Reis 2016-03-01 Full Text Available Despite the diversity of banana varieties in Brazil, only a few cultivars have the proper agronomic traits and fruit quality for commercial exploitation. This study aimed at evaluating the physicochemical traits and sensorial acceptance of banana genotypes, in order to identify those with potential for commercial growing. Six improved banana genotypes were assessed (BRS Maravilha, PC 0101, FHIA 18, TM 2803, YB 4203 and BRS Caipira, as well as three commercial cultivars (Grand Naine, Pacovan and Prata Anã. Analyses of peel and pulp color, peel thickness, pulp yield, moisture, pH, soluble solids, titratable acidity, total carotenoids and sensorial acceptance were performed. The BRS Maravilha, FHIA 18, YB 4203 and BRS Caipira genotypes presented physicochemical traits similar to the Grand Naine, Pacovan and Prata Anã commercial cultivars. The BRS Maravilha and TM 2803 genotypes had sensorial acceptance similar to the Prata Anã and Grand Naine cultivars, and are therefore promising for commercial growing, with the advantage of being resistant to the black Sigatoka and Panama disease. 19. Temperature effects on peel spotting in "Sucrier banana" fruit NARCIS (Netherlands) Trakulnaleumsai, C.; Ketsa, S.; Doorn, van W.G. 2006-01-01 Banana fruit of the cultivar Sucrier¿ (Musa acuminata, AA Group) develops peel spotting at a relatively early stage of development (when the peel is about as slightly more yellow than green). Holding ripening bananas at 15 and 18 °C instead of room temperature (26¿27 °C) only temporarily reduced 20. Determinants of market production of cooking banana in Nigeria ... African Journals Online (AJOL) The factors that influence farmers' decisions to produce cooking banana for market in southeast Nigeria were examined. Data were collected from a ... Results of the study indicate that about 80% of the farmers interviewed produce cooking banana both for household consumption and for sale. The proportion of cooking ... 1. Small scale banana farmers' awareness level and adoption of ... African Journals Online (AJOL) Descriptive statistics and binary logit regression were employed for data analyses. The results show that although majority of the farmers (96.67%) were aware of and had access to improved banana varieties, only 15.83% of them adopted the use of improved planting materials. Gros mitchel, Cavendish and sweet bananas ... 2. Farmer acceptance of introduced banana genotypes in Uganda ... African Journals Online (AJOL) The same cultivars were acceptable mainly as dessert but also as cooking bananas during food shortages in central and western parts, especially, in areas where the growing of traditional cultivars is progressively declining. There was little interest in the new bananas in western parts of the country. Major considerations for ... 3. Cultural control of banana weevils in Ntungamo, southwestern Uganda NARCIS (Netherlands) Okech, S.H.; Gold, C.S.; Bagamba, F.; Masanza, M.; Tushemereirwe, W.; Ssennyonga, J. 2005-01-01 The International Institute of Tropical Agriculture and the Uganda National Banana Research Programme tested and evaluated selected cultural management options for the banana weevil through on-farm farmer participatory research in Ntungamo district, Uganda between 1996 and 003. A farmer adoption 4. Effects of relative humidity on banana fruit drop NARCIS (Netherlands) Saengpook, C.; Ketsa, S.; Doorn, van W.G. 2007-01-01 Commercial ripening of banana fruit occurs at high relative humidity (RH), which prevents browning of damaged skin areas. In experiments with ripening at high RH (94 ± 1%) the individual fruit (fingers) of Sucrier¿ (Musa acuminata, AA Group) banana exhibited a high rate of drop. The falling off of 5. Fermentation of Foc TR4-infected bananas and Trichoderma spp. Science.gov (United States) Yang, J; Li, B; Liu, S W; Biswas, M K; Liu, S; Wei, Y R; Zuo, C W; Deng, G M; Kuang, R B; Hu, C H; Yi, G J; Li, C Y 2016-10-17 Fusarium wilt (also known as Panama disease) is one of the most destructive banana diseases, and greatly hampers the global production of bananas. Consequently, it has been very detrimental to the Chinese banana industry. An infected plant is one of the major causes of the spread of Fusarium wilt to nearby regions. It is essential to develop an efficient and environmentally sustainable disease control method to restrict the spread of Fusarium wilt. We isolated Trichoderma spp from the rhizosphere soil, roots, and pseudostems of banana plants that showed Fusarium wilt symptoms in the infected areas. Their cellulase activities were measured by endoglucanase activity, β-glucosidase activity, and filter paper activity assays. Safety analyses of the Trichoderma isolates were conducted by inoculating them into banana plantlets. The antagonistic effects of the Trichoderma spp on the Fusarium pathogen Foc tropical Race 4 (Foc TR4) were tested by the dual culture technique. Four isolates that had high cellulase activity, no observable pathogenicity to banana plants, and high antagonistic capability were identified. The isolates were used to biodegrade diseased banana plants infected with GFP-tagged Foc TR4, and the compost was tested for biological control of the infectious agent; the results showed that the fermentation suppressed the incidence of wilt and killed the pathogen. This study indicates that Trichoderma isolates have the potential to eliminate the transmission of Foc TR4, and may be developed into an environmentally sustainable treatment for controlling Fusarium wilt in banana plants. 6. REACTION OF Musa balbisiana TO BANANA BACTERIAL WILT ... African Journals Online (AJOL) Prof. Adipala Ekwamu 2Makerere University, Department of Agricultural Production, P. O. Box 7062, Kampala, Uganda. Corresponding author: [email protected]. (Received 7 February, 2012; accepted 3 September, 2013). ABSTRACT. Banana bacterial wilt (Xanthomonas campestris) is an emerging disease of bananas in Uganda. 7. Ethical perception of human gene in transgenic banana | Amin ... African Journals Online (AJOL) Transgenic banana has been developed to prevent hepatitis B through vaccination. Its production seems to be an ideal alternative for cheaper vaccines. The objective of this paper is to assess the ethical perception of transgenic banana which involved the transfer of human albumin gene, and to compare their ethical ... 8. Egg and banana sign of severe pulmonary arterial hypertension. Science.gov (United States) Veean, Satyam; Nixon, William; Keshavamurthy, Jayanth 2018-01-01 The egg and banana sign can be seen on chest computed tomography (CT) in patients with severe pulmonary arterial hypertension (PAH). It is identified by the presence of the pulmonary artery (PA) lateral to the aortic arch with the aortic arch being described as the banana and the PA as the egg. 9. Agronomic performance of five banana cultivars under protected cultivation Science.gov (United States) Banana has been grown both in open-field and protected cultivation in Turkey. So far protected cultivation is very popular due to the high yield and quality. The objective of the study was to evaluate agronomic performance of five new banana cultivars under plastic greenhouse. ‘MA 13’, ‘Williams’, ‘... 10. Leaf anatomy of genotypes of banana plant grown under coloured ... African Journals Online (AJOL) This study aimed to evaluate the effect of spectral light quality on different anatomical features of banana tree plantlets grown under coloured shade nets. Banana plants of five genotypes obtained from micropropagation, were grown under white, blue, red and black nets, with shade of 50%, in a completely randomized ... 11. Banana peel: A novel substrate for cellulase production under solid ... African Journals Online (AJOL) These results indicated that banana peel provided necessary nutrients for cell growth and cellulase synthesis. It can be used as a potential substrate for cellulase production by T. viride GIM 3.0010 under solid-state fermentation. To the best of our knowledge, this is the first report on cellulase production using banana peel. 12. Production and evaluation of precooked dehydrated unripe banana ... African Journals Online (AJOL) No significant change in total aerobic counts or yeasts and moulds counts occurred in dehydrated banana slices packaged in 250 gauge polyethylene bags and stored at ambient temperature for 3 months. The slices were found to be high in starch (~68.5%) and minerals. When shallow fried, the dehydrated banana slices ... 13. Development of an in vitro culture system adapted to banana ... African Journals Online (AJOL) STORAGESEVER 2009-06-17 Jun 17, 2009 ... The beneficial impact of arbuscular mycorrhizal (AM) fungi on banana nutrition and resistance against abiotic and biotic stresses is well documented. However, most studies were conducted under greenhouse or field conditions and none reported the life cycle of the AM fungi on banana roots. It is obvious ... 14. Development of an in vitro culture system adapted to banana ... African Journals Online (AJOL) The beneficial impact of arbuscular mycorrhizal (AM) fungi on banana nutrition and resistance against abiotic and biotic stresses is well documented. However, most studies were conducted under greenhouse or field conditions and none reported the life cycle of the AM fungi on banana roots. It is obvious that any system ... 15. Urban consumer willingness to pay for introduced dessert bananas ... African Journals Online (AJOL) ... therefore have a market potential. It is recommended that market development activities including organising and training farmers in improved agronomic methods, handling bananas for local markets; and promotional studies of the introduced dessert bananas among the urban consumers be done to widen their demand. 16. Banana Xanthomonas wilt in Ethiopia: Occurrence and insect vector ... African Journals Online (AJOL) Bacterial wilt caused by Xanthomonas vasicola pv. musacearum (Xvm) is an important disease of enset and banana in south and south-western Ethiopia where, the diversity of the insect fauna on banana inflorescences was unknown and the role of insects as vectors of the disease had not been studied. The objectives of ... 17. Preliminary evaluation of improved banana varieties in Mozambique ... African Journals Online (AJOL) Banana (Musa spp.) production in Mozambique is largely confined to the Cavendish variety that is eaten as a dessert. On the other hand, banana is a staple food crop in many countries in sub-Saharan Africa. The introduction of a range of high yielding and disease resistant cooking and dessert varieties in Mozambique ... 18. Eggshells – assisted hydrolysis of banana pulp for biogas production African Journals Online (AJOL) KARAKANA In this study, pretreatment of banana pulp using eggshells in both calcined and un-calcined forms to examine the ... Key words: Anaerobic digestion, banana pulp hydrolysis biogas, eggshells. .... obtain fine powder. ..... using pig waste and cassava peels. ... from bioethanol waste: the effect of pH and urea addition to biogas. 19. Genetic Diversity Among East African Highland Bananas For ... African Journals Online (AJOL) There are 84 distinct cultivars of highland bananas (Musa spp.) in Uganda, grouped in five clone sets and it is not known which among these are female fertile. The objective of the study reported herein was to identify female fertile highland bananas that can be used in a cross breeding program and to determine the ... 20. Screening of in vitro derived mutants of banana against nematodes ... African Journals Online (AJOL) The rest of the mutants namely Ro Im V4 6-1-2 and Si Im V4 6-2-5 were found to be susceptible to nematodes. The resistant and moderately resistant mutants of banana could be further used in breeding programmes as well as being recognized as potential cultivars of commerce. Key words: Banana, nematode, resistance, ... 1. 33 CFR 334.570 - Banana River near Orsino, Fla.; restricted area. Science.gov (United States) 2010-07-01 ... 33 Navigation and Navigable Waters 3 2010-07-01 2010-07-01 false Banana River near Orsino, Fla... THE ARMY, DEPARTMENT OF DEFENSE DANGER ZONE AND RESTRICTED AREA REGULATIONS § 334.570 Banana River near Orsino, Fla.; restricted area. (a) The area. That part of Banana River N of the NASA Banana River... 2. Arvutifirmade TOP 101 aastal 2004 Index Scriptorium Estoniae 2005-01-01 Arvutifirmade TOP 101; Käibe TOP 20; Käibe kasvu TOP 15; Kasumi TOP 15; Rentaabluse TOP 20; Kasumi kasvu TOP 15; Omakapiali tootlikkuse TOP 15; Eesti arvutifirmade finantsandmed; Arvutifirmade üldandmed 3. Jaekaubanduse TOP 100 aastal 2001 Index Scriptorium Estoniae 2002-01-01 TOP 100. Käibe TOP 30. Käibe kasvu TOP 30. Kasumi TOP 30. Kasumi kasvu TOP 30. Rentaabluse TOP 30. Varade tootlikkuse TOP 30. Jaekaubandusettevõtete finantsseadmed. Jaekaubandusettevõtete üldandmed 4. Jaekaubandusettevõtete TOP 70 Index Scriptorium Estoniae 2005-01-01 Jaekaubandusettevõtete TOP 70; Käibe TOP 25; Kasumi TOP 25; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Omakapitali tootlikkuse TOP 20; Jaekaubandusettevõtete üld- ja finantsandmed 5. Majutusasutuste TOP 40 aastal 2002 Index Scriptorium Estoniae 2003-01-01 Majutusasutuste TOP 40 aastal 2002. Käibe TOP 40. Kasumi TOP 40. Käibe kasvu TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Majutusasutuste üldandmed. Majutusasutuste finantsandmed 6. Identification of genes differentially expressed during ripening of banana. Science.gov (United States) Manrique-Trujillo, Sandra Mabel; Ramírez-López, Ana Cecilia; Ibarra-Laclette, Enrique; Gómez-Lim, Miguel Angel 2007-08-01 The banana (Musa acuminata, subgroup Cavendish 'Grand Nain') is a climacteric fruit of economic importance. A better understanding of the banana ripening process is needed to improve fruit quality and to extend shelf life. Eighty-four up-regulated unigenes were identified by differential screening of a banana fruit cDNA subtraction library at a late ripening stage. The ripening stages in this study were defined according to the peel color index (PCI). Unigene sequences were analyzed with different databases to assign a putative identification. The expression patterns of 36 transcripts confirmed as positive by differential screening were analyzed comparing the PCI 1, PCI 5 and PCI 7 ripening stages. Expression profiles were obtained for unigenes annotated as orcinol O-methyltransferase, putative alcohol dehydrogenase, ubiquitin-protein ligase, chorismate mutase and two unigenes with non-significant matches with any reported sequence. Similar expression profiles were observed in banana pulp and peel. Our results show differential expression of a group of genes involved in processes associated with fruit ripening, such as stress, detoxification, cytoskeleton and biosynthesis of volatile compounds. Some of the identified genes had not been characterized in banana fruit. Besides providing an overview of gene expression programs and metabolic pathways at late stages of banana fruit ripening, this study contributes to increasing the information available on banana fruit ESTs. 7. Development of environmental friendly lost circulation material from banana peel Science.gov (United States) Sauki, Arina; Hasan, Nur â.Izzati; Naimi, Fardelen Binti Md; Othman, Nur Hidayati 2017-12-01 Loss of expensive mud could lead to major financial problem in executing a drilling project and is one of the biggest problems that need to be tackled during drilling. Synthetic Based Mud (SBM) is the most stable state of the art drilling mud used in current drilling technologies. However, the problem with lost circulation is still inevitable. The focus of this project is to develop a new potential waste material from banana peel in order to combat lost circulation in SBM. Standard industrial Lost Circulation Material (LCM) is used to compare the performance of banana peel as LCM in SBM. The effects of different sizing of banana peels (600 micron, 300 micron and 100 micron) were studied on the rheological and filtration properties of SBM and the bridging performance of banana peel as LCM additive. The tests were conducted using viscometer, HTHP filter press and sand bed tester. Thermal analysis of banana peel was also studied using TGA. According to the results obtained, 300 and 100 micron size of banana peel LCM exhibited an improved bridging performance by 65% as compared to industrial LCM. However, banana peel LCM with the size of 600 micron failed to act as LCM due to the total invasion of mud into the sand bed. 8. Batch biomethanation of banana trash and coir path Energy Technology Data Exchange (ETDEWEB) Deivanai, K.; Bai, R.K. [Madurai Kamaraj Univ. (India) 1995-08-01 Anaerobic digestion of banana trash and coir pith was carried out for a period of one month by batch digestion. During biomethanation reduction of total- and volatile-solids was, respectively, 25.3 and 39.6% in banana trash and 13.6 and 21.6% in coir pith. A production of 9.22 l and 1.69 l (per kg TS added) of biogas with average methane content of 72 and 80% was achieved from banana trash and coir pith, respectively. (author) 9. Tissue culture regeneration and radiation induced mutagenesis in banana International Nuclear Information System (INIS) Kulkarni, V.M.; Ganapathi, T.R. 2009-01-01 Radiation induced mutagenesis is an important tool for banana genetic improvement. At BARC, protocols for shoo-tip multiplication of commercial banana varieties have been developed and transferred to user agencies for commercial production. Excellent embryogenic cell suspensions were established in banana cvs. Rasthali and Rajeli, and were maintained at low temperatures for long-term storage. Normal plantlets were successfully regenerated from these cell suspensions. The cell suspensions and shoot-tip cultures were gamma-irradiated for mutagenesis. The mutagenized populations were field screened and a few interesting mutants have been isolated. The existence of genetic variation was confirmed using DNA markers. Further evaluation of these mutants is in progress. (author) 10. The effects of banana peel preparations on the properties of banana peel dietary fibre concentrate Directory of Open Access Journals (Sweden) Phatcharaporn Wachirasiri 2010-01-01 Full Text Available Four different preparation methods of banana peel, dry milling, wet milling, wet milling and tap water washing, and wet milling and hot water washing were investigated on their effects on the chemical composition and properties of the banana peel dietary fibre concentrate (BDFC. The dry milling process gave the BDFC a significant higher fat, protein, and starch content than the wet milling process, resulting in a lower water holding capacity (WHC and oil holding capacity(OHC. Washing after wet milling could enhance the concentration of total dietary fibre by improving the removal of protein and fat. Washing with hot water after wet milling process caused a higher loss of soluble fibre fraction, resulting in a lower WHC and OHC of the obtained BDFC when compared to washing with tap water. Wet milling and tap water washing gave the BDFC the highest concentration of total and soluble dietary fibre, WHC and OHC. 11. Banana Musa tissue culture plants enhanced by endophytic fungi African Journals Online (AJOL) Mo Merging biotechnology with biological control: Banana Musa tissue culture plants enhanced by endophytic .... While working in the laminar flow cabinet, sterile filter papers were placed in ..... University of Bonn, Bonn, Germany. Niere, B., 2001. 12. Policy Issues in the Structure, Conduct and Performance of Banana ... African Journals Online (AJOL) Madukwe **Department of Agricultural Economics University of Nigeria, Nsukka ... stated that banana is playing a crucial food security role in many developing countries. ... upon the infrastructure and social services of a country's towns and cities ... 13. Influence of triadimefon on the growth and development of banana ... African Journals Online (AJOL) dessert banana cultivars (Hindi, Basrai and Williams) were affected compared to the control. The optimum culture conditions for root formation were obtained in the case of sub-culturing. The excised shoot cultures into Murashige and Skoog ... 14. Micropropagation of some Malaysian banana and plantain (Musa sp ... African Journals Online (AJOL) USER 2010-04-19 Apr 19, 2010 ... As one of the origins of bananas, Malaysia has a great variety of them, that is, ... reduces contamination rate during micropropagation as compared to soil ..... 42. Faostat (2005). Food and Agricultural Organization of the United. 15. urban consumer willingness to pay for introduced dessert bananas African Journals Online (AJOL) Administrator National Agricultural Research Organization (NARO) P.O. Box 7065, Kampala, ... the introduced dessert bananas among the urban consumers be done to widen their demand. Key Words: Fusarium wilt, Gros Michel, hedonic model, Musa spp. 16. Acetylation and characterization of banana (Musa paradisiaca) starch. Science.gov (United States) Bello-Pérez, L A; Contreras-Ramos, S M; Jìmenez-Aparicio, A; Paredes-López, O 2000-01-01 Banana native starch was acetylated and some of its functional properties were evaluated and compared to corn starch. In general, acetylated banana starch presented higher values in ash, protein and fat than corn acetylated starch. The modified starches had minor tendency to retrogradation assessed as % transmittance of starch pastes. At high temperature acetylated starches presented a water retention capacity similar to their native counterpart. The acetylation considerably increased the solubility of starches, and a similar behavior was found for swelling power. When freeze-thaw stability was studied, acetyl banana starch drained approximately 60% of water in the first and second cycles, but in the third and fourth cycles the percentage of separated water was low. However, acetyl corn starch showed lower freeze-thaw stability than the untreated sample. The modification increased the viscosity of banana starch pastes. 17. Influence of triadimefon on the growth and development of banana ... African Journals Online (AJOL) SAM 2014-04-16 Apr 16, 2014 ... cultures of the three-dessert banana cultivars (Hindi, Basrai and Williams) were affected compared to ..... includes reduce in surface area of leaves and .... reflect a type of particular stress conditions exhibit ... Emirates J. Food. 18. Characterisation of colletotrichum species associated with anthracnose of banana. Science.gov (United States) Zakaria, Latiffah; Sahak, Shamsiah; Zakaria, Maziah; Salleh, Baharuddin 2009-12-01 A total of 13 Colletotrichum isolates were obtained from different banana cultivars (Musa spp.) with symptoms of anthracnose. Colletotrichum isolates from anthracnose of guava (Psidium guajava) and water apple (Syzygium aqueum) were also included in this study. Based on cultural and morphological characteristics, isolates from banana and guava were identified as Colletotrichum musae and from water apple as Colletotrichum gloeosporiodes. Isolates of C. musae from banana and guava had similar banding patterns in a randomly amplified polymorphic DNA (RAPD) analysis with four random primers, and they clustered together in a UPGMA analysis. C. gloeosporiodes from water apple was clustered in a separate cluster. Based on the present study, C. musae was frequently isolated from anthracnose of different banana cultivars and the RAPD banding patterns of C. musae isolates were highly similar but showed intraspecific variations. 19. In Vivo Digestibility of Molasses-Treated Fresh Banana Leaves ... African Journals Online (AJOL) treated fresh banana leaves in West African Dwarf sheep was conducted at the Teaching and Research Farm and Animal Nutrition Laboratory of the University of Dschang between August and September 2009. For this, six sheep were used and ... 20. Sustainable Banana Production and Pesticides in Costa Rica | IDRC ... International Development Research Centre (IDRC) Digital Library (Canada) Large multinational producers employ thousands of workers, who live near ... in the banana industry is high and constitutes a health hazard for the farm workers, ... and assess the effects of these levels on the health and neurodevelopment of ... 1. Reisifirmade TOP 40 aastal 2002 Index Scriptorium Estoniae 2003-01-01 Reisifirmade TOP 40 aastal 2002. Reisifirmade TOP-i esikümme. Käibe TOP 40. Kasumi TOP 40. Käibe kasvu TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Reisifirmade üldandmed. Reisifirmade finantsandmed. Tehnilise käibe alusel arvutatud edetabelid: Reisifirmade TOP 25; Käibe TOP 40; Rentaabluse TOP 10; Käibe kasvu TOP 10 2. Visually suboptimal bananas: How ripeness affects consumer expectation and perception. Science.gov (United States) Symmank, Claudia; Zahn, Susann; Rohm, Harald 2018-01-01 One reason for the significant amount of food that is wasted in developed countries is that consumers often expect visually suboptimal food as being less palatable. Using bananas as example, the objective of this study was to determine how appearance affects consumer overall liking, the rating of sensory attributes, purchase intention, and the intended use of bananas. The ripeness degree (RD) of the samples was adjusted to RD 5 (control) and RD 7 (more ripened, visually suboptimal). After preliminary experiments, a total of 233 participants were asked to judge their satisfaction with the intensity of sensory attributes that referred to flavor, taste, and texture using just-about-right scales. Subjects who received peeled samples were asked after tasting, whereas subjects who received unpeeled bananas judged expectation and, after peeling and tasting, perception. Expected overall liking and purchase intention were significantly lower for RD 7 bananas. Purchase intention was still significantly different between RD 5 and RD 7 after tasting, whereas no difference in overall liking was observed. Significant differences between RD 5 and RD 7 were observed when asking participants for their intended use of the bananas. Concerning the sensory attributes, penalty analysis revealed that only the firmness of the RD 7 bananas was still not just-about-right after tasting. The importance that consumers attribute to the shelf-life of food had a pronounced impact on purchase intention of bananas with different ripeness degree. In the case of suboptimal bananas, the results demonstrate a positive relationship between the sensory perception and overall liking and purchase intention. Convincing consumers that visually suboptimal food is still tasty is of high relevance for recommending different ways of communication. Copyright © 2017 Elsevier Ltd. All rights reserved. 3. MECHANICAL CHARACTERIZATION AND ANALYSIS OF RANDOMLY DISTRIBUTED SHORT BANANA FIBER REINFORCED EPOXY COMPOSITES Directory of Open Access Journals (Sweden) R. K. Misra 2014-03-01 Full Text Available Short banana fiber reinforced composites have been prepared in laboratory to determine mechanical properties. It has been observed that as soon as the percentage of the banana fiber increases slightly there is a tremendous increase in ultimate tensile strength, % of strain and young modulus of elasticity. Reinforcement of banana fibers in epoxy resin increases stiffness and decreases damping properties of the composites. Therefore, 2.468% banana fiber reinforced composite plate stabilizes early as compared to 7.7135 % banana fiber reinforced composite plate but less stiff as compared to 7.7135 % banana fiber reinforced composite plate 4. Banana peel: an effective biosorbent for aflatoxins. Science.gov (United States) Shar, Zahid Hussain; Fletcher, Mary T; Sumbal, Gul Amer; Sherazi, Syed Tufail Hussain; Giles, Cindy; Bhanger, Muhammad Iqbal; Nizamani, Shafi Muhammad 2016-05-01 This work reports the application of banana peel as a novel bioadsorbent for in vitro removal of five mycotoxins (aflatoxins (AFB1, AFB2, AFG1, AFG2) and ochratoxin A). The effect of operational parameters including initial pH, adsorbent dose, contact time and temperature were studied in batch adsorption experiments. Scanning electron microscopy (SEM), Fourier transform infrared spectroscopy (FTIR) and point of zero charge (pHpzc) analysis were used to characterise the adsorbent material. Aflatoxins' adsorption equilibrium was achieved in 15 min, with highest adsorption at alkaline pH (6-8), while ochratoxin has not shown any significant adsorption due to surface charge repulsion. The experimental equilibrium data were tested by Langmuir, Freundlich and Hill isotherms. The Langmuir isotherm was found to be the best fitted model for aflatoxins, and the maximum monolayer coverage (Q0) was determined to be 8.4, 9.5, 0.4 and 1.1 ng mg(-1) for AFB1, AFB2, AFG1 and AFG2 respectively. Thermodynamic parameters including changes in free energy (ΔG), enthalpy (ΔH) and entropy (ΔS) were determined for the four aflatoxins. Free energy change and enthalpy change demonstrated that the adsorption process was exothermic and spontaneous. Adsorption and desorption study at different pH further demonstrated that the sorption of toxins was strong enough to sustain pH changes that would be experienced in the gastrointestinal tract. This study suggests that biosorption of aflatoxins by dried banana peel may be an effective low-cost decontamination method for incorporation in animal feed diets. 5. Biossorption of uranium on banana pith International Nuclear Information System (INIS) Boniolo, Milena Rodrigues 2008-01-01 Banana pith was characterized by Fourier Transformed Infrared Spectroscopy and Scanning Electron Microscopy, and investigated as a low cost bio sorbent for the removal of uranium ions from nitric solutions. Influences variable as were studied: adsorbent particle size, contact time, pH and temperature were studied. The removal percentage was increased from 13 to 57% when the particle size was decreased from 6.000 to 0.074 mm. The determined contact time was 40 minutes with 60% mean removal. The removal was increased from 40 to 55% when the pH varied from 2 to 5. The Langmuir and Freundlich linear isotherm models were applied to describe the adsorption equilibrium. The kinetic of the process was studied using the pseudo-first order and pseudo-second order models. Thermodynamics parameters such as ΔG, ΔS and ΔH were calculated. In concentration range of 50 - 500 mg.L -1 , the adsorption process was described better by the Freundlich equation. The adsorption capacity at equilibrium of uranium ions was determined from the Langmuir equation, and it was found 11.50 mg.g -1 at 25 ± 2 deg C. The kinetic was better represented by the pseudo-second order model. The bio sorption process for uranium removal from the solutions was considered an exothermic reaction, and the values of ΔH and ΔS obtained were -9.61 kJ.mol''- 1 and 1.37 J.mol''- 1 , respectively. The values of the Gibbs free energy changed from -10.03 to -10.06 kJ.mol -1 when the temperature was increased from 30 to 50 deg C. This study showed the potential application of the banana pith as bio sorbent of uranium ions. (author) 6. AMIDO RESISTENTE EM FARINHAS DE BANANA VERDE Directory of Open Access Journals (Sweden) Dayana Portes RAMOS 2010-03-01 Full Text Available Este trabalho teve por objetivo avaliar o teor de amido resistente (AR em farinhas de banana verde produzidas a partir de treze genótipos de bananeira. Para a produção da farinha foram separadas a 1ª, 3ª e 5ª pencas de cada genótipo, na qual cada penca correspondeu a uma repetição. Os frutos de cada penca no estádio 1 (casca completamente verde de maturação foram descascados manualmente, cortados em fatias circulares de 0,5 cm e desidratados em estufa com circulação de ar a 40ºC por 48 horas, sendo em seguida moídos. A análise de AR consistiu em um processo enzimático, calculando-se o conteúdo final pela concentração de glicose liberada. Os resultados foram submetidos à análise estatística e mostraram diferenças significativas para o teor de AR nas farinhas obtidas dos genótipos de bananeira, sendo que a farinha com maior teor de AR foi a produzida a partir do cultivar ‘Nam’ (40,25% e a menor pelo híbrido ‘Fhia 01’ (10,01%. Pode-se concluir que o conteúdo de AR varia em relação ao genótipo utilizado para a confecção da farinha e que a banana pode ser uma boa opção de estudo de alimento funcional. 7. Physical and biochemical properties of green banana flour. Science.gov (United States) Suntharalingam, S; Ravindran, G 1993-01-01 Banana flour prepared from two cooking banana varieties, namely 'Alukehel' and 'Monthan', were evaluated for their physical and biochemical characteristics. The yields of flour averaged 31.3% for 'Alukehel' and 25.5% for 'Monthan'. The pH of the flour ranged from 5.4 to 5.7. The bulk density and particle size distribution were also measured. The average chemical composition (% dry matter) of the flours were as follows: crude protein, 3.2; crude fat, 1.3; ash, 3.7; neutral detergent fiber, 8.9; acid detergent fiber, 3.8; cellulose, 3.1; lignin, 1.0 and hemicellulose, 5.0. Carbohydrate composition indicated the flour to contain 2.8% soluble sugars, 70.0% starch and 12.0% non-starch polysaccharides. Potassium is the predominant mineral in banana flour. Fresh green banana is a good source of vitamin C, but almost 65% is lost during the preparation of flour. Oxalate content (1.1-1.6%) of banana flour is probably nutritionally insignificant. The overall results are suggestive of the potential of green bananas as a source of flour. 8. Intergenerational Top Income Persistence DEFF Research Database (Denmark) Munk, Martin D.; Bonke, Jens; Hussain, M. Azhar 2016-01-01 In this paper, we investigate intergenerational top earnings and top income mobility in Denmark. Access to administrative registers allowed us to look at very small fractions of the population. We find that intergenerational mobility is lower in the top when including capital income in the income...... measure— for the rich top 0.1% fathers and sons the elasticity is 0.466. Compared with Sweden, however, the intergenerational top income persistence is about half the size in Denmark.... 9. Severity of banana leaf spot in an intercropping system in two cycles of banana Prata Anã Directory of Open Access Journals (Sweden) Valdeir Dias Gonçalves 2008-01-01 Full Text Available Prata Anã is the most planted banana cultivar in northern Minas Gerais, Brazil. It is however susceptible toseveral pathogens. This study was carried out to evaluate the disease severity of banana leaf spot in the Prata Anã cv. in thefirst and second cycle under six different planting systems. The randomized block experimental design was used with sixtreatments and four replications. In an evaluation of the severity of banana leaf spot, no disease symptoms were found onThap Maeo and Caipira. The evolution curve of the disease indicated seasonal effects in the first and second cycles. Theseverity of banana leaf spot was highest soon after the regional rainy period from November to March. A comparison of themeans of the evaluations indicated a reduction in disease severity from the first to the second cycle. 10. Top quark property measurements in single top CERN Document Server AUTHOR|(INSPIRE)INSPIRE-00386283; The ATLAS collaboration 2016-01-01 A review of the recent results on measurements of top quark properties in single top quark processes, performed at the LHC by ATLAS and CMS is presented. The measurements are in good agreement with predictions and no deviations from Standard Model expectations have been observed. 11. The effects of compost prepared from waste material of banana plants on the nutrient contents of banana leaves. Science.gov (United States) Doran, Ilhan; Sen, Bahtiyar; Kaya, Zülküf 2003-10-01 In this study, the possible utilization of removed shoots and plant parts of banana as compost after fruit harvest were investigated. Three doses (15-30-45 kg plan(-1)) of the compost prepared from the clone of Dwarf Cavendish banana were compared with Farmyard manure (50 kg plant(-1), Mineral fertilizers (180 g N + 150 g P + 335 g K plant(-1)) and Farmyard manure + Mineral fertilizers (25 kg FM + 180 g N + 150 g P + 335 g K plant(-1)) which determined positive effects on the nutrient contents of banana leaves. The banana plants were grown under a heated glasshouse and in a soil with physical and chemical properties suitable for banana growing. The contents of N, P, K and Mg in compost and in farmyard manure were found to be similar. Nitrogen, phosphorus and potassium contents of leaves in all applications except control, and Ca, Mg, Fe, Zn, Mn, Cu contents in all applications were determined between optimum levels of reference values. There were positive correlations among some nutrient contents of leaves, growth, yield and fruit quality characteristics. Farmyard manure, Farmyard manure + Mineral fertilizers and 45 kg plant(-1) of compost increased the nutrient contents of banana leaves. According to obtained results, 45 kg plant(-1) of compost was determined more suitable in terms of economical production and organic farming than the other fertiliser types. 12. Caracterização da farinha de banana verde Green banana flour characterization Directory of Open Access Journals (Sweden) Antonia de Maria Borges 2009-06-01 Full Text Available O presente trabalho objetivou a obtenção, a caracterização físico-química e o controle microbiológico durante o processamento da farinha de banana (Musa spp. verde, cv. Prata, visando o seu aproveitamento na panificação, produtos dietéticos e alimentos infantis. Para obtenção da farinha, os frutos foram cortados, imersos em metabissulfito de sódio, desidratados e triturados, sendo em seguida, feitas as seguintes determinações: umidade; extrato etéreo; proteína bruta; fibra bruta; cinzas; fração glicídica; amido; valor calórico; pH; acidez total titulável; vitamina C; macrominerais (K, P, Ca, Mg, S e N; microminerais (B, Cu, Mn, Zn e Fe; coliformes a 45 °C; fungos filamentosos e leveduras; Bacillus cereus; Salmonella sp.; Staphylococcus aureus; e contagem de aeróbios mesófilos. Os resultados indicaram que a banana 'Prata' verde é viável para o processo de obtenção da farinha de banana, tendo em vista que é rica em amido, proteína, potássio, fósforo, magnésio, zinco, cobre e tem um alto valor calórico. O pH, a acidez total titulável e a vitamina C estão compatíveis com os valores encontrados em outras farinhas. Quanto ao uso de boas práticas no processamento, a farinha encontra-se dentro do padrão microbiológico ideal e, portanto, está apta para o consumo.The objective of the present study was the physicochemical characterization and the microbiological control during the processing of the green banana flour (Musa spp., Prata cultivar, aiming at the use of the flour in bread-making, dietary products and children's food. To obtain the flour, the fruits were cut, immersed in sodium meta-bisulfite, dehydrated, and ground. The following criteria were determined: humidity; ethereal extract; raw protein; raw fiber; ash; glicidic fraction; starch; caloric value; pH; total titratable acidity; vitamin C; macrominerals (K, P, Ca, Mg, S and N; microminerals (B, Cu, Mn, Zn and Fe; coliforms at 45 °C; filamentous 13. Characterization of Heavy metals from banana farming soils Energy Technology Data Exchange (ETDEWEB) Lin, Dian; Huang, Cheng He; Huang, Dong Yi [College of Agronomy, Hainan University, Haikou City, Hainan Province (China); Ouyang, Ying [Department of Water Resources, St. Johns River Water Management District, Palatka, FL (United States) 2010-06-15 There is a growing public concern about the contamination of heavy metals in agricultural soils in China due to the increasingly applications of chemical fertilizers and pesticides during the last two decades. This study characterized the variability of heavy metals, including copper (Cu), zinc (Zn), lead (Pb), cadmium (Cd), and nickel (Ni), from the banana farming soils in western Hainan Island, China. Five banana farms from different locations in the island were selected to collect 69 mixed-soil samples in this study. Experimental data showed that concentrations of Cu ranged from 3.38 to 54.52, Zn from 24.0 to 189.8, Pb from 15.98 to 58.42, Cd from 0.43 to 3.21, and Ni from 3.47 to 121.86 mg kg{sup -1} dry wt. In general, concentrations of the heavy metals varied with metal species and changed from location to location, which occurred presumably due to the variations of soil parent materials and to a certain extent due to the use of different types of agrochemicals. Our study further revealed that concentrations of Cu and Zn were higher in the banana farming soils than in the natural (control) soils among all of the five locations, whereas mixed results were observed for Pb, Cd, and Ni in both the banana farming and control soils, depending on the locations. Comparisons of the heavy metal concentrations with the Chinese Soil Quality Standards (CSQSs) showed that Cu, Zn, and Pb contents were lower but Cd and Ni contents were higher in the banana farming soils than the Class II standard of the CSQSs. Results suggested that accumulation of Cu, Zn, and Pb in the soils is safe for banana fruit production, whereas accumulation of Cd and Ni in the same soils could potentially pose threats to banana fruit safety. (Abstract Copyright [2010], Wiley Periodicals, Inc.) 14. Audiitorfirmade TOP 50 aastal 2000 Index Scriptorium Estoniae 2001-01-01 Audiitorfirmade käibe TOP 50, käibe kasvu TOP 25, käibe languse TOP 15, kasumi TOP 50, kasumi kasvu TOP 10, kasumi languse TOP 10, audiitorfirmade finantsnäitajad. Rentaabluse TOP 50, varade tootlikkuse TOP 50 15. Top Production at LHCb CERN Multimedia Santana Rangel, Murilo 2015-01-01 Single and pair top production in the forward direction at the LHC allows for precision tests of the Standard Model. The observation of top quarks in 7 and 8 TeV data and prospects for precision measurements are shown. 16. Top Physics at Atlas CERN Document Server Romano, M; The ATLAS collaboration 2013-01-01 This talk is an overview of recent results on top-quark physics obtained by the ATLAS collaboration from the analysis of p-p collisions at 7 and 8 TeV at the Large Hadron Collider. Total and differential top pair cross section, single top cross section and mass measurements are presented. 17. A Simple Diffraction Experiment Using Banana Stem as a Natural Grating Science.gov (United States) Aji, Mahardika Prasetya; Karunawan, Jotti; Chasanah, Widyastuti Rochimatun; Nursuhud, Puji Iman; Wiguna, Pradita Ajeng; Sulhadi 2017-01-01 A simple diffraction experiment was designed using banana stem as natural grating. Coherent beams of lasers with wavelengths of 632.8 nm and 532 nm that pass through banana stem produce periodic diffraction patterns on a screen. The diffraction experiments were able to measure the distances between the slit of the banana stem, i.e. d = (28.76 ±… 18. Host plant odours enhance the responses of adult banana weevil to the synthetic aggregation pheromone Cosmolure+ NARCIS (Netherlands) Tinzaara, W.; Gold, C.S.; Dicke, M.; Huis, van A.; Ragama, P.E. 2007-01-01 Attraction of adult banana weevil, Cosmopolites sordidus to volatiles from banana pseudostem tissue and the synthetic pheromone Cosmolure+ presented singly or in combination, was studied in the laboratory and in the field. Olfactometric studies in the laboratory showed that 50 g of fermented banana 19. Evaluation of Information and Communication Technology Utilization by Small Holder Banana Farmers in Gatanga District, Kenya Science.gov (United States) Mwombe, Simon O. L.; Mugivane, Fred I.; Adolwa, Ivan S.; Nderitu, John H. 2014-01-01 Purpose: The study was carried out to identify information communication technologies (ICTs) used in production and marketing of bananas, to determine factors influencing intensity of use of ICT tools and to assess whether use of ICT has a significant influence on adoption of tissue culture bananas by small-scale banana farmers in Gatanga… 20. Production of Banana Fiber Yarns for Technical Textile Reinforced Composites Directory of Open Access Journals (Sweden) Zaida Ortega 2016-05-01 Full Text Available Natural fibers have been used as an alternative to synthetic ones for their greener character; banana fibers have the advantage of coming from an agricultural residue. Fibers have been extracted by mechanical means from banana tree pseudostems, as a strategy to valorize banana crops residues. To increase the mechanical properties of the composite, technical textiles can be used as reinforcement, instead of short fibers. To do so, fibers must be spun and woven. The aim of this paper is to show the viability of using banana fibers to obtain a yarn suitable to be woven, after an enzymatic treatment, which is more environmentally friendly. Extracted long fibers are cut to 50 mm length and then immersed into an enzymatic bath for their refining. Conditions of enzymatic treatment have been optimized to produce a textile grade of banana fibers, which have then been characterized. The optimum treating conditions were found with the use of Biopectinase K (100% related to fiber weight at 45 °C, pH 4.5 for 6 h, with bath renewal after three hours. The first spinning trials show that these fibers are suitable to be used for the production of yarns. The next step is the weaving process to obtain a technical fabric for composites production. 1. Comparative analysis of pigments in red and yellow banana fruit. Science.gov (United States) Fu, Xiumin; Cheng, Sihua; Liao, Yinyin; Huang, Bingzhi; Du, Bing; Zeng, Wei; Jiang, Yueming; Duan, Xuewu; Yang, Ziyin 2018-01-15 Color is an important characteristic determining the fruit value. Although ripe bananas usually have yellow peels, several banana cultivars have red peels. As details of the pigments in banana fruits are unknown, we investigated these pigments contents and compositions in the peel and pulp of red cultivar 'Hongjiaowang' and yellow cultivar 'Baxijiao' by UPLC-PDA-QTOF-MS and HPLC-PDA techniques. The 'Hongjiaowang' peel color was mainly determined by the presence of anthocyanin-containing epidermal cells. Rutinoside derivatives of cyanidin, peonidin, petunidin, and malvidin were unique to the red peel, and possibly responsible for the red color. 'Hongjiaowang' contained higher total content of carotenoids than 'Baxijiao' in both pulp and peel. Lutein, α-carotene, and β-carotene were main carotenoids, which might play a more important role than flavonoids in producing the yellow banana color owing to the properties and distribution in the fruit. The information will help us understand a complete profile of pigments in banana. Copyright © 2017 Elsevier Ltd. All rights reserved. 2. Ethanol production of banana shell and cassava starch International Nuclear Information System (INIS) Monsalve G, John F; Medina de Perez, Victoria Isabel; Ruiz colorado, Angela Adriana 2006-01-01 In this work the acid hydrolysis of the starch was evaluated in cassava and the cellulose shell banana and its later fermentation to ethanol, the means of fermentation were adjusted for the microorganisms saccharomyces cerevisiae nrrl y-2034 and zymomonas mobilis cp4. The banana shell has been characterized, which possesses a content of starch, cellulose and hemicelluloses that represent more than 80% of the shell deserve the study of this as source of carbon. The acid hydrolysis of the banana shell yield 20g/l reducing sugar was obtained as maximum concentration. For the cassava with 170 g/l of starch to ph 0.8 in 5 hours complete conversion is achieved to you reducing sugars and any inhibitory effect is not noticed on the part of the cultivations carried out with banana shell and cassava by the cyanide presence in the cassava and for the formation of toxic compounds in the acid hydrolysis the cellulose in banana shell. For the fermentation carried out with saccharomyces cerevisiae a concentration of ethanol of 7.92± 0.31% it is achieved and a considerable production of ethanol is not appreciated (smaller than 0.1 g/l) for none of the means fermented with zymomonas mobilis 3. Antioxidant and Antihyperglycemic Properties of Three Banana Cultivars (Musa spp.). Science.gov (United States) Adedayo, Bukola C; Oboh, Ganiyu; Oyeleye, Sunday I; Olasehinde, Tosin A 2016-01-01 Background . This study sought to investigate the antioxidant and antihyperglycemic properties of Musa sapientum (Latundan banana) (MSL), Musa acuminata (Cavendish banana) (MAC), and Musa acuminate (Red Dacca) (MAR). Materials and Methods. The sugar, starch, amylose, and amylopectin contents and glycemic index (GI) of the three banana cultivars were determined. Furthermore, total phenol and vitamin C contents and α -amylase and α -glucosidase inhibitory effects of banana samples were also determined. Results . MAC and MAR had the highest starch, amylose, and amylopectin contents and estimated glycemic index (eGI) with no significant different while MSL had the lowest. Furthermore, MAR (1.07 mg GAE/g) had a higher total phenol content than MAC (0.94 mg GAE/g) and MSL (0.96 mg GAE/g), while there was no significant difference in the vitamin C content. Furthermore, MAR had the highest α -amylase (IC 50 = 3.95 mg/mL) inhibitory activity while MAC had the least (IC 50 = 4.27 mg/mL). Moreover, MAC and MAR inhibited glucosidase activity better than MSL (IC 50 3.47 mg/mL). Conclusion . The low sugar, GI, amylose, and amylopectin contents of the three banana cultivars as well as their α -amylase and α -glucosidase inhibitory activities could be possible mechanisms and justification for their recommendation in the management of type-2 diabetes. 4. Production of bioethanol using agricultural waste: banana pseudo stem Directory of Open Access Journals (Sweden) Snehal Ingale 2014-09-01 Full Text Available India is amongst the largest banana (Musa acuminata producing countries and thus banana pseudo stem is commonly available agricultural waste to be used as lignocellulosic substrate. Present study focuses on exploitation of banana pseudo stem as a source for bioethanol production from the sugars released due to different chemical and biological pretreatments. Two fungal strains Aspergillus ellipticus and Aspergillus fumigatus reported to be producing cellulolytic enzymes on sugarcane bagasse were used under co-culture fermentation on banana pseudo stem to degrade holocellulose and facilitate maximum release of reducing sugars. The hydrolysate obtained after alkali and microbial treatments was fermented by Saccharomyces cerevisiae NCIM 3570 to produce ethanol. Fermentation of cellulosic hydrolysate (4.1 g% gave maximum ethanol (17.1 g/L with yield (84% and productivity (0.024 g%/h after 72 h. Some critical aspects of fungal pretreatment for saccharification of cellulosic substrate using A. ellipticus and A. fumigatus for ethanol production by S. cerevisiae NCIM 3570 have been explored in this study. It was observed that pretreated banana pseudo stem can be economically utilized as a cheaper substrate for ethanol production. 5. CDF Top Physics Science.gov (United States) Tartarelli, G. F.; CDF Collaboration 1996-05-01 The authors present the latest results about top physics obtained by the CDF experiment at the Fermilab Tevatron collider. The data sample used for these analysis (about 110 pb{sup{minus}1}) represents almost the entire statistics collected by CDF during four years (1992--95) of data taking. This large data size has allowed detailed studies of top production and decay properties. The results discussed here include the determination of the top quark mass, the measurement of the production cross section, the study of the kinematics of the top events and a look at top decays. 6. Evolution of endogenous sequences of banana streak virus: what can we learn from banana (Musa sp.) evolution? Science.gov (United States) Gayral, Philippe; Blondin, Laurence; Guidolin, Olivier; Carreel, Françoise; Hippolyte, Isabelle; Perrier, Xavier; Iskra-Caruana, Marie-Line 2010-07-01 Endogenous plant pararetroviruses (EPRVs) are viral sequences of the family Caulimoviridae integrated into the nuclear genome of numerous plant species. The ability of some endogenous sequences of Banana streak viruses (eBSVs) in the genome of banana (Musa sp.) to induce infections just like the virus itself was recently demonstrated (P. Gayral et al., J. Virol. 83:6697-6710, 2008). Although eBSVs probably arose from accidental events, infectious eBSVs constitute an extreme case of parasitism, as well as a newly described strategy for vertical virus transmission in plants. We investigated the early evolutionary stages of infectious eBSV for two distinct BSV species-GF (BSGFV) and Imové (BSImV)-through the study of their distribution, insertion polymorphism, and structure evolution among selected banana genotypes representative of the diversity of 60 wild Musa species and genotypes. To do so, the historical frame of host evolution was analyzed by inferring banana phylogeny from two chloroplast regions-matK and trnL-trnF-as well as from the nuclear genome, using 19 microsatellite loci. We demonstrated that both BSV species integrated recently in banana evolution, circa 640,000 years ago. The two infectious eBSVs were subjected to different selective pressures and showed distinct levels of rearrangement within their final structure. In addition, the molecular phylogenies of integrated and nonintegrated BSVs enabled us to establish the phylogenetic origins of eBSGFV and eBSImV. 7. Banana leaf and glucose mineralization and soil organic matter in microhabitats of banana plantations under long-term pesticide use. Science.gov (United States) Blume, Elena; Reichert, José Miguel 2015-06-01 Soil organic matter (SOM) and microbial activity are key components of soil quality and sustainability. In the humid tropics of Costa Rica 3 pesticide regimes were studied-fungicide (low input); fungicide and herbicide (medium input); and fungicide, herbicide, and nematicide (high input)-under continuous banana cultivation for 5 yr (young) or 20 yr (old) in 3 microhabitats-nematicide ring around plants, litter pile of harvested banana, and bare area between litter pile and nematicide ring. Soil samples were incubated sequentially in the laboratory: unamended, amended with glucose, and amended with ground banana leaves. Soil organic matter varied with microhabitat, being greatest in the litter pile, where microbes had the greatest basal respiration with ground banana leaf, whereas microbes in the nematicide ring had the greatest respiration with glucose. These results suggest that soil microbes adapt to specific microhabitats. Young banana plantations had similar SOM compared with old plantations, but the former had greater basal microbial respiration in unamended and in glucose-amended soil and greater first-order mineralization rates in glucose-amended soil, thus indicating soil biological quality decline over time. High pesticide input did not decrease microbial activity or mineralization rate in surface soil. In conclusion, microbial activity in tropical volcanic soil is highly adaptable to organic and inorganic inputs. © 2015 SETAC. 8. Green banana pasta: an alternative for gluten-free diets. Science.gov (United States) Zandonadi, Renata Puppin; Botelho, Raquel Braz Assunção; Gandolfi, Lenora; Ginani, Janini Selva; Montenegro, Flávio Martins; Pratesi, Riccardo 2012-07-01 The objective of this study was to develop and analyze a gluten-free pasta made with green banana flour. The study was divided into five steps: preparation/selection, chemical, sensory, technological, and statistical analysis. The modified sample presented greater acceptance (84.5% for celiac individuals and 61.2% for nonceliac) than standard samples (53.6% for nonceliac individuals). There was no significant difference between the modified and the standard samples in terms of appearance, aroma, flavor, and overall quality. The modified pastas presented approximately 98% less lipids. Green bananas are considered a subproduct of low commercial value with little industrial use. The possibility of developing gluten-free products with green banana flour can expand the product supply for people with celiac disease and contribute to a more diverse diet. Copyright © 2012 Academy of Nutrition and Dietetics. Published by Elsevier Inc. All rights reserved. 9. Hyperspectral imaging system for disease scanning on banana plants Science.gov (United States) Ochoa, Daniel; Cevallos, Juan; Vargas, German; Criollo, Ronald; Romero, Dennis; Castro, Rodrigo; Bayona, Oswaldo 2016-05-01 Black Sigatoka (BS) is a banana plant disease caused by the fungus Mycosphaerella fijiensis. BS symptoms can be observed at late infection stages. By that time, BS has probably spread to other plants. In this paper, we present our current work on building an hyper-spectral (HS) imaging system aimed at in-vivo detection of BS pre-symptomatic responses in banana leaves. The proposed imaging system comprises a motorized stage, a high-sensitivity VIS-NIR camera and an optical spectrograph. To capture images of the banana leaf, the stage's speed and camera's frame rate must be computed to reduce motion blur and to obtain the same resolution along both spatial dimensions of the resulting HS cube. Our continuous leaf scanning approach allows imaging leaves of arbitrary length with minimum frame loss. Once the images are captured, a denoising step is performed to improve HS image quality and spectral profile extraction. 10. Biosynthesis of CdS nanoparticles in banana peel extract. Science.gov (United States) Zhou, Guang Ju; Li, Shuo Hao; Zhang, Yu Cang; Fu, Yun Zhi 2014-06-01 Cadmium sulfide (CdS) nanoparticles (NPs) were synthesized by using banana peel extract as a convenient, non-toxic, eco-friendly 'green' capping agent. Cadmium nitrate and sodium sulfide are main reagents. A variety of CdS NPs are prepared through changing reaction conditions (banana extracts, the amount of banana peel extract, solution pH, concentration and reactive temperature). The prepared CdS colloid displays strong fluorescence spectrum. X-ray diffraction analysis demonstrates the successful formation of CdS NPs. Fourier transform infra-red (FTIR) spectrogram indicates the involvement of carboxyl, amine and hydroxyl groups in the formation of CdS NPs. Transmission electron microscope (TEM) result reveals that the average size of the NPs is around 1.48 nm. 11. Turismifirmade 2000. a. TOP 30 Index Scriptorium Estoniae 2001-01-01 Turismiettevõtete üldandmed: turismiettevõtete finantsnäitajad; käibe TOP 30; käibe kasvu TOP 10; kasumi TOP 20; kasumi kasvu TOP 10; kasumi languse TOP; rentaabluse TOP 10; varade tootlikkuse TOP 10 12. Ehitusmaterjalitootjate TOP 70 aastal 2003 Index Scriptorium Estoniae 2004-01-01 Ilmunud ka: Delovõje Vedomosti : Stroitelstvo, 29. sept. 2004, lk. 2,4. Ehitusmaterjalitootjate TOP 70; Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP. Ehitusmaterjalitootjate üldandmed 13. Gamma radiation effects on the viscosity of green banana flour International Nuclear Information System (INIS) Uehara, Vanessa B.; Inamura, Patricia Y.; Mastro, Nelida L. Del 2009-01-01 Banana (Musa sp) is a tropical fruits with great acceptability among consumers and produced in Brazil in a large scale. Bananas are not being as exploited as they could be in prepared food, and research could stimulate greater interest from industry. The viscosity characteristics and a product consistency can determine its acceptance by the consumer. Particularly the starch obtained from green banana had been studied from the nutritional point of view since the concept of Resistant Starch was introduced. Powder RS with high content of amylose was included in an approved food list with alleged functional properties in Brazilian legislation. Ionizing radiation can be used as a public health intervention measure for the control of food-borne diseases. Radiation is also a very convenient tool for polymer materials modification through degradation, grafting and crosslinking. In this work the influence of ionizing radiation on the rheological behavior of green banana pulp was investigated. Samples of green banana pulp flour were irradiated in a 60 Co Gammacell 220 (AECL) with doses of 0 kGy,1 kGy, 3 kGy, 5 kGy and 10 kGy in glass recipients. After irradiation 3% and 5% aqueous dilution were prepared and viscosity measurements performed in a Brooksfield, model DVIII viscometer using spindle SC4-18 and SC4-31. There was a reduction of the initial viscosity of the samples as a consequence of radiation processing, being the reduction inversely proportional to the flour concentration. The polysaccharide content of the banana starch seems to be degraded by radiation in solid state as shown by the reduction of viscosity as a function of radiation dose. (author) 14. Banana production systems: identification of alternative systems for more sustainable production. Science.gov (United States) Bellamy, Angelina Sanderson 2013-04-01 Large-scale, monoculture production systems dependent on synthetic fertilizers and pesticides, increase yields, but are costly and have deleterious impacts on human health and the environment. This research investigates variations in banana production practices in Costa Rica, to identify alternative systems that combine high productivity and profitability, with reduced reliance on agrochemicals. Farm workers were observed during daily production activities; 39 banana producers and 8 extension workers/researchers were interviewed; and a review of field experiments conducted by the National Banana Corporation between 1997 and 2002 was made. Correspondence analysis showed that there is no structured variation in large-scale banana producers' practices, but two other banana production systems were identified: a small-scale organic system and a small-scale conventional coffee-banana intercropped system. Field-scale research may reveal ways that these practices can be scaled up to achieve a productive and profitable system producing high-quality export bananas with fewer or no pesticides. 15. Chemical and physical characterization of Musa sepientum and Musa balbisiana fibers of banana tree International Nuclear Information System (INIS) Albinante, Sandra R.; Pacheco, Elen B.A.V.; Visconte, Leila L.Y.; Batista, Luciano do N. 2011-01-01 This study aimed to characterize the fibers of cavendish and silver banana trunks (Musa sepientum and Musa balbisiana, respectively) concerning their density, lignin and moisture contents, and chemical structure by using the techniques of infrared spectroscopy and low field solid state nuclear magnetic resonance, NMR. From NMR analysis, it was possible to observe the morphological differences between cavendish and silver types of banana fibers. FTIR technique did not allow the observation of any important difference in the banana fibers spectra. The cavendish banana fiber showed higher moisture and lignin contents than the silver banana fiber The NMR technique showed that relaxation times for silver banana fiber were higher than those for cavendish banana fiber, which can be credited to the lower moisture content values found in the silver fibers. (author) 16. TOP LINAC design; Progetto del TOP LINAC Energy Technology Data Exchange (ETDEWEB) Picardi, L; Ronsivalle, C; Vignati, A [ENEA, Centro Ricerche Frascati, Rome (Italy). Dip. Innovazione 1997-11-01 The report describes a linear accelerator for protons named TOP LINAC designed for the TOP (Terapia Oncologica con Protoni, Oncological Protontherapy) project launched by the Italian National Institute of Health (Istituto Superiore di Sanita, ISS) to explore in collaboration with the biggest Oncological Hospital in Rome (Istituto Regina Elena, IRE) the potentialities of the therapy with accelerated protons and establish guide lines for the application of this new type of radiotherapy in comparison with the more traditional electron and x-rays radiotherapy. The concept of a compact accelerator for protontherapy applications bore within the Italian Hadrontherapy Collaboration (TERA Collaboration) with the aim to diffuse the protontherapy on the National territory. The ISS program plans to use the TOP linac proton beam also for production of PET (Positron Emission Tomography) radioisotopes and radiobiology studies. Official agreements are in course between ISS and ENEA which provides its experience in the industrial and medical accelerators for the design and the construction of the TOP linac. The accelerator that will be the first 3 GHz proton linac in the world, will be composed of a 428.3 MHz 7 Me V RFQ + DTL injector followed by a 7-65 Me V section of a 3 GHz SCDTL structure and a 65 - 200 Me V variable energy SCL 3 GHz structure. In particular the SCDTL section uses a highly innovative accelerating structure patented by ENEA. In this report the clinical and physical requests are discussed and a preliminary design of the whole machine is given. 17. Biossorption of uranium on banana pith; Biossorcao de uranio nas cascas de banana Energy Technology Data Exchange (ETDEWEB) Boniolo, Milena Rodrigues 2008-07-01 Banana pith was characterized by Fourier Transformed Infrared Spectroscopy and Scanning Electron Microscopy, and investigated as a low cost bio sorbent for the removal of uranium ions from nitric solutions. Influences variable as were studied: adsorbent particle size, contact time, pH and temperature were studied. The removal percentage was increased from 13 to 57% when the particle size was decreased from 6.000 to 0.074 mm. The determined contact time was 40 minutes with 60% mean removal. The removal was increased from 40 to 55% when the pH varied from 2 to 5. The Langmuir and Freundlich linear isotherm models were applied to describe the adsorption equilibrium. The kinetic of the process was studied using the pseudo-first order and pseudo-second order models. Thermodynamics parameters such as {delta}G, {delta}S and {delta}H were calculated. In concentration range of 50 - 500 mg.L{sup -1}, the adsorption process was described better by the Freundlich equation. The adsorption capacity at equilibrium of uranium ions was determined from the Langmuir equation, and it was found 11.50 mg.g{sup -1} at 25 {+-} 2 deg C. The kinetic was better represented by the pseudo-second order model. The bio sorption process for uranium removal from the solutions was considered an exothermic reaction, and the values of {delta}H and {delta}S obtained were -9.61 kJ.mol''-{sup 1} and 1.37 J.mol''-{sup 1}, respectively. The values of the Gibbs free energy changed from -10.03 to -10.06 kJ.mol{sup -1} when the temperature was increased from 30 to 50 deg C. This study showed the potential application of the banana pith as bio sorbent of uranium ions. (author) 18. Biossorption of uranium on banana pith; Biossorcao de uranio nas cascas de banana Energy Technology Data Exchange (ETDEWEB) Boniolo, Milena Rodrigues 2008-07-01 Banana pith was characterized by Fourier Transformed Infrared Spectroscopy and Scanning Electron Microscopy, and investigated as a low cost bio sorbent for the removal of uranium ions from nitric solutions. Influences variable as were studied: adsorbent particle size, contact time, pH and temperature were studied. The removal percentage was increased from 13 to 57% when the particle size was decreased from 6.000 to 0.074 mm. The determined contact time was 40 minutes with 60% mean removal. The removal was increased from 40 to 55% when the pH varied from 2 to 5. The Langmuir and Freundlich linear isotherm models were applied to describe the adsorption equilibrium. The kinetic of the process was studied using the pseudo-first order and pseudo-second order models. Thermodynamics parameters such as {delta}G, {delta}S and {delta}H were calculated. In concentration range of 50 - 500 mg.L{sup -1}, the adsorption process was described better by the Freundlich equation. The adsorption capacity at equilibrium of uranium ions was determined from the Langmuir equation, and it was found 11.50 mg.g{sup -1} at 25 {+-} 2 deg C. The kinetic was better represented by the pseudo-second order model. The bio sorption process for uranium removal from the solutions was considered an exothermic reaction, and the values of {delta}H and {delta}S obtained were -9.61 kJ.mol''-{sup 1} and 1.37 J.mol''-{sup 1}, respectively. The values of the Gibbs free energy changed from -10.03 to -10.06 kJ.mol{sup -1} when the temperature was increased from 30 to 50 deg C. This study showed the potential application of the banana pith as bio sorbent of uranium ions. (author) 19. Ekspedeerimisettevõtete TOP 50 Index Scriptorium Estoniae 2006-01-01 Ilmunud ka: Delovõje Vedomosti : Transport i Logistika nr. 11, 29. nov. lk. 32-35. Ekspedeerimisettevõtete TOP. Vt. samas: Käibe TOP 10; Käibekasvu TOP 10; Kasumi TOP 10; Kasumikasvu TOP 10; Rentaabluse TOP 10; ROE TOP 10; Ekspedeerimisettevõtete üld- ja finantsandmed. Ajal. Delovõje Vedomosti : Transport i Logistika toodud ainult Ekspedeerimisettevõtete TOP 50 20. From crossbreeding to biotechnology-facilitated improvement of banana and plantain. Science.gov (United States) Ortiz, Rodomiro; Swennen, Rony 2014-01-01 The annual harvest of banana and plantain (Musa spp.) is approximately 145 million tons worldwide. About 85% of this global production comes from small plots and kitchen or backyard gardens from the developing world, and only 15% goes to the export trade. Musa acuminata and Musa balbisiana are the ancestors of several hundreds of parthenocarpic Musa diploid and polyploid cultivars, which show multiple origins through inter- and intra-specific hybridizations from these two wild diploid species. Generating hybrids combining host plant resistance to pathogens and pests, short growth cycles and height, high fruit yield, parthenocarpy, and desired quality from the cultivars remains a challenge for Musa crossbreeding, which started about one century ago in Trinidad. The success of Musa crossbreeding depends on the production of true hybrid seeds in a crop known for its high levels of female sterility, particularly among polyploid cultivars. All banana export cultivars grown today are, however, selections from somatic mutants of the group Cavendish and have a very narrow genetic base, while smallholders in sub-Saharan Africa, tropical Asia and Latin America use some bred-hybrids (mostly cooking types). Musa improvement goals need to shift to address emerging threats because of the changing climate. Innovative cell and molecular biology tools have the potential to enhance the pace and efficiency of genetic improvement in Musa. Micro-propagation has been successful for high throughput of clean planting materials while in vitro seed germination assists in obtaining seedlings after inter-specific and across ploidy hybridization. Flow cytometry protocols are used for checking ploidy among genebank accessions and breeding materials. DNA markers, the genetic maps based on them, and the recent sequencing of the banana genome offer means for gaining more insights in the genetics of the crops and to identifying genes that could lead to accelerating Musa betterment. Likewise, DNA 1. Assessment of insect invaders of decaying banana and plantain ... African Journals Online (AJOL) Insect invaders of decaying banana and plantain pseudo stems in Umuagwo, Ohaji-Egbema, Imo State were investigated in randomly selected crop plots near living homes (<500 m) and far from living homes (.500m). Investigation was done by the use of trapping systems, dissections of cut decaying pseudo stems and ... 2. Review on postharvest technology of banana fruit | Hailu | African ... African Journals Online (AJOL) These include disinfecting, packaging and storage temperature. Pre- and postharvest treatments were found to have an effect on postharvest quality of banana, suggesting that postharvest quality of produce subjected to preharvest treatments should be assessed from a quality improvement, maintenance and consumer ... 3. Banana peel: A novel substrate for cellulase production under solid ... African Journals Online (AJOL) use 2011-12-05 Dec 5, 2011 ... The feasibility of using banana peel for the production of cellulase by Trichoderma viride GIM 3.0010 in solid-state fermentation was evaluated in this study. The effect of incubation time, incubation temperature, initial moisture content of the medium, inoculum size and supplementation of carbon sources ... 4. Effect of gamma irradiation on Hom Tong banana International Nuclear Information System (INIS) 1971-01-01 This report contains research on the use of gamma irradiation to retard the ripening and extend the shelf life of bananas. The major concerns were the effects that irradiation would have on the nutritional content, the organoleptic properties and the pigment of the fruit 5. In vitro multiplication of banana (Musa sp.) cv. Grand Naine African Journals Online (AJOL) SAM 2014-07-02 Jul 2, 2014 ... with IAA 2.00 mg/l resulted in maximum establishment of cultures in lesser time. MS medium + BAP. 4.00 mg/l + ... is, Musa acuminata (Malaysia) and Musa balbsiana. (India) (Georget et al., 2000). Banana plantlets produced. 6. Banana (Musa spp.) Production Characteristics and Performance in Uganda NARCIS (Netherlands) Bagamba, F.; Burger, C.P.J.; Tushemereirwe, W.K. 2010-01-01 The highland cooking banana (Musa spp., AAA-EA genome) is the most important crop in the East African Great Lakes region. In Uganda, production has expanded and productivity increased in the country’s southwest and declined in the Central region where the crop has traditional roots. Analyzing crop 7. Household uses of the banana plant in eastern Democratic Republic ... African Journals Online (AJOL) SARAH 2015-11-30 Nov 30, 2015 ... use of other banana plant parts other than the fruit pulp, has been widely reported. This has not been ... prevalent uses across plant parts included use for mulch and compost, feed for livestock, construction materials, ropes for ... pharmaceutical and the food industry (Oliveira et al.,. 2007). In Malaysia and ... 8. Molecular diagnostics for the Sigatoka disease complex of banana NARCIS (Netherlands) Arzanlou, M.; Abeln, E.C.A.; Kema, G.H.J.; Waalwijk, C.; Carlier, J.; Crous, P.W. 2007-01-01 The Sigatoka disease complex of banana involves three related ascomycetous fungi, Mycosphaerella fijiensis, M. musicola, and M. eumusae. The exact distribution of these three species and their disease epidemiology remain unclear, because their symptoms and life cycles are rather similar. Disease 9. Sequencing the Major Mycosphaerella Pathogens of Wheat and Banana NARCIS (Netherlands) Kema, G.H.J. 2009-01-01 Mycosphaerella is one of the largest genera of plant-pathogenic fungi with more than 1,000 named species, many of which are important pathogens causing leaf spotting diseases in a wide variety of crops including cereals, citrus, banana, eucalypts, soft fruits and horticultural crops. A few species 10. Sequencing the Major Mycosphaerella Pathogens of Wheat and Banana NARCIS (Netherlands) Kema, G.H.; Dunkle, L.D.; Churchill, A.C.; Carlier, J.; James, A.; Souza, M.T.; Crous, P.W.; Roux, N.; Lee, T.A. van der; Wiitenberg, A.; Lindquist, E.; Grigoriev, I.; Bristow, J.; Goodwin, S.B. 2007-01-01 Mycosphaerella is one of the largest genera of plant pathogenic fungi with more than 1,000 named species, many of which are important pathogens causing leaf spotting diseases in a wide variety of crops including cereals, citrus, banana, eucalypts, soft fruits, and horticultural crops. A few species 11. Factors influencing the diffusion of cooking banana in Nigeria ... African Journals Online (AJOL) As an interim measure in combating the incidence of black Sigatoka disease on plantain, the International Institute of Tropical Agriculture introduced cooking banana in Southeast Nigeria in the late 1980s. This was multiplied and distributed to farmers through the extension systems of both governmental and ... 12. Did backcrossing contribute to the origin of hybrid edible bananas? Czech Academy of Sciences Publication Activity Database De Langhe, E.; Hřibová, Eva; Carpentier, S.; Doležel, Jaroslav; Swennen, R. 2010-01-01 Roč. 106, č. 6 (2010), s. 849-857 ISSN 0305-7364 R&D Projects: GA AV ČR IAA600380703 Institutional research plan: CEZ:AV0Z50380511 Keywords : Backcrossing * banana * breeding Subject RIV: EF - Botanics Impact factor: 3.388, year: 2010 13. Factors driving the adoption of cooking banana processing and ... African Journals Online (AJOL) As part of efforts in realising her aim of introducing cooking banana into Nigeria, the International Institute of Tropical Agriculture (IITA) mounted training and awareness campaigns on its utilisation in collaboration with Shell and Agip Oil companies between 1991 and 1997. This study looked into the adoption profile of the ... 14. Banana-shaped molecules derived from substituted isophthalic acids Indian Academy of Sciences (India) Different precursors inducing the bending angle of the banana-shaped molecules. Figure 2. ... Pramana – J. Phys., Vol. 61, No. 2, August ... isotropic liquid; N: nematic; SmA: smectic A; SmC: smectic C. For other phase assign- ments, see text. 15. Household uses of the banana plant in eastern Democratic Republic ... African Journals Online (AJOL) SARAH 2015-11-30 Nov 30, 2015 ... ... food, feed and medicine. Other uses included extraction of fibre, arts and crafts and use as medicine. .... used banana peels as livestock feed was not significantly different (P ≤ 0.05). ..... Glass Fiber Woven Fabric-Reinforced Polyester. Composite. Journal ... chemical resistance of natural fibres reinforced. 16. Nutrient-enhancement of Matooke banana for improved nutrient ... African Journals Online (AJOL) A total of 173 PLHIVregistered with Rakai Health Science Project were chosen and interviewed using structured questionnaires to determine the current contribution of banana to the household food security. Nutrient intake data were collected using Gibson s 24-hour recall method and food frequency questionnaires. 17. Physiological and biochemical changes during banana ripening and finger drop NARCIS (Netherlands) Imsabai, W.; Ketsa, S.; Doorn, van W.G. 2006-01-01 Fruit drop of banana is due to breaking at the junction of the pedicel and pulp, and we found no true abscission zone. The breakage seems therefore due to weakening of the peel. We investigated pectin hydrolysis and some properties at the rupture zone, using Hom Thong (Musa acuminata, AAA Group) 18. Hot water treatments delay cold-induced banana peel blackening NARCIS (Netherlands) Promyou, S.; Ketsa, S.; Doorn, van W.G. 2008-01-01 Banana fruit of cv. Gros Michel (Musa acuminata, AAA Group, locally called cv. Hom Thong) and cv. Namwa (Musa x paradisiaca, ABB Group) were immersed for 5, 10 and 15 min in water at 42 degrees C, or in water at 25 degrees C (control), and were then stored at 4 degrees C. Hot water treatment for 15 19. A molecular diagnostic for tropical race 4 of the banana NARCIS (Netherlands) Dita Rodriguez, M.A.; Waalwijk, C.; Buddenhagen, I.W.; Souza, M.T.; Kema, G.H.J. 2010-01-01 This study analysed genomic variation of the translation elongation factor 1 (TEF-1) and the intergenic spacer region (IGS) of the nuclear ribosomal operon of Fusarium oxysporum f. sp. cubense (Foc) isolates, from different banana production areas, representing strains within the known races, 20. Effects of potassium deficiency, drought and weevils on banana ... African Journals Online (AJOL) Mo yields, but the mean benefit-cost ratio is still low (1.6) when N and P fertilizer costs ... high. We recommend testing the use of mulch to decrease drought risk and enhance fertilizer use efficiency. ... Bananas are a major food and cash crop in. 1. Banana-shaped molecules derived from substituted isophthalic acids Indian Academy of Sciences (India) In this paper we present a review of five-rings banana-shaped molecules derived from isophthalic acids. This study deals with about a hundred compounds and most of them have not been published. By a combination of several linking groups and different selected substituents either on the outer rings or on the central core ... 2. Agronomic performance of introduced banana varieties in lowlands ... African Journals Online (AJOL) Most of Rwandan banana cultivars are low-yielding and susceptible to pests and diseases. High yielding and pest/disease resistant varieties have been obtained in advanced breeding centers recently. Introduction, evaluation and adoption of such varieties by local producers may be one of the options to boost yields. 3. Lipophilic phytochemicals from banana fruits of several Musa species. Science.gov (United States) Vilela, Carla; Santos, Sónia A O; Villaverde, Juan J; Oliveira, Lúcia; Nunes, Alberto; Cordeiro, Nereida; Freire, Carmen S R; Silvestre, Armando J D 2014-11-01 The chemical composition of the lipophilic extract of ripe pulp of banana fruit from several banana cultivars belonging to the Musa acuminata and Musa balbisiana species (namely 'Chinese Cavendish', 'Giant Cavendish', 'Dwarf Red', 'Grand Nain', 'Eilon', 'Gruesa', 'Silver', 'Ricasa', 'Williams' and 'Zelig') was studied by gas chromatography-mass spectrometry for the first time. The banana cultivars showed similar amounts of lipophilic extractives (ca. 0.4% of dry material weight) as well as qualitative chemical compositions. The major groups of compounds identified in these fractions were fatty acids and sterols making up 68.6-84.3% and 11.1-28.0%, respectively, of the total amount of lipophilic components. Smaller amounts of long chain aliphatic alcohols and α-tocopherol were also identified. These results are a relevant contribution for the valorisation of these banana cultivars as sources of valuable phytochemicals (ω-3 and ω-6 fatty acids, and sterols) with well-established beneficial nutritional and health effects. Copyright © 2014 Elsevier Ltd. All rights reserved. 4. Evaluation of genetic diversity between 27 banana cultivars (Musa ... African Journals Online (AJOL) Cultivated bananas (Musa spp.) are mostly diploid or triploid cultivars with various combinations of the A and B genomes inherited from their diploid ancestors Musa acuminata Colla. and Musa balbisiana. Colla. respectively. Random amplified polymorphic DNA (RAPD) markers were used to establish the relatedness of 27 ... 5. Factors Affecting Utilization of Cooking Banana among Households ... African Journals Online (AJOL) The study investigated factors affecting utilization of cooking banana among households in Oguta area of Imo State, Nigeria. Data were collected from 84 randomly selected respondents from six communities in the study area who were administered with structured questionnaire. Data analysis was by use of descriptive ... 6. Prospects and determinants of adoption of IITA plantain and banana ... African Journals Online (AJOL) High yielding and disease resistant plantain and banana hybrids and its associated technologies generated by IITA to combat the menace of black Sigatoka disease (Mycosphaerella fijiensis) were massively disseminated in year 2000. Since the hybrids were slightly different from the existing varieties in fruit size there was ... 7. Efficient regeneration of the endangered banana cultivar 'Virupakshi ... African Journals Online (AJOL) Plantlets of the banana cultivar 'Virupakshi' (AAB) were regenerated from somatic embryos derived from embryogenic cells of calli from immature male flower explants. Induction of calli from explants was favored by a relatively moderate concentration of 2,4-dichlorophenoxyacetic acid (2,4-D) (4 mg/L), high concentrations of ... 8. Anthelmintic effects of dried ground banana plant leaves ( Musa spp ... African Journals Online (AJOL) Background: Helminths is a endoparasites that cause the major losses for profitable sheep production in Brazil. The increased development of resistant strains of endoparasites have enforced the search for sustainable alternatives. The aim of this paper was to provide information about endoparasites control with banana ... 9. Characteristics of micro-propagated banana (Musa spp.) cultures ... African Journals Online (AJOL) Administrator 2011-05-23 May 23, 2011 ... was conducted to assess the effect of NaCl and PEG separately as well as in combination on plant micro- propagation efficiency of banana (Musa spp.) cv., Basrai. In this experiment, 4-weeks old plantlets of the 3rd sub- culture with well propagation on MS2b nutrient were sub- cultured on three differentially ... 10. Optimization of biogas production from banana peels: Effect of ... African Journals Online (AJOL) PROMOTING ACCESS TO AFRICAN RESEARCH ... The matooke processing industry being set up by the Presidential Initiative on Banana Industrial Development ... a solution to that waste, but information on the pre-treatment of the matooke peel waste is inadequate. ... EMAIL FREE FULL TEXT EMAIL FREE FULL TEXT 11. Root activity pattern of banana under irrigated and rain conditions International Nuclear Information System (INIS) Sobhana, A.; Aravindakshan, M.; Wahid, P.A. 1989-01-01 Root morphology by excavation method and root activity pattern by 32 P soil-injection technique have been studied in banana var., Nendran under rainfed/irrigated conditions. The number of roots, length and diameter of roots and dry weight of roots were found to be more for the rainfed banana crop compared to the irrigated. The results of the radiotracer studies indicated that about 60 per cent of the active roots of irrigated banana lie within 20 cm distance and about 90 per cent of the total root activity is found within 40 cm distance from the plant. In the case of rainfed crop about 85 per cent of the active roots were found within a radius of 40 cm around the plant. Active roots were found to be more concentrated at 15 to 30 cm depth under rainfed conditions while the density of active roots was more or less uniform along the profile upto a dpeth of 60 cm in irrigated banana. (author). 4 refs., 3 figs 12. Strategy to increase Barangan Banana production in Kabupaten Deli Serdang Science.gov (United States) Adhany, I.; Chalil, D.; Ginting, R. 2018-02-01 This study was conducted to analyze internal and external factors in increasing Barangan Banana production in Kabupaten Deli Serdang. Samples were determined by snowball sampling technique and purposive sampling method. Using SWOT analysis method, this study found that there were 6 internal strategic factors and 9 external strategic factors. Among that strategic factors, support for production facilities appears as the most important internal strategic factor, while the demand for Barangan Banana. as the most important external strategic factor. Based on the importance and existing condition of these strategic factors, using support for production facilities and realization of supporting facilities with farming experience are the strategies covering strength-opportunity (SO), organizing mentoring to meet the demand for Barangan Banana are the strategies covering weakness-opportunity (WO), making use of funding support and subsidies to widen the land, using tissue culture seeds and facilities and infrastructures are the strategies covering strength-threat (ST), increas the funding support to widen the land, the use of tissue culture seeds and facilities and infrastructures are the strategies covering weakness-threat (WT) are discussed and proposed to increase Barangan Banana productivity in Kabupaten Deli Serdang. 13. Sheep fed with banana leaf hay reduce ruminal protozoa population. Science.gov (United States) Freitas, Cláudio Eduardo Silva; Duarte, Eduardo Robson; Alves, Dorismar David; Martinele, Isabel; D'Agosto, Marta; Cedrola, Franciane; de Moura Freitas, Angélica Alves; Dos Santos Soares, Franklin Delano; Beltran, Makenzi 2017-04-01 A ciliate protozoa suppression can reduce methane production increasing the energy efficiency utilization by ruminants. The physicochemical characteristics of rumen fluid and the profile of the rumen protozoa populations were evaluated for sheep fed banana leaf hay in replacement of the Cynodon dactylon cv. vaqueiro hay. A total of 30 male sheep were raised in intensive system during 15 days of adaptation and 63 days of experimental period. The animals were distributed in a completely randomized design that included six replicates of five treatments with replacement levels (0, 25, 50, 75, and 100%) of the grass vaquero for the banana leaf hay. Samples of fluid were collected directly from the rumen with sterile catheters. Color, odor, viscosity, and the methylene blue reduction potential (MBRP) were evaluated and pH estimated using a digital potentiometer. After decimal dilutions, counts of genus protozoa were performed in Sedgewick Rafter chambers. The averages of pH, MBRP, color, odor, and viscosity were not influenced by the inclusion of the banana leaf hay. However, the total number of protozoa and Entodinium spp. population significantly decreased at 75 and 100% inclusions of banana leaf hay as roughage. 14. Investigation of waste banana peels and radish leaves for their ... African Journals Online (AJOL) This article is mainly based on the production of biodiesel and bioethanol from waste banana peels and radish leaves. The oily content from both the samples were converted to biodiesel by acid catalyzed and base catalyzed transesterification using methanol and ethanol. The biodiesel so obtained was subjected to ... 15. Optimization of in vitro multiplication for exotic banana (Musa spp ... African Journals Online (AJOL) win7 2015-06-17 Jun 17, 2015 ... drought stress and plants diseases, the banana production in Pakistan has ..... evaluated that hormonal response is specific genotypic dependent. ... BAP 4.0 mg/L-1 in interaction with 1.0 mg/L-1 IAA was suggested as optimal ... 16. Metal analyses of ash derived alkalis from banana and plantain ... African Journals Online (AJOL) The objective of this work was to determine the metal content of plantain and banana peels ash derived alkali and the possibility of using it as alternate and cheap source of alkali in soap industry. This was done by ashing the peels and dissolving it in de-ionised water to achieve the corresponding hydroxides with pH above ... 17. Coculture fermentation of banana agro-waste to ethanol by ... African Journals Online (AJOL) Banana is a major cash crop of many regions generating good amount of waste after harvest. This agro waste which is left for natural degradation is used as substrate for single step ethanol fermentation by thermophilic, cellulolytic, ethanologenic Clostridium thermocellum CT2, a new culture isolated from elephant ... 18. Top Quark Physics International Nuclear Information System (INIS) Larios, F. 2006-01-01 We give an overview of the physics of the Top quark, from the experimental discovery to the studies of its properties. We review some of the work done on the Electroweak and Flavor Changing couplings associated with the Top quark in the Standard Model and beyond. We will focus on the specific contribution of phycisits working in Mexico and Mexican physicists working abroad 19. Top quark theory NARCIS (Netherlands) Laenen, E. 2012-01-01 The theoretical aspects of a number of top quark properties such as its mass and its couplings are reviewed. Essential aspects in the theoretical description of top quark production, singly, in pairs and in association, as well as its decay related to spin and angular correlations are discussed. 20. Top physics at CDF Energy Technology Data Exchange (ETDEWEB) Hughes, R.E. [Univ. of Rochester, NY (United States) 1997-01-01 We report on top physics results using a 100 pb{sup -1} data sample of p{bar p} collisions at {radical}s = 1.8 TeV collected with the Collider Detector at Fermilab (CDF). We have identified top signals in a variety of decay channels, and used these channels to extract a measurement of the top mass and production cross section. A subset of the data (67 pb{sup -1}) is used to determine M{sub top} = 176 {+-} 8(stat) {+-} 10(syst) and {sigma}(tt) = 7.6 {sub -2.0}{sup +2.4} pb. We present studies of the kinematics of t{bar t} events and extract the first direct measurement of V{sub tb}. Finally, we indicate prospects for future study of top physics at the Tevatron. 1. Top Quark Mass CERN Document Server Mulders, Martijn 2016-01-01 Ever since the discovery of the top quark at the Tevatron collider in 1995 the measurement of its mass has been a high priority. As one of the fundamental parameters of the Standard Theory of particle physics, the precise value of the top quark mass together with other inputs provides a test for the self-consistency of the theory, and has consequences for the stability of the Higgs field that permeates the Universe. In this review I will briefly summarize the experimental techniques used at the Tevatron and the LHC experiments throughout the years to measure the top quark mass with ever improving accuracy, and highlight the recent progress in combining all measurements in a single world average combination. As experimental measurements became more precise, the question of their theoretical interpretation has become important. The difficulty of relating the measured quantity to the fundamental top mass parameter has inspired alternative measurement methods that extract the top mass in complementary ways. I wil... 2. Application of Cold Storage for Raja Sere Banana (Musa acuminata colla) Science.gov (United States) Crismas, S. R. S.; Purwanto, Y. A.; Sutrisno 2018-05-01 Raja Sere is one of the indigenous banana cultivars in Indonesia. This cultivar has a yellow color when ripen, small size and sweet taste. Traditionally, the growers market this banana cultivar to the market without any treatment to delay the ripening process. Banana fruits are commonly being harvested at the condition of hard green mature. At this condition of hard green mature, banana fruits can be stored for a long-term period. The objective of this study was to examine the effect of cold storage on the quality of raja sere banana that stored at 13°C. Banana fruits cultivar Raja Sere were harvested from local farmer field at the condition of hard green mature (about 14 weeks age after the flower bloom). Fifteen bunches of banana were stored in cold storage with a temperature of 13°C for 0, 3, 6, 9, and 12 days, respectively. For the control, room temperature storage (28°C) was used. At a storage period, samples of banana fruits ripened in the ripening chamber by injecting 100 ppm of ethylene gas at 25°C for 24 hours. The quality parameters namely respiration rate, hardness, total soluble solids (TSS), change in color, and weight loss were measured. For those banana fruits stored at room temperature, the shelf-life of banana was only reached up to 6 days. For those banana fruits stored in cold storage, the condition of banana fruits was reached up to 12 days. After cold storage and ripening, the third day measurement was the optimal time for bananas to be consumed which indicated by the yellow color (lightness value = 68.51, a* = 4.74 and value b* = 62.63), TSS 24.30 °Brix and hardness 0.48 kgf, weight loss about 7.53-16.45% and CO2 respiration rate of 100.37 mLCO2 / kg.hr. 3. Distribution, timing of attack, and oviposition of the banana weevil, Cosmopolites sordidus, on banana crop residues in Uganda NARCIS (Netherlands) Masanza, M.; Gold, C.S.; Huis, van A. 2005-01-01 Crop sanitation (removal and chopping of residue corms and pseudostems following plant harvest) has been recommended as a 'best bet' means of reducing banana weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae), populations. However, it has been unclear when such practices should be 4. Phylogeny of Banana Streak Virus reveals recent and repetitive endogenization in the genome of its banana host (Musa sp.). Science.gov (United States) Gayral, Philippe; Iskra-Caruana, Marie-Line 2009-07-01 Banana streak virus (BSV) is a plant dsDNA pararetrovirus (family Caulimoviridae, genus badnavirus). Although integration is not an essential step in the BSV replication cycle, the nuclear genome of banana (Musa sp.) contains BSV endogenous pararetrovirus sequences (BSV EPRVs). Some BSV EPRVs are infectious by reconstituting a functional viral genome. Recent studies revealed a large molecular diversity of episomal BSV viruses (i.e., nonintegrated) while others focused on BSV EPRV sequences only. In this study, the evolutionary history of badnavirus integration in banana was inferred from phylogenetic relationships between BSV and BSV EPRVs. The relative evolution rates and selective pressures (d(N)/d(S) ratio) were also compared between endogenous and episomal viral sequences. At least 27 recent independent integration events occurred after the divergence of three banana species, indicating that viral integration is a recent and frequent phenomenon. Relaxation of selective pressure on badnaviral sequences that experienced neutral evolution after integration in the plant genome was recorded. Additionally, a significant decrease (35%) in the EPRV evolution rate was observed compared to BSV, reflecting the difference in the evolution rate between episomal dsDNA viruses and plant genome. The comparison of our results with the evolution rate of the Musa genome and other reverse-transcribing viruses suggests that EPRVs play an active role in episomal BSV diversity and evolution. 5. Top quark mass measurement International Nuclear Information System (INIS) Maki, Tuula; Helsinki Inst. of Phys.; Helsinki U. of Tech. 2008-01-01 The top quark is the heaviest elementary particle. Its mass is one of the fundamental parameters of the standard model of particle physics, and an important input to precision electroweak tests. This thesis describes three measurements of the top-quark mass in the dilepton decay channel. The dilepton events have two neutrinos in the final state; neutrinos are weakly interacting particles that cannot be detected with a multipurpose experiment. Therefore, the signal of dilepton events consists of a large amount of missing energy and momentum carried off by the neutrinos. The top-quark mass is reconstructed for each event by assuming an additional constraint from a top mass independent distribution. Template distributions are constructed from simulated samples of signal and background events, and parameterized to form continuous probability density functions. The final top-quark mass is derived using a likelihood fit to compare the reconstructed top mass distribution from data to the parameterized templates. One of the analyses uses a novel technique to add top mass information from the observed number of events by including a cross-section-constraint in the likelihood function. All measurements use data samples collected by the CDF II detector 6. Top physics in WHIZARD Energy Technology Data Exchange (ETDEWEB) Reuter, Juergen; Chokoufe Nejad, Bijan [DESY, Hamburg (Germany). Theory Group; Bach, Fabian [European Commission, Eurostat, Luxembourg (Luxembourg); Hoang, Andre [Vienna Univ. (Austria). Faculty of Physics; Vienna Univ. (Austria). Erwin Schroedinger International Inst. for Mathematical Physics; Kilian, Wolfgang [Siegen Univ. (Germany); Stahlhofen, Maximilian [Mainz Univ. (Germany). PRISMA Cluster of Excellence; Teubner, Thomas [Liverpool Univ. (United Kingdom). Dept. of Mathematical Sciences; Weiss, Christian [DESY, Hamburg (Germany). Theory Group; Siegen Univ. (Germany) 2016-03-15 In this talk we summarize the top physics setup in the event generator WHIZARD with a main focus on lepton colliders. This includes full six-, eight- and ten-fermion processes, factorized processes and spin correlations. For lepton colliders, QCD NLO processes for top quark physics are available and discussed. A special focus is on the top-quark pair threshold, where a special implementation combines a non-relativistic effective field theory calculation augmented by a next-to-leading threshold logarithm resummation with a continuum relativistic fixed-order QCD NLO simulation. 7. Top physics in WHIZARD International Nuclear Information System (INIS) Reuter, Juergen; Chokoufe Nejad, Bijan; Hoang, Andre; Stahlhofen, Maximilian 2016-03-01 In this talk we summarize the top physics setup in the event generator WHIZARD with a main focus on lepton colliders. This includes full six-, eight- and ten-fermion processes, factorized processes and spin correlations. For lepton colliders, QCD NLO processes for top quark physics are available and discussed. A special focus is on the top-quark pair threshold, where a special implementation combines a non-relativistic effective field theory calculation augmented by a next-to-leading threshold logarithm resummation with a continuum relativistic fixed-order QCD NLO simulation. 8. Top quark discovered International Nuclear Information System (INIS) Anon. 1995-01-01 Nine months after a careful announcement of tentative evidence for the long-awaited sixth 'top' quark, physicists from the CDF and DO experiments at Fermilab's Tevatron proton-antiproton collider declared on 2 March that they had finally discovered the top quark. Last year (June 1994, page 1), the CDF experiment at the Tevatron reported a dozen candidate top events. These, said CDF, had all the characteristics expected of top, but the difficulties of extracting the tiny signal from a trillion proton-antiproton collisions made them shy of claiming a discovery. For its part, the companion DO Tevatron experiment reported a few similar events but were even more guarded about their interpretation as top quarks. Just after these hesitant announcements, performance at the Tevatron improved dramatically last summer. After the commissioning of a new linear accelerator and a magnet realignment, the machine reached a new world record proton-antiproton collision luminosity of 1.28 x 10 31 per sq cm per s, ten times that originally planned. Data began to pour in at an unprecedented rate and the data sample grew to six trillion collisions. Luminosity has subsequently climbed to 1.7 x 10 31 . The top quark is the final letter in the alphabet of Standard Model particles. According to this picture, all matter is composed of six stronglyinteracting subnuclear particles, the quarks, and six weakly interacting particles, the leptons. Both sextets are neatly arranged as three pairs in order of increasing mass. The fifth quark, the 'beauty' or 'b' quark, was also discovered at Fermilab, back in 1977. Since then physicists have been eagerly waiting for the top to turn up, but have been frustrated by its heaviness - the top is some 40 times the mass of its 'beautiful' partner. Not only is the top quark the heaviest by far, but it is the only quark which has been actively hunted. After the quarry was glimpsed last year, the net has now been 9. Audiitorite TOP 50 aastal 2005 Index Scriptorium Estoniae 2006-01-01 Audiitorite TOP. Vt. samas: Käibe TOP 10; Käibe kasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; ROE TOP 10; Ketlin Priilinn. Kasvu tagavad lojaalsed kliendid; Klient on audiitori parim müügimees; Teeli Remmelg. Kliendid vaatavad pigem kvaliteeti kui madalat hinda. Kommenteerivad Signe Keernik ja Kalle Lahe. Tabel: Audiitoriettevõtete üld- ja finantsandmed 10. Biomass waste-to-energy valorisation technologies: a review case for banana processing in Uganda. Science.gov (United States) Gumisiriza, Robert; Hawumba, Joseph Funa; Okure, Mackay; Hensel, Oliver 2017-01-01 Uganda's banana industry is heavily impeded by the lack of cheap, reliable and sustainable energy mainly needed for processing of banana fruit into pulp and subsequent drying into chips before milling into banana flour that has several uses in the bakery industry, among others. Uganda has one of the lowest electricity access levels, estimated at only 2-3% in rural areas where most of the banana growing is located. In addition, most banana farmers have limited financial capacity to access modern solar energy technologies that can generate sufficient energy for industrial processing. Besides energy scarcity and unreliability, banana production, marketing and industrial processing generate large quantities of organic wastes that are disposed of majorly by unregulated dumping in places such as swamps, thereby forming huge putrefying biomass that emit green house gases (methane and carbon dioxide). On the other hand, the energy content of banana waste, if harnessed through appropriate waste-to-energy technologies, would not only solve the energy requirement for processing of banana pulp, but would also offer an additional benefit of avoiding fossil fuels through the use of renewable energy. The potential waste-to-energy technologies that can be used in valorisation of banana waste can be grouped into three: Thermal (Direct combustion and Incineration), Thermo-chemical (Torrefaction, Plasma treatment, Gasification and Pyrolysis) and Biochemical (Composting, Ethanol fermentation and Anaerobic Digestion). However, due to high moisture content of banana waste, direct application of either thermal or thermo-chemical waste-to-energy technologies is challenging. Although, supercritical water gasification does not require drying of feedstock beforehand and can be a promising thermo-chemical technology for gasification of wet biomass such as banana waste, it is an expensive technology that may not be adopted by banana farmers in Uganda. Biochemical conversion technologies are 11. Evaluation of Banana Hypersensitivity Among a Group of Atopic Egyptian Children: Relation to Parental/Self Reports OpenAIRE El-Sayed, Zeinab A.; El-Ghoneimy, Dalia H.; El-Shennawy, Dina; Nasser, Manar W. 2013-01-01 Purpose To evaluate the frequency of banana sensitization and allergy among a group of atopic Egyptian children in relation to parental/self reports. Methods This is a case-control study included 2 groups of allergic children with and without history of banana allergy, each included 40 patients. They were subjected to skin prick test (SPT) using commercial banana allergen extract and prick-prick test (PPT) using raw banana, in addition to measuring the serum banana-specific IgE. Oral banana c... 12. Jõgevamaa ettevõtete TOP 55 Index Scriptorium Estoniae 2004-01-01 Jõgevamaa ettevõtete TOP 55 aastal 2003. Käibe TOP 40. Kasumi TOP 40. Käibe kasvu TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Jõgevamaa firmade üld- ja finantsandmed 13. Põlvamaa ettevõtete TOP 50 Index Scriptorium Estoniae 2004-01-01 Ettevõtete TOP 50. Käibe TOP 40. Kasumi TOP 40. Käibe kasvu TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Põlvamaa firmade üldandmed. Põlvamaa firmade finantsandmed 14. Viljandimaa ettevõtete TOP 50 Index Scriptorium Estoniae 2005-01-01 Viljandimaa ettevõtete TOP 50; Käibe TOP 35; Kasumi TOP 35; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Omakapitali tootlikkuse TOP 20; Viljandimaa ettevõtete üld- ja finantsandmed 15. Jõgevamaa ettevõtete TOP 50 Index Scriptorium Estoniae 2005-01-01 Jõgevamaa ettevõtete TOP 50; Käibe TOP 35; Kasumi TOP 35; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Omakapitali tootlikkuse TOP 20; Jõgevamaa ettevõtete üld- ja finantsandmed 16. Characterization of a new pathovar of Agrobacterium vitis causing banana leaf blight in China. Science.gov (United States) Huang, Siliang; Long, Mengling; Fu, Gang; Lin, Shanhai; Qin, Liping; Hu, Chunjin; Cen, Zhenlu; Lu, Jie; Li, Qiqin 2015-01-01 A new banana leaf blight was found in Nanning city, China, during a 7-year survey (2003-2009) of the bacterial diseases on banana plants. Eight bacterial strains were isolated from affected banana leaves, and identified as an intraspecific taxon of Agrobacterium vitis based on their 16S rDNA sequence similarities with those of 37 randomly selected bacterial strains registered in GenBank database. The representative strain Ag-1 was virulent on banana leaves and shared similar growth and biochemical reactions with the reference strain IAM14140 of A. vitis. The strains causing banana leaf blight were denominated as A. vitis pv. musae. The traditional A. vitis strains virulent to grapevines were proposed to be revised as A. vitis pv. vitis. This is the first record of a new type of A. vitis causing banana leaf blight in China. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim. 17. Boosted tops at ATLAS CERN Document Server Villaplana, M; The ATLAS collaboration 2011-01-01 A sample of candidate events for highly boosted top quarks is selected following the standard ATLAS selection for semi-leptonic ttbar events plus a requirement that the invariant mass of the reconstructed ttbar pair is greater than 700 GeV. Event displays are presented for the most promising candidates, as well as quantitative results for observables designed to isolate a boosted top quark signal. 18. Telekommunikatsiooni & arvutitootjate TOP Index Scriptorium Estoniae 2006-01-01 Telekommunikatsiooni TOP 30. Tabelid. Vt. samas: Indrek Kald. Elisa ka tänavu esikohal; Väike Tele2 võidab paindlikkusest; Kernel läks üle piiri. Diagrammid. Arvutifirmade TOP 100. Tabelid. Vt. samas: Enn Heinsoo. Esikoha tõi kontori müük; Proeksperdil suund välisturgudele;¡ Ida-Eesti turg arenevale firmale kitsas. Diagrammid. Nimekiri: Omanikud 19. Ky'osimba Onaanya: Understanding Productivity of East African Highland Banana OpenAIRE Taulya, G. 2016-01-01 Over 30 million people in East Africa depend on East African highland bananas for food and income. The bananas are grown with limited additions of nutrients and no irrigation, despite widespread poor soil fertility and regular dry seasons. This thesis describes the effect of increasing rainfall and application of potassium and nitrogen fertilizers on banana growth and yields. In areas that receive less than 1100 mm of rainfall per year, additional rainfall increases yields by 65%. Application... 20. The influence of gamma irradiation on shelf-life extension of banana International Nuclear Information System (INIS) Tiravat, K. 1971-01-01 Effect of various doses of gamma radiation on shelf-life extension of Hom Tong banana stored at 17 0 C, 20 0 C, and 23 0 C was described. Irradiated banana stored at 20 0 C and 23 0 C did not show any retardation in ripening. Doses from 20-40 Krad appeared to delay ripening of the banana stored at 17 0 C for 3-5 days. No significant difference in weight losses between irradiated and non-irradiated banana was detected during storage 1. Nitrogen and potassium fertilization on 'Caipira' and 'BRS Princesa' bananas in the Ribeira Valley. OpenAIRE NOMURA, E. S.; CUQUEL, F. L.; DAMATTO JUNIOR, E. R.; FUZITANI, E. J.; BORGES, A. L.; SAES, L. A. 2016-01-01 ABSTRACT ‘BRS Princesa’ (AAAB) and ‘Caipira’ (AAA) banana cultivars have similar sensorial features in comparison to the ‘Maçã’ banana. They are resistant to Panama disease, which allows them to grow in the Ribeira Valley, the largest banana plantation area in the São Paulo State. However, there is no information on how to fertilize crop under these edaphoclimatic conditions. This study aimed to evaluate the development and production of ‘Caipira’ and ‘BRS Princesa’ bananas, by applying four ... 2. Tobacco arabinogalactan protein NtEPc can promote banana (Musa AAA) somatic embryogenesis. Science.gov (United States) Shu, H; Xu, L; Li, Z; Li, J; Jin, Z; Chang, S 2014-12-01 Banana is an important tropical fruit worldwide. Parthenocarpy and female sterility made it impossible to improve banana varieties through common hybridization. Genetic transformation for banana improvement is imperative. But the low rate that banana embryogenic callus was induced made the transformation cannot be performed in many laboratories. Finding ways to promote banana somatic embryogenesis is critical for banana genetic transformation. After tobacco arabinogalactan protein gene NtEPc was transformed into Escherichia coli (DE3), the recombinant protein was purified and filter-sterilized. A series of the sterilized protein was added into tissue culture medium. It was found that the number of banana immature male flowers developing embryogenic calli increased significantly in the presence of NtEPc protein compared with the effect of the control medium. Among the treatments, explants cultured on medium containing 10 mg/l of NtEPc protein had the highest chance to develop embryogenic calli. The percentage of lines that developed embryogenic calli on this medium was about 12.5 %. These demonstrated that NtEPc protein can be used to promote banana embryogenesis. This is the first paper that reported that foreign arabinogalactan protein (AGP) could be used to improve banana somatic embryogenesis. 3. Biochemical effects of gamma irradiation on banana fruits International Nuclear Information System (INIS) El-Motaium, R.A. 1980-01-01 It is of important to study the extension of shelf-life at ambient temperature. This study would be of significant in the case of non- refrigerated transport, practices within the country and transhipment to distant countries. studies have therefore extended to assess the shelf-life of irradiated banana stored under-room temperature. Extension of shelf -life have been achieved by many methods, the most modern one is using gamma irradiation as a promising technology for developing nations. the aim of this investigation is to study the biochemical effects of gamma irradiation on G ros Michel m ature green banana fruits and also to determine the optimum dose level and the optimum storage conditions which resulted in, keeping the organoleptic qualities as it is and maximum extension in shelf-life 4. High oxygen levels promote peel spotting in banana fruit NARCIS (Netherlands) Maneenuam, T.; Ketsa, S.; Doorn, van W.G. 2007-01-01 We studied the effect of high oxygen on early peel spotting in Sucrier¿ bananas held at 25 °C and 90% RH. Fruit first ripened to colour index 3¿4 (about as yellow as green) and were then held in containers with a continuous gas flow of 18 ± 2 kPa (control) or 90 ± 2 kPa oxygen. High oxygen promoted 5. Biological control of banana black Sigatoka disease with Trichoderma OpenAIRE Poholl Adan Sagratzki Cavero; Rogério Eiji Hanada; Luadir Gasparotto; Rosalee Albuquerque Coelho Neto; Jorge Teodoro de Souza 2015-01-01 Black Sigatoka disease caused by Mycosphaerella fijiensis is the most severe banana disease worldwide. The pathogen is in an invasive phase in Brazil and is already present in most States of the country. The potential of 29 isolates of Trichoderma spp. was studied for the control of black Sigatoka disease under field conditions. Four isolates were able to significantly reduce disease severity and were further tested in a second field experiment. Isolate 2.047 showed the best results in both f... 6. Interaction of Spatially Localized LHW with Banana Particles Czech Academy of Sciences Publication Activity Database Krlín, Ladislav; Fuchs, Vladimír; Pánek, Radomír; Papřok, Richard; Seidl, Jakub 2014-01-01 Roč. 1, č. 3 (2014), s. 166-168 ISSN 2336-2626. [SPPT 2014 - 26th Symposium on Plasma Physics and Technology/26./. Prague, 16.06.2014-19.06.2014] R&D Projects: GA ČR GAP205/11/2341 Institutional support: RVO:61389021 Keywords : anomalous acceleration * stochasticity * lower hybrid waves * banana particles * tokamaks Subject RIV: BL - Plasma and Gas Discharge Physics http://fyzika.feld.cvut.cz/misc/ppt/articles/2014/krlin.pdf 7. Manual transportation within the plot and physical damages to bananas Directory of Open Access Journals (Sweden) Magalhães Mário Jorge Maia de 2004-01-01 Full Text Available The manual transportation of banana bunches within plots provokes physical damages to fruits compromising their quality. To assess the influence of the distance banana bunches travel on the shoulders of harvesters within the plot, on the incidence of physical damages present on the peel of fruits of the Nanicão cultivar, two experiments were carried out in the Vale do Ribeira region (SP, in sites with slope < 1%. Each experiment divided the plot in different distance bands, two of which were included in this study: one located far away from the collection roads (30-50 m and 80-100 m distance bands and another in an intermediate position (70-80 m and 130-150 m distance bands. For each distance band, six banana bunches of 36 mm gauged fruits were randomly sampled. Four banana hands were cut from the middle region of each bunch and ten fruits were assessed per hand, totaling 240 fruits per treatment. Bunches were harvested at the same maturity degree and those served as control were not transported. A total of 1440 fruits was assessed in the two experiments. The physical damages on the fruit surface were graded on a scale with 6 divisions: 0-0.25 cm²; 0.25-0.5 cm²; 0.5-1.0 cm²; 1.0-1.5 cm²; 1.5-2.0 cm²; 2.0-2.5 cm². The bunches transported on the shoulders of harvesters on distances over 70 m suffered increased (P < 0.01 damaged area. Most damages presented areas up to 0.5 cm². 8. The radurisation of bananas under commercial conditions: Pt. 1 International Nuclear Information System (INIS) Brodrick, H.T.; Strydom, G.J. 1984-01-01 In a large scale trial with bananas, a doubling of storage life in the ripening rooms (from 14 to 29 d) was achieved using irradiation treatment to an average dose of 0,85 kGy. Both colour development and fruit softening were significantly reduced in the irradiated fruits compared with the untreated batch. A slight phytotoxic effect to the fruit was noticed at the dose of 0,85 kGy 9. Molecular diagnostics for the sigatoka disease complex of banana. Science.gov (United States) Arzanlou, Mahdi; Abeln, Edwin C A; Kema, Gert H J; Waalwijk, Cees; Carlier, Jean; Vries, Ineke de; Guzmán, Mauricio; Crous, Pedro W 2007-09-01 ABSTRACT The Sigatoka disease complex of banana involves three related ascomycetous fungi, Mycosphaerella fijiensis, M. musicola, and M. eumusae. The exact distribution of these three species and their disease epidemiology remain unclear, because their symptoms and life cycles are rather similar. Disease diagnosis in the Mycosphaerella complex of banana is based on the presence of host symptoms and fungal fruiting structures, which hamper preventive management strategies. In the present study, we have developed rapid and robust species-specific molecular-based diagnostic tools for detection and quantification of M. fijiensis, M. musicola, and M. eumusae. Conventional species-specific polymerase chain reaction (PCR) primers were developed based on the actin gene that detected DNA at as little as 100, 1, and 10 pg/mul from M. fijiensis, M. musicola, and M. eumusae, respectively. Furthermore, TaqMan real-time quantitative PCR assays were developed based on the beta-tubulin gene and detected quantities of DNA as low as 1 pg/mul for each Mycosphaerella sp. from pure cultures and DNA at 1.6 pg/mul per milligram of dry leaf tissue for M. fijiensis that was validated using naturally infected banana leaves. 10. Ripening influences banana and plantain peels composition and energy content. Science.gov (United States) Emaga, Thomas Happi; Bindelle, Jérôme; Agneesens, Richard; Buldgen, André; Wathelet, Bernard; Paquot, Michel 2011-01-01 Musa sp. peels are widely used by smallholders as complementary feeds for cattle in the tropics. A study of the influence of the variety and the maturation stage of the fruit on fermentability and metabolisable energy (ME) content of the peels was performed using banana (Yangambi Km5) and plantain (Big Ebanga) peels at three stages of maturation in an in vitro model of the rumen. Peel samples were analysed for starch, free sugars and fibre composition. Samples were incubated in the presence of rumen fluid. Kinetics of gas production were modelled, ME content was calculated using prediction equation and short-chain fatty acids production and molar ratio were measured after 72 h of fermentation. Final gas production was higher in plantain (269-339 ml g(-1)) compared to banana (237-328 ml g(-1)) and plantain exhibited higher ME contents (8.9-9.7 MJ/kg of dry matter, DM) compared to banana (7.7-8.8 MJ/kg of DM). Butyrate molar ratio decreased with maturity of the peels. The main influence of the variety and the stage of maturation on all fermentation parameters as well as ME contents of the peels was correlated to changes in the carbohydrate fraction of the peels, including starch and fibre. 11. Banana fertigation with treated sanitary wastewater: postharvest and microbiological quality Directory of Open Access Journals (Sweden) Pablo Fernando Santos Alves 2017-06-01 Full Text Available Sewage may serve as a source of water and nutrients for plants. In this study, the effects of fertigation with treated sanitary wastewater from Janaúba Sewage Treatment Plant were evaluated on the postharvest and microbiological quality of ‘Prata-Anã’ banana. A randomized block experimental design was used. Four concentrations of wastewater were tested (70, 130, 170, and 200% of 150 kg ha-1 sodium. A wastewater-free control treatment was used for comparison. Two crop cycles were assessed for postharvest and microbiological quality. The parameters measured included total soluble solids, titratable acidity, total soluble solids/titratable acidity ratio, pH, total coliforms, and fecal coliforms on both the peel and the pulp. In the first crop cycle, both soluble solids and fruit pulp pH decreased as wastewater level increased up to a maximum of 141.5%. These correlations were not observed in the second cycle. Wastewater management did not affect the titratable acidity of the soluble solids. The agricultural application of treated sanitary wastewater provided banana fruits with a microbiological profile similar to that obtained with the control (pure water and with mineral fertilizers. A microbial balance is necessary to maintain the nutritional status of the banana crop. 12. PROTOCOL FOR HARVESTING ‘BRS PRINCESS’ BANANA FRUITS Directory of Open Access Journals (Sweden) LUIZ FERNANDO GANASSALI DE OLIVEIRA JUNIOR Full Text Available ABSTRACT The aim of this study was to develop a protocol to determine the ideal harvest time for ‘BRS Princess’ banana, using the number of aborted bracts. The bananas were selected according to the number of aborted bracts since the flowering until the time of harvest, yield clusters with 90, 95, 100 and 105 aborted bracts. The physical and chemical analyzes were performed every 3 days on fruits: soluble solids, titratable acidity, weight loss, length and diameter, pH, firmness, skin color (CIELab and pectin enzyme activity. The statistical design was completely randomized in a 4x5 factorial, with 4 points and 5 periods of harvest analysis and data were evaluated using analysis of variance and regression. For all parameters, fruits harvested at 90 and 105 aborted bracts had unwanted changes in its metabolism when compared to the other treatments, while fruits harvested at 95 and 100 aborted bracts had the best post-harvest characteristics. This method was effective in determining the point of harvest in ‘BRS Princess’ banana fruits, since it allows to obtain fruit quality after storage, and is a simple and objective method. 13. Mechanical properties of woven banana fibre reinforced epoxy composites International Nuclear Information System (INIS) Sapuan, S.M.; Leenie, A.; Harimi, M.; Beng, Y.K. 2006-01-01 In this paper, the experiments of tensile and flexural (three-point bending) tests were carried out using natural fibre with composite materials (Musaceae/epoxy). Three samples prepared from woven banana fibre composites of different geometries were used in this research. From the results obtained, it was found that the maximum value of stress in x-direction is 14.14 MN/m 2 , meanwhile the maximum value of stress in y-direction is 3.398 MN/m 2 . For the Young's modulus, the value of 0.976 GN/m 2 in x-direction and 0.863 GN/m 2 in y-direction were computed. As for the case of three-point bending (flexural), the maximum load applied is 36.25 N to get the deflection of woven banana fibre specimen beam of 0.5 mm. The maximum stress and Young's modulus in x-direction was recorded to be 26.181 MN/m 2 and 2.685 GN/m 2 , respectively. Statistical analysis using ANOVA-one way has showed that the differences of results obtained from those three samples are not significant, which confirm a very stable mechanical behaviour of the composites under different tests. This shows the importance of this product and allows many researchers to develop an adequate system for producing a good quality of woven banana fibre composite which maybe used for household utilities 14. Chemical compositions and glycemic responses to banana varieties. Science.gov (United States) Hettiaratchi, U P K; Ekanayake, S; Welihinda, J 2011-06-01 Chemical compositions and glycemic indices of four varieties of banana (Musa spp.) (kolikuttu-Silk AAB, embul-Mysore AAB, anamalu-Gros Michel AAA, seeni kesel-Pisang Awak ABB) were determined. Silk, Gros Michel, Pisang Awak and Mysore contained the highest percentages of starch (14%), sucrose (38%), free glucose (29%) and fructose (58%) as a percentage of the total available carbohydrate content respectively. Total dietary fiber contents of four varieties ranged from 2.7 to 5.3%. Glycemic indices of Silk, Mysore, Gros Michel and Pisang Awak were 61 ± 5, 61 ± 6, 67 ± 7, 69 ± 9 and can be categorized as low against white bread as the standard. A single banana of the four varieties elicited a low glycemic load. Thus, consumption of a banana from any of these varieties can be recommended as a snack for healthy or diabetic patients who are under dietary management or pharmacological drugs to regulate blood glucose responses in between meals. 15. Adsorption of Cu, As, Pb and Zn by Banana Trunk International Nuclear Information System (INIS) Nurzulaifa Shaheera Erne Mohd Yasim; Zitty Sarah Ismail; Suhanom Mohd Zaki; Mohd Fahmi Abd Azis 2016-01-01 The purpose of this study is to investigate the effectiveness of banana trunk as an adsorbent in removal of heavy metals in aqueous solution. Functional groups of adsorbent were determined using Fourier Transform Infrared spectroscopy (FTIR). Batch experiments were conducted to determine the adsorption percentage of heavy metals (Cu, As, Pb and Zn). The optimum adsorption using banana trunk was based on pH difference, contact time and dosage. Adsorption percentage was found to be proportional to pH, contact time and dosage. Maximum adsorption percentage of Cu, As, Pb and Zn at pH 6, 100 minutes and 8 gram of dosage are 95.80 %, 75.40 %, 99.36 % and 97.24 %, respectively. Langmuir and Freundlich isotherms were used to determine the equilibrium state for heavy metals ion adsorption experiments. All equilibrium heavy metals were well explained by the Freundlich isotherm model with R"2= 0.9441, R"2= 0.8671, R"2= 0.9489 and R"2= 0.9375 for Cu, As, Pb and Zn respectively. It is concluded that banana trunk has considerable potential for the removal of heavy metals from aqueous solution. (author) 16. The use of aggregation pheromone to enhance dissemination of Beauveria bassiana for the control of the banana weevil in Uganda NARCIS (Netherlands) Tinzaara, W.; Gold, C.S.; Dicke, M.; Huis, van A.; Nankinga, C.M.; Kagezi, G.H.; Ragama, P.E. 2007-01-01 Candidate strains of Beauveria bassiana were identified for use in integrated pest management of the banana weevil Cosmopolites sordidus. Horizontal field transmission of B. bassiana between banana weevils using different delivery systems, including aggregation pheromones, was investigated. We 17. TOP2017 Experimental summary CERN Document Server Giammanco, Andrea 2017-01-01 Thanks to the unprecedentedly fast accumulation of high-energy data at the LHC during the ongoing Run~2, most of the traditional top-quark analyses are experiencing the luxury of having to worry about how to punch through the Systematics Wall'', and think about new ways to maximize the utility of their data. New processes involving top quarks are being studied for the first time, and the good old pair-production processes are being explored in unusual settings, such as collisions involving heavy ions, or reference data'' collected by the LHC at relatively low centre-of-mass energy. The TOP2017 conference featured 37 talks delivered by experimental physicists, including seven in the Young Scientists Forum'' where young colleagues were given the opportunity to elaborate more deeply than usual on their own work. As it is impossible to do justice to all the experimental resu... 18. Top-ophilia Energy Technology Data Exchange (ETDEWEB) Quigg, Chris; /Fermilab 2008-01-01 Almost from the moment in June 1977 when the discovery of the Upsilon resonance revealed the existence of what we now call the bottom quark, physicists began searching for its partner. Through the years, as we established the electric charge and weak isospin of the b-quark, and detected the virtual influence of its mate, it became clear that the top quark must exist. Exactly at what mass, we couldn't say, but we knew just how top events would look. We also knew that top events would be rare--if the Tevatron could make them at all--and that picking out the events would pose a real challenge for the experimenters and their detectors. 19. The influence of crop management on banana weevil, Cosmopolites sordidus (Coleoptera: Curculionidae) populations and yield of highland cooking banana (cv. Atwalira) in Uganda. Science.gov (United States) Rukazambuga, N D T M; Gold, C S; Gowen, S R; Ragama, P 2002-10-01 A field study was undertaken in Uganda using highland cooking banana (cv. Atwalira) to test the hypothesis that bananas grown under stressed conditions are more susceptible to attack by Cosmopolites sordidus (Germar). Four banana treatments were employed to create different levels of host-plant vitality: (1) high stress: intercrop with finger millet; (2) moderate stress: monoculture without soil amendments; (3) low stress: monoculture with manure; (4) high vigour: monoculture with continuous mulch and manure. Adult C. sordidus were released at the base of banana mats 11 months after planting and populations were monitored for three years using mark and recapture methods. Cosmopolites sordidus density was greatest in the mulched plots which may have reflected increased longevity and/or longer tenure time in moist soils. Lowest C. sordidus numbers were found in intercropped banana. Damage, estimated as percentage corm tissue consumed by larvae, was similar among treatments. However, the total amount of tissue consumed was greater in mulched banana than in other systems. Plants supporting the heaviest levels of C. sordidus damage displayed bunch size reductions of 40-55%. Banana yield losses ranged from 14-20% per plot with similar levels in the intercropped and mulched systems. Yield reductions, reported as t ha-1, were twice as high in the mulched system as in the intercrop. The results from this study indicate that C. sordidus problems are not confined to stressed banana systems or those with low levels of management, but that the weevil can also attain pest status in well-managed and productive banana stands. 20. Cooking Banana Consumption Patterns in the Plantain-growing Area of Southeastern Nigeria Directory of Open Access Journals (Sweden) Tshiunza, M. 2001-01-01 Full Text Available Cooking bananas (Musa spp., ABB genome were intro-duced into Southeastern Nigeria by the International Institute of Tropical Agriculture (IITA in the mid-1980s as an interim measure to reduce the incidence of black sigatoka disease (caused by the fungus Mycosphaerel-la fijiensis Morelet on plantain. However, the people of this region were not familiar with their utilisation methods. To address this lack of the knowledge and thereby sustain cooking banana cultivation, IITA, in collaboration with the Shell Petroleum Development Company (SPDC and the Nigeria Agip OU Company (NAOC commenced a training campaign on cooking banana processing methods. This study examined the patterns of utilisation of cooking bananas ten years after the training took place and compared them with plantain. About 95 % of the households interviewed are consuming cooking banana, indicating a broad acceptance of the crop in the region. Overall, two ripening stages termed green and ripe are the most popular ripening stages for the consumption of both plantain and cooking banana, followed by partially ripe maturity stage. The most common forms of consumption for green plantain are, in decreasing order of importance, pottage, boiled, roasted, and fried. Green cooking banana is also mostly eaten in pottage and boiled forms, and less frequently in fried and pounded forms. Ripe plantain is mostly eaten in fried and pottage forms, while ripe cooking banana is mostly eaten in fried and raw forms. Partially ripe plantain is mostly eaten in pottage, fried, boiled, and roasted forms, while partially ripe cooking banana is eaten in fried, pottage and boiled forms. These results indicate that the consumption patterns of plantain and cooking banana are very similar. This similarity has greatly contributed to the rapid integration of cooking banana within the existing plantain consumption and cropping systems. 1. Pectinase production by Aspergillus niger using banana (Musa balbisiana) peel as substrate and its effect on clarification of banana juice. Science.gov (United States) Barman, Sumi; Sit, Nandan; Badwaik, Laxmikant S; Deka, Sankar C 2015-06-01 Optimization of substrate concentration, time of incubation and temperature for crude pectinase production from A. niger was carried out using Bhimkol banana (Musa balbisiana) peel as substrate. The crude pectinase produced was partially purified using ethanol and effectiveness of crude and partially purified pectinase was studied for banana juice clarification. The optimum substrate concentration, incubation time and temperature of incubation were 8.07 %, 65.82 h and 32.37 °C respectively, and the polygalacturonase (PG) activity achieved was 6.6 U/ml for crude pectinase. The partially purified enzyme showed more than 3 times of polygalacturonase activity as compared to the crude enzyme. The SDS-PAGE profile showed that the molecular weight of proteins present in the different pectinases varied from 34 to 42 kDa. The study further revealed that highest clarification was achieved when raw banana juice was incubated for 60 min with 2 % concentration of partially purified pectinase and the absorbance obtained was 0.10. 2. TOP LINAC design International Nuclear Information System (INIS) Picardi, L.; Ronsivalle, C.; Vignati, A. 1997-11-01 The report describes a linear accelerator for protons named TOP LINAC designed for the TOP (Terapia Oncologica con Protoni, Oncological Protontherapy) project launched by the Italian National Institute of Health (Istituto Superiore di Sanita', ISS) to explore in collaboration with the biggest Oncological Hospital in Rome (Istituto Regina Elena, IRE) the potentialities of the therapy with accelerated protons and establish guide lines for the application of this new type of radiotherapy in comparison with the more traditional electron and x-rays radiotherapy. The concept of a compact accelerator for protontherapy applications bore within the Italian Hadrontherapy Collaboration (TERA Collaboration) with the aim to diffuse the protontherapy on the National territory. The ISS program plans to use the TOP linac proton beam also for production of PET (Positron Emission Tomography) radioisotopes and radiobiology studies. Official agreements are in course between ISS and ENEA which provides its experience in the industrial and medical accelerators for the design and the construction of the TOP linac. The accelerator that will be the first 3 GHz proton linac in the world, will be composed of a 428.3 MHz 7 Me V RFQ + DTL injector followed by a 7-65 Me V section of a 3 GHz SCDTL structure and a 65 - 200 Me V variable energy SCL 3 GHz structure. In particular the SCDTL section uses a highly innovative accelerating structure patented by ENEA. In this report the clinical and physical requests are discussed and a preliminary design of the whole machine is given 3. Top emitting white OLEDs Energy Technology Data Exchange (ETDEWEB) Freitag, Patricia; Luessem, Bjoern; Leo, Karl [Technische Universitaet Dresden, Institut fuer Angewandte Photophysik, George-Baehr-Strasse 1, 01069 Dresden (Germany) 2009-07-01 Top emitting organic light emitting diodes (TOLEDs) provide a number of interesting opportunities for new applications, such as the opportunity to fabricate ITO-free devices by using opaque substrates. This makes it possible to manufacture low cost OLEDs for signage and lighting applications. A general top emitting device consists of highly reflecting metal contacts as anode and semitransparent cathode, the latter one for better outcouling reasons. In between several organic materials are deposited as charge transporting, blocking, and emission layers. Here, we show a top emitting white organic light emitting diode with silver electrodes arranged in a p-i-n structure with p- and n-doped charge transport layers. The centrical emission layer consists of two phosphorescent (red and green) and one fluorescent (blue) emitter systems separated by an ambipolar interlayer to avoid mutual exciton quenching. By adding an additional dielectric capping layer on top of the device stack, we achieve a reduction of the strong microcavity effects which appear due to the high reflection of both metal electrodes. Therefore, the outcoupled light shows broad and nearly angle-independent emission spectra, which is essential for white light emitting diodes. 4. Top quark theory Indian Academy of Sciences (India) 2012-10-04 Oct 4, 2012 ... The theoretical aspects of a number of top quark properties such as ... to the quadratic divergences of the Higgs self-energy, while yet, ..... given in the literature, each with the aim of recovering a well-behaved expansion in αs. 5. Top quark properties Indian Academy of Sciences (India) eter for the tests of the electroweak theory, since radiative corrections to many ... The uncertainty due to jet energy scale (JES) is the dominating systematic .... In the Standard Model, the charge of the top quark is predicted to be that of a normal up- ..... non-negative and f+ + f0 < 1, and the star marks the expectation from the ... 6. Anomalous top magnetic couplings Indian Academy of Sciences (India) 2012-11-09 Nov 9, 2012 ... Corresponding author. E-mail: [email protected]. Abstract. The real and imaginary parts of the one-loop electroweak contributions to the left and right tensorial anomalous couplings of the tbW vertex in the Standard Model (SM) are computed. Keywords. Top; anomalous. PACS Nos 14.65.Ha; 12.15 ... 7. Evidence for the presence of a female produced sex pheromone in the banana weevil, Cosmopolites sordidus Germar (Coleoptera: Curculionidae) Science.gov (United States) Behavior-modifying chemicals such as pheromones and kairomones have great potential in pest management. Studies reported here investigated chemical cues involved in mating and aggregation behavior of banana weevil, Cosmopolites sordidus, a major insect pest of banana in every country where bananas a... 8. "The Rotten Banana" Fires Back: The Story of a Danish Discourse of "Inclusive" Rurality in the Making Science.gov (United States) Winther, Malene Brandt; Svendsen, Gunnar Lind Haase 2012-01-01 The popularity of a particular term--the Rotten Banana--has paralleled the one-sided centralisation of public services since the Danish Municipal Reform of 2007. The Rotten Banana denotes peripheral Denmark, which takes a geographically curved form that resembles a banana, and it symbolises the belief that rural areas are backward and (too)… 9. CPT analysis with top physics Energy Technology Data Exchange (ETDEWEB) Cembranos, Jose A. R., E-mail: [email protected] [Universidad Complutense de Madrid, Departamento de Fisica Teorica I (Spain) 2013-03-15 We discuss the possibility of observing CPT violation from top anti-top production in hadronic colliders. We study a general approach by analyzing constraints on the mass difference between the top and anti-top quarks. We present current bounds from Tevatron data, and comment on the prospects for improving these bounds at the LHC and the ILC. 10. Top Quark Properties at Tevatron Energy Technology Data Exchange (ETDEWEB) Lysák, Roman [Prague, Inst. Phys. 2017-11-27 The latest CDF and D0 experiment measurements of the top quark properties except the top quark mass are presented. The final combination of the CDF and D0 forward-backward asymmetry measurements is shown together with the D0 measurements of the inclusive top quark pair cross-section as well as the top quark polarization. 11. Effect of thidiazuron on in vivo shoot proliferation of popular banana ... African Journals Online (AJOL) SARAH 2014-09-30 Sep 30, 2014 ... Mzuzu underscore the need for further studies to determine alternative best cytokine-based growth regulators. Key words: Thidiazuron, in vivo proliferation, Sucker growth, Banana. INTRODUCTION. In vivo macropropagation is an alternative technique for mass production of banana planting materials. 12. Effect of age, female mating status and density on the banana weevil response to aggregation pheromone NARCIS (Netherlands) Tinzaara, W.; Gold, C.S.; Dicke, M.; Huis, van A.; Ragama, P.E. 2011-01-01 The banana (Musa spp.) weevil (Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) is a major pest in East Africa causing yield losses of up to 14 metric tonnes per hectare annually. Laboratory and field experiments were conducted to determine whether the response of the banana weevil, 13. Effect of mulching on banana weevil movement relative to pheromone traps NARCIS (Netherlands) Tinzaara, W.; Gold, C.S.; Dicke, M.; Huis, van A.; Ragama, P.E. 2008-01-01 Banana weevil (Cosmopolites sordidus) is a major pest in East Africa causing yield losses of up to 14 metric tonnes per hectare annually. A study was conducted in Uganda to determine the effect of mulching on banana (Musa spp. L.) weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae), 14. Arabidopsis and Musa cyclin D2 expressed in banana (cv. “Sukali ... African Journals Online (AJOL) Genetic transformation of banana is important because of its polyploidy, sterility and long generation time of most cultivars which limit conventional breeding. However, transformability and regeneration of transgenic lines remains low in bananas. This research reports on the potential of CycD2 genes to improve ... 15. Effect of age, female mating status and density on the banana weevil ... African Journals Online (AJOL) The banana (Musa spp.) weevil (Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) is a major pest in East Africa causing yield losses of up to 14 metric tonnes per hectare annually. Laboratory and field experiments were conducted to determine whether the response of the banana weevil, Cosmopolites sordidus ... 16. Anti-nutrients and heavy metals in some new plantain and banana ... African Journals Online (AJOL) Plantain and banana flour are important raw material in the baking and confectionery industry, and complementary food formulation. Five new plantain and banana hybrids developed by the International Institute of Tropical Agriculture (IITA) at Highrainfall Station, Onne, Nigeria were screened for certain anti-nutritional ... 17. 33 CFR 334.560 - Banana River at Patrick Air Force Base, Fla.; restricted area. Science.gov (United States) 2010-07-01 ... 33 Navigation and Navigable Waters 3 2010-07-01 2010-07-01 false Banana River at Patrick Air Force Base, Fla.; restricted area. 334.560 Section 334.560 Navigation and Navigable Waters CORPS OF ENGINEERS, DEPARTMENT OF THE ARMY, DEPARTMENT OF DEFENSE DANGER ZONE AND RESTRICTED AREA REGULATIONS § 334.560 Banana... 18. The Draft Genome Sequence of Mycosphaerella fijiensis, the Black Sigatoka Pathogen of Banana Science.gov (United States) Mycosphaerella fijiensis is a fungal pathogen of banana and the causal agent of the devastating Black Sigatoka or black leaf streak disease. Its control requires weekly fungicide applications when bananas are grown under disease-conducive conditions, which mostly represent precarious tropical enviro... 19. Exploring the Potential of Banana SAP as Dye for the Adinkra ... African Journals Online (AJOL) A study was carried out to explore the potential of banana sap as a dye for the Adinkra industry in Ghana. Pseudostem extract of banana and stem bark extract of Bridelia micratha were compared as dyeing stuff. A consumer preference study was also conducted to assess the acceptability of the products developed. 20. Impacts of Climate Variability and Change on Banana Yields in the ... African Journals Online (AJOL) Climate variability and change are existing sets of conditions which affect crop productivity. An evaluation of their impacts on banana yield in the CDC-DelMonte Banana Project at Tiko is fundamental in conceiving adaptation strategies towards coping with, and minimizing their deleterious impacts for maximum productivity ... 1. Attitudes, perceptions, and trust. Insights from a consumer survey regarding genetically modified banana in Uganda NARCIS (Netherlands) Kikulwe, E.M.; Wesseler, J.H.H.; Falck-Zepeda, J. 2011-01-01 Genetically modified (GM) crops and food are still controversial. This paper analyzes consumers’ perceptions and institutional awareness and trust toward GM banana regulation in Uganda. Results are based on a study conducted among 421 banana-consuming households between July and August 2007. Results 2. The banana (Musa acuminata) genome and the evolution of monocotyledonous plants NARCIS (Netherlands) Hont, D' A.; Denoeud, F.; Aury, J.M.; Kema, G.H.J.; Dita Rodriguez, M.A.; Waalwijk, C. 2012-01-01 Bananas (Musa spp.), including dessert and cooking types, are giant perennial monocotyledonous herbs of the order Zingiberales, a sister group to the well-studied Poales, which include cereals. Bananas are vital for food security in many tropical and subtropical countries and the most popular fruit 3. Effect of crop sanitation on banana weevil Cosmopolites sordidus (Germar) populations and associated damage NARCIS (Netherlands) Masanza, M. 2003-01-01 The banana weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) is a serious pest of bananas. However, its ecology is not well elucidated especially in East Africa where plantations are up to 50 years old and are under various management and cropping systems. No single 4. Ky’osimba Onaanya: understanding productivity of East African Highland banana NARCIS (Netherlands) Taulya, G. 2015-01-01 Over 30 million people in East Africa depend on East African highland bananas for food and income. The bananas are grown with limited additions of nutrients and no irrigation, despite widespread poor soil fertility and regular dry seasons. This thesis describes the effect of increasing rainfall 5. Cereal bars produced with banana peel flour: evaluation of acceptability and sensory profile. Science.gov (United States) Carvalho, Vania Silva; Conti-Silva, Ana Carolina 2018-01-01 A mixture design was used to investigate the effects of banana peel flour, rice flakes and oat flour on sensory acceptability of cereal bars, with subsequent evaluation of sensory profile of products identified as having high acceptability. Regions of greater response for acceptability of the cereal bars, which are dependent on the three investigated components, were found. Although having good acceptability, sensory profiles of cereal bars were different. A cereal bar with the lowest quantity of banana peel flour was described as having a higher amount of rice flakes, chewiness and crispness, while formulations with intermediate and highest quantities of banana peel flour were described by darker color, higher banana aroma and bitter taste. Contrary to expectations, banana flavor of cereal bar with highest quantity of banana peel flour was lower than cereal bars with intermediate quantities. Cereal bars were not different in terms of hardness and adhesiveness and they also had a similar sweet taste and oat flavor. The use of banana peel flour in production of cereal bars is feasible and, even with different sensory profiles, cereal bars with banana peel flour are acceptable, which may favor the development of new products for different market niches. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry. 6. Effect of green banana pulp on physicochemical and sensory properties of probiotic yoghurt Directory of Open Access Journals (Sweden) Elizabete Lourenço da COSTA Full Text Available Abstract In order to investigate the potential of the green banana as a prebiotic, and for its content of resistant starch, fermented yogurts were produced by cultures composed of Lactobacillus delbrueckii, Streptococcus thermophilus, Bifidobacterium bifidum and Lactobacillus acidophilus as well as being enriched with three concentrations of industrialized green banana pulp (GBP (3%, 5% and 10% w/v. The green banana pulp added to the yogurt stimulated the multiplication of L. acidophilus after the first day of fermentation and B. bifidum after seven days in cold storage compared to the control that consisted of yogurt without the addition of green banana pulp. The dose-response effect was not observed; however, the results show that the green banana pulp has a prebiotic potential without interfering with either the physicochemical or sensorial characteristics. 7. It is only a banana-Traveltime sensitivity kernels using the unwrapped phase KAUST Repository Djebbi, Ramzi; Alkhalifah, Tariq Ali 2012-01-01 Traveltime sensitivity kernels for finite-frequency traveltimes computed using the Born or Rytov approximations admits hallow banana shaped responses in the plane of propagation and a circular doughnut shaped responses in the cross section. This suggests that finite-frequency traveltimes are insensitive to velocity information along the infinite-frequency ray path, which is obviously inaccurate and creates a disconnect in the traveltime dependency on frequency. Using the instantaneous traveltime of the wavefield, which is capable of unwrapping the phase function, we obtain traveltime sensitivity kernels that have plain banana shape responses, with the thickness of the banana governed by the investigated frequency. This result confirms that the hallow banana shape is simply a result of the wrapping of the phase of the wavefield, in which Born nor Rytov approximations can properly deal with. The instantaneous traveltime can, thus, mitigate the nonlinearity problem encountered in finite-frequency traveltime inversions that may arise from these hallow banana sensitivity kernels. 8. Effect of physiological harvest stages on the composition of bioactive compounds in Cavendish bananas* Science.gov (United States) Bruno Bonnet, Christelle; Hubert, Olivier; Mbeguie-A-Mbeguie, Didier; Pallet, Dominique; Hiol, Abel; Reynes, Max; Poucheret, Patrick 2013-01-01 The combined influence of maturation, ripening, and climate on the profile of bioactive compounds was studied in banana (Musa acuminata, AAA, Cavendish, cv. Grande Naine). Their bioactive compounds were determined by the Folin-Ciocalteu assay and high-performance thin layer chromatographic (HPTLC) method. The polyphenol content of bananas harvested after 400 degree days remained unchanged during ripening, while bananas harvested after 600 and 900 degree days exhibited a significant polyphenol increase. Although dopamine was the polyphenol with the highest concentration in banana peels during the green developmental stage and ripening, its kinetics differed from the total polyphenol profile. Our results showed that this matrix of choice (maturation, ripening, and climate) may allow selection of the banana (M. acuminata, AAA, Cavendish, cv. Grande Naine) status that will produce optimal concentrations of identified compounds with human health relevance. PMID:23549844 9. Effects of Increasing Levels of Dietary Cooked and Uncooked Banana Meal on Growth Performance and Carcass Parameters of Broiler Chicken Directory of Open Access Journals (Sweden) N.S.B.M Atapattu* and T.S.M.S. Senevirathne 2013-04-01 Full Text Available Discarded banana is a valuable feed ingredient for poultry feed formulations. However, due to the presence of resistant starches, inclusion of more than 10% banana meal in poultry rations reduces the growth performance. The objective of this study was to determine whether higher levels of banana meal could be included in broiler diets if raw banana is cooked before being processed into meal. Discarded banana (Cavendish collected at harvesting was processed into two types of banana meals. Cooked banana meal was prepared by cooking banana at 100oC for 15 minutes and subsequent drying. Uncooked banana meal was prepared by drying at 800C for three days. Giving a 2 x 4 factorial arrangement, 144 broiler chicks in 48 cages received one of the eight experimental diets containing either cooked or uncooked banana meal at 0, 10, 20 or 30% ad libitum from day 21-42. Birds fed cooked banana meal were significantly heavier on day 28 and 35. Live weight on day 42, weight gain, feed intake or feed conversion efficiency were not affected either by the type or level of banana meal and their interaction. Cooked banana meal increased the weights of the crop and liver significantly. Weight of the small intestine, proventriculus, gizzard abdominal fat pad and the fat free tibia ash contents were not affected by the dietary treatments. It was concluded that uncooked banana meal produced using peeled raw banana can be included up to 30% in nutritionally balanced broiler finisher diets without any adverse effects on performance. 10. Landfill Top Covers DEFF Research Database (Denmark) Scheutz, Charlotte; Kjeldsen, Peter 2011-01-01 The purpose of the final cover of a landfill is to contain the waste and to provide for a physical separation between the waste and the environment for protection of public health. Most landfill covers are designed with the primary goal to reduce or prevent infiltration of precipitation...... into the landfill in order to minimize leachate generation. In addition the cover also has to control the release of gases produced in the landfill so the gas can be ventilated, collected and utilized, or oxidized in situ. The landfill cover should also minimize erosion and support vegetation. Finally the cover...... is landscaped in order to fit into the surrounding area/environment or meet specific plans for the final use of the landfill. To fulfill the above listed requirements landfill covers are often multicomponent systems which are placed directly on top of the waste. The top cover may be placed immediately after... 11. Computing Z-top International Nuclear Information System (INIS) Kashani-Poor, A.K. 2014-01-01 The topological string presents an arena in which many features of string theory proper, such as the interplay between world-sheet and target space descriptions or open-closed duality, can be distilled into computational techniques which yield results beyond perturbation theory. In this thesis, I will summarize my research activity in this area. The presentation is organized around computations of the topological string partition function Z-top based on various perspectives on the topological string. (author) 12. Top of the list International Nuclear Information System (INIS) Cameron, A.; Vries, E. de 2006-01-01 In this article the authors look at how the major turbine suppliers fared in the year 2005, and look forward to 2006 which could be the best ever for the wind industry. The world wind turbine market continues to be dominated by ten major companies, who together account for almost 100% of the total global market. Although there are some new companies on the horizon it is the ten major companies that the authors concentrate on, in particular the top three, Vestas, Gamesa and Enercon who between them control 66% of the market. They give their overall report on business in 2005 and give their predictions for 2006. By far the leading wind turbine supplier is the Danish Company Vestas, accounting for over 30% of the global market. However all of the top ten companies report a successful year and all are expecting an even better 2006. This paper also reports on potential newcomers to the industry, one of these is California-based Clipper Windpower which commenced manufacturing the 2.5 MW Liberty turbine in Iowa. Chinese manufacturer Goldwind is also looking to enter the market having acquired a licence to produce the Vensys 1.2 MW turbine. It may well be that changes are on the way, there will be jockeying for position at the top and newcomers will want to make an impression on the market 13. DISTRIBUTION AND INCIDENCE LEAF DISEASES OF BANANA IN SEVERAL BANANA PRODUCTION CENTERS IN NORTH SUMATRA, WEST SUMATRA BENGKULU AND WEST JAVA Directory of Open Access Journals (Sweden) Sahlan 2011-06-01 Full Text Available The research was aimed to determine the type, the distribution and the incidence of banana leaf diseases in several production centers in West Sumatra, Bengkulu, North Sumatra and West Java. Direct observations on banana orchards were conducted in some districts in Simalungun, Deli Serdang and Medan (North Sumatra, Tanah Datar, Limapuluh Kota, Agam, Pariaman and Pasaman (West Sumatra, Rejang Lebong and Kepahyang (Bengkulu, Sukabumi, Purwakarta and Subang (West Java from November to December 2006. Two banana orchards were randomly selected in each district. Plant population at the selected orchard was at least 100 plants. From each sampled orchard, if banana population consisted of similar or only one variety, 10 plants were randomly chosen according to wind direction. Meanwhile, when the banana varieties were varied, five plants were randomly selected. The result showed that Black Sigatoka and Eumusae leaf spot were found in West Sumatra, Bengkulu and North Sumatra at severity level of between 15 % to 62.31%, whilst speckle disease was mainly found in North Sumatra and in parts of West Sumatra at severity level of between 72,72% to 100% and 15 to 30%, respectively. Banana varieties that were primarily attacked by leaf diseases were Cavendish, Telor, Barangan and Emas. 14. Top quark measurements at ATLAS CERN Document Server Grancagnolo, Sergio; The ATLAS collaboration 2017-01-01 The top quark is the heaviest known fundamental particle. As it is the only quark that decays before it hadronizes, this gives us the unique opportunity to probe the properties of bare quarks at the Large Hadron Collider. This talk will present highlights of a few recent precision measurements by the ATLAS Collaboration of the top quark using 13 TeV and 8 TeV collision data: top-quark pair and single top production cross sections including differential distributions will be presented alongside top quark properties measurements. These measurements, including results using boosted top quarks, probe our understanding of top quark production in the TeV regime. Measurements of the top quark mass and searches for rare top quark decays are also presented. 15. Top quark measurements at ATLAS CERN Document Server AUTHOR|(INSPIRE)INSPIRE-00041686; The ATLAS collaboration 2017-01-01 The top quark is the heaviest known fundamental particle. As it is the only quark that decays before it hadronizes, it allows us to probe the properties of bare quarks at the Large Hadron Collider. Highlights of a few recent precision measurements by the ATLAS Collaboration of the top quark using 13 TeV and 8 TeV collision data will be presented: top-quark pair and single top production cross sections including differential distributions will be presented alongside measurements of top-quark properties, including results using boosted top quarks, probe our understanding of top-quark production in the TeV regime. Measurements of the top-quark mass and searches for rare top quark decays are also presented. 16. Mechanical properties of woven banana fibre reinforced epoxy composites Energy Technology Data Exchange (ETDEWEB) Sapuan, S.M. [Department of Mechanical and Manufacturing Engineering, Universiti Putra Malaysia, 43400 Serdang, Selangor (Malaysia)]. E-mail: [email protected]; Leenie, A. [Department of Mechanical and Manufacturing Engineering, Universiti Putra Malaysia, 43400 Serdang, Selangor (Malaysia); Harimi, M. [School of Engineering and Information Technology, Universiti Malaysia Sabah, 88999 Kota Kinabalu, Sabah (Malaysia); Beng, Y.K. [School of Engineering and Information Technology, Universiti Malaysia Sabah, 88999 Kota Kinabalu, Sabah (Malaysia) 2006-07-01 In this paper, the experiments of tensile and flexural (three-point bending) tests were carried out using natural fibre with composite materials (Musaceae/epoxy). Three samples prepared from woven banana fibre composites of different geometries were used in this research. From the results obtained, it was found that the maximum value of stress in x-direction is 14.14 MN/m{sup 2}, meanwhile the maximum value of stress in y-direction is 3.398 MN/m{sup 2}. For the Young's modulus, the value of 0.976 GN/m{sup 2} in x-direction and 0.863 GN/m{sup 2} in y-direction were computed. As for the case of three-point bending (flexural), the maximum load applied is 36.25 N to get the deflection of woven banana fibre specimen beam of 0.5 mm. The maximum stress and Young's modulus in x-direction was recorded to be 26.181 MN/m{sup 2} and 2.685 GN/m{sup 2}, respectively. Statistical analysis using ANOVA-one way has showed that the differences of results obtained from those three samples are not significant, which confirm a very stable mechanical behaviour of the composites under different tests. This shows the importance of this product and allows many researchers to develop an adequate system for producing a good quality of woven banana fibre composite which maybe used for household utilities. 17. Naisjuhtidega ettevõtete TOP 100 Index Scriptorium Estoniae 2004-01-01 Naisjuhtidega ettevõtete TOP 100. Käibe TOP 20. Käibe kasvu TOP 20. Kasumi TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Riigi- ja kohaliku omavalitsuse asutuste naisjuhtide TOP 25. Riigi- ja kohaliku omavalitsuse asutuste eelarve TOP 25. Riigi- ja kohaliku omavalitsuse asutuste töötajate arvu TOP 25. Riigi- ja kohaliku omavalitsuse asutuste naisjuhtide palga TOP 25 18. Boron toxicity in banana (Musa AAA) plantations of Costa Rica International Nuclear Information System (INIS) Vargas, Alfonso; Serrano, Edgardo; Arias, Fulvio; Arias M, Oscar 2007-01-01 A marginal, irregular and continuous necrosis was observed in the leaves of in banana plants (Musa AAA, cvs. Grande Naine and Valery), This necrosis was developed from an irregular chlorotic area, from the edge towards the internal part of the leaf blade. The central portion of the leaf kept the original green color. Soil and foliar analyses showed that symptoms were caused by high boron concentrations, probably due to excessive soil or foliage applications of the nutriment, or to the effect of very frequent applications of boron during fertigation, combined with a decrease of calcium in the leaf. (author) [es 19. Assessment of RNAi-induced silencing in banana (Musa spp.). Science.gov (United States) Dang, Tuong Vi T; Windelinckx, Saskia; Henry, Isabelle M; De Coninck, Barbara; Cammue, Bruno P A; Swennen, Rony; Remy, Serge 2014-09-18 In plants, RNA- based gene silencing mediated by small RNAs functions at the transcriptional or post-transcriptional level to negatively regulate target genes, repetitive sequences, viral RNAs and/or transposon elements. Post-transcriptional gene silencing (PTGS) or the RNA interference (RNAi) approach has been achieved in a wide range of plant species for inhibiting the expression of target genes by generating double-stranded RNA (dsRNA). However, to our knowledge, successful RNAi-application to knock-down endogenous genes has not been reported in the important staple food crop banana. Using embryogenic cell suspension (ECS) transformed with ß-glucuronidase (GUS) as a model system, we assessed silencing of gusAINT using three intron-spliced hairpin RNA (ihpRNA) constructs containing gusAINT sequences of 299-nt, 26-nt and 19-nt, respectively. Their silencing potential was analysed in 2 different experimental set-ups. In the first, Agrobacterium-mediated co-transformation of banana ECS with a gusAINT containing vector and an ihpRNA construct resulted in a significantly reduced GUS enzyme activity 6-8 days after co-cultivation with either the 299-nt and 19-nt ihpRNA vectors. In the second approach, these ihpRNA constructs were transferred to stable GUS-expressing ECS and their silencing potential was evaluated in the regenerated in vitro plants. In comparison to control plants, transgenic plants transformed with the 299-nt gusAINT targeting sequence showed a 4.5 fold down-regulated gusA mRNA expression level, while GUS enzyme activity was reduced by 9 fold. Histochemical staining of plant tissues confirmed these findings. Northern blotting used to detect the expression of siRNA in the 299-nt ihpRNA vector transgenic in vitro plants revealed a negative relationship between siRNA expression and GUS enzyme activity. In contrast, no reduction in GUS activity or GUS mRNA expression occurred in the regenerated lines transformed with either of the two gusAINT oligo target 20. Banana regime pressure anisotropy in a bumpy cylinder magnetic field International Nuclear Information System (INIS) Garcia-Perciante, A.L.; Callen, J.D.; Shaing, K.C.; Hegna, C.C. 2006-01-01 The pressure anisotropy is calculated for a plasma in a bumpy cylindrical magnetic field in the low collisionality (banana) regime for small magnetic-field modulations (ε≡ΔB/2B parallel is then calculated and is shown to exceed the flux-surface-averaged parallel viscous force parallel > by a factor of O(1/ε). A high-frequency limit (ω>>ν) for the pressure anisotropy is also determined and the calculation is then extended to include the full frequency dependence by using an expansion in Cordey eigenfunctions 1. Effect of banana pulp and peel flour on physicochemical properties and in vitro starch digestibility of yellow alkaline noodles. Science.gov (United States) Ramli, Saifullah; Alkarkhi, Abbas F M; Shin Yong, Yeoh; Min-Tze, Liong; Easa, Azhar Mat 2009-01-01 The present study describes the utilization of banana--Cavendish (Musa acuminata L., cv cavendshii) and Dream (Musa acuminata colla. AAA, cv 'Berangan')--pulp and peel flours as functional ingredients in yellow alkaline noodles. Noodles were prepared by partial substitution of wheat flour with ripe banana pulp or peel flours. In most cases, the starch hydrolysis index, predicted glycaemic index (pGI) and physicochemical properties of cooked noodles were affected by banana flour addition. In general, the pGI values of cooked noodles were in the order; banana peel noodles banana pulp noodles peel flour was higher in total dietary fibre but lower in resistant starch contents than the pulp flour, the low pGI of banana peel noodles was mainly due to its high dietary fibre content. In conclusion, banana pulp and peel flour could be useful for controlling starch hydrolysis of yellow noodles, even though some physicochemical properties of the noodles were altered. 2. Influence of ripeness of banana on the blood glucose and insulin response in type 2 diabetic subjects. Science.gov (United States) Hermansen, K; Rasmussen, O; Gregersen, S; Larsen, S 1992-10-01 Banana is a popular and tasty fruit which often is restricted in the diet prescribed for diabetic patients owing to the high content of free sugars. However, in under-ripe bananas starch constitutes 80-90% of the carbohydrate content, which as the banana ripens changes into free sugars. To study the effect of ripening on the postprandial blood glucose and insulin responses to banana, 10 type 2 (non-insulin-dependent) diabetic subjects consumed three meals, consisting of 120 g under-ripe banana, 120 g over-ripe banana or 40 g white bread on separate days. The mean postprandial blood glucose response area to white bread (181 +/- 45 mmol l-1 x 240 min) was significantly higher compared with under-ripe banana (62 +/- 17 mmol l-1 x 240 min: p alfa-amylase in humans.(ABSTRACT TRUNCATED AT 250 WORDS) 3. Study of top and anti-top mass difference CERN Document Server Leedumrongwatthanakun, Saroch 2013-01-01 The invariance of the standard model under CPT transformations leads to the equality of particle and antiparticle masses. The recent measurements performed by the CMS experiment on the top anti-top mass difference are a test of such symmetry. In this work non-perturbative QCD effects, which may eventually lead to an apparent difference in the mass of a top and anti-top quark, are studied. 4. Tekstiilitööstuse TOP 47 Index Scriptorium Estoniae 2005-01-01 Tekstiilitööstuse TOP 47; Käibe TOP 35; Kasumi TOP 35; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Varade tootlikkuse TOP 20; Tekstiilitööstuse TOP-i firmade üld- ja finantsandmed 5. Rõivatööstuse TOP 50 Index Scriptorium Estoniae 2005-01-01 Rõivatööstuse TOP 50; Käibe TOP 35; Kasumi TOP 35; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Omakapitali tootlikkuse TOP 20; Rõivatööstuse TOP-i firmade üld- ja finantsandmed 6. Eesti Ettevõtete TOP 100 Index Scriptorium Estoniae 2002-01-01 TOP 100. Käibe TOP 500. Käibe kasvu TOP 100. Kasumi TOP 100. Kasumi kasvu TOP 100. Rentaabluse TOP 100. Omakapitali tootlikkuse TOP 100. Eesti edukamate ettevõtete üldandmed. Eesti edukamate ettevõtete finantsnäitajad. Valdkonna ja maakonna TOP-ide edukamate ettevõtete finantsnäitajad 7. Effect of LED irradiation on the ripening and nutritional quality of postharvest banana fruit. Science.gov (United States) Huang, Jen-Yi; Xu, Fengying; Zhou, Weibiao 2018-04-24 With the ability to tailor wavelengths necessary to the photosynthetically active radiation spectrum of plant pigments, light-emitting diodes (LEDs) offer vast possibilities in horticultural lighting. The influence of LED light irradiation on major postharvest features of banana was investigated. Mature green bananas were treated daily with selected blue (464-474 nm), green (515-525 nm) and red (617-627 nm) LED lights for 8 days, and compared with non-illuminated control. The positive effect of LED lighting on the acceleration of ripening in bananas was greatest for blue, followed by red and green. Under the irradiation of LED lights, faster peel de-greening and flesh softening, and increased ethylene production and respiration rate in bananas were observed during storage. Furthermore, the accumulations of ascorbic acid, total phenols, and total sugars in banana fruit were enhanced by LED light exposure. LED light treatment can induce the ripening of bananas and improve their quality and nutrition potential. These findings might provide new chemical-free strategies to shorten the time to ripen banana after harvest by using LED light source. This article is protected by copyright. All rights reserved. 8. Bacterial Diseases of Bananas and Enset: Current State of Knowledge and Integrated Approaches Toward Sustainable Management Directory of Open Access Journals (Sweden) Guy Blomme 2017-07-01 Full Text Available Bacterial diseases of bananas and enset have not received, until recently, an equal amount of attention compared to other major threats to banana production such as the fungal diseases black leaf streak (Mycosphaerella fijiensis and Fusarium wilt (Fusarium oxysporum f. sp. cubense. However, bacteria cause significant impacts on bananas globally and management practices are not always well known or adopted by farmers. Bacterial diseases in bananas and enset can be divided into three groups: (1 Ralstonia-associated diseases (Moko/Bugtok disease caused by Ralstonia solanacearum and banana blood disease caused by R. syzygii subsp. celebesensis; (2 Xanthomonas wilt of banana and enset, caused by Xanthomonas campestris pv. musacearum and (3 Erwinia-associated diseases (bacterial head rot or tip-over disease Erwinia carotovora ssp. carotovora and E. chrysanthemi, bacterial rhizome and pseudostem wet rot (Dickeya paradisiaca formerly E. chrysanthemi pv. paradisiaca. Other bacterial diseases of less widespread importance include: bacterial wilt of abaca, Javanese vascular wilt and bacterial fingertip rot (probably caused by Ralstonia spp., unconfirmed. This review describes global distribution, symptoms, pathogenic diversity, epidemiology and the state of the art for sustainable disease management of the major bacterial wilts currently affecting banana and enset. 9. Effect of Banana Stalk Organic Fertilizer on the Growth of Chinese Cabbage Institute of Scientific and Technical Information of China (English) Zheli; DING; Lina; HAN; Zhiqiang; JIN; Bizun; WANG; Huicai; ZENG; Wei; ZHENG; Yingdui; HE; Xiaoping; ZANG 2016-01-01 In order to solve the problem of waste disposal after banana harvest,we use banana stalk to produce banana stalk organic fertilizer,through a plot experiment. We compare the influence of normal organic fertilizer( Wanlubao) and banana stalk organic fertilizer as base fertilizers on Chinese cabbage growth,and evaluate the economic benefits of banana stalk organic fertilizer. The results show that organic fertilizer has little effect on water content and nutrient content of Chinese cabbage,but has significant effect on plant height and leaf width. Using organic fertilizer can increase the production of Chinese cabbage by 22. 50%- 43. 10%. With 6750 kg / ha normal organic fertilizer,Chinese cabbage gets the highest yield,which reaches 30135 kg / ha,followed by the treatment of 6750 kg / ha stalk organic fertilizer. At farmers’ conventional fertilization level( 4500 kg / ha),stalk organic fertilizer can increase the yield by more than 3. 50% in comparison with the normal organic fertilizer,and the economic benefit increases by 1800 yuan / ha. As a kind of banana waste cycling product,banana stalk organic fertilizer is of low cost and good effect,and can be used instead of normal organic fertilizer. 10. Lipophilic extracts from banana fruit residues: a source of valuable phytosterols. Science.gov (United States) Oliveira, Lúcia; Freire, Carmen S R; Silvestre, Armando J D; Cordeiro, Nereida 2008-10-22 The chemical composition of the lipophilic extracts of unripe pulp and peel of banana fruit 'Dwarf Cavendish' was studied by gas chromatography-mass spectrometry. Fatty acids, sterols, and steryl esters are the major families of lipophilic components present in banana tissues, followed by diacylglycerols, steryl glucosides, long chain fatty alcohols, and aromatic compounds. Fatty acids are more abundant in the banana pulp (29-90% of the total amount of lipophilic extract), with linoleic, linolenic, and oleic acids as the major compounds of this family. In banana peel, sterols represent about 49-71% of the lipophilic extract with two triterpenic ketones (31-norcyclolaudenone and cycloeucalenone) as the major components. The detection of high amounts of steryl esters (469-24405 mg/kg) and diacylglycerols (119-878 mg/kg), mainly present in the banana peel extract, explains the increase in the abundance of fatty acids and sterols after alkaline hydrolysis. Several steryl glucosides were also found in significative amounts (273-888 mg/kg), particularly in banana pulp (888 mg/kg). The high content of sterols (and their derivatives) in the 'Dwarf Cavendish' fruit can open new strategies for the valorization of the banana residues as a potential source of high-value phytochemicals with nutraceutical and functional food additive applications. 11. Prediction of textural attributes using color values of banana (Musa sapientum) during ripening. Science.gov (United States) Jaiswal, Pranita; Jha, Shyam Narayan; Kaur, Poonam Preet; Bhardwaj, Rishi; Singh, Ashish Kumar; Wadhawan, Vishakha 2014-06-01 Banana is an important sub-tropical fruit in international trade. It undergoes significant textural and color transformations during ripening process, which in turn influence the eating quality of the fruit. In present study, color ('L', 'a' and 'b' value) and textural attributes of bananas (peel, fruit and pulp firmness; pulp toughness; stickiness) were studied simultaneously using Hunter Color Lab and Texture Analyser, respectively, during ripening period of 10 days at ambient atmosphere. There was significant effect of ripening period on all the considered textural characteristics and color properties of bananas except color value 'b'. In general, textural descriptors (peel, fruit and pulp firmness; and pulp toughness) decreased during ripening except stickiness, while color values viz 'a' and 'b' increased with ripening barring 'L' value. Among various textural attributes, peel toughness and pulp firmness showed highest correlation (r) with 'a' value of banana peel. In order to predict textural properties using color values of banana, five types of equations (linear/polynomial/exponential/logarithmic/power) were fitted. Among them, polynomial equation was found to be the best fit (highest coefficient of determination, R(2)) for prediction of texture using color properties for bananas. The pulp firmness, peel toughness and pulp toughness showed R(2) above 0.84 with indicating its potentiality of the fitted equations for prediction of textural profile of bananas non-destructively using 'a' value. 12. A study on effect of ATH on Euphorbia coagulum modified polyester banana fiber composite Science.gov (United States) Kumari, Sanju; Rai, Bhuvneshwar; Kumar, Gulshan 2018-02-01 Fiber reinforced polymer composites are used for building and structural applications due to their high strength. In conventional composites both the binder and the reinforcing fibers are synthetic or either one of the material is natural. In the present study coagulum of Euphorbia royleana has been used for replacing polyester resinas binder in polyester banana composite. Euphorbia coagulum (driedlatex) is rich in resinous mass (60-80%), which are terpenes and polyisoprene (10-20%). Effect of varying percentage of coagulum content on various physico-mechanical properties of polyester-banana composites has been studied. Since banana fiber is sensitive to water due to presence of polar group, banana composite undergoes delamination and deterioration under humid condition. Alkali treated banana fiber along with coagulum content has improved overall mechanical properties and reduction in water absorption. The best physico-mechanical properties have been achieved on replacing 40% of polyester resin by coagulum. An increase of 50% in bending strength, 30% bending modulus and 45% impact strength as well as 68% decrease in water absorption was observed. Incorporation of 20% ATH as flame retardant in coagulum modified banana polyester composite enhanced limiting oxygen index from 20.6 to 26.8% and smoke density reduced up to 40%. This study presents the possibility of utilization of renewable materials for environmental friendly composite development as well as to find out alternative feedstock for petroleum products. Developed Euphorbia latex modified banana polyester composites can have potential utility in hardboard, partition panel, plywood and automotive etc. 13. Inventory of Musa paradisiaca L. (banana kepok in Lumajang regency, Malang regency, and Magelang regency Directory of Open Access Journals (Sweden) Suhadi Suhadi 2010-12-01 Full Text Available Banana is fruit containing fairly high nutrition and provides quick reserve enegy. The crop grows in tropical area with average rainfall all the year and banana produces at any season. One of the bananas which has high value sale and high competable potency is subvariety of kepok banana. Kepok banana has various subvarieties, these subvarieties have the same morphologies but have different texture appearances thus uneasy to differenciate among them. The texture appearance determines the quality and price of the banana. Often the buyer makes a mistake in choosing subvariety of kepok he wants to, whereas the seller gives him the cheapiest subvariety of kepok. Methods we used was method of exploration using free exploration technique step by step without any certain path. There were two phases in the research namely the fi rst phase was carried out in field and the second phase was done in the laboratory. Subvarieties of kepok found in Lumajang Regency are 4 subcultivars, Malang Regency there are 3 and Magelang Regency are subcultivars subcultivars, The sequence of the qulity of kapok subcultivars are as follows, red kepok, yellow kepok, big (gede, gilo, gembrot kepok, and white kepok. Sugestion, organic ferlitilizer should be used in the fertilization of banana cultivation, and conservation of red kepok is highly required. 14. Treasury Offset Program (TOP) MI Data.gov (United States) Social Security Administration — The TOP MI helps OPSOS coordinate TOP case processing in the regions. The MI also helped communicate our progress and findings to BFQM and ORDP, as well as the ACOSS. 15. Top Quark Physics with CMS CERN Multimedia CERN. Geneva 2011-01-01 Higgs mechanism. There are various hints at deviations from the Standard Model expectation which have been observed recently by Tevatron experiments in top final states. Several signatures of new physics accessible at the LHC either suffer from top-quark production as a significant background or contain top quarks themselves. In this talk, we present results on top quark physics obtained from the first LHC data collected by the CMS experiment.They include measurements of the top pair production cross section in various channels and their combination, measurements of the top quark mass, the single top cross section, a search for new particles decaying into top pairs, and a first look at the charge asymmetry. 16. Black leaf streak disease affects starch metabolism in banana fruit. Science.gov (United States) Saraiva, Lorenzo de Amorim; Castelan, Florence Polegato; Shitakubo, Renata; Hassimotto, Neuza Mariko Aymoto; Purgatto, Eduardo; Chillet, Marc; Cordenunsi, Beatriz Rosana 2013-06-12 Black leaf streak disease (BLSD), also known as black sigatoka, represents the main foliar disease in Brazilian banana plantations. In addition to photosynthetic leaf area losses and yield losses, this disease causes an alteration in the pre- and postharvest behavior of the fruit. The aim of this work was to investigate the starch metabolism of fruits during fruit ripening from plants infected with BLSD by evaluating carbohydrate content (i.e., starch, soluble sugars, oligosaccharides, amylose), phenolic compound content, phytohormones, enzymatic activities (i.e., starch phosphorylases, α- and β-amylase), and starch granules. The results indicated that the starch metabolism in banana fruit ripening is affected by BLSD infection. Fruit from infested plots contained unusual amounts of soluble sugars in the green stage and smaller starch granules and showed a different pattern of superficial degradation. Enzymatic activities linked to starch degradation were also altered by the disease. Moreover, the levels of indole-acetic acid and phenolic compounds indicated an advanced fruit physiological age for fruits from infested plots. 17. Resistance selection on banana CV. Ambon Kuning Against Fusarium Wilt International Nuclear Information System (INIS) Sutarto, Ismiyarti; Meldia, Yeni; Jumjunidang 1998-01-01 This research was conducted in order to study the occurrence of mutation on irradiated plantlets and their resistance of plants of banana cv. Ambon Kuning against Fusarium wilt. Plantlets of banana cv. Ambon Kuning sized 5 cm were exposed to gamma rays at the doses 5 - 35 Gy intervals, then were subcultured for obtained M 1 V 5 plantlets. More over, the planlets were acclimatized and were planted in the field was already infected by Fasarium (f).culbense (FOC). The result indicated that irradiated plantlets of the doses 20 - 35 Gy were not able to survive up to 6 months after exposing to gamma rays. Abnormalities of M 1 V 5 plantlets originated from irradiated plantlets at the doses 10 and 15 Gy were shown on rossette plantlets with rigid and dark green leaves, and the formation of smooth mass morphologically shaped like calculus. The appearance of plant height and number of suckers of suckers of M 1 V 5 plants in the field was quite various. The number of survival plants after 8 moths planting was 8, 7, 15, and 28, respectively originated from untreated plants and irradiated plantlets at the doses 5, 10, and 15 Gy. After one year planting , only 2 plants were able to survive from irradiated plantlet at the dose 15 Gy. The plants could produce 27 plantlets obtained from culturing their shoot tips. Further study of these plantlets was needed in order create the stability of their resistance to FOC. (author) 18. Post harvest changes gamma-irradiated banana Prata International Nuclear Information System (INIS) Vilas Boas, E.V. de; Chitarra, A.B.; Chitarra, M.I.F. 1996-01-01 The effect of the gamma-irradiation was evaluated at 0.25 and 0.50 kGy, on the development of peel coloration, CO 2 and ethylene evolution, conversion of starch to sugars, pulp-to-peel ratio, pectic solubilization and activities of enzymes of the cell wall, pectin methylesterase (PME), and polygalacturonase (PG), during maturation of 'Prata' bananas. The gamma-irradiation did not affect the normal colour development of the fruits. An increase in the ethylene peak and a decrease in the CO 2 peak was observed. The gamma-irradiation did not affect the degradation of starch, while a delay in soluble sugar accumulation was noted on the 6 and 7 colour grades. The fruits subjected to 0.25 kGy had the highest increase in the pulp-to-peel relation, beginning with colour grade 5, due to a possible stress effect of that dose. An increase of pectin solubilization was observed. Higher PME activities were exhibited by irradiated fruits, although the gamma-irradiation suppressed the PG activity throughout the maturation period. The gamma-irradiation did not extend the post-harvest life of 'Prata' bananas. (author) [pt 19. Top production at hadron colliders Indian Academy of Sciences (India) New results on top quark production are presented from four hadron collider experiments: CDF and D0 at the Tevatron, and ATLAS and CMS at the LHC. Cross-sections for single top and top pair production are discussed, as well as results on the top–antitop production asymmetry and searches for new physics including ... 20. Single top t-channel CERN Document Server Faltermann, Nils 2017-01-01 The production of single top quarks allows to study the interplay of top quark physics and the electroweak sector of the standard model. Deviations from predictions can be a hint for physics beyond the standard model. The t-channel is the dominant production mode for single top quarks at the LHC. This talk presents the latest measurements from the ATLAS and CMS collaborations. 1. Forward Top Physics at LHCb CERN Multimedia CERN. Geneva 2018-01-01 The first Run 2 measurement of top pair production in the dilepton channel at 13 TeV will be presented, along with previous Run 1 measurements in final states accessible to both single top and top pair production. Heavy flavour tagging strategies at LHCb will also be discussed. 2. Chemical composition and physicochemical properties of green banana (Musa acuminata x balbisiana Colla cv. Awak) flour. Science.gov (United States) Haslinda, W H; Cheng, L H; Chong, L C; Noor Aziah, A A 2009-01-01 Flour was prepared from peeled and unpeeled banana Awak ABB. Samples prepared were subjected to analysis for determination of chemical composition, mineral, dietary fibre, starch and total phenolics content, antioxidant activity and pasting properties. In general, flour prepared from unpeeled banana was found to show enhanced nutrition values with higher contents of mineral, dietary fibre and total phenolics. Hence, flour fortified with peel showed relatively higher antioxidant activity. On the other hand, better pasting properties were shown when banana flour was blended with peel. It was found that a relatively lower pasting temperature, peak viscosity, breakdown, final viscosity and setback were evident in a sample blended with peel. 3. K40 y Cs137 in bananas exported from Costa Rica International Nuclear Information System (INIS) Loria, L. G.; Mora, P.; Badilla, M. 1999-01-01 Using low level gamma spectroscopy, the specific activity of K 4 0 and Cs 1 37 in banana samples is quantified during the period 1996-1998. The bananas were supplied by the export companies that operate in Costa Rica. The calculated derived intervention level (DIL) for Cs 1 37 was 4000 times greater than the specific activity measured in the fruit due to worldwide nuclear events. This result permits its free commercialization. Banana is an excellent source of potassium, since it was determined that each kg of the fruit has 3.8 g of this element. (Author) [es 4. The top ten. Science.gov (United States) Davis, C 1985-10-01 3 lists were compiled to answer the question of what countries have experienced the most population growth from World War II to the present. The 1st list includes the 10 countries which accounted for the largest gain in world population. The 2nd list shows the 10 countries with the highest growth rates of the 150 or so most populous countries in the world. The last list also shows countries ranked by growth rate but is limited to those with a current population of at least 10 million. To deal with the fact that some countries did not exist at 1 of the reference points, a set of estimates for 1940 provided a convenient starting point. China and India headed the list of countries ranked by absolute gains. They contributed half again as much growth as the next 8 nations combined. Altogether Asia, home to almost 60% of the world's population, captured 5 of the top 10 slots. The US and the USSR are the only developed countries on the list. The Soviet Union also presents the 1st case on the list of a country with a lower ranking in population gain (6th) than in absolute size in 1985 (3rd). Many of the population giants of 1940 (Japan, Great Britain, Germany, France, and Italy) are not on the list. Despite large base populations, their growth has been slow in the intervening years. All of the countries with the highest growth rates experienced over 3% annual average growth during the last 45 years: Kuwait, United Arab Emirates, Libya, Venezuela, Costa Rica, Kenya, Ivory Coast, Mexico, Honduras, and Syria. The influence of immigration is much in evidence in this list and explains otherwise unachievable rates of growth. The top 4 countries in annual rates of growth are all oil exporters. The top 3 countries in the 3rd list -- Venezuela, Kenya, and Mexico -- are carried over from the 2nd list, sharing the characteristics of rapid population growth and substantial base populations. The geographic mix of this list is notable. Asia, Africa, and Latin America each have at least 5. Marketing de banana: preferências do consumidor quanto aos atributos de qualidade dos frutos Marketing of banana: consumer preferences relating to fruit quality attributes Directory of Open Access Journals (Sweden) Fernando César Akira Urbano Matsuura 2004-04-01 Full Text Available O Brasil produz aproximadamente seis milhões de toneladas por ano de banana (Musa spp., com consumo médio da ordem de 35 kg/ habitante / ano. A aceitação da banana deve-se, principalmente, a seus aspectos sensoriais, valor nutricional e conveniência. A identificação das necessidades e desejos dos clientes consiste em uma atividade crítica do marketing. O objetivo deste trabalho foi o de pesquisar as preferências do consumidor de um mercado local (município de Cruz das Almas - Estado da Bahia considerando os atributos de qualidade dos frutos frescos de banana madura. A metodologia utilizada foi a da pesquisa descritiva por método estatístico. Os dados foram coletados por questionário, na forma de entrevista pessoal com 400 pessoas. Os atributos de qualidade (variáveis questionados e avaliados foram relacionados com a aparência, cor, textura, aroma, sabor e vida útil esperada dos frutos de banana. De acordo com a preferência dos consumidores entrevistados, o fruto de banana maduro ideal deve apresentar características como: penca contendo 10 a 12 dedos (frutos, dedos de tamanho médio ou grande, diâmetro médio, quina presente, ausência de pintas pretas na casca, cor da polpa amarelo-clara ou média, textura firme, aroma e sabor de intensidade média, mediamente doce e vida útil de 7 a 10 dias em condição ambiente. O sabor, vida útil e aparência dos frutos de banana são considerados os mais importantes atributos na escolha ou compra da banana, segundo os consumidores entrevistados.Brazil has an approximate production of six million annual tons of banana (Musa spp., with a consumption close to 35 kg / inhabitant / year. The acceptance of the banana fruit is due, mainly, to its sensorial aspects, nutritional value and convenience. The identification of the customers' needs and desires consists of a critical activity of the marketing. The objective of this work was to research the consumer preferences of a local market (Cruz 6. Top quark physics International Nuclear Information System (INIS) Menzione, A. 1995-10-01 Most of the material presented in this report, comes from contributions to the parallel session PL20 of this conference. We summarise the experimental results of direct production of Top quarks, coming from the CDF and C0 Collaborations at Fermilab, and compare these results to what one expects within current theoretical understanding. Particular attention is given to new results such as all hadronic modes of t bar t decay. As far as the mass is concerned, a comparison is made with precision measurements of related quantities, coming from LEP and other experiments. An attempt is made to look at the medium-term future and understand which variables and with what accuracy one can measure them with increased integrated luminosity 7. Top quark pair production and top quark properties at CDF Energy Technology Data Exchange (ETDEWEB) Moon, Chang-Seong [INFN, Pisa 2016-06-02 We present the most recent measurements of top quark pairs production and top quark properties in proton-antiproton collisions with center-of-mass energy of 1.96 TeV using CDF II detector at the Tevatron. The combination of top pair production cross section measurements and the direct measurement of top quark width are reported. The test of Standard Model predictions for top quark decaying intob$-quarks, performed by measuring the ratio$R$between the top quark branching fraction to$b$-quark and the branching fraction to any type of down quark is shown. The extraction of the CKM matrix element$|V_{tb}|$from the ratio$R$is discussed. We also present the latest measurements on the forward-backward asymmetry ($A_{FB}$) in top anti-top quark production. With the full CDF Run II data set, the measurements are performed in top anti-top decaying to final states that contain one or two charged leptons (electrons or muons). In addition, we combine the results of the leptonic forward-backward asymmetry in$t\\bar t$system between the two final states. All the results show deviations from the next-to-leading order (NLO) standard model (SM) calculation. 8. ''In vivo'' methodology for mutation induction in banana, cultivar ''Maca'' International Nuclear Information System (INIS) Tulmann Neto, A.; Domingues, E.T.; Alvarez, A.L.F.; Mendez, B.M.J.; Ando, A. 1990-01-01 Full text: The ''Maca'' cultivar is a banana of high acceptability in the south west of Brazil. However, it is very susceptible to several diseases. Due to the difficulties in the application of the traditional plant breeding methods, the Radiation Genetics Section of CENA is utilising the ''in vivo'', and the ''in vitro'' mutation breeding approach. The ''in vivo'' methodology is based on the work of HAMILTON. This method is being utilised in Brazil for rapid banana propagation. Rhizomes (20 cm diameter) were obtained from young field grown plants before flower differentiation. In these rhizomes, only 5-6 leaf sheaths were retained, the others being removed. The rhizomes were maintained in a greenhouse in boxes with vermiculite, covered with plastic. After one week, all leaf sheaths were removed, until the exposure of the meristematic apex with about 2 mm size. This apex was cut off with a scalpel and a cross shaped cut (2,5 cm) was made. This stimulates the development of lateral buds. After four months, the meristematic apices of these new buds were cut off in the same way and immediately the rhizomes were irradiated with gamma rays. Around the eliminated lateral buds callus developed and new lateral buds were formed. The LD 50 in relation to the number of these new buds produced was around 30 Gy. According to the author of the original method, from the callus one can obtain axillary or adventitious buds. In the early stages it is possible, based on the shape, to distinguish both types. The advantage of utilising adventitious buds originating from only one cell to avoid chimerism is well known in mutation breeding. However, it is not certain whether this is the case in the present method. After detachment from rhizomes and rooting in soil, plants with 15-20 cm height were inoculated with Fusarium oxysporum f.sp. cubense. After 3 weeks the plants showed symptoms of the Panama disease and screening could be done at this stage. The total time between the removal of 9. ''In vivo'' methodology for mutation induction in banana, cultivar ''Maca'' Energy Technology Data Exchange (ETDEWEB) Tulmann Neto, A; Domingues, E T; Alvarez, A L.F.; Mendez, B M.J.; Ando, A [Centre for Nuclear Energy in Agriculture (CENA), University of Sao Paulo, Piracicaba, SP. (Brazil) 1990-07-01 Full text: The ''Maca'' cultivar is a banana of high acceptability in the south west of Brazil. However, it is very susceptible to several diseases. Due to the difficulties in the application of the traditional plant breeding methods, the Radiation Genetics Section of CENA is utilising the ''in vivo'', and the ''in vitro'' mutation breeding approach. The ''in vivo'' methodology is based on the work of HAMILTON. This method is being utilised in Brazil for rapid banana propagation. Rhizomes (20 cm diameter) were obtained from young field grown plants before flower differentiation. In these rhizomes, only 5-6 leaf sheaths were retained, the others being removed. The rhizomes were maintained in a greenhouse in boxes with vermiculite, covered with plastic. After one week, all leaf sheaths were removed, until the exposure of the meristematic apex with about 2 mm size. This apex was cut off with a scalpel and a cross shaped cut (2,5 cm) was made. This stimulates the development of lateral buds. After four months, the meristematic apices of these new buds were cut off in the same way and immediately the rhizomes were irradiated with gamma rays. Around the eliminated lateral buds callus developed and new lateral buds were formed. The LD{sub 50} in relation to the number of these new buds produced was around 30 Gy. According to the author of the original method, from the callus one can obtain axillary or adventitious buds. In the early stages it is possible, based on the shape, to distinguish both types. The advantage of utilising adventitious buds originating from only one cell to avoid chimerism is well known in mutation breeding. However, it is not certain whether this is the case in the present method. After detachment from rhizomes and rooting in soil, plants with 15-20 cm height were inoculated with Fusarium oxysporum f.sp. cubense. After 3 weeks the plants showed symptoms of the Panama disease and screening could be done at this stage. The total time between the 10. Toitlustusettevõtete TOP 30 aastal 2002 Index Scriptorium Estoniae 2003-01-01 Toitlustusettevõtete TOP 30 aastal 2002. Käibe TOP 30. Kasumi TOP 30. Käibe kasvu TOP 30. Kasumi kasvu TOP 30. Rentaabluse TOP 30. Omakapitali tootlikkuse TOP 30. Toitlustusettevõtete üldandmed. Toitlustusettevõtete finantsandmed 11. Hiiumaa ettevõtete TOP 50 Index Scriptorium Estoniae 2005-01-01 Hiiumaa ettevõtete TOP 50; Käibe TOP 35; Kasumi TOP 35; Hiiumaa ettevõtete üld- ja finantsandmed; Käibe kasvu TOP 20; Kasumi kasvu TOP 20; Rentaabluse TOP 20; Omakapitali tootlikkuse TOP 20. Vt samas: Teeli Remmalg: Hiiumaal jätkub plastitööstuse võidumarss 12. The search for the top quark International Nuclear Information System (INIS) Barbaro-Galtieri, A. 1989-03-01 This paper discusses the following topics: top search in the near future, general remarks, top search at HERA, searching for the top quarks at the Z 0 machines, finding the top at Lep II, top search in UA2, top search in UA1, and top search at CDF. 58 refs., 38 figs 13. Toiduainetööstuste TOP 100 Index Scriptorium Estoniae 2006-01-01 Toiduainetööstuse TOP. Vt. samas: Käibe TOP 10; Käibekasvu TOP 10; Kasumi TOP 10; Kasumi kasvu TOP 10; Rentaabluse TOP 10; Omakapitali tootlikkuse TOP 10; Toiduainetööstuse üld- ja finantsandmed 14. Online-kaubanduse TOP aastal 2003 Index Scriptorium Estoniae 2004-01-01 Online-kaubanduse TOP aastal 2003. Käibe TOP 5. Käibe kasvu TOP 5. Rentaabluse TOP 5. Kasumi TOP 5. Kasumi kasvu TOP 5. Omakapitali tootlikkuse TOP 5. Lisa: TOPi koostamise metoodika. Online-kaubanduse firmade üldandmed. Online kaubanduse firmade finantsandmed 15. TopView - ATLAS top physics analysis package CERN Document Server Shibata, A 2007-01-01 TopView is a common analysis package which is widely used in the ATLAS top physics working group. The package is fully based on the official ATLAS software Athena and EventView and playing a central role in the collaborative analysis model. It is a functional package which accounts for a broad range issues in implementing physics analysis. As well as being a modular framework suitable as a common workplace for collaborators, TopView implements numerous analysis tools including a complete top-antitop reconstruction and single top reconstruction. The package is currently used to produce common ntuple from Monte Carlo production and future use cases are under rapid development. In this paper, the design and ideas behind TopView and the performance of the analyses implemented in the package are presented with detailed documentation of the contents and instruction for using the package. 16. Transgenic Cavendish bananas with resistance to Fusarium wilt tropical race 4. Science.gov (United States) Dale, James; James, Anthony; Paul, Jean-Yves; Khanna, Harjeet; Smith, Mark; Peraza-Echeverria, Santy; Garcia-Bastidas, Fernando; Kema, Gert; Waterhouse, Peter; Mengersen, Kerrie; Harding, Robert 2017-11-14 Banana (Musa spp.) is a staple food for more than 400 million people. Over 40% of world production and virtually all the export trade is based on Cavendish banana. However, Cavendish banana is under threat from a virulent fungus, Fusarium oxysporum f. sp. cubense tropical race 4 (TR4) for which no acceptable resistant replacement has been identified. Here we report the identification of transgenic Cavendish with resistance to TR4. In our 3-year field trial, two lines of transgenic Cavendish, one transformed with RGA2, a gene isolated from a TR4-resistant diploid banana, and the other with a nematode-derived gene, Ced9, remain disease free. Transgene expression in the RGA2 lines is strongly correlated with resistance. Endogenous RGA2 homologs are also present in Cavendish but are expressed tenfold lower than that in our most resistant transgenic line. The expression of these homologs can potentially be elevated through gene editing, to provide non-transgenic resistance. 17. Detection of DNA methylation changes in micropropagated banana plants using methylation-sensitive amplification polymorphism (MSAP). Science.gov (United States) Peraza-Echeverria, S; Herrera-Valencia, V A.; Kay, A -J. 2001-07-01 The extent of DNA methylation polymorphisms was evaluated in micropropagated banana (Musa AAA cv. 'Grand Naine') derived from either the vegetative apex of the sucker or the floral apex of the male inflorescence using the methylation-sensitive amplification polymorphism (MSAP) technique. In all, 465 fragments, each representing a recognition site cleaved by either or both of the isoschizomers were amplified using eight combinations of primers. A total of 107 sites (23%) were found to be methylated at cytosine in the genome of micropropagated banana plants. In plants micropropagated from the male inflorescence explant 14 (3%) DNA methylation events were polymorphic, while plants micropropagated from the sucker explant produced 8 (1.7%) polymorphisms. No DNA methylation polymorphisms were detected in conventionally propagated banana plants. These results demonstrated the usefulness of MSAP to detect DNA methylation events in micropropagated banana plants and indicate that DNA methylation polymorphisms are associated with micropropagation. 18. In vitro mutation induction for resistance to Fusarium wilt in the banana Energy Technology Data Exchange (ETDEWEB) Tulmann Neto, A; Mendes, B M.J.; Latado, R [Centro de Energia Nuclear na Agricultura, Piracicaba, SP (Brazil); Cesar Santos, P dos; Boliani, A [Universidade Estadual Paulista, Ilha Solteira, SP (Brazil). Faculdade de Agronomia 1995-11-01 In Brazil, which is one of the world`s principal banana production regions, almost all production is consumed within the country. Consumers show high preference for the cultivar Maca (AAB group). However, it is becoming increasingly difficult to produce bananas of this type because of their high susceptibility to Fusarium wilt, caused by Fusarium oxysporum f. sp. cubense. Sexual breeding, which consists of recombination and selection, is limited in the banana because of polyploidy and sterility. Spontaneous somatic mutations are an important source of new cultirvars, and mutation breeding might be particularly important to generate genetic variation. Because of this, the mutation breeding approach has been used in Brazil. The objective of this research was to induce gamma ray mutations for resistance or to increase the level of tolerance to Fusarium wilt in the banana cultivar Maca on the basis of screening under field conditions. 4 refs. 19. Discrete Dynamical Systems Meet the Classic Monkey-and-the-Bananas Problem. Science.gov (United States) Gannon, Gerald E.; Martelli, Mario U. 2001-01-01 Presents a solution of the three-sailors-and-the-bananas problem and attempts a generalization. Introduces an interesting way of looking at the mathematics with an idea drawn from discrete dynamical systems. (KHR) 20. Focused ion beam analysis of banana peel and its application for arsenate ion removal International Nuclear Information System (INIS) Memon, J.R. 2015-01-01 Banana peel, a common fruit waste, has been investigated for its ability to remove arsenate ions from ground water as a function of pH, contact time, and initial metal ion concentration. Focused ion beam (FIB) analysis revealed the internal morphology of the banana peels. Arsenate ions were entered into micropores of banana peel. pH was seen to have no effect on the sorption process. Retained species were eluted using 5 mL of 2 M H/sub 2/SO/sub 4/. The kinetics of sorption were observed to follow the pseudo first order rate equation. The sorption data followed Freundlich and D-R isotherms. The energy value obtained from the D-R isotherms indicated that the sorption was physical in nature for arsenate species. Our study has shown that banana peel has the ability to remove arsenate species from ground water samples. (author) 1. Mechanical and water absorption behaviour of banana/sisal reinforced hybrid composites International Nuclear Information System (INIS) Venkateshwaran, N.; ElayaPerumal, A.; Alavudeen, A.; Thiruchitrambalam, M. 2011-01-01 Highlights: → It explores the utilization of waste banana fiber. → Improving the mechanical property by hybridization. → Results show its usefulness to low cost application. -- Abstract: The tensile, flexural, impact and water absorption tests were carried out using banana/epoxy composite material. Initially, optimum fiber length and weight percentage were determined. To improve the mechanical properties, banana fiber was hybridised with sisal fiber. This study showed that addition of sisal fiber in banana/epoxy composites of up to 50% by weight results in increasing the mechanical properties and decreasing the moisture absorption property. Morphological analysis was carried out to observe fracture behaviour and fiber pull-out of the samples using scanning electron microscope. 2. Bio solids Application on Banana Production: Soil Chemical Properties and Plant Nutrition International Nuclear Information System (INIS) Teixeira, L.A.J; Berton, R.S.B; Coscione, A.R; Saes, L.A 2011-01-01 Bio solids are relatively rich in N, P, and S and could be used to substitute mineral fertilization for banana crop. A field experiment was carried out in a Yellow Oxisol to investigate the effects of bio solids application on soil chemical properties and on banana leaf's nutrient concentration during the first cropping cycle. Soil analysis (ph, organic matter, resin P, exchangeable Ca and K, available B, DTPA-extracted micro nutrients, and heavy metals) and index-leaf analysis (B, Cu, Fe, Mn, Zn, Cd, Cr, Ni, and Pb) were evaluated. Bio solids can completely substitute mineral N and P fertilizer to banana growth. Soil exchangeable K and leaf-K concentration must be monitored in order to avoid K deficiency in banana plants. No risk of heavy metal (Cr, Ni, Pb, and Cd) concentration increase in the index leaf was observed when bio solids were applied at the recommended N rate. 3. Molecular Diagnostics in the Mycosphaerella Leaf Spot Disease Complex of Banana and for Radopholus similis NARCIS (Netherlands) Arzanlou, M.; Kema, G.H.J.; Waalwijk, C.; Carlier, I.; Vries, de P.M.; Guzmán, M.; Araya Vargas, M.; Helder, J.; Crous, P.W. 2009-01-01 Mycosphaerella leaf spots and nematodes threaten banana cultivation worldwide. The Mycosphaerella disease complex involves three related ascomycetous fungi: Mycosphaerella fijiensis, M. musicola and M. eumusae. The exact distribution of these three species and their disease epidemiology remain 4. Biochemical and molecular tools reveal two diverse Xanthomonas groups in bananas DEFF Research Database (Denmark) Adriko, John; Aritua, V.; Mortensen, Carmen Nieves 2016-01-01 Xanthomonas campestris pv. musacearum (Xcm) causing the banana Xanthomonas wilt (BXW) disease has been the main xanthomonad associated with bananas in East and Central Africa based on phenotypic and biochemical characteristics. However, biochemical methods cannot effectively distinguish between...... pathogenic and non-pathogenic xanthomonads. In this study, gram-negative and yellow-pigmented mucoid bacteria were isolated from BXW symptomatic and symptomless bananas collected from different parts of Uganda. Biolog, Xcm-specific (GspDm), Xanthomonas vasicola species-specific (NZ085) and Xanthomonas genus......-specific (X1623) primers in PCR, and sequencing of ITS region were used to identify and characterize the isolates. Biolog tests revealed several isolates as xanthomonads. The GspDm and NZ085 primers accurately identified three isolates from diseased bananas as Xcm and these were pathogenic when re... 5. Focused Ion Beam Analysis of Banana Peel and Its Application for Arsenate Ion Removal Directory of Open Access Journals (Sweden) Jamil R. Memon 2015-06-01 Full Text Available Banana peel, a common fruit waste, has been investigated for its ability to remove arsenate ions from ground water as a function of pH, contact time, and initial metal ion concentration. Focused ion beam (FIB analysis revealed the internal morphology of the banana peels. Arsenate ions were entered into micropores of banana peel. pH was seen to have no effect on the sorption process. Retained species were eluted using 5 mL of 2 M H2SO4. The kinetics of sorption were observed to follow the pseudo first order rate equation. The sorption data followed Freundlich and D-R isotherms. The energy value obtained from the D-R isotherms indicated that the sorption was physical in nature for arsenate species. Our study has shown that banana peel has the ability to remove arsenate species from ground water samples. 6. [Banana peel: a possible source of infection in the treatment of nipple fissures]. Science.gov (United States) Novak, Franz Reis; de Almeida, João Aprígio Guerra; de Souza e Silva, Rosana 2003-01-01 To study the microbiology of banana peel being sold in the city of Rio de Janeiro, in an attempt to determine the possibility that the peel may represent a source of infection for women who use it to treat nipple fissures. The following microorganisms were studied in 20 banana peel samples: mesophiles, total coliforms, fecal coliforms, Pseudomonas aeruginosa, lipolytic and proteolytic microorganisms, molds and yeasts, lactic bacteria, and coagulase-positive staphylococcus. The microbiological analyses revealed the occurrence of several typical groups of microorganisms, with the following distribution of positive results being detected in banana peel samples: mesophiles, 100%; total coliforms, 20%; coagulase-positive staphylococcus, 25%; molds and yeasts, 30%; proteolytic microorganisms, 70%; lipolytic microorganisms, 30%, and lactic bacteria, 95%. Fecal coliforms and Pseudomonas aeruginosa were not isolated. The results show the presence of potentially pathogenic microorganisms in levels which could compromise the microbiological quality of the banana peel. Its use for the treatment of nipple fissures can initiate an infectious process. 7. Physicochemical quality and antioxidant changes in ‘Leb Mue Nang’ banana fruit during ripening Directory of Open Access Journals (Sweden) Pannipa Youryon 2017-02-01 Full Text Available The physicochemical and antioxidant changes of ‘Kluai Leb Mue Nang’ banana fruit (Musa AA group were investigated during ripening. The visual appearance, peel and pulp color, firmness, total soluble solids concentration (TSS, total acidity (TA and bioactive compounds of the fruit at three stages of ripening (mature green, ripe and overripe were monitored. Changes in both the peel and pulp color, texture, TSS and TA contents during banana ripening were similar to those of other banana fruits. Interestingly, the highest total antioxidants capacity and total phenols concentration were found in the ripe banana fruit. 2,2-Diphenyl-1-picrylhydrazyl radical scavenging activity remained constant and the highest total flavonoids concentration was found in the mature green fruit. 8. Studies on physico-chemical changes during artificial ripening of banana (Musa sp) variety 'Robusta'. Science.gov (United States) Kulkarni, Shyamrao Gururao; Kudachikar, V B; Keshava Prakash, M N 2011-12-01 Banana (Musa sp var 'Robusta') fruits harvested at 75-80% maturity were dip treated with different concentrations of ethrel (250-1,000 ppm) solution for 5 min. Ethrel at 500 ppm induced uniform ripening without impairing taste and flavour of banana. Untreated control banana fruits remained shriveled, green and failed to ripen evenly even after 8 days of storage. Fruits treated with 500 ppm of ethrel ripened well in 6 days at 20 ± 1 °C. Changes in total soluble solids, acidity, total sugars and total carotenoids showed increasing trends up to 6 days during ripening whereas fruit shear force values, pulp pH and total chlorophyll in peel showed decreasing trends. Sensory quality of ethrel treated banana fruits (fully ripe) were excellent with respect to external colour, taste, flavour and overall quality. 9. Bacillus spp as a biological control agent against panama disease in banana CSIR Research Space (South Africa) Gumede, WHN 2006-02-01 Full Text Available The decreased productivity levels in crop production as a consequence of disease infection have been a great concern amongst agricultural communities. A similar threat is facing the banana-cultivating community due to Panama disease. Panama disease... 10. Treatment of banana and potato plants with a new antifungal composition (European patent specification) NARCIS (Netherlands) Stark, J.; Rijn, van F.T.J.; Krieken, van der W.M.; Stevens, L.H. 2010-01-01 International publication number: WO 2009/077613 (25.06.2009 Gazette 2009/26) The present invention relates to the treatment of banana and potato plants with a composition containing natamycin and at least one phosphite containing compound 11. Natural Variation in Banana Varieties Highlights the Role of Melatonin in Postharvest Ripening and Quality. Science.gov (United States) Hu, Wei; Yang, Hai; Tie, Weiwei; Yan, Yan; Ding, Zehong; Liu, Yang; Wu, Chunlai; Wang, Jiashui; Reiter, Russel J; Tan, Dun-Xian; Shi, Haitao; Xu, Biyu; Jin, Zhiqiang 2017-11-22 This study aimed to investigate the role of melatonin in postharvest ripening and quality in various banana varieties with contrasting ripening periods. During the postharvest life, endogenous melatonin showed similar performance with ethylene in connection to ripening. In comparison to ethylene, melatonin was more correlated with postharvest banana ripening. Exogenous application of melatonin resulted in a delay of postharvest banana ripening. Moreover, this effect is concentration-dependent, with 200 and 500 μM treatments more effective than the 50 μM treatment. Exogenous melatonin also led to elevated endogenous melatonin content, reduced ethylene production through regulation of the expression of MaACO1 and MaACS1, and delayed sharp changes of quality indices. Taken together, this study highlights that melatonin is an indicator for banana fruit ripening in various varieties, and the repression of ethylene biosynthesis and postharvest ripening by melatonin can be used for biological control of postharvest fruit ripening and quality. 12. Ion thermal conductivity and ion distribution function in the banana regime International Nuclear Information System (INIS) Taguchi, Masayoshi 1988-01-01 A method for calculating the ion thermal conductivity and the ion distribution function in the banana regime is formulated for an axisymmetric toroidal plasma of arbitrary aspect ratio. A simple expression for this conductivity is also derived. (author) 13. Extraction and characterization of proteins from banana (Musa Sapientum L) flower and evaluation of antimicrobial activities. Science.gov (United States) Sitthiya, Kewalee; Devkota, Lavaraj; Sadiq, Muhammad Bilal; Anal, Anil Kumar 2018-02-01 Ultrasonic assisted alkaline extraction of protein from banana flower was optimized using response surface methodology. The extracted proteins were characterized by Fourier transform infrared spectroscopy and molecular weight distribution was determined by gel electrophoresis. The maximum protein yield of 252.25 mg/g was obtained under optimized extraction conditions: temperature 50 °C, 30 min extraction time and 1 M NaOH concentration. The alkaline extraction produced a significantly high protein yield compared to enzymatic extraction of banana flower. Chemical finger printing of proteins showed the presence of tyrosine, tryptophan and amide bonds in extracted protein. Alkaline and pepsin assisted extracted banana flower proteins showed characteristic bands at 40 and 10 kDA, respectively. The extracted proteins showed antibacterial effects against both gram positive and gram negative bacteria. The high protein content and antimicrobial activity indicate the potential applications of banana flower in the food and feed industry. 14. Attitudes, perceptions, and trust. Insights from a consumer survey regarding genetically modified banana in Uganda. Science.gov (United States) Kikulwe, Enoch M; Wesseler, Justus; Falck-Zepeda, Jose 2011-10-01 Genetically modified (GM) crops and food are still controversial. This paper analyzes consumers' perceptions and institutional awareness and trust toward GM banana regulation in Uganda. Results are based on a study conducted among 421 banana-consuming households between July and August 2007. Results show a high willingness to purchase GM banana among consumers. An explanatory factor analysis is conducted to identify the perceptions toward genetic modification. The identified factors are used in a cluster analysis that grouped consumers into segments of GM skepticism, government trust, health safety concern, and food and environmental safety concern. Socioeconomic characteristics differed significantly across segments. Consumer characteristics and perception factors influence consumers' willingness to purchase GM banana. The institutional awareness and trust varied significantly across segments as well. The findings would be essential to policy makers when designing risk-communication strategies targeting different consumer segments to ensure proper discussion and addressing potential concerns about GM technology. Copyright © 2011 Elsevier Ltd. All rights reserved. 15. Effect of gamma radiation treatment on some fungi causing storage diseases of banana fruits International Nuclear Information System (INIS) EL-Ashmawi, A.M.M. 1982-01-01 Banana is one of the most popular fruits in many tropical and sub-tropical countries. in recent years, the quality of egyptian banana markedly declined. A major factor contributing to this decline is the development of fruit rot, which is the most widely occurring disease either in the field or in storage. Different fungi attack banana fruits causing considerable losses. Most of the fungi responsible for post harvest rots of banana are usually carried from the field, on the surface of the fruit itself or in injured and rotting fruits causing severe rats during storage. These rots make the fruits difficult to handle and undesirable to the consumers. Botryodiplodia theobromae is known to be the most important pathogen responsible for the infection in storage 16. FRUIT JUICES AS AN ALTERNATIVE TECHNIQUE FOR CONSERVATION OF FRESH-CUT BANANA OpenAIRE ANDERSON ADRIANO MARTINS MELO; LEONARDO THOMAZ DINIZ; ADRIANO DO NASCIMENTO SIMÕES; ROLF PUSCHMANN 2014-01-01 Browning discoloration after cutting is detrimental for the quality of a number of fruits and vegetables, such as banana, apple, pear, potato, and some roots such as cassava, yam, and others. Browning and softening compromise banana after cut shelf-life in a few hours under cold storage. Therefore, anti-browning compounds have been applied to slices before packing. Some commonly used substances are calcium chloride, ascorbic acid, cysteine and citric acid, in immersed inchemical mixtures. Rec... 17. Species of beetles (Coleoptera; Scarabaeidae associated to banana (Musa spp. in Ceballos, Ciego de Avila, Cuba Directory of Open Access Journals (Sweden) Maria Luisa Sisne Luis 2016-03-01 Full Text Available A white light trap was placed in bananas plantations, according to Sisne, 2009 and MINAG, 1985, in the Citric enterprise of Ciego de Ávila during the period between May and July of 2010 with the objective of determining the composition of genus and species of the order Coleoptera family Scarabaeidae associated to the agroecosystem. The species Cyclocephala cubana Chapin, Phyllophaga puberula Duval, and Phyllophaga patruelis Chev. are associated to bananas crops in these areas. 18. Effectiveness of ambon banana stem juice as immunostimulatory against Aeromonas hydrophila infections in catfish Clarias gariepinus OpenAIRE Qorie Astria; Sri Nuryati; Kukuh Nirmala; Alimuddin Alimuddin 2017-01-01 ABSTRACT Outbreaks of infectious diseases due to Aeromonas hydrophila in catfish can cause high death rates (80–100%). Fish disease control can be done using phytopharmaceutical to prevent or treat diseases of fish. One of the phytopharmaceutical that known to prevent the fish diseases is ambon banana stem Musa cavendishii var. dwarf Paxton. This study was conducted to test the effectiveness of catfish immersion using banana stem juice as an immunostimulant against bacterial infections A. hyd... 19. Anti-nutrients and heavy metals in some new plantain and banana ... African Journals Online (AJOL) Plantain and banana flour are important raw material in the baking and confectionery industry, and complementary food formulation. Five new plantain and banana ... PITA\\'s 14, 24 and Agbagba contained 4.0x10-5% cyanide, while 4.0x10-5 was obtained for PITA\\'s 17, 26 and BITA 3. The phenolic content of PITA\\'s 14 and ... 20. A Socio-Economic History of the International Banana Trade, 1870-1930 OpenAIRE ABBOTT, Roderick 2009-01-01 The genesis, and even more the growth, of the international banana industry is intimately bound up with the development of steamships (from 1850 onwards) and with the spread of railway construction around the world. The coming of steam, which ensured consistent and swifter passage from the Caribbean and Central America to the United States, and later to Europe, meant that bananas could be delivered in good condition rather than rotten, as had happened in earlier days. Later, when refrigerated... 1. Sulphate Potasium Extraction From Banana Stem Ash with Bleaching Earth Waste Liquid OpenAIRE Edahwati, Luluk 2010-01-01 Sulphate potassium is one of the important chemistry compound for industrial at our country, usually for fertilizerindustry. Therefore, necessary done sulphate potassium maker watchfulness from various ingredient that containsulphate potassium compound. Among others with extract banana stem ash. Watchfulness methodology that is withmix banana stem ash with aquadest (20 gram ash/1 water litre) in temperature 100oc during 30 minutes and stirringrotation 300 rpm. Ash extract that got reaction wi... 2. Repetitive part of the banana (Musa acuminata) genome investigated by low-depth 454 sequencing. Science.gov (United States) Hribová, Eva; Neumann, Pavel; Matsumoto, Takashi; Roux, Nicolas; Macas, Jirí; Dolezel, Jaroslav 2010-09-16 Bananas and plantains (Musa spp.) are grown in more than a hundred tropical and subtropical countries and provide staple food for hundreds of millions of people. They are seed-sterile crops propagated clonally and this makes them vulnerable to a rapid spread of devastating diseases and at the same time hampers breeding improved cultivars. Although the socio-economic importance of bananas and plantains cannot be overestimated, they remain outside the focus of major research programs. This slows down the study of nuclear genome and the development of molecular tools to facilitate banana improvement. In this work, we report on the first thorough characterization of the repeat component of the banana (M. acuminata cv. 'Calcutta 4') genome. Analysis of almost 100 Mb of sequence data (0.15× genome coverage) permitted partial sequence reconstruction and characterization of repetitive DNA, making up about 30% of the genome. The results showed that the banana repeats are predominantly made of various types of Ty1/copia and Ty3/gypsy retroelements representing 16 and 7% of the genome respectively. On the other hand, DNA transposons were found to be rare. In addition to new families of transposable elements, two new satellite repeats were discovered and found useful as cytogenetic markers. To help in banana sequence annotation, a specific Musa repeat database was created, and its utility was demonstrated by analyzing the repeat composition of 62 genomic BAC clones. A low-depth 454 sequencing of banana nuclear genome provided the largest amount of DNA sequence data available until now for Musa and permitted reconstruction of most of the major types of DNA repeats. The information obtained in this study improves the knowledge of the long-range organization of banana chromosomes, and provides sequence resources needed for repeat masking and annotation during the Musa genome sequencing project. It also provides sequence data for isolation of DNA markers to be used in genetic 3. Banana NAC transcription factor MusaNAC042 is positively associated with drought and salinity tolerance. Science.gov (United States) Tak, Himanshu; Negi, Sanjana; Ganapathi, T R 2017-03-01 Banana is an important fruit crop and its yield is hampered by multiple abiotic stress conditions encountered during its growth. The NAC (NAM, ATAF, and CUC) transcription factors are involved in plant response to biotic and abiotic stresses. In the present study, we studied the induction of banana NAC042 transcription factor in drought and high salinity conditions and its overexpression in transgenic banana to improve drought and salinity tolerance. MusaNAC042 expression was positively associated with stress conditions like salinity and drought and it encoded a nuclear localized protein. Transgenic lines of banana cultivar Rasthali overexpressing MusaNAC042 were generated by Agrobacterium-mediated transformation of banana embryogenic cells and T-DNA insertion was confirmed by PCR and Southern blot analysis. Our results using leaf disc assay indicated that transgenic banana lines were able to tolerate drought and high salinity stress better than the control plants and retained higher level of total chlorophyll and lower level of MDA content (malondialdehyde). Transgenic lines analyzed for salinity (250 mM NaCl) and drought (Soil gravimetric water content 0.15) tolerance showed higher proline content, better Fv/Fm ratio, and lower levels of MDA content than control suggesting that MusaNAC042 may be involved in responses to higher salinity and drought stresses in banana. Expression of several abiotic stress-related genes like those coding for CBF/DREB, LEA, and WRKY factors was altered in transgenic lines indicating that MusaNAC042 is an efficient modulator of abiotic stress response in banana. 4. Banana Cultivar Mapping and Constraints Identification With Farmers in Southwest Kenya International Nuclear Information System (INIS) Onyago, M.; Makworo, S.; Njue, E.; Rees, D. 1999-01-01 Southwest Kenya is one of the major banana growing ares of Kenya and contributes about 40% of the country's production. Banana is an important crop in Kenya providing the small-scale resource poor farmers with food and the much-needed income. In southwest Kenya, Banana is a priority crop that is grown in a wide range of farms from backyard gardens to medium farms. Rapid Rural Appraisal (RRP) and Participatory Rural Appraisal (PRA) were conducted between January, 1996 and December 1997. The overall objective was to identify major constraints in specific areas, rank them with farmers and determine suitable interventions. The RRA and PRA revealed that farmers grow banana as a pure stand, intercropped with crops like maize, beans and groundnuts, in addition along the contour lines for soil conservation purpose. Diverse banana cultivars are grown in the region, of which some are localised while others are widespread. The major cultivars grown include 'Ekeganda', Kisukari', 'Ng'ombe' and 'Pisang Awak'. Pisang Awak was found prevalent in drier areas of Kendu Bay (LM 3 ) and very cold areas of Bomet (LH 1 ). Ekeganda is the most popular cooking type while Kisukari is a dessert type and is the second most popular banana cultivar in the region. Some of the constraints cut across the AEZs and include a complex of pests and diseases, poor orchard management, soil exhaustion because of land pressure, lack of superior cultivars and socioeconomic factors such as poor infrastructure. Panama disease, nematodes and banana weevil are prevalent at the medium and lower AEZs while moles are menace in the upper midlands and highland areas. To address some of the major constraints, on-farm banana cultivar evaluation trials were initiated in four AEZs namely LM 3-4 , LH 1 , UM 1 and LM 2 , respectively. Monitoring and evaluation of these cultivars by both researchers, extension and farmers is on-going 5. Naisjuhtidega ettevõtete TOP 70 Index Scriptorium Estoniae 2003-01-01 Naisjuhtidega ettevõtete TOP 70. Käibe TOP 30. Käibe kasvu TOP 30. Kasumi TOP 30. Kasumi kasvu TOP 30. Rentaabluse TOP 30. Omakapitali tootlikkuse TOP 30. Nais- ja meesjuhtidega ettevõtted 2001. aasta 500 edukama ettevõtte hulgas. Naistegevjuhtidega firmade TOP 2001 ja 2002. Riigi- ja kohaliku omavalitsuse asutuste naisjuhtide TOP 20. Riigi- ja kohaliku omavalitsuse asutuste eelarve TOP 20. Riigi- ja kohaliku omavalitsuse asutuste töötajate arvu TOP 20. Naistegevjuhtidega firmade osakaal 2001. aastal. Riigi- ja kohaliku omavalitsuse asutuste naisjuhtide aastapalga TOP 20. Kommenteerib Tiina Raitviir. 6. Holographic entanglement entropy for hollow cones and banana shaped regions Energy Technology Data Exchange (ETDEWEB) Dorn, Harald [Institut für Physik und IRIS Adlershof, Humboldt-Universität zu Berlin,Zum Großen Windkanal 6, D-12489 Berlin (Germany) 2016-06-09 We consider banana shaped regions as examples of compact regions, whose boundary has two conical singularities. Their regularised holographic entropy is calculated with all divergent as well as finite terms. The coefficient of the squared logarithmic divergence, also in such a case with internally curved boundary, agrees with that calculated in the literature for infinite circular cones with their internally flat boundary. For the otherwise conformally invariant coefficient of the ordinary logarithmic divergence an anomaly under exceptional conformal transformations is observed. The construction of minimal submanifolds, needed for the entanglement entropy of cones, requires fine-tuning of Cauchy data. Perturbations of such fine-tuning leads to solutions relevant for hollow cones. The divergent parts for the entanglement entropy of hollow cones are calculated. Increasing the difference between the opening angles of their outer and inner boundary, one finds a transition between connected solutions for small differences to disconnected solutions for larger ones. 7. Structure analysis and laxative effects of oligosaccharides isolated from bananas. Science.gov (United States) Wang, Juan; Huang, Hui Hua; Cheng, Yan Feng; Yang, Gong Ming 2012-10-01 Banana oligosaccharides (BOS) were extracted with water, and then separated and purified using column chromatography. Gel penetration chromatography was used to determine the molecular weights. Thin layer chromatogram and capillary electrophoresis were employed to analyze the monosaccharide composition. The indican bond and structure of the BOS molecule were determined using Fourier transform infrared spectroscopy and nuclear magnetic resonance. Results showed that BOS were probably composed of eight β-D-pyran glucose units linked with 1→6 indican bonds. The laxative effects of BOS were investigated in mice using the method described in "Handbook of Technical Standards for Testing and Assessment of Health Food in China." The length of the small intestine over which a carbon suspension solution advanced in mice treated with low-, middle-, and high-dose BOS was significantly greater than that in the model group, suggesting that BOS are effective in accelerating the movement of the small intestine. 8. Biological control of banana black Sigatoka disease with Trichoderma Directory of Open Access Journals (Sweden) Poholl Adan Sagratzki Cavero 2015-06-01 Full Text Available Black Sigatoka disease caused by Mycosphaerella fijiensis is the most severe banana disease worldwide. The pathogen is in an invasive phase in Brazil and is already present in most States of the country. The potential of 29 isolates of Trichoderma spp. was studied for the control of black Sigatoka disease under field conditions. Four isolates were able to significantly reduce disease severity and were further tested in a second field experiment. Isolate 2.047 showed the best results in both field experiments and was selected for fungicide sensitivity tests and mass production. This isolate was identified as Trichoderma atroviride by sequencing fragments of the ITS region of the rDNA and tef-1α of the RNA polymerase. Trichoderma atroviride was as effective as the fungicide Azoxystrobin, which is recommended for controlling black Sigatoka. This biocontrol agent has potential to control the disease and may be scaled-up for field applications on rice-based solid fermentation 9. Improvement of bananas (Musa cvs.) through in vitro anther culture Energy Technology Data Exchange (ETDEWEB) Perea Dallos, M [Universidad Nacional de Colombia, Bogota (Colombia). Dept. de Biologia 1995-06-01 Agricultural products play a great role in the Colombian economy, and the banana is one of the most important. Since 1981, one of the more serious problems effecting production of this crop is the fungus Mycosphaerella fijensis sp. difformis, that causes black sigatoka disease. Most of the recent efforts to control this disease have been directed towards the identification of clones tolerant or resistant to this disease. One alternative approach is the use of anther culture to obtain resistant haploid plants. Diploid clones (Musa - AA) have been used as a model in this study. The results presented here identify the most appropriate stage of anther development for callus induction and proliferation, and treatments that reduce tissue browning. (author). 12 refs, 1 fig., 1 tab. 10. Production and storage stability of non alcoholic banana beverage powder. Science.gov (United States) Mugula, J K; Lyimo, M H; Kessy, F L 1994-02-01 Powder for an instant, non-alcoholic beverage formulation was manufactured by sundrying and ovendrying of a popular dessert ('silk') banana variety. The reconstituted beverage was organoleptically acceptable. The effect of traditional sundrying on mats and ovendrying methods on product quality was investigated. Sundrying resulted in losses of Vitamin A, C and total sugar contents by 74, 91 and 63%, while ovendrying losses were 73, 90 and 62%, respectively. Nutrient losses during storage for three months in transparent polythene bags reached 93, 93 and 70% in sundried samples and 84, 99 and 55% in ovendried samples, respectively. The moisture content of sundried and ovendried samples increased by 12 and 17%, respectively, during storage. The increase in microbial load in this period was higher in sundried samples. 11. 32-Channel banana-avg montage is better than 16-channel double banana montage to detect epileptiform discharges in routine EEGs. Science.gov (United States) Ochoa, Juan; Gonzalez, Walter; Bautista, Ramon; DeCerce, John 2008-10-01 We designed a study, comparing the yield of standard 16-channel longitudinal bipolar montage (double banana) versus a combined 32-channel longitudinal bipolar plus average referential montage (banana-plus), to detect epileptic abnormalities. We selected 25 consecutive routine EEG samples with a diagnosis of spike or sharp waves in the temporal regions and 25 consecutive focal slowing and 50 normal EEGs. A total of 100 samples were printed in both montages and randomized for reading. Thirty independent EEG readers blinded from the EEG diagnosis were invited to participate. Twenty-two readers successfully completed the test for a total of 4400 answers collected for analysis. The average sensitivity to detect epileptiform discharges for 16 and 32-channel montages was 36.5% and 61%, respectively (Pdouble banana montage. Residents and EEG fellows could improve EEG-reading accuracy if taught on a combined 32-channel montage. 12. Natural Ecosystem Surrounding a Conventional Banana Crop Improves Plant Health and Fruit Quality Directory of Open Access Journals (Sweden) Florence P. Castelan 2018-06-01 Full Text Available Natural ecosystems near agricultural landscapes may provide rich environments for growing crops. However, the effect of a natural ecosystem on crop health and fruit quality is poorly understood. In the present study, it was investigated whether the presence of a natural ecosystem surrounding a crop area influences banana plant health and fruit postharvest behavior. Plants from two conventional banana crop areas with identical planting time and cultural practices were used; the only difference between banana crop areas is that one area was surrounded by a natural forest (Atlantic forest fragment (Near-NF, while the other area was inserted at the center of a conventional banana crop (Distant-NF. Results showed that bananas harvested from Near-NF showed higher greenlife and a more homogeneous profile during ripening compared to fruits harvested from Distant-NF. Differences in quality parameters including greenlife, carbohydrate profile, and pulp firmness between fruits harvested from Near-NF and Distant-NF are explained, at least partly, by differences in the balance of plant growth regulators (indole-3-acetic acid and abscisic acid in bananas during ripening. Furthermore, plants from Near-NF showed a lower severity index of black leaf streak disease (BLSD and higher levels of phenolic compounds in leaves compared to plants from Distant-NF. Together, the results provide additional evidence on how the maintenance of natural ecosystems near conventional crop areas could be a promising tool to improve plant health and fruit quality. 13. Effect of chitosan coating and bamboo FSC (fruit storage chamber) to expand banana shelf life Science.gov (United States) Pratiwi, Aksarani'Sa; Dwivany, Fenny M.; Larasati, Dwinita; Islamia, Hana Cahya; Martien, Ronny 2015-09-01 Chitosan has been widely used as fruit preserver and proven to extend the shelf life of many fruits, such as banana. However, banana producers and many industries in Indonesia still facing storage problems which may lead to mechanical damage of the fruits and ripening acceleration. Therefore, we have designed food storage chamber (FSC) based on bamboo material. Bamboo was selected because of material abundance in Indonesia, economically effective, and not causing an autocatalytic reaction to the ethylene gas produced by the banana. In this research, Cavendish banana that has reached the maturity level of mature green were coated with 1% chitosan and placed inside the FSC. As control treatments, uncoated banana was also placed inside the FSC as well as uncoated banana that were placed at open space. All of the treatments were placed at 25°C temperature and observed for 9 days. Water produced by respiration was reduced by the addition of charcoal inside a fabric pouch. The result showed that treatment using FSC and chitosan can delay ripening process. 14. Identification of genes encoding granule-bound starch synthase involved in amylose metabolism in banana fruit. Directory of Open Access Journals (Sweden) Hongxia Miao Full Text Available Granule-bound starch synthase (GBSS is responsible for amylose synthesis, but the role of GBSS genes and their encoded proteins remains poorly understood in banana. In this study, amylose content and GBSS activity gradually increased during development of the banana fruit, and decreased during storage of the mature fruit. GBSS protein in banana starch granules was approximately 55.0 kDa. The protein was up-regulated expression during development while it was down-regulated expression during storage. Six genes, designated as MaGBSSI-1, MaGBSSI-2, MaGBSSI-3, MaGBSSI-4, MaGBSSII-1, and MaGBSSII-2, were cloned and characterized from banana fruit. Among the six genes, the expression pattern of MaGBSSI-3 was the most consistent with the changes in amylose content, GBSS enzyme activity, GBSS protein levels, and the quantity or size of starch granules in banana fruit. These results suggest that MaGBSSI-3 might regulate amylose metabolism by affecting the variation of GBSS levels and the quantity or size of starch granules in banana fruit during development or storage. 15. Fermentative characteristics and nutritional value of elephant grass silage added with dehydrated banana peel Directory of Open Access Journals (Sweden) Lara Maria Santos Brant 2017-04-01 Full Text Available The aim of this study was to evaluate the fermentative losses and nutritional value of elephant grass silages with the increasing of inclusion levels of dehydrated banana peel. The experiment was conducted in a completely randomized design, with six treatments and four replications, being the silage exclusively from elephant grass, and five levels of inclusion of banana peel to the elephant grass silage, as the following - 5; 10; 15; 20 and 25%, being added based on natural matter. The addition of the banana peel in the silage reduced linearly (p < 0.05 the pH, the ammoniacal nitrogen and the losses of the fermentative process. In addition, the inclusion of banana peel increased linearly (p < 0.05 the dry matter and non-fibrous carbohydrates. On the other hand, the neutral detergent fiber and the acid detergent fiber were linearly reduced with the inclusion of the banana peel (p < 0.05, but there was no change in the dry matter digestibility in situ. The inclusion of dehydrated banana peel in elephant grass silage reduces the losses of the fermentation process with more consistent results at the 25% inclusion level, however, it reduces the silage nutritional value due to fibrous and protein quality. 16. Dietary fibre components and pectin chemical features of peels during ripening in banana and plantain varieties. Science.gov (United States) Happi Emaga, Thomas; Robert, Christelle; Ronkart, Sébastien N; Wathelet, Bernard; Paquot, Michel 2008-07-01 The effects of the ripeness stage of banana (Musa AAA) and plantain (Musa AAB) peels on neutral detergent fibre, acid detergent fibre, cellulose, hemicelluloses, lignin, pectin contents, and pectin chemical features were studied. Plantain peels contained a higher amount of lignin but had a lower hemicellulose content than banana peels. A sequential extraction of pectins showed that acid extraction was the most efficient to isolate banana peel pectins, whereas an ammonium oxalate extraction was more appropriate for plantain peels. In all the stages of maturation, the pectin content in banana peels was higher compared to plantain peels. Moreover, the galacturonic acid and methoxy group contents in banana peels were higher than in plantain peels. The average molecular weights of the extracted pectins were in the range of 132.6-573.8 kDa and were not dependant on peel variety, while the stage of maturation did not affect the dietary fibre yields and the composition in pectic polysaccharides in a consistent manner. This study has showed that banana peels are a potential source of dietary fibres and pectins. 17. Use of Banana (Musa acuminata Colla AAA) Peel Extract as an Antioxidant Source in Orange Juices. Science.gov (United States) Ortiz, Lucía; Dorta, Eva; Gloria Lobo, M; González-Mendoza, L Antonio; Díaz, Carlos; González, Mónica 2017-03-01 Using banana peel extract as an antioxidant in freshly squeezed orange juices and juices from concentrate was evaluated. Free radical scavenging capacity increased by adding banana peel extracts to both types of orange juice. In addition, remarkable increases in antioxidant capacity using 2,2'-azino-bis-(3-ethylbenzothiazoline)-6-sulfonic acid (ABTS) radical were observed when equal or greater than 5 mg of banana peel extract per ml of freshly squeezed juice was added. No clear effects were observed in the capacity to inhibit lipid peroxidation. Adding 5 mg banana peel extract per ml of orange juice did not substantially modify the physicochemical and sensory characteristics of either type of juice. However, undesirable changes in the sensory characteristics (in-mouth sensations and colour) were detected when equal or greater than 10 mg banana peel extract per ml of orange juice was added. These results confirm that banana peel is a promising natural additive that increases the capacity to scavenge free radicals of orange juice with acceptable sensory and physicochemical characteristics for the consumer. 18. Agroforestry leads to shifts within the gammaproteobacterial microbiome of banana plants cultivated in Central America. Science.gov (United States) Köberl, Martina; Dita, Miguel; Martinuz, Alfonso; Staver, Charles; Berg, Gabriele 2015-01-01 Bananas (Musa spp.) belong to the most important global food commodities, and their cultivation represents the world's largest monoculture. Although the plant-associated microbiome has substantial influence on plant growth and health, there is a lack of knowledge of the banana microbiome and its influencing factors. We studied the impact of (i) biogeography, and (ii) agroforestry on the banana-associated gammaproteobacterial microbiome analyzing plants grown in smallholder farms in Nicaragua and Costa Rica. Profiles of 16S rRNA genes revealed high abundances of Pseudomonadales, Enterobacteriales, Xanthomonadales, and Legionellales. An extraordinary high diversity of the gammaproteobacterial microbiota was observed within the endophytic microenvironments (endorhiza and pseudostem), which was similar in both countries. Enterobacteria were identified as dominant group of above-ground plant parts (pseudostem and leaves). Neither biogeography nor agroforestry showed a statistically significant impact on the gammaproteobacterial banana microbiome in general. However, indicator species for each microenvironment and country, as well as for plants grown in Coffea intercropping systems with and without agri-silvicultural production of different Fabaceae trees (Inga spp. in Nicaragua and Erythrina poeppigiana in Costa Rica) could be identified. For example, banana plants grown in agroforestry systems were characterized by an increase of potential plant-beneficial bacteria, like Pseudomonas and Stenotrophomonas, and on the other side by a decrease of Erwinia. Hence, this study could show that as a result of legume-based agroforestry the indigenous banana-associated gammaproteobacterial community noticeably shifted. 19. Effects of Green Banana Flour on the Physical, Chemical and Sensory Properties of Ice Cream Directory of Open Access Journals (Sweden) Filiz Yangılar 2015-01-01 Full Text Available In the present study, possible eff ects of the addition of banana flour at different mass fractions (1 and 2 % are investigated on physical (overrun, viscosity, chemical (dry matter, fat and ash content, acidity, pH, water and oil holding capacity and colour, mineral content (Ca, K, Na, P, S, Mg, Fe, Mn, Zn and Ni and sensory properties of ice cream. Fibre-rich banana pieces were found to contain 66.8 g per 100 g of total dietary fibre, 58.6 g per 100 g of which were insoluble dietary fi bre, while 8.2 g per 100 g were soluble dietary fi bre. It can be concluded from these results that banana is a valuable dietary fi bre source which can be used in food production. Flour obtained from green banana pulp and peel was found to have signifi cant (p<0.05 effect on the chemical composition of ice creams. Sulphur content increased while calcium content decreased in ice cream depending on banana flour content. Sensory results indicated that ice cream sample containing 2 % of green banana pulp flour received the highest score from panellists. 20. Hyperkalemia and hyperdopaminemia induced by an obsessive eating of banana in an anorexia nervosa adolescent. Science.gov (United States) Tazoe, Mami; Narita, Masaaki; Sakuta, Ryoichi; Nagai, Toshiro; Narita, Naoko 2007-07-01 Banana is known as a dopamine-rich and potassium-rich food, however no previous data regarding biochemical or psychological alteration induced by excess intake of banana has been reported. We have experienced an adolescent female case of Anorexia nervosa (AN) who denied eating anything but maximum 20 bananas and less than 500 ml mineral water per day for more than two years. During the period of massive banana eating habit, she showed increase of serum potassium (from 4.7 mEq/l to 6.1 mEq/l) and whole blood dopamine (from 11 ng/ml to 210 ng/ml; normal range 0.5-6.2 ng/ml), and obvious dysthymia that is inexplicable only by the pathology of AN. When the patient resumed other food ingestion after 26 months of obsessive and restricted eating of banana, the abnormalities in her blood data and her psychological state were all corrected toward normal. We conclude that in this case, the obsessive and restricted habit of banana ingestion resulted in hyperkalemia, hyperdopaminemia, and psychological change. 1. Socioeconomic importance of the banana tree (Musa spp.) in the Guinean Highland Savannah agroforests. Science.gov (United States) Mapongmetsem, Pierre Marie; Nkongmeneck, Bernard Aloys; Gubbuk, Hamide 2012-01-01 Home gardens are defined as less complex agroforests which look like and function as natural forest ecosystems but are integrated into agricultural management systems located around houses. Investigations were carried out in 187 households. The aim of the study was to identify the different types of banana home gardens existing in the periurban zone of Ngaoundere town. The results showed that the majority of home gardens in the area were very young (less than 15 years old) and very small in size (less than 1 ha). Eleven types of home gardens were found in the periurban area of Ngaoundere town. The different home garden types showed important variations in all their structural characteristics. Two local species of banana are cultivated in the systems, Musa sinensis and Musa paradisiaca. The total banana production is 3.57 tons per year. The total quantity of banana consumed in the periurban zone was 3.54 tons (93.5%) whereas 1.01 tons were sold in local or urban markets. The main banana producers belonged to home gardens 2, 4, 7, and 9. The quantity of banana offered to relatives was more than what the farmers received from others. Farmers, rely on agroforests because the flow of their products helps them consolidate friendship and conserve biodiversity at the same time. 2. Socioeconomic Importance of the Banana Tree (Musa Spp. in the Guinean Highland Savannah Agroforests Directory of Open Access Journals (Sweden) Pierre Marie Mapongmetsem 2012-01-01 Full Text Available Home gardens are defined as less complex agroforests which look like and function as natural forest ecosystems but are integrated into agricultural management systems located around houses. Investigations were carried out in 187 households. The aim of the study was to identify the different types of banana home gardens existing in the periurban zone of Ngaoundere town. The results showed that the majority of home gardens in the area were very young (less than 15 years old and very small in size (less than 1 ha. Eleven types of home gardens were found in the periurban area of Ngaoundere town. The different home garden types showed important variations in all their structural characteristics. Two local species of banana are cultivated in the systems, Musa sinensis and Musa paradisiaca. The total banana production is 3.57 tons per year. The total quantity of banana consumed in the periurban zone was 3.54 tons (93.5% whereas 1.01 tons were sold in local or urban markets. The main banana producers belonged to home gardens 2, 4, 7, and 9. The quantity of banana offered to relatives was more than what the farmers received from others. Farmers, rely on agroforests because the flow of their products helps them consolidate friendship and conserve biodiversity at the same time. 3. Identification of Genes Encoding Granule-Bound Starch Synthase Involved in Amylose Metabolism in Banana Fruit Science.gov (United States) Liu, Weixin; Xu, Biyu; Jin, Zhiqiang 2014-01-01 Granule-bound starch synthase (GBSS) is responsible for amylose synthesis, but the role of GBSS genes and their encoded proteins remains poorly understood in banana. In this study, amylose content and GBSS activity gradually increased during development of the banana fruit, and decreased during storage of the mature fruit. GBSS protein in banana starch granules was approximately 55.0 kDa. The protein was up-regulated expression during development while it was down-regulated expression during storage. Six genes, designated as MaGBSSI-1, MaGBSSI-2, MaGBSSI-3, MaGBSSI-4, MaGBSSII-1, and MaGBSSII-2, were cloned and characterized from banana fruit. Among the six genes, the expression pattern of MaGBSSI-3 was the most consistent with the changes in amylose content, GBSS enzyme activity, GBSS protein levels, and the quantity or size of starch granules in banana fruit. These results suggest that MaGBSSI-3 might regulate amylose metabolism by affecting the variation of GBSS levels and the quantity or size of starch granules in banana fruit during development or storage. PMID:24505384 4. Continous application of bioorganic fertilizer induced resilient culturable bacteria community associated with banana Fusarium wilt suppression Science.gov (United States) Fu, Lin; Ruan, Yunze; Tao, Chengyuan; Li, Rong; Shen, Qirong 2016-06-01 Fusarium wilt of banana always drives farmers to find new land for banana cultivation due to the comeback of the disease after a few cropping years. A novel idea for solving this problem is the continuous application of bioorganic fertilizer (BIO), which should be practiced from the beginning of banana planting. In this study, BIO was applied in newly reclaimed fields to pre-control banana Fusarium wilt and the culturable rhizobacteria community were evaluated using Biolog Ecoplates and culture-dependent denaturing gradient gel electrophoresis (CD-DGGE). The results showed that BIO application significantly reduced disease incidences and increased crop yields, respectivly. And the stabilized general bacterial metabolic potential, especially for the utilization of carbohydrates, carboxylic acids and phenolic compounds, was induced by BIO application. DGGE profiles demonstrated that resilient community structure of culturable rhizobacteria with higher richness and diversity were observed in BIO treated soils. Morever, enriched culturable bacteria affiliated with Firmicutes, Gammaproteobacteria and Actinobacteria were also detected. In total, continuous application of BIO effectively suppressed Fusarium wilt disease by stabilizing culturable bacterial metabolic potential and community structure. This study revealed a new method to control Fusarium wilt of banana for long term banana cultivation. 5. Ripe Banana Flour as a Source of Antioxidants in Layer and Sponge Cakes. Science.gov (United States) Segundo, Cristina; Román, Laura; Lobo, Manuel; Martinez, Mario M; Gómez, Manuel 2017-12-01 About one-fifth of all bananas harvested become culls that are normally disposed of improperly. However, ripe banana pulp contains significant amounts of fibre and polyphenol compounds as well as a high content of simple sugars (61.06 g/100 g), making it suitable for sucrose replacement in bakery products. This work studied the feasibility of incorporating ripe banana flour (20 and 40% of replacement) in cake formulation. Physical, nutritional and sensory attributes of sponge and layer cakes were evaluated. The inclusion of ripe banana flour generally led to an increased batter consistency that hindered cake expansion, resulting in a slightly lower specific volume and higher hardness. This effect was minimised in layer cakes where differences in volume were only evident with the higher level of replacement. The lower volume and higher hardness contributed to the decline of the acceptability observed in the sensory test. Unlike physical attributes, the banana flour inclusion significantly improved the nutritional properties of the cakes, bringing about an enhancement in dietary fibre, polyphenols and antioxidant capacity (up to a three-fold improvement in antioxidant capacity performance). Therefore, results showed that sugar replacement by ripe banana flour enhanced the nutritional properties of cakes, but attention should be paid to its inclusion level. 6. Effects of Green Banana Flour on the Physical, Chemical and Sensory Properties of Ice Cream. Science.gov (United States) Yangılar, Filiz 2015-09-01 In the present study, possible effects of the addition of banana flour at different mass fractions (1 and 2%) are investigated on physical (overrun, viscosity), chemical (dry matter, fat and ash content, acidity, pH, water and oil holding capacity and colour), mineral content (Ca, K, Na, P, S, Mg, Fe, Mn, Zn and Ni) and sensory properties of ice cream. Fibre--rich banana pieces were found to contain 66.8 g per 100 g of total dietary fibre, 58.6 g per 100 g of which were insoluble dietary fibre, while 8.2 g per 100 g were soluble dietary fibre. It can be concluded from these results that banana is a valuable dietary fibre source which can be used in food production. Flour obtained from green banana pulp and peel was found to have significant (pice creams. Sulphur content increased while calcium content decreased in ice cream depending on banana flour content. Sensory results indicated that ice cream sample containing 2% of green banana pulp flour received the highest score from panellists. 7. Standardization of a molecular diagnostic method for Cucumber mosaic virus (cmv in Ecuadorian bananas Directory of Open Access Journals (Sweden) Johanna Liseth Buitrón-Bustamante 2017-01-01 Full Text Available Several pests and diseases affect banana crop in Ecuador and Cucumber mosaic virus (cmv is one of the most important pathogens. The aim of this research was to standardize a new molecular approach to achieve a sensitive and highly specific detection of cmv in Ecuadorian bananas. Specific primers were designed from the sequence encodingResumoA cultura da banana no Equador vê-se afetada por uma série de doenças, das quais o cucumber mosaic vírus(cmv é um dos fitopatógenos mais impor-tantes. Com este estudo procurou-se padronizar uma técnica molecular para a detecção sensível e altamente específica deste agente viral na banana equatoriana. Para este fim, realizou-se o desenho de primers específicos, a partir da sequência que se codifica para a proteína da cápside do vírus. for the virus capsid protein. PC-F1, PC-R D1 and K-F primers, obtained from cDNA replicated from R NA of infected banana, allowed accurate virus detection by Reverse transcription and Hemi-nested PCR. Virus detection was possible even in asymptomatic plants, providing a tech-nology with potential use for the Ecuadorian banana producers. 8. Financial top 100 International Nuclear Information System (INIS) Jaremko, G. 1999-01-01 The way in which the oil industry has rebounded from a difficult financial year was discussed. In 1998, oil prices fell by about 30 per cent from 1997 prices to average US$14.40. In the second quarter of 1999, the rebounding international markets raised oil prices by about one-third to an average US17.60. Recent surveys have shown that the oil and gas industry will experience major changes in the decade ahead. It was forecasted that in 25 years, total Canadian oil production could near three million barrels per day, an increase of 36 per cent from today's 2.2 million. However, as conventional reserves are running out, production of conventional oil is expected to be diminished by about two-thirds. Naturally-occurring, light refinery-ready oil is forecasted to come mostly from the east coast, whereas western output will be dominated by oil sands. Natural gas will continue to be a huge growth opportunity in the west because of new pipeline expansions and exports to the United States. This paper presented a list of Canada's financial top 100 oil and gas companies which emerged with the best endurance. The list included statistics from 1997 and 1998 for gross revenues, total assets, net income, cash/operating activities, capital expenditures, closing share prices (with percentage change), and market capitalization. 1 fig., 1 tab 9. Our top priority CERN Multimedia 2012-01-01 After three years of LHC running, we are still at the beginning of a long research programme with our flagship facility, and hopefully 4 July 2012 will go down in history as the date of one of many landmark discoveries spanning several years. CERN’s top priority for the next decade and more is the full exploitation of the LHC. With speculation about potential future facilities mounting in the light of the discovery of a new Higgs-like particle, it’s important to state that most clearly. Of course, this will rely on continued global collaboration, and it’s important that CERN engage constructively with other regions. It is important to plan ahead, particularly since the lead times for new projects in particle physics are long, and our field is increasingly global in nature. That’s why the European particle physics community is currently engaged in updating its long-term strategy. Planning ahead allowed us to be ready technologically to build the LHC whe... 10. Participative Design With Top Management DEFF Research Database (Denmark) Simonsen, Jesper 2004-01-01 meetings aimed at aligning top management with the suppliers analysis. The article describes the MUST methods anchoring principle and the technique of problem mapping supporting this principle. This participatory approach resulted in mutual learning processes with top management which is rarely reported...... on in the PD community. Top management participated by reviewing, challenging, and reformulating the IT designers central suppositions, assumptions, and hypotheses related to the causal relation between identified problems and suggested solutions.... 11. Ig Nobel Prize-winning episode: Trip from a slip on a banana peel to the mysterious world of mucus OpenAIRE K. Mabuchi; R. Sakai; M. Honna; M. Ujihira 2016-01-01 Slip on a banana peel is not only a gag seed but also a genuinely tribological phenomenon. We measured the frictional coefficient under banana skin on floor material. The measured frictional coefficient was much lower than the value on common materials and similar one on well lubricated surfaces. Some deductions on mystery of organics were leaded from the similarity of gel function in banana peels and in articular joints. Every polymers are only synthesized by organisms. Furthermore, viscous ... 12. Sensory and chemical interactions of food pairings (basmati rice, bacon and extra virgin olive oil) with banana OpenAIRE Traynor, Mark; Burke, Roisin; O'Sullivan, Maurice G; Hannon, John; Barry-Ryan, Catherine 2013-01-01 The aim of this study aimed to investigate food pairings as an important sensory phenomenon in order to determine how different components in the selected food pairings affect and interact with other components. Three novel food pairings (banana and bacon, banana and olive oil, and banana and rice) were selected. A conjoint approach utilising qualitative (organic volatile analysis and descriptive sensory analysis) and quantitative (comparable semi quantitative organic volatile analysis and af... 13. Boosted top: experimental tools overview CERN Document Server Usai, Emanuele 2015-01-01 An overview of tools and methods for the reconstruction of high-boost top quark decays at the LHC is given in this report. The focus is on hadronic decays, in particular an overview of the current status of top quark taggers in physics analyses is presented. The most widely used jet substructure techniques, normally used in combination with top quark taggers, are reviewed. Special techniques to treat pileup in large cone jets are described, along with a comparison of the performance of several boosted top quark reconstruction techniques. 14. Could stops lighten the top? International Nuclear Information System (INIS) Bilal, A.; Ellis, J.; Fogli, G.L. 1990-01-01 The analysis of the presently available electroweak data including radiative corrections in the standard model suggests that the top quark weighs more than the Z 0 . We examine whether squark loops in the minimal supersymmetric model, particularly those involving stops (partners of the top quark), could reduce substantially the preferred range of top quark masses. Given the present lower bounds on squark masses, we find that stop effects can reduce the central value of m t by at most a few GeV, although they do make a very heavy top quark increasingly unlikely. (orig.) 15. (2R,5S)-Theaspirane Identified as the Kairomone for the Banana Weevil, Cosmopolites sordidus, from Attractive Senesced Leaves of the Host Banana, Musa spp. Science.gov (United States) Abagale, Samson A; Woodcock, Christine M; Hooper, Antony M; Caulfield, John C; Withall, David; Chamberlain, Keith; Acquaah, Samuel O; Van Emden, Helmut; Braimah, Haruna; Pickett, John A; Birkett, Michael A 2018-04-12 The principal active component produced by highly attractive senesced host banana leaves, Musa spp., for the banana weevil, Cosmopolites sordidus, is shown by coupled gas chromatography-electroantennography (GC-EAG), coupled GC-mass spectrometry (GC-MS), chemical synthesis and coupled enantioselective (chiral) GC-EAG to be (2R,5S)-theaspirane. In laboratory behaviour tests, the synthetic compound is as attractive as natural host leaf material and presents a new opportunity for pest control. © 2018 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim. 16. Utilization Of The Visceral Organsof Catfish (Pangasius Hypoptalmus) Added To Banana Peel (Musa Acuminata Balbisiana) To Produceliquid Organic Fertilizer OpenAIRE Fryathama, Ilham; Sukmiwati, Mery; ', Sumarto 2017-01-01 This study aimed to obtain liquid organic fertilizer produced from the visceral organs of catfish added to banana skin for enriching the macro elements N, P, and K.The method used in this study was the experimental. Parameters used were the value of pH, nitrogen, phosphorus and potassium. The experimentwas composed as completely randomized design (CRD), and the treatment was addition of banana peel which varied into 4 different amoun, namely: without the banana peel (K0), banana peel 100 g (K... 17. Detection of antimicrobial activity of banana peel (Musa paradisiaca L.) on Porphyromonas gingivalis and Aggregatibacter actinomycetemcomitans: An in vitro study. Science.gov (United States) Kapadia, Suraj Premal; Pudakalkatti, Pushpa S; Shivanaikar, Sachin 2015-01-01 Banana is used widely because of its nutritional values. In past, there are studies that show banana plant parts, and their fruits can be used to treat the human diseases. Banana peel is a part of banana fruit that also has the antibacterial activity against microorganisms but has not been studied extensively. Since, there are no studies that relate the antibacterial activity of banana peel against periodontal pathogens. Hence, the aim of this study is to determine the antimicrobial activity of banana peel extract on Porphyromonas gingivalis (P. gingivalis) and Aggregatibacter actinomycetemcomitans (A. actinomycetemcomitans). Standard strains of P. gingivalis and A. actinomycetemcomitans were used in this study which was obtained from the in-house bacterial bank of Department of Molecular Biology and Immunology at Maratha Mandal's Nathajirao G. Halgekar Institute of Dental Sciences and Research Centre. The banana peel extract was prepared, and the antibacterial activity was assessed using well agar diffusion method and minimum inhibitory concentration was assessed using serial broth dilution method. In the current study, both the tested microorganisms showed antibacterial activity. In well diffusion method, P. gingivalis and A. actinomycetemcomitans showed 15 mm and 12 mm inhibition zone against an alcoholic extract of banana peel, respectively. In serial broth dilution method P. gingivalis and A. actinomycetemcomitans were sensitive until 31.25 μg/ml dilutions. From results of the study, it is suggested that an alcoholic extract of banana peel has antimicrobial activity against P. gingivalis and A. actinomycetemcomitans. 18. Genome-Wide Identification and Expression Analyses of Aquaporin Gene Family during Development and Abiotic Stress in Banana Science.gov (United States) Hu, Wei; Hou, Xiaowan; Huang, Chao; Yan, Yan; Tie, Weiwei; Ding, Zehong; Wei, Yunxie; Liu, Juhua; Miao, Hongxia; Lu, Zhiwei; Li, Meiying; Xu, Biyu; Jin, Zhiqiang 2015-01-01 Aquaporins (AQPs) function to selectively control the flow of water and other small molecules through biological membranes, playing crucial roles in various biological processes. However, little information is available on the AQP gene family in bananas. In this study, we identified 47 banana AQP genes based on the banana genome sequence. Evolutionary analysis of AQPs from banana, Arabidopsis, poplar, and rice indicated that banana AQPs (MaAQPs) were clustered into four subfamilies. Conserved motif analysis showed that all banana AQPs contained the typical AQP-like or major intrinsic protein (MIP) domain. Gene structure analysis suggested the majority of MaAQPs had two to four introns with a highly specific number and length for each subfamily. Expression analysis of MaAQP genes during fruit development and postharvest ripening showed that some MaAQP genes exhibited high expression levels during these stages, indicating the involvement of MaAQP genes in banana fruit development and ripening. Additionally, some MaAQP genes showed strong induction after stress treatment and therefore, may represent potential candidates for improving banana resistance to abiotic stress. Taken together, this study identified some excellent tissue-specific, fruit development- and ripening-dependent, and abiotic stress-responsive candidate MaAQP genes, which could lay a solid foundation for genetic improvement of banana cultivars. PMID:26307965 19. Banana infecting fungus, Fusarium musae, is also an opportunistic human pathogen: are bananas potential carriers and source of fusariosis? Science.gov (United States) Triest, David; Stubbe, Dirk; De Cremer, Koen; Piérard, Denis; Detandt, Monique; Hendrickx, Marijke 2015-01-01 During re-identification of Fusarium strains in the BCCM™/IHEM fungal collection by multilocus sequence-analysis we observed that five strains, previously identified as Fusarium verticillioides, were Fusarium musae, a species described in 2011 from banana fruits. Four strains were isolated from blood samples or biopsies of immune-suppressed patients and one was isolated from the clinical environment, all originating from different hospitals in Belgium or France, 2001-2008. The F. musae identity of our isolates was confirmed by phylogenetic analysis using reference sequences of type material. Absence of the gene cluster necessary for fumonisin biosynthesis, characteristic to F. musae, was also the case for our isolates. In vitro antifungal susceptibility testing revealed no important differences in their susceptibility compared to clinical F. verticillioides strains and terbinafine was the most effective drug. Additional clinical F. musae strains were searched by performing BLAST queries in GenBank. Eight strains were found, of which six were keratitis cases from the U.S. multistate contact lens-associated outbreak in 2005 and 2006. The two other strains were also from the U.S., causing either a skin infection or sinusitis. This report is the first to describe F. musae as causative agent of superficial and opportunistic, disseminated infections in humans. Imported bananas might act as carriers of F. musae spores and be a potential source of infection with F. musae in humans. An alternative hypothesis is that the natural distribution of F. musae is geographically a lot broader than originally suspected and F. musae is present on different plant hosts. © 2015 by The Mycological Society of America. 20. Effect of packaging materials on shelf life and quality of banana cultivars (Musa spp.). Science.gov (United States) Hailu, M; Seyoum Workneh, T; Belew, D 2014-11-01 This study was carried out to evaluate the effect of packaging materials on the shelf life of three banana cultivars. Four packaging materials, namely, perforated low density polyethylene bag, perforated high density polyethylene bag, dried banana leaf, teff straw and no packaging materials (control) were used with three banana cultivars, locally known as, Poyo, Giant Cavendish and Williams I. The experiment was carried out in Randomized Complete Block Design in a factorial combination with three replications. Physical parameters including weight loss, peel colour, peel thickness, pulp thickness, pulp to peel ratio, pulp firmness, pulp dry matter, decay, loss percent of marketability were assessed every 3 days. Banana remained marketable for 36 days in the high density polyethylene and low density polyethylene bags, and for 18 days in banana leaf and teff straw packaging treatments. Unpackaged fruits remained marketable for 15 days only. Fruits that were not packaged lost their weight by 24.0 % whereas fruits packaged in banana leaf and teff straw became unmarketable with final weight loss of 19.8 % and 20.9 %, respectively. Packaged fruits remained well until 36th days of storage with final weight loss of only 8.2 % and 9.20 %, respectively. Starting from green mature stage, the colour of the banana peel changed to yellow and this process was found to be fast for unpackaged fruits. Packaging maintained the peel and the pulp thickness, firmness, dry matter and pulp to peel ratio was kept lower. Decay loss for unpackaged banana fruits was16 % at the end of date 15, whereas the decay loss of fruits packaged using high density and low density polyethylene bags were 43.0 % and 41.2 %, respectively at the end of the 36th day of the experiment. It can, thus, be concluded that packaging of banana fruits in high density and low density polyethylene bags resulted in longer shelf life and improved quality of the produce followed by packaging in dried banana leaf 1. Origins and Domestication of Cultivated Banana Inferred from Chloroplast and Nuclear Genes Science.gov (United States) Zhang, Cui; Wang, Xin-Feng; Shi, Feng-Xue; Chen, Wen-Na; Ge, Xue-Jun 2013-01-01 Background Cultivated bananas are large, vegetatively-propagated members of the genus Musa. More than 1,000 cultivars are grown worldwide and they are major economic and food resources in numerous developing countries. It has been suggested that cultivated bananas originated from the islands of Southeast Asia (ISEA) and have been developed through complex geodomestication pathways. However, the maternal and parental donors of most cultivars are unknown, and the pattern of nucleotide diversity in domesticated banana has not been fully resolved. Methodology/Principal Findings We studied the genetics of 16 cultivated and 18 wild Musa accessions using two single-copy nuclear (granule-bound starch synthase I, GBSS I, also known as Waxy, and alcohol dehydrogenase 1, Adh1) and two chloroplast (maturase K, matK, and the trnL-F gene cluster) genes. The results of phylogenetic analyses showed that all A-genome haplotypes of cultivated bananas were grouped together with those of ISEA subspecies of M. acuminata (A-genome). Similarly, the B- and S-genome haplotypes of cultivated bananas clustered with the wild species M. balbisiana (B-genome) and M. schizocarpa (S-genome), respectively. Notably, it has been shown that distinct haplotypes of each cultivar (A-genome group) were nested together to different ISEA subspecies M. acuminata. Analyses of nucleotide polymorphism in the Waxy and Adh1 genes revealed that, in comparison to the wild relatives, cultivated banana exhibited slightly lower nucleotide diversity both across all sites and specifically at silent sites. However, dramatically reduced nucleotide diversity was found at nonsynonymous sites for cultivated bananas. Conclusions/Significance Our study not only confirmed the origin of cultivated banana as arising from multiple intra- and inter-specific hybridization events, but also showed that cultivated banana may have not suffered a severe genetic bottleneck during the domestication process. Importantly, our findings 2. Heat and mass transfer coefficients and modeling of infrared drying of banana slices Directory of Open Access Journals (Sweden) Fernanda Machado Baptestini Full Text Available ABSTRACT Banana is one of the most consumed fruits in the world, having a large part of its production performed in tropical countries. This product possesses a wide range of vitamins and minerals, being an important component of the alimentation worldwide. However, the shelf life of bananas is short, thus requiring procedures to prevent the quality loss and increase the shelf life. One of these procedures widely used is drying. This work aimed to study the infrared drying process of banana slices (cv. Prata and determine the heat and mass transfer coefficients of this process. In addition, effective diffusion coefficient and relationship between ripening stages of banana and drying were obtained. Banana slices at four different ripening stages were dried using a dryer with infrared heating source with four different temperatures (65, 75, 85, and 95 ºC. Midilli model was the one that best represented infrared drying of banana slices. Heat and mass transfer coefficients varied, respectively, between 46.84 and 70.54 W m-2 K-1 and 0.040 to 0.0632 m s-1 for temperature range, at the different ripening stages. Effective diffusion coefficient ranged from 1.96 to 3.59 × 10-15 m² s-1. Activation energy encountered were 16.392, 29.531, 23.194, and 25.206 kJ mol-1 for 2nd, 3rd, 5th, and 7th ripening stages, respectively. Ripening stages did not affect the infrared drying of bananas. 3. Global Transcriptomic Analysis of Targeted Silencing of Two Paralogous ACC Oxidase Genes in Banana Science.gov (United States) Xia, Yan; Kuan, Chi; Chiu, Chien-Hsiang; Chen, Xiao-Jing; Do, Yi-Yin; Huang, Pung-Ling 2016-01-01 Among 18 1-aminocyclopropane-1-carboxylic acid (ACC) oxidase homologous genes existing in the banana genome there are two genes, Mh-ACO1 and Mh-ACO2, that participate in banana fruit ripening. To better understand the physiological functions of Mh-ACO1 and Mh-ACO2, two hairpin-type siRNA expression vectors targeting both the Mh-ACO1 and Mh-ACO2 were constructed and incorporated into the banana genome by Agrobacterium-mediated transformation. The generation of Mh-ACO1 and Mh-ACO2 RNAi transgenic banana plants was confirmed by Southern blot analysis. To gain insights into the functional diversity and complexity between Mh-ACO1 and Mh-ACO2, transcriptome sequencing of banana fruits using the Illumina next-generation sequencer was performed. A total of 32,093,976 reads, assembled into 88,031 unigenes for 123,617 transcripts were obtained. Significantly enriched Gene Oncology (GO) terms and the number of differentially expressed genes (DEGs) with GO annotation were ‘catalytic activity’ (1327, 56.4%), ‘heme binding’ (65, 2.76%), ‘tetrapyrrole binding’ (66, 2.81%), and ‘oxidoreductase activity’ (287, 12.21%). Real-time RT-PCR was further performed with mRNAs from both peel and pulp of banana fruits in Mh-ACO1 and Mh-ACO2 RNAi transgenic plants. The results showed that expression levels of genes related to ethylene signaling in ripening banana fruits were strongly influenced by the expression of genes associated with ethylene biosynthesis. PMID:27681726 4. Production of transgenic banana plants conferring tolerance to salt stress (abstract) International Nuclear Information System (INIS) Ismail, I.A.; Salama, M.; Hamid, A.A.; Sadiq, A.S. 2005-01-01 Production of bananas is limited in areas that have soils with excess sodium. In this study, a transformation system in banana Grand Nain cultivar was established using the apical meristem explant and plasmid pAB6 containing the herbicide-resistant gene (bar) as a selectable marker and gus reporter gene. The micro projectile bombardment transformation system using 650 psi was successfully used for introducing the studied genes in banana explants. The expression of the introduced genes was detected using leaf painting and GUS histochemical tests, respectively. The present results showed that among the selection stage, 36.5% of the bombarded explants survived on the BI3 medium supplemented with 3 mg/L bialaphos, while, 26.6% of the tested explants showed a positive reaction in the GUS assay. To detect the presence of bar and gus genes the PCR was successfully used. These results encourage the idea of possibility of banana crop improvement using in vitro technique through micro projectile bombardment. Therefore, the plasmid pNM1 that carries the bar and P5CS (delta 1 l-pyrroline-5-carboxylate synthetase for proline accumulation) genes was introduced in banana Grand Nain cultivar to produce transgenic plants expressing the salt tolerance gene. Results showed that the majority of herbicide-resistant banana plaptlets were successfully acclimatized. In studying the effects of different salt concentrations on the produced transgenic banana plants, results showed lower decrease in the percentage of survived plants, pseudostem diameter and leaf area with an increase of salt concentrations in case of transgenic plants compared with the controls. (author) 5. Análise do comércio de bananas em Lavras: Minas Gerais Analysis of banana trade in Lavras: Minas Gerais Directory of Open Access Journals (Sweden) Lair Victor Pereira 2009-06-01 Full Text Available A participação de Lavras na oferta de banana no mercado local é muito pequena, considerando-se que o Brasil é o segundo país maior produtor com 6,6 milhões de toneladas e Minas Gerais é o quarto entre os Estados produtores dessa fruta. Visando a quantificar a participação de Lavras e região na oferta de banana no mercado local, realizou-se esse trabalho em duas etapas: 2002/2003 e 2004/2005. A aplicação mensal de questionários nos principais estabelecimentos comerciais de hortifruti e feiras - livre de Lavras, permitiu conhecer o volume comercializado, procedência e perdas das principais cultivares de banana. Os resultados obtidos mostram que em 2002/2003 foram comercializados 945,24 t e em 2004/2005 foi de 1.001,98 t. Desse volume, 6,56% em 2002/2003 e 14,62% em 2004/2005 tiveram como origem Lavras. O consumo per capita anual manteve-se em torno de 11,8 kg nos dois períodos pesquisados. As bananas tipo 'Prata', foram as mais comercializadas nas duas etapas, 54,7% no primeiro período e 58,7% no segundo, sendo que 7,91% e 18,35% , respectivamente, tiveram como origem Lavras. O volume de banana 'Marmelo' e do tipo 'Nanicão', foram de 1,91% e 28,4%, respectivamente, sendo que 84,0% da 'Marmelo' e 3,43% da tipo 'Nanicão' na segunda etapa foram procedentes de Lavras. A banana 'Maçã' teve uma redução de 125,30 t para 107,47 t, correspondendo a 13,26%, sendo que a oferta dessa cultivar, originada de Lavras, manteve-se em 13,8%. As bananas 'Maçã' e 'Marmelo' apresentaram as menores perdas, 3,56% e 4,78% e as dos tipos 'Prata'e 'Nanicão'as maiores perdas, 9,39% e 10, 75%, respectivamente.The participation of Lavras in the banana production offered to the local commerce is still very low considering that Brazil is the second banana producer of the world, with a production around 6.6 ton/year and per-capita consumption of 24.4 kg/year. Minas Gerais ranks in the fifth place among the most important Brazilian state producers. This 6. Wave Engine Topping Cycle Assessment Science.gov (United States) Welch, Gerard E. 1996-01-01 The performance benefits derived by topping a gas turbine engine with a wave engine are assessed. The wave engine is a wave rotor that produces shaft power by exploiting gas dynamic energy exchange and flow turning. The wave engine is added to the baseline turboshaft engine while keeping high-pressure-turbine inlet conditions, compressor pressure ratio, engine mass flow rate, and cooling flow fractions fixed. Related work has focused on topping with pressure-exchangers (i.e., wave rotors that provide pressure gain with zero net shaft power output); however, more energy can be added to a wave-engine-topped cycle leading to greater engine specific-power-enhancement The energy addition occurs at a lower pressure in the wave-engine-topped cycle; thus the specific-fuel-consumption-enhancement effected by ideal wave engine topping is slightly lower than that effected by ideal pressure-exchanger topping. At a component level, however, flow turning affords the wave engine a degree-of-freedom relative to the pressure-exchanger that enables a more efficient match with the baseline engine. In some cases, therefore, the SFC-enhancement by wave engine topping is greater than that by pressure-exchanger topping. An ideal wave-rotor-characteristic is used to identify key wave engine design parameters and to contrast the wave engine and pressure-exchanger topping approaches. An aerodynamic design procedure is described in which wave engine design-point performance levels are computed using a one-dimensional wave rotor model. Wave engines using various wave cycles are considered including two-port cycles with on-rotor combustion (valved-combustors) and reverse-flow and through-flow four-port cycles with heat addition in conventional burners. A through-flow wave cycle design with symmetric blading is used to assess engine performance benefits. The wave-engine-topped turboshaft engine produces 16% more power than does a pressure-exchanger-topped engine under the specified topping 7. Status of the top quark: Top production cross section and top properties Energy Technology Data Exchange (ETDEWEB) Boisvert, V.; /Rochester U. 2006-08-01 This report describes the latest cross section and property measurements associated with the top quark at the Tevatron Run II. The largest data sample used is 760 pb{sup -1} of integrated luminosity. Due to its large mass, the top quark might be involved in the process of electroweak symmetry breaking, making it a useful probe for signs of new physics. 8. New directions in top physics Energy Technology Data Exchange (ETDEWEB) Schell, Torben Karl 2017-02-01 The top quark plays an important role for many aspects of particle physics. The coupling of the Higgs boson to top quarks is a key parameter to probe electroweak symmetry breaking and is important for the evolution of the Higgs potential to high energies. In addition, many models of physics beyond the Standard Model predict heavy particles that decay to top-quark pairs. Furthermore, the unexplained hierarchy of fermion masses culminates in the large top-quark mass. In this thesis, we consider resonance searches based on top quarks in the fully hadronic final state. We employ multivariate techniques in form of boosted decision trees and add several improvements to the original HEPTopTagger algorithm. These modifications and extensions result in the new HEPTopTagger2. The achieved improvements are used to estimate the precision to which the top Yukawa coupling can be measured at a future 100 TeV proton-proton collider in the fully hadronic final state of t anti tH production. We find that at such a collider a precision measurement of the top Yukawa coupling to 1% should be possible. The statistical precision is backed up by demonstrating that in the ratio σ(t anti tH)/σ(t anti tZ) theoretical uncertainties cancel to below-percent level. Finally, we propose a Froggatt-Nielsen-type model to address the hierarchy of fermion masses in the Standard Model and determine current and projected bounds on the available parameter space. 9. Development of VNTR Markers to Assess Genetic Diversity of Mycosphaerella Fijiensis, the Causal Agent of Black Leaf Streak Disease in Bananas (Musa spp.) Science.gov (United States) Mycosphaerella fijiensis is the causal agent of black leaf streak (BLS) disease in bananas. This pathogen threatens global banana production as the main export cultivars are highly susceptible. As a consequence, commercial banana plantations must be protected chemically with fungicides; up to 40 app... 10. 33 CFR 334.540 - Banana River at the Eastern Range, 45th Space Wing, Cape Canaveral Air Force Station, FL... Science.gov (United States) 2010-07-01 ... 33 Navigation and Navigable Waters 3 2010-07-01 2010-07-01 false Banana River at the Eastern Range... AND RESTRICTED AREA REGULATIONS § 334.540 Banana River at the Eastern Range, 45th Space Wing, Cape... navigable waters of the United States, as defined at 33 CFR part 329, within the Banana River contiguous to... 11. SUSCETIBILIDADE VARIETAL DE FRUTOS DE BANANEIRA AO FRIO COLD DAMAGE IN BANANAS Directory of Open Access Journals (Sweden) LUIZ ALBERTO LICHTEMBERG 2001-12-01 Full Text Available No Sul do Brasil, os danos causados pelo frio depreciam a qualidade da banana que permanece no campo durante o outono e inverno, dificultando a sua comercialização. Visando a verificar diferenças entre cultivares quanto à resistência ao frio no campo e em pós-colheita, foram realizados três experimentos em Itajaí-SC. No primeiro, foram avaliados os danos de frio em 13 cultivares do grupo AAA, 7 cultivares do grupo AAB, 6 híbridos do grupo AAAB e 1 cultivar do grupo ABB, em cachos colhidos em outubro de 1997. No segundo experimento, foram avaliados danos de frio em cultivares dos grupos AAA, AAB, ABB e AAAB, em cachos colhidos de 07-05-99 a 27-08-99. No terceiro experimento, foram avaliados danos de frio em bananas de quatro cultivares, armazenadas a 10°C, durante 5, 10 e 20 dias. O genoma B conferiu maior resistência da fruta às baixas temperaturas, tanto a campo quanto na armazenagem. Verificaram-se diferenças quanto a danos de frio tanto entre grupos genômicos, quanto entre cultivares do mesmo grupo. A maior resistência às baixas temperaturas pode permitir o transporte de bananas dos grupos AAB, ABB e AAAB a longas distâncias, em temperaturas inferiores a 12°C.In Southern Brazil, banana bunches that remain in the field during the Fall and Winter are subject to cold damage. Three experiments were carried out in Itajaí, SC, Brazil, with the purpose of investigating the differences among banana cultivars as to their resistance to cold damage. In the first experiment, 13 AAA group cultivars, 7 AAB group cultivars, 6 AAAB hybrids and 1 ABB group cultivar were evaluated as to the level of cold damage. The second experiment evaluated banana bunches of cultivars of AAA, AAB, ABB, and AAAB groups, harvested from May 07/99 to August 27/99. The third experiment examined cold damage to banana clusters stored at 10°C during 5, 10 and 20 days. The B genome appeared to confer the greatest cold resistance to banana fruits, both in the 12. SOIL CHEMICAL ATTRIBUTES AND LEAF NUTRIENTS OF ‘PACOVAN’ BANANA UNDER TWO COVER CROPS Directory of Open Access Journals (Sweden) JOSÉ EGÍDIO FLORI 2016-01-01 Full Text Available Banana is one of the most consumed fruits in the world, which is grown in most tropical countries. The objective of this work was to evaluate the main attributes of soil fertility in a banana crop under two cover crops and two root development locations. The work was conducted in Curaçá, BA, Brazil, between October 2011 and May 2013, using a randomized block design in split plot with five repetitions. Two cover crops were assessed in the plots, the cover 1 consisting of Pueraria phaseoloides, and the cover 2 consisting of a crop mix with Sorghum bicolor, Ricinus communis L., Canavalia ensiformis, Mucuna aterrima and Zea mays, and two soil sampling locations in the subplots, between plants in the banana rows (location 1 and between the banana rows (location 2. There were significant and independent effects for the cover crop and sampling location factors for the variables organic matter, Ca and P, and significant effects for the interaction between cover crops and sampling locations for the variables potassium, magnesium and total exchangeable bases. The cover crop mix and the between-row location presented the highest organic matter content. Potassium was the nutrient with the highest negative variation from the initial content and its leaf content was below the reference value, however not reducing the crop yield. The banana crop associated with crop cover using the crop mix provided greater availability of nutrients in the soil compared to the coverage with tropical kudzu. 13. FRUIT JUICES AS AN ALTERNATIVE TECHNIQUE FOR CONSERVATION OF FRESH-CUT BANANA Directory of Open Access Journals (Sweden) ANDERSON ADRIANO MARTINS MELO 2014-01-01 Full Text Available Browning discoloration after cutting is detrimental for the quality of a number of fruits and vegetables, such as banana, apple, pear, potato, and some roots such as cassava, yam, and others. Browning and softening compromise banana after cut shelf-life in a few hours under cold storage. Therefore, anti-browning compounds have been applied to slices before packing. Some commonly used substances are calcium chloride, ascorbic acid, cysteine and citric acid, in immersed inchemical mixtures. Recent studies have demonstrated the possibility of preserving fresh-cut banana immersed in sweetened fruit juice for relatively longer periods, favoring commercialization. This type of conservation, although widely used in Brazil for fruit salads, consists of a more complex system in a physiological basis, requiring adjustment of the solution parameters, such as sugar concentration, pH and acidity, considering the viability and freshness of the plant tissue. In this short review, we discuss some experimental data and present a new method for preserving fresh-cut banana. Reduction of enzymatic activity, either in temporary dipping treatment or permanent immersion of banana slices is regarded as a key factor for maintaining its quality during cold storage. 14. Kinetic study on ferulic acid production from banana stem waste via mechanical extraction Science.gov (United States) Zainol, Norazwina; Masngut, Nasratun; Khairi Jusup, Muhamad 2018-04-01 Banana is the tropical plants associated with lots of medicinal properties. It has been reported to be a potential source of phenolic compounds such as ferulic acid (FA). FA has excellent antioxidant properties higher than vitamin C and E. FA also have a wide range of biological activities, such as antioxidant activities and anti-microbial activities. This paper presents an experimental and kinetic study on ferulic acid (FA) production from banana stem waste (BSW) via mechanical extraction. The objective of this research is to determine the kinetic parameters in the ferulic acid production. The banana stem waste was randomly collected from the local banana plantation in Felda Lepar Hilir, Pahang. The banana stem juice was mechanically extracted by using sugarcane press machine (KR3176) and further analyzed in high performance liquid chromatography. The differential and integral method was applied to determine the kinetic parameter of the extraction process and the data obtained were fitted into the 0th, 1st and 2nd order of extraction process. Based on the results, the kinetic parameter and R2 value from were 0.05 and 0.93, respectively. It was determined that the 0th kinetic order fitted the reaction processes to best represent the mechanical extraction. 15. Effects of gamma radiation on banana 'nanica' (Musa sp., group AAA) irradiated in pre climacteric phase International Nuclear Information System (INIS) Silva, Simone Faria; Dionisio, Ana Paula; Walder, Julio Marcos Melges 2007-01-01 The present work verified the effect of gamma radiation on physical and chemical parameters of banana 'nanica', analyzing possible alterations on the period of conservation and the possibility of commercial irradiation aiming the exportation. The results had demonstrated that the radiations had not produced effect on pH and total acidity. However, the bananas of the 'control group' and those that had received 0,75 kGy, had presented greater maturation degree and, radiated with 0,30 kGy, had presented greater firmness. In accordance with the results of the organoleptic analysis, can be perceived that the bananas most mature, especially of the 'control group', had had greater acceptance. The bananas of treatments 0,30 and 0,60 kGy had had minors notes for presenting minor maturation stadium. Knowing that the irradiation in adequate dose and fruits of good quality brings benefits to the storage and the process of exportation, we conclude that the dose most appropriate for the control of the maturation of the 'nanica' banana is 0,30 kGy. (author) 16. Image analysis to evaluate the browning degree of banana (Musa spp.) peel. Science.gov (United States) Cho, Jeong-Seok; Lee, Hyeon-Jeong; Park, Jung-Hoon; Sung, Jun-Hyung; Choi, Ji-Young; Moon, Kwang-Deog 2016-03-01 Image analysis was applied to examine banana peel browning. The banana samples were divided into 3 treatment groups: no treatment and normal packaging (Cont); CO2 gas exchange packaging (CO); normal packaging with an ethylene generator (ET). We confirmed that the browning of banana peels developed more quickly in the CO group than the other groups based on sensory test and enzyme assay. The G (green) and CIE L(∗), a(∗), and b(∗) values obtained from the image analysis sharply increased or decreased in the CO group. And these colour values showed high correlation coefficients (>0.9) with the sensory test results. CIE L(∗)a(∗)b(∗) values using a colorimeter also showed high correlation coefficients but comparatively lower than those of image analysis. Based on this analysis, browning of the banana occurred more quickly for CO2 gas exchange packaging, and image analysis can be used to evaluate the browning of banana peels. Copyright © 2015 Elsevier Ltd. All rights reserved. 17. Effects of Green Banana Flour on the Physical, Chemical and Sensory Properties of Ice Cream Science.gov (United States) 2015-01-01 Summary In the present study, possible effects of the addition of banana flour at different mass fractions (1 and 2%) are investigated on physical (overrun, viscosity), chemical (dry matter, fat and ash content, acidity, pH, water and oil holding capacity and colour), mineral content (Ca, K, Na, P, S, Mg, Fe, Mn, Zn and Ni) and sensory properties of ice cream. Fibre--rich banana pieces were found to contain 66.8 g per 100 g of total dietary fibre, 58.6 g per 100 g of which were insoluble dietary fibre, while 8.2 g per 100 g were soluble dietary fibre. It can be concluded from these results that banana is a valuable dietary fibre source which can be used in food production. Flour obtained from green banana pulp and peel was found to have significant (pbanana flour content. Sensory results indicated that ice cream sample containing 2% of green banana pulp flour received the highest score from panellists. PMID:27904363 18. Antioxidant effcacy of unripe banana (Musa acuminata Colla) peel extracts in sunflower oil during accelerated storage. Science.gov (United States) Ling, Stella Sye Chee; Chang, Sui Kiat; Sia, Winne Chiaw Mei; Yim, Hip Seng 2015-01-01 Sunflower oil is prone to oxidation during storage time, leading to production of toxic compounds that might affect human health. Synthetic antioxidants are used to prevent lipid oxidation. Spreading interest in the replacement of synthetic food antioxidants by natural ones has fostered research on fruit and vegetables for new antioxidants. In this study, the efficacy of unripe banana peel extracts (100, 200 and 300 ppm) in stabilizing sunflower oil was tested under accelerated storage (65°C) for a period of 24 days. BHA and α-tocopherol served as comparative standards besides the control. Established parameters such as peroxide value (PV), iodine value (IV), p-anisidine value (p-AnV), total oxidation value (TOTOX), thiobarbituric acid reactive substances (TBARS) and free fatty acid (FFA) content were used to assess the extent of oil deterioration. After 24 days storage at 65°C, sunflower oil containing 200 and 300 ppm extract of unripe banana peel showed significantly lower PV and TOTOX compared to BHA and α-tocopherol. TBARS, p-AnV and FFA values of sunflower oil containing 200 and 300 ppm of unripe banana peel extract exhibited comparable inhibitory effects with BHA. Unripe banana peel extract at 200 and 300 ppm demonstrated inhibitory effect against both primary and secondary oxidation up to 24 days under accelerated storage conditions. Unripe banana peel extract may be used as a potential source of natural antioxidants in the application of food industry to suppress lipid oxidation. 19. Vomiting, abdominal distention and early feeding of banana (Musa paradisiaca) in neonates. Science.gov (United States) Wiryo, Hananto; Hakimi, M; Wahab, A Samik; Soeparto, Pitono 2003-09-01 The objective of this cohort study was to assess the relationship between banana given as early solid food with the symptoms of intestinal obstruction (SIO) among neonates, in a rural community in West Lombok District, West Nusa Tenggara Province, Indonesia. Mothers having newborn infants were interviewed and 3,420 neonates were followed for 28 days. Compared with infants who were not given solid food, the relative risk (RR) for infants given food other than banana as early solid food was 1.87, 95% CI 0.48-8.24, p=0.4, while for infants given banana only as early solid food the RR was 9.15, 95% CI 1.96-42.58, p 0.0005. After adjustment for birthweight, colostrum, and breastfeeding, the odds ratio for infants given banana and the appearance of SIO was 2.99, 95% CI 2.65-5.14; p=0.0012. These data indicate that banana given as early solid food is an important risk factor for the appearance of SIO in neonates. 20. CARCASS YIELD OF BROILER CHICKENS FED BANANA (Musa paradisiaca LEAVES FERMENTED WITH Trichoderma viride Directory of Open Access Journals (Sweden) J. S. Mandey 2016-03-01 Full Text Available The study was conducted to evaluate the effect of level of banana (Musa paradisiaca leaves fermented with Trichoderma viride at different days on the carcass yield of broiler chickens. A hundred and eighty 3-weeks-old broiler chicks were used in this present experiment based on factorial design (3×4. The birds were randomly allocated into three experimental diets containing of 5, 10 and 15% of banana leaves fermented within 0, 5, 10 and 15 days. Each treatment was divided into three replicates of five chicks in each. The experiment was terminated after 4 weeks or when the birds were 7-weeks-old. Feed intake, body weight gain, feed efficiency and carcass yield were measured during the study. The data were subjected to the analysis of variance test followed by least significant difference (LSD test. Results showed that daily feed intake was significantly affected (P˂0.01 by the dietary treatments, in which feed intake was highest in broilers fed diet containing 10% banana leaves fermented for 10 days. The daily weight gain, feed efficiency and carcass yield were significantly affected (P˂0.01 by the treatments, in which the highest values of daily weight gain, feed efficiency, and carcass yield were observed in birds fed diet containing 10% banana leaves fermented for 10 days. It can be concluded that diet containing 10% banana leaves fermented for 10 days can be included in broiler ration without detrimental effects on the performance and carcass yield. 1. In vitro colonic fermentation and glycemic response of different kinds of unripe banana flour. Science.gov (United States) Menezes, Elizabete Wenzel; Dan, Milana C T; Cardenette, Giselli H L; Goñi, Isabel; Bello-Pérez, Luis Arturo; Lajolo, Franco M 2010-12-01 This work aimed to study the in vitro colonic fermentation profile of unavailable carbohydrates of two different kinds of unripe banana flour and to evaluate their postprandial glycemic responses. The unripe banana mass (UBM), obtained from the cooked pulp of unripe bananas (Musa acuminata, Nanicão variety), and the unripe banana starch (UBS), obtained from isolated starch of unripe banana, plantain type (Musa paradisiaca) in natura, were studied. The fermentability of the flours was evaluated by different parameters, using rat inoculum, as well as the glycemic response produced after the ingestion by healthy volunteers. The flours presented high concentration of unavailable carbohydrates, which varied in the content of resistant starch, dietary fiber and indigestible fraction (IF). The in vitro colonic fermentation of the flours was high, 98% for the UBS and 75% for the UBM when expressed by the total amount of SCFA such as acetate, butyrate and propionate in relation to lactulose. The increase in the area under the glycemic curve after ingestion of the flours was 90% lower for the UBS and 40% lower for the UBM than the increase produced after bread intake. These characteristics highlight the potential of UBM and UBS as functional ingredients. However, in vivo studies are necessary in order to evaluate the possible benefit effects of the fermentation on intestinal health. 2. Ethanol production process from banana fruit and its lignocellulosic residues: Energy analysis Energy Technology Data Exchange (ETDEWEB) Velasquez-Arredondo, H.I. [Grupo de Investigacion Bioprocesos y Flujos Reactivos, Universidad Nacional de Colombia, Sede Medellin, Calle 59 A N 63-20 (Colombia); Departamento de Engenharia Mecanica, Escola Politecnica, Universidade de Sao Paulo, Avenida Professor Mello Moraes 2231 (Brazil); Ruiz-Colorado, A.A. [Grupo de Investigacion Bioprocesos y Flujos Reactivos, Universidad Nacional de Colombia, Sede Medellin, Calle 59 A N 63-20 (Colombia); De Oliveira, S. Jr. [Departamento de Engenharia Mecanica, Escola Politecnica, Universidade de Sao Paulo, Avenida Professor Mello Moraes 2231 (Brazil) 2010-07-15 Tropical countries, such as Brazil and Colombia, have the possibility of using agricultural lands for growing biomass to produce bio-fuels such as biodiesel and ethanol. This study applies an energy analysis to the production process of anhydrous ethanol obtained from the hydrolysis of starch and cellulosic and hemicellulosic material present in the banana fruit and its residual biomass. Four different production routes were analyzed: acid hydrolysis of amylaceous material (banana pulp and banana fruit) and enzymatic hydrolysis of lignocellulosic material (flower stalk and banana skin). The analysis considered banana plant cultivation, feedstock transport, hydrolysis, fermentation, distillation, dehydration, residue treatment and utility plant. The best indexes were obtained for amylaceous material for which mass performance varied from 346.5 L/t to 388.7 L/t, Net Energy Value (NEV) ranged from 9.86 MJ/L to 9.94 MJ/L and the energy ratio was 1.9 MJ/MJ. For lignocellulosic materials, the figures were less favorable; mass performance varied from 86.1 to 123.5 L/t, NEV from 5.24 to 8.79 MJ/L and energy ratio from 1.3 to 1.6 MJ/MJ. The analysis showed, however, that both processes can be considered energetically feasible. (author) 3. Phenotypic and molecular characterization of Colletotrichum species associated with anthracnose of banana (Musa spp) in Malaysia. Science.gov (United States) Intan Sakinah, M A; Suzianti, I V; Latiffah, Z 2014-05-09 Anthracnose caused by Colletotrichum species is a common postharvest disease of banana fruit. We investigated and identified Colletotrichum species associated with anthracnose in several local banana cultivars based on morphological characteristics and sequencing of ITS regions and of the β-tubulin gene. Thirty-eight Colletotrichum isolates were encountered in anthracnose lesions of five local banana cultivars, 'berangan', 'mas', 'awak', 'rastali', and 'nangka'. Based on morphological characteristics, 32 isolates were identified as Colletotrichum gloeosporioides and 6 isolates as C. musae. C. gloeosporioides isolates were divided into two morphotypes, with differences in colony color, shape of the conidia and growth rate. Based on ITS regions and β-tubulin sequences, 35 of the isolates were identified as C. gloeosporioides and only 3 isolates as C. musae; the percentage of similarity from BLAST ranged from 95-100% for ITS regions and 97-100% for β-tubulin. C. gloeosporioides isolates were more prevalent compared to C. musae. This is the first record of C. gloeosporioides associated with banana anthracnose in Malaysia. In a phylogenetic analysis of the combined dataset of ITS regions and β-tubulin using a maximum likelihood method, C. gloeosporioides and C. musae isolates were clearly separated into two groups. We concluded that C. gloeosporioides and C. musae isolates are associated with anthracnose in the local banana cultivars and that C. gloeosporioides is more prevalent than C. musae. 4. Improved tolerance toward fungal diseases in transgenic Cavendish banana (Musa spp. AAA group) cv. Grand Nain. Science.gov (United States) Vishnevetsky, Jane; White, Thomas L; Palmateer, Aaron J; Flaishman, Moshe; Cohen, Yuval; Elad, Yigal; Velcheva, Margarita; Hanania, Uri; Sahar, Nachman; Dgani, Oded; Perl, Avihai 2011-02-01 The most devastating disease currently threatening to destroy the banana industry worldwide is undoubtedly Sigatoka Leaf spot disease caused by Mycosphaerella fijiensis. In this study, we developed a transformation system for banana and expressed the endochitinase gene ThEn-42 from Trichoderma harzianum together with the grape stilbene synthase (StSy) gene in transgenic banana plants under the control of the 35S promoter and the inducible PR-10 promoter, respectively. The superoxide dismutase gene Cu,Zn-SOD from tomato, under control of the ubiquitin promoter, was added to this cassette to improve scavenging of free radicals generated during fungal attack. A 4-year field trial demonstrated several transgenic banana lines with improved tolerance to Sigatoka. As the genes conferring Sigatoka tolerance may have a wide range of anti-fungal activities we also inoculated the regenerated banana plants with Botrytis cinerea. The best transgenic lines exhibiting Sigatoka tolerance were also found to have tolerance to B. cinerea in laboratory assays. 5. Genetic resources in Musa bananas and improvement of their disease resistance International Nuclear Information System (INIS) Borges Fuentes, O.L. 1977-01-01 The cultivated bananas belong to the genus Musa and it is the wild species Musa acuminata and Musa balbisiana which contributed to the origin of the sorts used as food. Most of these are triploids and possess a high degree of sterility. The sources of variation that are of importance for genetic improvement of the bananas are (1) hereditary differences between the cultivated clones; (2) difference between species and sub-species; (3) differences between the primitive cultivars derived from Musa acuminata, and (4) mutations that can be artiificially induced. The bananas are attacked by many diseases. Their vulnerability to certain diseases is highly significant in view of the extreme genetic uniformity of the commercial crops and the absence of resistant genes. In the past the wild species and the diploids used as food served as sources of resistance. However, efforts to induce resistance in the cultivated triploid bananas have not been successful. The use of mutagenic agents is proposed as a possible way of improving genetic variability in banana cultivation. (author) 6. Antioxidant activity test on ambonese banana stem sap (Musa parasidiaca var. sapientum Directory of Open Access Journals (Sweden) Hendrik Setia Budi 2015-12-01 Full Text Available Background: Polymorphonuclear cells (PMN release oxygen free radicals or reactive oxygen species (ROS during inflammation. As a result, ROS level is higher than antioxidant level in our body during oxidative stress leading to prolong inflammation or continuous tissue damage. Indonesia, on the other hand, is a country with various herbal medicines. For instance, ambonese banana (Musa parasidiaca var. sapientum is often used as herbal medicine. Ambonese banana, moreover, has flavonoid, polyphenol, tannin, and saponin as antioxidants to reduce free radicals by transferring their hydrogen atom. Medicine used to reduce the impact of free radicals is known as antioxidant. Antioxidant is proved to accelerate wound healing. Purpose: This research aims to analyze the effects of the antioxidant activity of Ambonese banana stem sap extract. Method: Antioxidant activities in this research were examined with 1,1-Diphenyl-2-picryl-hidrazyl (DPPH method by reacting with stable radical compounds. Spectrophotometry with a wavelength of 517 nm was used to measure absorption results shown in purple. The absorption results then were calculated by IC50 reduction activity. Result: There were significant differences of Ambonese banana stem sap antioxidant activity (p50%. Conclusion: Ambonese banana stem sap extract has antioxidant activities. 7. Carbon footprint of premium quality export bananas: case study in Ecuador, the world's largest exporter. Science.gov (United States) Iriarte, Alfredo; Almeida, Maria Gabriela; Villalobos, Pablo 2014-02-15 Nowadays, the new international market demands challenge the food producing countries to include the measurement of the environmental impact generated along the production process for their products. In order to comply with the environmentally responsible market requests the measurement of the greenhouse gas emissions of Ecuadorian agricultural goods has been promoted employing the carbon footprint concept. Ecuador is the largest exporter of bananas in the world. Within this context, this study is a first assessment of the carbon footprint of the Ecuadorian premium export banana (Musa AAA) using a considerable amount of field data. The system boundaries considered from agricultural production to delivery in a European destination port. The data collected over three years permitted identifying the hot spot stages. For the calculation, the CCaLC V3.0 software developed by the University of Manchester is used. The carbon footprint of the Ecuadorian export banana ranged from 0.45 to 1.04 kg CO2-equivalent/kg banana depending on the international overseas transport employed. The principal contributors to the carbon footprint are the on farm production and overseas transport stages. Mitigation and reduction strategies were suggested for the main emission sources in order to achieve sustainable banana production. Copyright © 2013 Elsevier B.V. All rights reserved. 8. Nanocomposites of rice and banana flours blend with montmorillonite: Partial characterization Energy Technology Data Exchange (ETDEWEB) Rodríguez-Marín, María L.; Bello-Pérez, Luis A. [Centro de Desarrollo de Productos Bióticos, Instituto Politécnico Nacional, Km 6 carr Yautepec-Jojutla, Calle Ceprobi No. 8, Colonia San Isidro, Apartado Postal 24, C.P 62731, Yautepec, Morelos (Mexico); Yee-Madeira, Hernani [Departamento de Física, Escuela Superior de Física y Matemáticas-IPN, Edificio 9, U.P., ‘Adolfo López Mateos’ Col. Lindavista, C.P. 07738, México, D. F. (Mexico); Zhong, Qixin [Department of Food science and Technology, the University of Tennessee, Knoxville (United States); González-Soto, Rosalía A., E-mail: [email protected] [Centro de Desarrollo de Productos Bióticos, Instituto Politécnico Nacional, Km 6 carr Yautepec-Jojutla, Calle Ceprobi No. 8, Colonia San Isidro, Apartado Postal 24, C.P 62731, Yautepec, Morelos (Mexico) 2013-10-15 Rice and banana flours are inexpensive starchy materials that can form films with more improved properties than those made with their starch because flour and starch present different hydrophobicity. Montmorillonite (MMT) can be used to further improve the properties of starch-based films, which has not received much research attention for starchy flours. The aim of this work was to evaluate the mechanical and barrier properties of nanocomposite films of banana and rice flours as matrix material with addition of MMT as a nanofiller. MMT was modified using citric acid to produce intercalated structures, as verified by the X-ray diffraction pattern. The intercalated MMT was blended with flour slurries, and films were prepared by casting. Nanocomposite films of banana and rice flours presented an increase in the tensile at break and elongation percentage, respectively, more than their respective control films without MMT. This study showed that banana and rice flours could be alternative raw materials to use in making nanocomposite films. - Highlights: • Flour films presented adequate mechanical and barrier properties. • Addition of montmorillonite modified the mechanical and barrier properties of flour films. • The mechanical properties of the films were influenced by the different components of the flours. • The fiber of the banana flour improved the mechanical properties of the films. 9. Nanocomposites of rice and banana flours blend with montmorillonite: Partial characterization International Nuclear Information System (INIS) Rodríguez-Marín, María L.; Bello-Pérez, Luis A.; Yee-Madeira, Hernani; Zhong, Qixin; González-Soto, Rosalía A. 2013-01-01 Rice and banana flours are inexpensive starchy materials that can form films with more improved properties than those made with their starch because flour and starch present different hydrophobicity. Montmorillonite (MMT) can be used to further improve the properties of starch-based films, which has not received much research attention for starchy flours. The aim of this work was to evaluate the mechanical and barrier properties of nanocomposite films of banana and rice flours as matrix material with addition of MMT as a nanofiller. MMT was modified using citric acid to produce intercalated structures, as verified by the X-ray diffraction pattern. The intercalated MMT was blended with flour slurries, and films were prepared by casting. Nanocomposite films of banana and rice flours presented an increase in the tensile at break and elongation percentage, respectively, more than their respective control films without MMT. This study showed that banana and rice flours could be alternative raw materials to use in making nanocomposite films. - Highlights: • Flour films presented adequate mechanical and barrier properties. • Addition of montmorillonite modified the mechanical and barrier properties of flour films. • The mechanical properties of the films were influenced by the different components of the flours. • The fiber of the banana flour improved the mechanical properties of the films 10. TRADE ENHANCEMENT CHARACTERISTICS OF DESSERT BANANA FRUITS AND ESTIMATES OF TRANSACTION COSTS IN OKIGWE METROPOLIS, IMO STATE NIGERIA Directory of Open Access Journals (Sweden) C. Ogbonna Emerole 2013-07-01 Full Text Available This study on trade enhancement Characteristics of sweet (dessert banana fruit and estimation of transaction costs was conducted in Okigwe Metropolis of Imo State, Nigeria. Stratified random sampling technique was adopted in selecting 80 respondents comprising 40 dessert banana traders (panelists and 40 dessert banana consumers. Monthly trade data was collected from the respondents using pretested semi-structured questionnaire during dry season (November-April and rain season (May-October for the year 2012. Data collected were subjected to descriptive statistical analysis; with transaction costs estimated as ex ante and ex post components. Hedonic pricing regression model was used in determining buyer socioeconomic/banana attributes that influenced willingness to pay price. Fruit characteristics that significantly enhanced trade of sweet banana in descending order were taste (3.83, fruit variety (3.57, and fruit skin colour (3.50. Other significant factors were level of ripeness (3.49, availability in off-season (3.46, fruit size (3.20 and cleanliness (3.20. Mean ex-ante transaction costs for sweet banana was N77, 800.00/trader and its mean ex-post transaction cost was N25,080.00/trader. We recommended that traders should take advantage of Global Mobile System (GSM to overcome information barriers on banana trading. Government and health institutions should intensify consumer safety education, and encourage horticultural unions to heighten postharvest monitoring of stored and displayed dessert banana fruits to enforce observance of ripening standards. 11. Enhancing dissemination of Beauveria bassiana with host plant base incision trapfor the management of the banana weevil Cosmopolites sordidus NARCIS (Netherlands) Tinzaara, W.; Emudong, P.; Nankinga, C.; Tushemereirwe, W.; Kagezi, G.H.; Gold, C.S.; Dicke, M.; Huis, van A.; Karamura, E. 2015-01-01 The banana weevil, Cosmopolites sordidus (Germar) (Coleoptera: Curculionidae) is an important pest of highland banana in East and central Africa. It causes yield loss of up to 100% in heavily infested fields. Studies were carried out in Uganda to evaluate the efficacy of the the plant base incision 12. Endophytic control of Cosmopolites sordidus and Radopholus similis using Fusarium oxysporum V5w2 in tissue culture banana NARCIS (Netherlands) Ochieno, D.M.W. 2010-01-01 Banana plants are being inoculated with Fusarium oxysporum V5w2 and Beauveria bassiana G41 for endophytic control of pests. The effects of F. oxysporum V5w2 and B. bassiana G41, soil sterility, fertilizer, and mulching, on Cosmopolites sordidus and Radopholus similis in banana plants, are 13. Identification and evaluation of two diagnostic markers linked to Fusarium wilt resistance (race 4) in banana (Musa spp.). Science.gov (United States) Wang, Wei; Hu, Yulin; Sun, Dequan; Staehelin, Christian; Xin, Dawei; Xie, Jianghui 2012-01-01 Fusarium wilt caused by the fungus Fusarium oxysporum f. sp. cubense race 4 (FOC4) results in vascular tissue damage and ultimately death of banana (Musa spp.) plants. Somaclonal variants of in vitro micropropagated banana can hamper success in propagation of genotypes resistant to FOC4. Early identification of FOC4 resistance in micropropagated banana plantlets is difficult, however. In this study, we identified sequence-characterized amplified region (SCAR) markers of banana associated with resistance to FOC4. Using pooled DNA from resistant or susceptible genotypes and 500 arbitrary 10-mer oligonucleotide primers, 24 random amplified polymorphic DNA (RAPD) products were identified. Two of these RAPD markers were successfully converted to SCAR markers, called ScaU1001 (GenBank accession number HQ613949) and ScaS0901 (GenBank accession number HQ613950). ScaS0901 and ScaU1001 could be amplified in FOC4-resistant banana genotypes ("Williams 8818-1" and Goldfinger), but not in five tested banana cultivars susceptible to FOC4. The two SCAR markers were then used to identify a somaclonal variant of the genotype "Williams 8818-1", which lost resistance to FOC4. Hence, the identified SCAR markers can be applied for a rapid quality control of FOC4-resistant banana plantlets immediately after the in vitro micropropagation stage. Furthermore, ScaU1001 and ScaS0901 will facilitate marker-assisted selection of new banana cultivars resistant to FOC4. 14. Effect of 1-Methylcyclopropene coupled with controlled atmosphere storage on the ripening and quality of ‘Cavendish’ bananas Science.gov (United States) Fresh-fruit banana is well known to have a short-life after harvest. A short pre-pilot study was carried out to test the effect of atmospheric condition exposure to 1-MCP on the quality, limited to cosmetic and peel appearance, and shelf life of fresh-fruit bananas. Low level of O2 (3 kPa) and high ... 15. Production and Characterization of a Distilled Alcoholic Beverage Obtained by Fermentation of Banana Waste (Musa cavendishii from Selected Yeast Directory of Open Access Journals (Sweden) Mara Eli de Matos 2017-11-01 Full Text Available Banana is one of the most important fruits in the Brazilian diet and is mainly consumed naturally. Losses from crop to final consumer are high and estimated in about 30%. The aim of this work was to elaborate a distilled alcoholic beverage from discarded banana and to compare with commercial trademarks. Initially, yeast strains were isolated from banana fruit and characterized by their production of volatile aroma compounds. The highest aroma-producing yeast isolate was identified by ITS-rRNA gene sequencing as Pichia kluyveri. Pasteurized banana pulp and peel was fermented by the selected P. kluyveri at approximately 107 cells/mL. The sugars were converted quickly, and a high ethanol concentration (413 mg/L was achieved after 24 h of fermentation. The fermented banana must was distilled in a Femel Alambic, and the head, heart and tail fractions were collected. The banana brandy produced had highest concentration of volatile compounds compared to trademarks, such as isoamyl acetate (13.5 mg/L, ethyl hexanoate (0.8 mg/L and others. The results showed that whole banana must could be a good substrate for fermentation and distillation, and the sensory analysis performed revealed that the produced beverage had good acceptance by the tasters. This study demonstrates the potential of banana as a possible alternative to reduce waste and increase income to farmers. 16. Hyperspectral Surface Analysis for Ripeness Estimation and Quick UV-C Surface Treatments for Preservation of Bananas Science.gov (United States) Zhao, W.; Yang, Zh.; Chen, Zh.; Liu, J.; Wang, W. Ch.; Zheng, W. Yu. 2016-05-01 This study aimed to determine the ripeness of bananas using hyperspectral surface analysis and how a rapid UV-C (ultraviolet-C light) surface treatment could reduce decay. The surface of the banana fruit and its stages of maturity were studied using a hyperspectral imaging technique in the visible and near infrared (370-1000 nm) regions. The vselected color ratios from these spectral images were used for classifying the whole banana into immature, ripe, half-ripe and overripe stages. By using a BP neural network, models based on the wavelengths were developed to predict quality attributes. The mean discrimination rate was 98.17%. The surface of the fresh bananas was treated with UV-C at dosages from 15-55 μW/cm2. The visual qualities with or without UV-C treatment were compared using the image, the chromatic aberration test, the firmness test and the area of black spot on the banana skin. The results showed that high dosages of UV-C damaged the banana skin, while low dosages were more efficient at delaying changes in the relative brightness of the skin. The maximum UV-C treatment dose for satisfactory banana preservation was between 21 and 24 μW/cm2. These results could help to improve the visual quality of bananas and to classify their ripeness more easily. 17. Radiosensitivity of in vitro Cultured Banana Shoot-Tips International Nuclear Information System (INIS) Elagamawy, M.R. 2002-01-01 Longitudinally dissected shoot apices of Grand Nain, Williams and Maghrabi banana cultivars were exposed to gamma irradiation with Cobalt 60 source at the doses of 0, 20, 40 and 60 Gy and immediately placed into proliferation medium. A number of micropropagation cycles after irradiation were necessary up to M1 V2 to M1 V4 stage to let mutated sectors developing into non-chimeric shoots. Radiosensitivity was evaluated by the rate of shoot proliferation and by the shoot fresh weight increase. Increasing gamma doses caused reduction in survival rates and average number of shoots. Grand Nain exhibited the highest multiplication- rate (3.1). The lower dose (20 Gy) enhanced shoot multiplication ratio specially in Williams and Maghrabi, which however decreased with increased doses. The doses of 20-40 Gy yielded Ld50, with sensible degree of shoot multiplication, which occurred hardly ever beyond 40 Gy. The dose of 60 Gy resulted 80% lethal shoot growth. Linear decrease in fresh weight was observed in post-irradiation recovery, notably in the Maghrabi. In contradictory vulnerable damage was observed in Williams which showed the highest fresh weight value. Shoot proliferation appeared generally on the surface of the corm. Root formation was observed without additional hormone. The roots were dark colored and was decreased with the increased of dosage 18. Some factors affecting the in vitro culture of banana International Nuclear Information System (INIS) Zadi, T.A.N.; Khan, N.H.; Rehman, Z.U. 2006-01-01 Factors affecting in vitro regeneration of shoots in shoot tip explant cultures of banana cultivar 'Basrai', such as solid and liquid media, growth regulators, vitamins, and antioxidants were studied. Three-quarters strength of MS liquid medium supplemented with 17.75 micro m 6-benzyladenine (BA), 11.42 micro M indole-3-acetic acid (IAA) and 205 micro M adenine sulphate induced the formation of mean number of 12.3 shoots, with the mean length of 3.0 cm, after three weeks of culture. Maximum shoot multiplication (14.33) occurred in liquid medium containing 22.19 micro M BA. Addition of 2.0% activated charcoal (AC) to the liquid medium improved quality of the regenerated plants with expanded and glossy leaves, though the number of shoots was reduced (13.66). Profuse formation of roots was characteristically induced by AC. Addition of citric acid (CA) to the medium caused decline in morphogenetic expression of the cultures. (author) 19. A sarabande of tropical fruit proteomics: Avocado, banana, and mango. Science.gov (United States) Righetti, Pier Giorgio; Esteve, Clara; D'Amato, Alfonsina; Fasoli, Elisa; Luisa Marina, María; Concepción García, María 2015-05-01 The present review highlights the progress made in plant proteomics via the introduction of combinatorial peptide ligand libraries (CPLL) for detecting low-abundance species. Thanks to a novel approach to the CPLL methodology, namely, that of performing the capture both under native and denaturing conditions, identifying plant species in the order of thousands, rather than hundreds, is now possible. We report here data on a trio of tropical fruits, namely, banana, avocado, and mango. The first two are classified as "recalcitrant" tissues since minute amounts of proteins (in the order of 1%) are embedded on a very large matrix of plant-specific material (e.g., polysaccharides and other plant polymers). Yet, even under these adverse conditions we could report, in a single sweep, from 1000 to 3000 unique gene products. In the case of mango the investigation has been extended to the peel too, since this skin is popularly used to flavor dishes in Far East cuisine. Even in this tough peel 330 proteins could be identified, whereas in soft peels, such as in lemons, one thousand unique species could be detected. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim. 20. Banana peel extract mediated synthesis of gold nanoparticles. Science.gov (United States) Bankar, Ashok; Joshi, Bhagyashree; Kumar, Ameeta Ravi; Zinjarde, Smita 2010-10-01 Gold nanoparticles were synthesized by using banana peel extract (BPE) as a simple, non-toxic, eco-friendly 'green material'. The boiled, crushed, acetone precipitated, air-dried peel powder was used to reduce chloroauric acid. A variety of nanoparticles were formed when the reaction conditions were altered with respect to pH, BPE content, chloroauric acid concentration and temperature of incubation. The reaction mixtures displayed vivid colors and UV-vis spectra characteristic of gold nanoparticles. Dynamic light scattering (DLS) studies revealed that the average size of the nanoparticles under standard synthetic conditions was around 300nm. Scanning electron microscopy and energy dispersive spectrometry (EDS) confirmed these results. A coffee ring phenomenon, led to the aggregation of the nanoparticles into microcubes and microwire networks towards the periphery of the air-dried samples. X-ray diffraction studies of the samples revealed spectra that were characteristic for gold. Fourier transform infra red (FTIR) spectroscopy indicated the involvement of carboxyl, amine and hydroxyl groups in the synthetic process. The BPE mediated nanoparticles displayed efficient antimicrobial activity towards most of the tested fungal and bacterial cultures. 1. Improvement of banana through biotechnology and mutation breeding International Nuclear Information System (INIS) Rao, P.S.; Ganapathi, T.R.; Bapat, V.A.; Kulkarni, V.M.; Suprasanna, P. 1998-01-01 Protocols were standardized for in vitro propagation of several elite and diverse banana accessions using shoot tip explants. Tissue culture raised plants were field planted at multiple locations. Studies were undertaken for the induction of mutations using multiple shoot cultures of six selected cultivars, Shreemanti (AAA), Basrai (AAA), Lal Kela (AAA), Rasthali (AAB), Karibale Monthan (ABB) and a wild diploid (BB). These shoot cultures were irradiated at different doses of gamma rays (0-100 Gy) and subcultured thrice (up to M 1 V 3 ) to separate shimeras, followed by induction of rooting (M 1 V 4 ). In general, the rate of multiplication had a negative association with the dose of gamma rays. Enhanced multiplication of shoots was noticed at lower doses. The proliferation of shoots was arrested beyond 50 Gy and a dose of 70 Gy was completely lethal for all the genotypes studied. The rooted plantlets were hardened in the green house and in the early stages of field growth, a few cholorophyll and morphological variants have been noticed. Preliminary studies have been made with DNA samples of different varieties and variants for DNA quality and restriction digestion. Studies are underway to characterize these using PCR based methods. (author) 2. Experimental Investigations on Thermal Conductivity of Fenugreek and Banana Composites Science.gov (United States) Pujari, Satish; Venkatesh, Talari; Seeli, Hepsiba 2018-04-01 The use of composite materials in manufacturing has significantly increased in the past decade. Research is being done to identify natural fibers that can be used as composites. Several natural fibers are already being used in the industry as composites. The appealing advantages of using natural fibers are reflected in lower density when compared to synthetic fibers and also in saving costs. This research paper highlights the experiment that analyses the use of biodegradable fenugreek composite as natural fiber and concludes that fenugreek natural fibers are an excellent substitute to the synthetic fibers in terms of reinforcement properties for the polymers. These fenugreek fibers are naturally sourced, renewable, cost effective and bio-friendly. In thermal energy storage systems as well as in air conditioning systems, thermal insulators are predominantly used to enhance the storage properties. An experiment was created to investigate the thermal properties of fenugreek banana composites for different fiber concentrations. The experimental results showed that the thermal conductivity of the composites decrease with an increase in the fiber content. The experimental results were compared with the theoretical models to describe the variation of thermal conductivity with the volume fraction of the fiber. Good agreement between theoretical and experimental results was observed. 3. Results of Ageing Tests on the Forward MSGC 'BANANA' Prototype CERN Document Server Boimska, B; Capéans-Garrido, M; Claes, S; Hoch, Michael; Million, Gilbert; Ropelewski, Leszek; Sharma, Abhishek; Shekhtman, L I; Van Doninck, Walter; Van Lancker, Luc 1996-01-01 The forward tracker of the CMS experiment at CERN will consist of a large number of MicroStrip Gas Chambers ( MSGCs), expected to operate for several years in the high luminosity environment of LHC with an accumulated dose rate of ~ 10 mC/yr per cm of strip. They are planned to be arranged around the interaction point in modular wheels which contain several MSGCs. The 'open'option of a forward CMS 'banana' module includes the electronics inside the gas volume. Long term tests of one such prototype containing two plates, one with gold and the other with chromium strips on diamond-like coated D263 glass with representative electronics and surface mount components inside the gas volume have been performed; the satisfactory results are reported here. A first test has been made with a gas flow rate of 60 cc/min in the prototype ( corresponding to 5 gas renewals per hour) and has shown no gain drop up to 85 mC/cm of accumulated charge per strip. A second test has been performed with a gas flow rate of 6 cc/min corr... 4. ECONOMIC VIABILITY OF NUTRIENT MANAGEMENT PRACTICES FOR BANANA CROP VIABILIDADE ECONÔMICA DE MANEJOS NUTRICIONAIS NA CULTURA DE BANANA Directory of Open Access Journals (Sweden) Maura Seiko Tsutsui Esperancini 2011-04-01 Full Text Available Proper fertilizing management, in order to optimize fruit quality and yield, is a relevant stage on the production process to the rural entrepreneur profitability. So, the aim of this study was to analyze the economic feasibility of five nutrient management practices for banana crop, Cavendish cultivars, in the Médio Paranapanema region, São Paulo State, Brazil, in 2009/2010. The effective operational cost (EOC and total operational cost (TOC structures and three profitability indicators were used. Significant differences were observed among the management systems, and the system that resulted in major economic advantage to the producer provided an average profit rate 25.6% higher than other treatments, with total net revenue about 29.5% higher than other management types. The unitary cost was lower for the most profitable nutrient management practice, although the yield was 9.5% lower than the management system that presented the highest yield levels. KEY-WORDS: Musa sp.; banana crop; fertilization; production cost; profitability. O manejo adequado da adubação, visando à otimização da produtividade e qualidade dos frutos, é uma etapa representativa do processo produtivo para rentabilidade do empreendedor rural. Neste contexto, objetivou-se analisar a viabilidade econômica de cinco tipos de manejo nutricional de bananeiras de cultivares do subgrupo Cavendish, na região do Médio Paranapanema (SP, em 2009/2010. Foram utilizadas estruturas do custo operacional efetivo (COE e custo operacional total (COT e três indicadores de rentabilidade. Diferenças significativas foram observadas entre os manejos, sendo que o sistema de manejo que resultou em maior vantagem econômica para o produtor proporcionou índice médio de lucratividade 25,6% superior aos demais 5. Top quark properties at ATLAS CERN Document Server Dilip, Jana 2008-01-01 The ATLAS potential for the study of the top quark properties and physics beyond the Standard Model in the top quark sector, is described. The measurements of the top quark charge, the spin and spin correlations, the Standard Model decay (t-> bW), rare top quark decays associated to flavour changing neutral currents (t-> qX with X = gluon, Z, photon) and ttbar resonances are discussed. The sensitivity of the ATLAS experiment is estimated for an expected luminosity of 1fb-1 at the LHC. The full simulation of the ATLAS detector is used. For the Standard Model measurements the expected precision is presented. For the tests of physics beyond the Standard Model, the 5 sigma discovery potential (in the presence of a signal) and the 95% Confidence Level (CL) limit (in the absence of a signal) are given. 6. Võrumaa ettevõtete TOP 50 aastal 2003 Index Scriptorium Estoniae 2004-01-01 Võrumaa ettevõtete TOP 50 aastal 2003. Käibe TOP 40. Kasumi TOP 40. Käibe kasvu TOP 20. Kasumi kasvu TOP 20. Rentaabluse TOP 20. Omakapitali tootlikkuse TOP 20. Võrumaa firmade üld- ja finantsandmed 7. Uncovering the single top: observation of electroweak top quark production Energy Technology Data Exchange (ETDEWEB) Benitez, Jorge Armando [Michigan State Univ., East Lansing, MI (United States) 2009-01-01 The top quark is generally produced in quark and anti-quark pairs. However, the Standard Model also predicts the production of only one top quark which is mediated by the electroweak interaction, known as 'Single Top'. Single Top quark production is important because it provides a unique and direct way to measure the CKM matrix element Vtb, and can be used to explore physics possibilities beyond the Standard Model predictions. This dissertation presents the results of the observation of Single Top using 2.3 fb-1 of Data collected with the D0 detector at the Fermilab Tevatron collider. The analysis includes the Single Top muon+jets and electron+jets final states and employs Boosted Decision Tress as a method to separate the signal from the background. The resulting Single Top cross section measurement is: (1) σ(p\\bar{p}$→ tb + X, tqb + X) = 3.74-0.74+0.95 pb, where the errors include both statistical and systematic uncertainties. The probability to measure a cross section at this value or higher in the absence of signal is p = 1.9 x 10-6. This corresponds to a standard deviation Gaussian equivalence of 4.6. When combining this result with two other analysis methods, the resulting cross section measurement is: (2) σ(p$\\bar{p}$→ tb + X, tqb + X) = 3.94 ± 0.88 pb, and the corresponding measurement significance is 5.0 standard deviations. 8. Search for Top in TOTEM CERN Document Server Corrales, Alonso 2017-01-01 The work assigned to me during this summer (from June 19th to August 11th) was in the "Search for Top in TOTEM" project, a CMG project in collaboration with TOTEM, under the supervision of Martijn Mulders and Laurent Forthomme. The goal is to determine physical properties of the top anti top quark pair (tt) created by central exclusive processes, particularly the invariant mass. For that it is necessary to relate the measurements of the reconstructed top pair kinematics in the central detector of CMS with the detected protons in the Roman Pots of TOTEM just outside the interaction site, about 200 m from it. CMS only considers decay products from the proton collision, it does not regard the forward protons that participated in the collision. On the contrary, TOTEM does consider these protons. In this research measurements of both TOTEM and CMS were combined Figure 1 shows a diagram of the production of the top anti top pair from the interaction of the two protons. A central exclusive process occurs when the pr... 9. Increasing sustainability through the use of organic matters/manures in banana production. cv. harichal International Nuclear Information System (INIS) Miano, T.F.; Baloch, M.A.; Baloch, A.F.; Miano, T.F. 2005-01-01 A banana experiment was carried out with cv. Harichal under the ecological conditions of Tando Jam to study the effect of organic manures/matter on the growth and bunch weight (yield) of banana. The treatments applied were ; FYM, Dry leaves, Stalk of the banana bunch and control with constant doses of NPK (136g + 57g + 148g per plant). Minimum days (490.33) from planting to harvest were observed under the treatment of FYM followed by stalk of the bunch and dried leaves. The highest single fruit weight (107 g), fruit length( 18.30 cm) bunch weight (25.46 kg) and calculated yield per hectare (33.80 tons) were observed under FYM with NPK fertilizer followed by stalk of the bunch and dried leaves. (author) 10. Growth Performance of Pekin Ducks Fed with Golden Snail and Fresh Banana Peelings Directory of Open Access Journals (Sweden) Ulep, LJL. 1995-01-01 Full Text Available The growth performance and economics of feeding confined Pekin ducks with three different levels of golden snail fresh meat and banana peelings in equal percentage for replacing 50 %, 70 % or 90 % of the commercial feed of the diet was studied. Body weight gains and feed consumption of ducks, cost of feed and profit above feed and stock cost different significantly among treatments. Feed conversion varied during the first month of feeding but became comparable after the second month. Ducks fed the diet with 45 % banana peel and 45 % golden snail meat gave the best performance, were the most economical and yielded the highest profit. Snail meat and banana peeling utilization as replacement to commercial diet for ducks is advantageaous in terms of growth performance and cost benefit. 11. Fungi associated with fruit crown rot in organic banana (Musa spp. L. in Piura, Peru Directory of Open Access Journals (Sweden) René Aguilar Anccota 2013-05-01 Full Text Available The department of Piura is the principal banana-producing zone in Peru, sharing 87% of exportations. In this zone, one of the most important postharvest diseases is crown rot. The economic loses attributed to this disease are estimated to be between 25 and 30% of organic bananas exported. The objective of this study was to identify the causal agents associated with this disease. Samples taken refrigerated fruit from the areas of Querecotillo, Salitral and Mallares were taken and selected after the fact. Thielaviopsis paradoxa, Lasiodiplodia theobromae, Colletotrichum musae and Fusarium verticilloides. In order to demonstrate the pathogenicity of the isolated species, inoculations were given in the area of the crown of the fruit on healthy bananas. These fungi caused symptoms of infection in different proportions, concluding that crown rot is a disease with a complex etiology. 12. The Effect of Consuming Ambon Banana (Musa paradisiaca Var. Sapientum on Sleep Latency of Elderly Hypertension Directory of Open Access Journals (Sweden) Selvi Ria Ristania 2017-08-01 Full Text Available Elderly hypertension often reported that their latency elongated was less compared to healthy elderly. The increase of latency impact on health, it causes susceptibility to illness, stress, confusion, disorientation, mood disorders, less fresh, decrease ability to make decisions. The aim of this research was to explain the effect of consuming Ambon banana on sleep latency of elderly hypertension in Mulyorejo, Surabaya. Time series one group pre-test post test design was used in this research. Affordable population in this research was elderly hypertension in RW 2 and RW 3 Mulyorejo Surabaya. Total sample was 15 respondents and taken by total sampling technique. The independent variable was consuming Ambon banana, and dependent variable was sleep latency of elderly hypertension. Every day the latency and blood pressure on elderly was monitored. Data was collected using questionnaire, and then analyzed using Wilcoxon Signed Rank Test. The result showed that consuming Ambon banana affect sleep latency (p=0.009. 13. BANANAS: providing child care services to a multi-ethnic community. Science.gov (United States) Vu, Catherine M; Schwartz, Sara L; Austin, Michael J 2011-01-01 BANANAS, Inc. is a nonprofit organization that has provided child care resource and referral services for over 35 years. BANANAS emerged as a grassroots effort initiated by a group of female volunteers who sought to build a network of women with children who needed childcare. As the organization developed, its leaders recognized and responded to additional needs, including resource and information sharing, workshops and classes, and political advocacy. Beginning as a collective, BANANAS has grown into a multifaceted service delivery and advocacy nonprofit operating with an annual budget of$12 million. This history of the agency reflects the development of a unique community-based effort, its challenges and rewards, and the multiple successes that this pioneering nonprofit has experienced.
14. Combined effects of pectic enzymes on the degradation of pectin polysaccharides of banana fruit
International Nuclear Information System (INIS)
Jheng, G.; Jiang, Y.; Ghen, Y.; Yang, S.
2011-01-01
Pectin polysaccharide is one of the major components of the primary cellular wall in the middle lamella of plant tissues. The degradation of pectin polysaccharide contributes to fruit softening. In this study, water-soluble pectin (WSP) and acid-soluble pectin (ASP) were isolated from pulp tissues of banana fruit at various ripening stages, and combinations of the enzymes such as polygalcturonase (PG), pectin methylesterase (PME) and beta-galactosidase (beta-Gal) were used to investigate the effect on the degradation of WSP and ASP. PG promoted the degradation of pectin polysaccharides, especially in ASP. An enhanced effect of the degradation of WSP and ASP from various ripening banana fruit was observed in the presence of PME. In addition, beta-Gal accelerated slightly the degradation of WSP and ASP in the presence of PG. Overall, PG, PME and beta-Gal can coordinate to promote the degradation of pectin polysaccharides of banana fruit, resulting in fruit softening. (author)
15. Banana fluxes in the plateau regime for a nonaxisymmetrically confined plasma
International Nuclear Information System (INIS)
Balescu, R.; Fantechi, S.
1990-01-01
The banana (or banana-plateau) fluxes, related to the generalized stresses left-angle B·∇·π α(n) right-angle, left-angle B T ·∇·π α(n) right-angle have been determined in the plateau regime, for a plasma confined by a toroidal magnetic field of arbitrary geometry. The complete set of transport coefficients for both the ''parallel'' (ambipolar) and ''toroidal'' (nonambipolar) banana fluxes was obtained in the 13-moment (13M) approximation, going beyond the previously known expressions in the nonaxisymmetric case. The main emphasis is laid on the structure of the transport matrix and of its coefficients. It is shown that the Onsager symmetry of this matrix partly breaks down (for the mixed electron--ion coefficients) in a nonaxisymmetrically confined plasma
16. Suitability of combination of calcium propionate and chitosan for preserving minimally processed banana quality.
Science.gov (United States)
Mirshekari, Amin; Madani, Babak; Golding, John B
2017-08-01
The marketability of fresh-cut banana slices is limited by the rapid rate of fruit softening and browning. However, there is no scientific literature available about the role of postharvest calcium propionate and chitosan treatment on the quality attributes of fresh-cut banana. Therefore, the aim of the present study was to investigate these effects. The application of calcium propionate plus chitosan (CaP+Chit) retained higher firmness, higher ascorbic acid content, higher total antioxidant activity and higher total phenolic compounds, along with lower browning, lower polyphenol oxidase, lower peroxidase, lower polygalacturonase and lower pectin methyl esterase activities and microbial growth, compared to control banana slices after 5 days of cold storage. The results of the present study show that CaP+Chit could be used to slow the loss of quality at the same time as maintaining quality and inhibiting microbial loads. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry.
17. Process Engineering App lied to the Production of Bioethanol Using Banana Rejection Urabá
Directory of Open Access Journals (Sweden)
2012-06-01
Full Text Available Most countries either financially or in advanced stages of development are faced with the problem of disposal and treatment of waste and organic waste; these can be treated in different ways, for example by reducing its volume or processing any useful substance using physicochemical transformation processesbananas in bioethanol and analyzing environmental impacts to meet sanitary standards.The objective of this study was to evaluate the potential for ethanol production from banana and the study and application of process engineering in ethanol production using banana rejection as feedstock in the region of Urabá.Bioethanol is obtained by fermentation and distillation of rejected bananas where the results are reflected in the operational controls and leaf pattern obtained in the standardization process.
18. Optimization of process parameter for graft copolymerization of glycidyl methacrylate onto delignified banana fibers
International Nuclear Information System (INIS)
Selambakkannu, S.; Nor Azillah Fatimah Othman; Siti Fatahiyah Mohamad
2016-01-01
This paper focused on pre-treated banana fibers as a trunk polymer for optimization of radiation-induced graft copolymerization process parameters. Pre-treated banana fiber was grafted with glycidyl methacrylate (GMA) via electron beam irradiation. Optimization of grafting parameters in term of grafting yield was analyzed at numerous radiation dose, monomer concentration and reaction time. Grafting yield had been calculated gravimetrically against all the process parameters. The grafting yield at 40 kGy had increases from 14 % to 22.5 % at 1 h and 24 h of reaction time respectively. Grafting yield at 1 % of GMA was about 58 % and it increases to 187 % at 3 % GMA. The grafting of GMA onto pre-treated banana fibers confirmed with the characterization using FTIR, SEM and TGA. Grafting of GMA onto pre-treated fibers was successfully carried out and it was confirmed by the results obtained via the characterization. (author)
19. Banana peel: a green and economical sorbent for Cr(III) removal
International Nuclear Information System (INIS)
Memon, J.R.; Bhanger, M.I.; Memon, S.Q.
2008-01-01
Banana peel, a common fruit waste has been investigated to remove and preconcentrate Cr(III) from industrial wastewater. It was characterized by FT-IR spectroscopy. The parameters pH, contact time, initial metal ion concentration and temperature were investigated and the maximum sorption was found to be 95%. The binding of metal ions was found to be pH dependent with the optimal sorption occurring at pH 4. The retained species were eluted using 5 ml of 2 M HNO/sub 3/. The mechanism for the binding of Cr(III) on the banana peel surface was also studied in detail. The Langmuir and Dubinin-Radushkevich (D-R) isotherms were used to describe the partitioning behavior for the system at different temperatures. Kinetic and thermodynamic measurements of the banana peel for chromium ions were also studied. The method was applied for the removal and preconcentration of Cr(III) from industrial wastewater. (author)
20. REGULATION OF BLOOD PRESSURE IN PATIENTS WITH PRIMARY HYPERTENSION WITH SMOOTHIE BANANA (MUSA PARADISIACA
Directory of Open Access Journals (Sweden)
Eni Puji Lestari
2017-04-01
Full Text Available Introduction: Hypertension is a major problem that often happen in Indonesia. Hypertension can cause many complications. In Indonesia almost patients with hypertension got farmacologic therapy, but there is no difference. Banana smoothie is one of nonfarmacologic therapy that can be used to lower blood pressure. The purpose of this study was to analyze the effect of banana smoothie on regulation in patients with primary hypertension. Method: This study used quasy experimental design. The population in this study were patients with primary hypertension in Kedungturi village Taman Sidoarjo. The sampling technique used nonprobability sampling type of purposive sampling. The total number of sample were 16 respondents who were selected based on inclusion and exclusion criteria. Result:The Result of paired t-test at the systolic blood pressure and diastolic blood pressure in experiment group showed p value = 0.000. Independent t test between experiment group post-test and control group post-test showed p value = 0.000 for systolic blood pressure and p value = 0.002 for diastolic blood pressure. This result showed that there was a difference value of pretest and post-test systolic and diastolic blood pressure. With the result of independen t-test we know that there is a difference value between exsperiment and control blood pressure. Discussion: This study explain that there was significant effect of banana smoothie to regulate blood pressure in patients with primary hypertention. Banana smoothie can regulate the blood pressure because of high kalium substance. The function of kalium is to reduce the effect of natrium so the blood pressure can down. It can be conclude that banana smoothie can regulate the blood pressure in patients with primary hypertention. In further day patients with hypertension can choose banana smoothie to regulate their blood pressure.
1. Co-extrusion of food grains-banana pulp for nutritious snacks: optimization of process variables.
Science.gov (United States)
Mridula, D; Sethi, Swati; Tushir, Surya; Bhadwal, Sheetal; Gupta, R K; Nanda, S K
2017-08-01
Present study was undertaken to optimize the process conditions for development of food grains (maize, defatted soy flour, sesame seed)-banana based nutritious expanded snacks using extrusion processing. Experiments were designed using Box-Behnken design with banana pulp (8-24 g), screw speed (300-350 rpm) and feed moisture (14-16% w.b.). Seven responses viz. expansion ratio (ER), bulk density (BD), water absorption index (WAI), protein, minerals, iron and sensory acceptability were considered for optimizing independent parameters. ER, BD, WAI, protein content, total minerals, iron content, and overall acceptability ranged 2.69-3.36, 153.43-238.83 kg/m 3 , 4.56-4.88 g/g, 15.19-15.52%, 2.06-2.27%, 4.39-4.67 mg/100 g (w.b.) and 6.76-7.36, respectively. ER was significantly affected by all three process variables while BD was influenced by banana pulp and screw speed only. Studied process variables did not affected colour quality except 'a' value with banana pulp and screw speed. Banana pulp had positive correlation with water solubility index, total minerals and iron content and negative with WAI, protein and overall acceptability. Based upon multiple response analysis, optimized conditions were 8 g banana pulp, 350 rpm screw speed and 14% feed moisture indicating the protein, calorie, iron content and overall sensory acceptability in sample as 15.46%, 401 kcal/100 g, 4.48 mg/100 g and 7.6 respectively.
2. Deleterious effects of plant cystatins against the banana weevil Cosmopolites sordidus.
Science.gov (United States)
Kiggundu, Andrew; Muchwezi, Josephine; Van der Vyver, Christell; Viljoen, Altus; Vorster, Juan; Schlüter, Urte; Kunert, Karl; Michaud, Dominique
2010-02-01
The general potential of plant cystatins for the development of insect-resistant transgenic plants still remains to be established given the natural ability of several insects to compensate for the loss of digestive cysteine protease activities. Here we assessed the potential of cystatins for the development of banana lines resistant to the banana weevil Cosmopolites sordidus, a major pest of banana and plantain in Africa. Protease inhibitory assays were conducted with protein and methylcoumarin (MCA) peptide substrates to measure the inhibitory efficiency of different cystatins in vitro, followed by a diet assay with cystatin-infiltrated banana stem disks to monitor the impact of two plant cystatins, oryzacystatin I (OC-I, or OsCYS1) and papaya cystatin (CpCYS1), on the overall growth rate of weevil larvae. As observed earlier for other Coleoptera, banana weevils produce a variety of proteases for dietary protein digestion, including in particular Z-Phe-Arg-MCA-hydrolyzing (cathepsin L-like) and Z-Arg-Arg-MCA-hydrolyzing (cathepsin B-like) proteases active in mildly acidic conditions. Both enzyme populations were sensitive to the cysteine protease inhibitor E-64 and to different plant cystatins including OsCYS1. In line with the broad inhibitory effects of cystatins, OsCYS1 and CpCYS1 caused an important growth delay in young larvae developing for 10 days in cystatin-infiltrated banana stem disks. These promising results, which illustrate the susceptibility of C. sordidus to plant cystatins, are discussed in the light of recent hypotheses suggesting a key role for cathepsin B-like enzymes as a determinant for resistance or susceptibility to plant cystatins in Coleoptera. 2009 Wiley Periodicals, Inc.
3. Variable content and distribution of arabinogalactan proteins in banana (Musa spp.) under low temperature stress.
Science.gov (United States)
Yan, Yonglian; Takáč, Tomáš; Li, Xiaoquan; Chen, Houbin; Wang, Yingying; Xu, Enfeng; Xie, Ling; Su, Zhaohua; Šamaj, Jozef; Xu, Chunxiang
2015-01-01
Information on the spatial distribution of arabinogalactan proteins (AGPs) in plant organs and tissues during plant reactions to low temperature (LT) is limited. In this study, the extracellular distribution of AGPs in banana leaves and roots, and their changes under LT stress were investigated in two genotypes differing in chilling tolerance, by immuno-techniques using 17 monoclonal antibodies against different AGP epitopes. Changes in total classical AGPs in banana leaves were also tested. The results showed that AGP epitopes recognized by JIM4, JIM14, JIM16, and CCRC-M32 antibodies were primarily distributed in leaf veins, while those recognized by JIM8, JIM13, JIM15, and PN16.4B4 antibodies exhibited predominant sclerenchymal localization. Epitopes recognized by LM2, LM14, and MAC207 antibodies were distributed in both epidermal and mesophyll cells. Both genotypes accumulated classical AGPs in leaves under LT treatment, and the chilling tolerant genotype contained higher classical AGPs at each temperature treatment. The abundance of JIM4 and JIM16 epitopes in the chilling-sensitive genotype decreased slightly after LT treatment, and this trend was opposite for the tolerant one. LT induced accumulation of LM2- and LM14-immunoreactive AGPs in the tolerant genotype compared to the sensitive one, especially in phloem and mesophyll cells. These epitopes thus might play important roles in banana LT tolerance. Different AGP components also showed differential distribution patterns in banana roots. In general, banana roots started to accumulate AGPs under LT treatment earlier than leaves. The levels of AGPs recognized by MAC207 and JIM13 antibodies in the control roots of the tolerant genotype were higher than in the chilling sensitive one. Furthermore, the chilling tolerant genotype showed high immuno-reactivity against JIM13 antibody. These results indicate that several AGPs are likely involved in banana tolerance to chilling injury.
4. Molecular characterization of Banana streak virus isolate from Musa Acuminata in China.
Science.gov (United States)
Zhuang, Jun; Wang, Jian-Hua; Zhang, Xin; Liu, Zhi-Xin
2011-12-01
Banana streak virus (BSV), a member of genus Badnavirus, is a causal agent of banana streak disease throughout the world. The genetic diversity of BSVs from different regions of banana plantations has previously been investigated, but there are relatively few reports of the genetic characteristic of episomal (non-integrated) BSV genomes isolated from China. Here, the complete genome, a total of 7722bp (GenBank accession number DQ092436), of an isolate of Banana streak virus (BSV) on cultivar Cavendish (BSAcYNV) in Yunnan, China was determined. The genome organises in the typical manner of badnaviruses. The intergenic region of genomic DNA contains a large stem-loop, which may contribute to the ribosome shift into the following open reading frames (ORFs). The coding region of BSAcYNV consists of three overlapping ORFs, ORF1 with a non-AUG start codon and ORF2 encoding two small proteins are individually involved in viral movement and ORF3 encodes a polyprotein. Besides the complete genome, a defective genome lacking the whole RNA leader region and a majority of ORF1 and which encompasses 6525bp was also isolated and sequenced from this BSV DNA reservoir in infected banana plants. Sequence analyses showed that BSAcYNV has closest similarity in terms of genome organization and the coding assignments with an BSV isolate from Vietnam (BSAcVNV). The corresponding coding regions shared identities of 88% and -95% at nucleotide and amino acid levels, respectively. Phylogenetic analysis also indicated BSAcYNV shared the closest geographical evolutionary relationship to BSAcVNV among sequenced banana streak badnaviruses.
5. Extraction of Cellulose from Kepok Banana Peel (Musa parasidiaca L. for Adsorption Procion Dye
Directory of Open Access Journals (Sweden)
Poedji Loekitowati Hariani
2016-05-01
cellulose. The morphology of cellulose more homogenous than kepok banana peel powder. It was observed that the optimum adsorption of Procion dye by cellulose was on the initial concentration of 30 mg/L, pH solution of 5 and contact time within 30 minutes. The obtained result that cellulose has removal percentage to adsorp Procion dye more higher than kepok banana peel powder. The adsorption equilibrium showed the Langmuir isotherm was described well for adsorption process (R2 = 0.991 than Freundlich isotherm (R2 = 0.922.
6. Differential response of banana cultivars to F. oxysporum f. sp. cubense infection for Chitinase activity
International Nuclear Information System (INIS)
Morpurgo, R.; Duren, M. Van; Grasso, G.; Afza, R.
1997-01-01
Six banana clones with varying levels of resistance were inoculated with conidial suspension of races 1 and 4 of Fusarium oxysporum f. sp. cubense. Chitanase activity in the corm and root tissues was monitored before and after infection to relate with the field resistance or susceptibility of banana cultivars. Resistant clones showed high constitutive chitinase activity in roots and a rapid response to infection. The results suggest that chitinase could be considered as part of a complex mechanism leading to disease resistance. (author). 5 refs, 8 figs
7. Banana peel extract suppressed prostate gland enlargement in testosterone-treated mice.
Science.gov (United States)
Akamine, Kiichiro; Koyama, Tomoyuki; Yazawa, Kazunaga
2009-09-01
A methanol extract of banana peel (BPEx, 200 mg/kg, p.o.) significantly suppressed the regrowth of ventral prostates and seminal vesicles induced by testosterone in castrated mice. Further studies in the androgen-responsive LNCaP human prostate cancer cell line showed that BPEx inhibited dose-dependently testosterone-induced cell growth, while the inhibitory activities of BPEx did not appear against dehydrotestosterone-induced cell growth. These results indicate that methanol extract of banana peel can inhibit 5alpha-reductase and might be useful in the treatment of benign prostate hyperplasia.
8. Effects of temperature anisotropy on neoclassical transport in the plateau and banana-plateau regimes
International Nuclear Information System (INIS)
Taguchi, Masayoshi
1999-01-01
The neoclassical transport theory in a presence of temperature anisotropy is investigated in the low to the intermediate collision frequency regimes for a large aspect-ratio tokamak plasma. The standard procedure for an isotropic plasma in the plateau regime is extended to an anisotropic plasma, and the neoclassical transport coefficients in this regime are explicitly calculated. By interpolating the results in the plateau regime and the previously obtained ones in the banana regime, the expressions for the neoclassical transport coefficients which are continuously valid from the banana to the plateau regimes are presented. (author)
9. Differential response of banana cultivars to F. oxysporum f. sp. cubense infection for Chitinase activity
Energy Technology Data Exchange (ETDEWEB)
Morpurgo, R; Duren, M Van; Grasso, G; Afza, R [Agriculture and Biotechnology Laboratory, International Atomic Energy Agency, Seibersdorf (Austria)
1997-07-01
Six banana clones with varying levels of resistance were inoculated with conidial suspension of races 1 and 4 of Fusarium oxysporum f. sp. cubense. Chitanase activity in the corm and root tissues was monitored before and after infection to relate with the field resistance or susceptibility of banana cultivars. Resistant clones showed high constitutive chitinase activity in roots and a rapid response to infection. The results suggest that chitinase could be considered as part of a complex mechanism leading to disease resistance. (author). 5 refs, 8 figs.
10. Genome-wide identification, phylogeny, and expression analyses of the 14-3-3 family reveal their involvement in the development, ripening and abiotic stress response in banana
Directory of Open Access Journals (Sweden)
meiying li
2016-09-01
Full Text Available Plant 14-3-3 proteins act as critical components of various cellular signaling processes and play an important role in regulating multiple physiological processes. However, less information is known about the 14-3-3 gene family in banana. In this study, 25 14-3-3 genes were identified from the banana genome. Based on the evolutionary analysis, banana 14-3-3 proteins were clustered into ε and non-ε groups. Conserved motif analysis showed that all identified banana 14-3-3 genes had the typical 14-3-3 motif. The gene structure of banana 14-3-3 genes showed distinct class-specific divergence between the ε group and the non-ε group. Most banana 14-3-3 genes showed strong transcript accumulation changes during fruit development and postharvest ripening in two banana varieties, indicating that they might be involved in regulating fruit development and ripening. Moreover, some 14-3-3 genes also showed great changes after osmotic, cold, and salt treatments in two banana varieties, suggested their potential role in regulating banana response to abiotic stress. Taken together, this systemic analysis reveals the involvement of banana 14-3-3 genes in fruit development, postharvest ripening, and response to abiotic stress and provides useful information for understanding the functions of 14-3-3 genes in banana.
11. The theory of the top
CERN Document Server
Klein, Felix; Nagem, Raymond J; Sandri, Guido
The Theory of the Top. Volume IV. Technical Applications of the Theory of the Top is the fourth and final volume in a series of self-contained English translations of the classic and definitive treatment of rigid body motion. Key features: * Complete and unabridged presentation with recent advances and additional notes; * Annotations by the translators provide insights into the nature of science and mathematics in the late 19th century; * Each volume interweaves theory and applications. The Theory of the Top was originally presented by Felix Klein as an 1895 lecture at Göttingen University that was broadened in scope and clarified as a result of collaboration with Arnold Sommerfeld. Graduate students and researchers interested in theoretical and applied mechanics will find this series of books a thorough and insightful account. Other volumes in the series include Introduction to the Kinematics and Kinetics of the Top, Development of the Theory in the Case of the Heavy Symmetric Top, and Perturbation...
12. Kinematic top analyses at CDF
Energy Technology Data Exchange (ETDEWEB)
Grassmann, H.; CDF Collaboration
1995-03-01
We present an update of the top quark analysis using kinematic techniques in p{bar p} collisions at {radical}s = 1.8 TeV with the Collider Detector at Fermilab (CDF). We reported before on a study which used 19.3 pb{sup {minus}1} of data from the 1992--1993 collider run, but now we use a larger data sample of 67 pb{sup {minus}1}. First, we analyze the total transverse energy of the hard collision in W+{ge}3 jet events, showing the likely presence of a t{bar t} component in the event sample. Next, we compare in more detail the kinematic structure of W+ {ge}3 jet events with expectations for top pair production and with background processes, predominantly direct W+ jet production. We again find W+ {ge} 3 jet events which cannot be explained in terms of background, but show kinematic features as expected from top. These events also show evidence for beauty quarks, in agreement with expectations from top, but not compatible with expectations from backgrounds. The findings confirm the observation of top events made earlier in the data of the 1992--1993 collider run.
13. Pemanfaatan Kulit Pisang Hasil Fermentasi Rhyzopus oligosporus dalam Ransum Terhadap Pertumbuhan Ayam Pedaging (UTILIZING OF FERMENTED BANANA PEELS BY RHYZOPUS OLIGOSPORUS IN RATION ON GROWTH OF BROILER
Directory of Open Access Journals (Sweden)
Theresia Nur Indah Koni
2013-11-01
Full Text Available An experiment has conducted to investigate the effect level of fermented banana peels by Rhyzopusoligosporus in ration to growth of broiler. The experiment was designed use Completely RandomizedDesign with four treatments: as a control ration (0% fermented banana peel; 5% fermented banana peel;10% fermented banana peel; 15% fermented banana peel, with four replicates. There were six ayampedagings a weeks with average weight 112.30 ± 2.74 g in each replicate. The result there was a 15%fermented banana peels in broiler ration was not significant (P>0,05 to feed consumption, but itsignificantly (P<0,01 effected gain, and feed conversion . Fermented banana peels by R. oligosporus couldbe used as much as 10% in broiler ration.
14. Strategi Perancangan Mutu Ripe Banana Chip (RBC Berbasis Harapan Konsumen
Directory of Open Access Journals (Sweden)
Bambang Herry P
2016-11-01
Full Text Available Ripe Banana Chip (RBC, merupakan salah satu jenis kripik yang dibuat dari pisang masak. RBC dapat dilakukan dengan menggunakan teknologi penggorengan vakum (vacum frying ataupun pembekuan (freezing. RBC pisang mas cukup banyak memiliki keunggulan dari nilai gizinya. Produk ini masih belum diketahui tingkat kesesuain mutunya dengan keinginan konsumen dilihat dari aspek fisik ataupun organoleptik. Tujuan penelitian ini agar dapat mengetahui rancangan mutu yang tepat untuk produk RBC pisang mas. Penelitian ini menggunakan tiga metode, diantaranya metode skoring, Customer Satisfaction Index (CSI, dan metode diagram tulang ikan. Hasil dari penelitian, yaitu : warna RBC dengan intensitas 5,56 (kuning dan skor 3,862 (suka ; informasi kemasan dengan intensitas 4,5 (setuju dan skor 3,534 (suka ; kenyamanan kemasan dengan intensitas 8,58 (nyaman dan skor 4,155 (suka ; keamanan kemasan dengan intensitas 8,15 (sangat aman dan skor 4,086 (suka ; ketebalan dengan intensitas 8,31(sangat tebal dan skor 1,604 (sangat tidak suka ; Oil dengan intensitas 4,86 (banyak dan skor 2,483 (tidak suka ; kerenyahan dengan intensitas 7,32 (keras dan skor 2,843 (cukup suka ; Easy of breaking dengan intensitas 5,17 (mudah dipatahkan dan skor 3,158 (cukup suka ; rasa manis dengan intensitas 7,89 (manis dan skor 3,208 (cukup suka ; rasa asam dengan intensitas 4,84 (agak asam dan skor 3,309 (cukup suka. Nilai CSI yang dihasilkan, yaitu sebesar 64% (puas. Strategi untuk meningkatkan mutu RBC diantaranya: mengurangi ukuran ketebalan bahan baku sebelum diproses; lebih dipertimbangkan lagi tingkat kematangan pisang; memberi pengarahan dan motivasi kepada pekerja; melakukan pengawasan saat produksi berlangsung; lebih memperhatikan keseragaman ukuran pisang, dan menggunakan alternatif lain dalam penirisan minyak, misalnya sentrifuse agar keberadaan minyak berkurang.
15. Caracterização microclimática em cultivo consorciado café/banana Microclimatic characterization in coffee and banana intercrop
Directory of Open Access Journals (Sweden)
José R. M. Pezzopane
2007-06-01
16. The Effects of Treatments on Batu Banana Flour and Percentage of Wheat Substitution on The Resistant Starch, In Vitro Starch Digestibility Content and Palatability of Cookies Made with Banana (Musa balbisiana Colla) Flour
Science.gov (United States)
Ratnasari, D.; Rustanti, N.; Arifan, F.; Afifah, DN
2018-02-01
Diabetes mellitus (DM) is the most common endocrine disease worldwide. Resistant starch is polysaccharide that is recommended for DM patient diets. One of the staple crops containing resistant starch is banana. It is the fourth most important staple crop in the world and critical for food security, best suited plant in warm, frost-free, and coastal climates area. Among banana varieties, Batu bananas (Musa balbisiana Colla) had the highest content of resistant starch (~39%), but its use as a food ingredient is limited. Inclusion of Batu banana flour into cookies manufacturing would both increase the economic value of Batu bananas and provide alternative snacks for DM patients. Here we sought to examine whether cookies made with modified Batu banana flour would be a suitable snack for DM patients. This study used a completely randomized design with two factors: substitution of Batu banana flour (25%, 50%,75%) for wheat-based flour and Batu banana flour treatment methods (no treatment, autoclaving-cooling, autoclaving-cooling-spontaneous fermentation). The resistant starch and in vitro starch digestibility levels were analyzed using two-way ANOVA and Tukey test, whereas the acceptance level was analyzed by Friedman and Wilcoxon tests. The content of resistant starch and in vitro starch digestibility of the different treatments ranged from 3.10 to 15.79% and 16.03 to 52.59%, respectively. Both factors differed significantly (p0.05). Meanwhile, palatability in terms of color, aroma, texture, and flavor differed significantly among the different treatments and starch contents (ppatients. Keywords: Batu banana, cookies, resistant starch, in vitro starch digestibility
17. WordPress Top Plugins
CERN Document Server
Corbin, Brandon
2010-01-01
Time flies when you're having fun. This is the right way to describe this WordPress Top Plugins book by Brandon Corbin. With real world examples and by showing you the perks of having these plugins installed on your websites, the author is all set to captivate your interest from start to end. Regardless of whether this is your first time working with WordPress, or you're a seasoned WordPress coding ninja, WordPress Top Plugins will walk you through finding and installing the best plugins for generating and sharing content, building communities and reader base, and generating real advertising r
18. Propagating quality planting material to improve plant health and crop performance, key practices for dessert banana, plantain and cooking banana: illustrated guide
OpenAIRE
Staver, Charles; Lescot, Thierry
2015-01-01
Available in English, French, Spanish and Arabic, on line and on CD-ROM, this illustrated guide summarizes the key practices for producing clean planting material of banana with a high yield potential for smallholders, depending on the pests and diseases which are present. The guide is also designed to contribute to better planning of the propagation of planting material for rural development and disaster relief projects. (Résumé d'auteur)
19. Physical Characteristics, Chemical Composition, Organoleptic Test And The Number Of Microbes In The Biscuits With Addition Of Flour Banana Peels
Science.gov (United States)
Hernawati; Aryani, A.; Shintawati, R.
2017-02-01
The purpose of this study to analyze the physical characteristics, chemical composition and organoleptic test of biscuit flour with the addition of flour banana peel. Materials used are banana peels Kepok. Kepok banana peel has been found to contain high fiber food. Biscuit-making stage includes the formation of cream, adding flour and wheat flour dietary fiber from banana peels to concentrations of 0% as control, 25%, 50% and 75% of 100 grams of wheat flour; mixing; molding; baking in the oven for 20-25 minutes with a temperature of 180°C. Parameters to be measured, namely the physical characteristics include: hardness, softness, consistency, crispness. Furthermore, the biscuits were tested by chemical analysis (proximate). Organoleptic test include: aroma, taste, mouthfeel, aftertaste. Data were analyzed statistically using SAS computing programs. Physical and organoleptic test results biscuits with the addition of flour banana peels has sufficient level of preference between like-liked. Based on the results of the proximate analysis of biscuits with the addition of flour banana peels has generally been in accordance with the National Standards of Indonesia (SNI). Conclusion of the study that the addition of flour banana peels in biscuits has the potential to become functional foods that contain high fiber.
20. Effect of banana peel cellulose as a dietary fiber supplement on baking and sensory qualities of butter cake
Directory of Open Access Journals (Sweden)
Chiraporn Sodchit
2013-12-01
Full Text Available Banana peels are a waste product of the banana industry that have caused an environmental problem. Conversion of banana peels to a food ingredient might be an alternative way of value-adding to this waste. This study aimed to extract cellulose from banana peels and use it as an ingredient in butter cake to increase dietary fiber content and to improve cake quality. The selection and optimization of extraction conditions of cellulose from banana peels employed chemical extractions. Banana peel cellulose (BPC was added to butter cake at 3 levels; 1.5, 3.0 and 4.5% w/w of flour compared with 3.0% commercial cellulose (CC and the control (no cellulose added. The sensory, chemical, physical and microbiological properties of the butter cakes were then determined. The odor, tenderness and moistness acceptance scores of the butter cake by 50 panelists ranged from “like moderately” to “like very much”, indicating that addition of BPC improved the sensory quality of the cake. The butter cake with added CC and BPC had significantly higher (pd”0.05 moisture and fiber contents than those of the control. Microorganism levels found in the butter cake conformed to the butter cake standard (OTOP standard product of Thailand 459/2549. The optimum concentration of added BPC was 1.5%. Thus, the addition of BPC extracted from banana peels to butter cake increased the fiber content and improve the cake quality.
1. The use of green banana (Musa balbisiana pulp and peel flour as an ingredient for tagliatelle pasta
Directory of Open Access Journals (Sweden)
Vanessa Naciuk Castelo-Branco
2017-08-01
Full Text Available Abstract Green banana flour shows good potential as a functional ingredient for special-purpose foods, but there are no data in the literature concerning the use of a green banana pulp and peel flour for the development of products such as pasta. The aim of the present study was to develop tagliatelle pasta substituting the wheat flour with different concentrations of a green banana mixed pulp and peel flour. The pasta formulations were prepared replacing the wheat flour by the green banana mixed pulp and peel flour in two concentrations: 15% and 30%. A control formulation with wheat flour was also prepared. The green banana mixed pulp and peel flour presented higher ash, total fibre and total phenolic compound contents than traditional wheat flour. The pasta formulation with the addition of 15% green banana flour showed the highest ash content and the best sensory acceptability of all the formulations. It was concluded that it was possible to develop a tagliatelle pasta with satisfactory acceptance replacing the wheat flour by a green banana mixed pulp and peel flour.
2. The Efficacies of Banana Stem Extract as a Candidate of Coccidiostat Against Rabbit Eimeria Stiedaio Ocysts: an in Vitro Analysis
Directory of Open Access Journals (Sweden)
Diana Indrasanti
2015-09-01
Full Text Available The objective of this research was to investigatethe ability of banana stem (Musa paradisiaca to inhibitsporulation of Eimeria stiedaioocystsderived fromrabbit by in vitroanalysis.Analyze the active substance proximate analysis and active substancesin this research were performed too. Banana stem extract were used in this experiment andsulfaquinoxalline(Coxy ®was run as acontrol. The Eimeria stiedaioocystswere incubated prior the presence of different concentration from banana stem extract 0%, 1%, 2%, 4%, 8%for 1, 2 and 3 daysat 26°C. In addition,Factorial patterned Completely Randomized Design (CRD with five replicates wasapplied on the experiment. Result analysis was performed by using Analysis of Variance and following by Honestly Significant Difference (HSD post hoc test. Here, we identified that banana stem extract contain different type of active substance such as tannin, saponin, and alkaloid. Banana stem extract significantly affected the oocysts sporulation included the amount of sporulatedoocysts (P<0.01, unsporulatedoocysts (P<0.01, and transformed oocysts (P<0.01. In conclusion banana stem could inhibit the development of Eimeria stiedaioocysts on in vitroexperiment. HSD test showed that the optimum potential efficacy of banana stem toinhibit sporulation was at 4% and 8% concentration during three days incubation.
3. Banana (Musa spp) from peel to pulp: ethnopharmacology, source of bioactive compounds and its relevance for human health.
Science.gov (United States)
Pereira, Aline; Maraschin, Marcelo
2015-02-03
Banana is a fruit with nutritional properties and also with acclaimed therapeutic uses, cultivated widely throughout the tropics as source of food and income for people. Banana peel is known by its local and traditional use to promote wound healing mainly from burns and to help overcome or prevent a substantial number of illnesses, as depression. This review critically assessed the phytochemical properties and biological activities of Musa spp fruit pulp and peel. A survey on the literature on banana (Musa spp, Musaceae) covering its botanical classification and nomenclature, as well as the local and traditional use of its pulp and peel was performed. Besides, the current state of art on banana fruit pulp and peel as interesting complex matrices sources of high-value compounds from secondary metabolism was also approached. Dessert bananas and plantains are systematic classified into four sections, Eumusa, Rhodochlamys, Australimusa, and Callimusa, according to the number of chromosomes. The fruits differ only in their ploidy arrangement and a single scientific name can be given to all the edible bananas, i.e., Musa spp. The chemical composition of banana's peel and pulp comprise mostly carotenoids, phenolic compounds, and biogenic amines. The biological potential of those biomasses is directly related to their chemical composition, particularly as pro-vitamin A supplementation, as potential antioxidants attributed to their phenolic constituents, as well as in the treatment of Parkinson's disease considering their contents in l-dopa and dopamine. Banana's pulp and peel can be used as natural sources of antioxidants and pro-vitamin A due to their contents in carotenoids, phenolics, and amine compounds, for instance. For the development of a phytomedicine or even an allopathic medicine, e.g., banana fruit pulp and peel could be of interest as raw materials riches in beneficial bioactive compounds. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
4. Detection of antimicrobial activity of banana peel (Musa paradisiaca L. on Porphyromonas gingivalis and Aggregatibacter actinomycetemcomitans: An in vitro study
Directory of Open Access Journals (Sweden)
2015-01-01
Full Text Available Introduction and Aim: Banana is used widely because of its nutritional values. In past, there are studies that show banana plant parts, and their fruits can be used to treat the human diseases. Banana peel is a part of banana fruit that also has the antibacterial activity against microorganisms but has not been studied extensively. Since, there are no studies that relate the antibacterial activity of banana peel against periodontal pathogens. Hence, the aim of this study is to determine the antimicrobial activity of banana peel extract on Porphyromonas gingivalis (P. gingivalis and Aggregatibacter actinomycetemcomitans (A. actinomycetemcomitans. Material and Methods: Standard strains of P. gingivalis and A. actinomycetemcomitans were used in this study which was obtained from the in-house bacterial bank of Department of Molecular Biology and Immunology at Maratha Mandal's Nathajirao G. Halgekar Institute of Dental Sciences and Research Centre. The banana peel extract was prepared, and the antibacterial activity was assessed using well agar diffusion method and minimum inhibitory concentration was assessed using serial broth dilution method. Results: In the current study, both the tested microorganisms showed antibacterial activity. In well diffusion method, P. gingivalis and A. actinomycetemcomitans showed 15 mm and 12 mm inhibition zone against an alcoholic extract of banana peel, respectively. In serial broth dilution method P. gingivalis and A. actinomycetemcomitans were sensitive until 31.25 μg/ml dilutions. Conclusion: From results of the study, it is suggested that an alcoholic extract of banana peel has antimicrobial activity against P. gingivalis and A. actinomycetemcomitans.
5. Phosphate fertilization changes the characteristics of 'Maçã' banana starch.
Science.gov (United States)
Mesquita, Camila de Barros; Garcia, Émerson Loli; Bolfarini, Ana Carolina Batista; Leonel, Sarita; Franco, Célia Maria L; Leonel, Magali
2018-06-01
The unripe banana has been studied as a potential source of starch for use in various applications. Considering the importance of phosphorus in the biosynthesis of the starch and also the interference of this mineral in starch properties, in this study it was evaluated the effect of rates of phosphate fertilizer applied in the cultivation of 'Maçã' banana on the characteristics of the starch. Starches extracted from fruits from different treatments were analyzed for morphological characteristics, X-ray diffraction pattern, relative crystallinity, granule size, amylose, resistant starch and phosphorus levels, as well as, for pasting and thermal properties. Results showed that the phosphate fertilization has interference on the characteristics of the banana starch led to increase of phosphorus content and size of the granules, reduction of crystallinity and resistant starch content, decrease of viscosity peak, breakdown, final viscosity, setback, transitions temperatures and enthalpy. These changes caused by phosphate fertilizer conditions can be increase the applications of the 'Maçã' banana starch. Copyright © 2018. Published by Elsevier B.V.
6. Register of new fruit and nut cultivars list 48. Banana, cacao, plantain
Science.gov (United States)
The Register of New Fruit and Nut Varieties 48 is a compilation of descriptions of new fruit and nut cultivars from around the world. In this edition, newly released banana, plantain, and cacao cultivars are described in terms of their origins, important fruit traits and yield. ...
7. Pesticides in surface waters in areas influenced by banana production in Costa Rica
International Nuclear Information System (INIS)
Castillo, L.E.; Ruepert, C.; Solis, E.
1999-01-01
Banana production in Costa Rica is highly dependent on pesticide use. However, only a few studies have been undertaken regarding the presence and environmental impact of the agrochemical substances used in the banana culture on the aquatic ecosystem of the Atlantic Region of Costa Rica. This study was, therefore, undertaken in Rio Suerte Basin that drains into the 'Nature Conservation Area' of Tortuguero in the Atlantic lowlands of the country from June 1993 to December 1994. In order to investigate further the occurrence of pesticides in the water bodies located near the possible sources especially during worst-case situations, water samples were analysed following pesticide applications during 1995-1997. Pesticide residues were determined by GC equipped with an electron capture detector (ECD) and a nitrogen phosphorous detector (NPD). The study targeted 11 of the 21 pesticides used in banana production, the others were not analyzed. The most frequently found compounds during the 1993-94 survey were the fungicide propiconazole and the nematocide cadusafos. Maximum concentrations measured after the pesticide applications were found in the main drainage canal and these were 2.1 ug/L carbofuran, 1.2 ug/L terbufos and 0.48 ug/L cadusafos. The peak concentration found shortly after the aerial application of the fungicide propiconazole was 13 ug/L in the creek leaving the banana plantation. (author)
8. Inhibition of bananas ripening by gamma radiation: physical, chemical and sensory aspects
International Nuclear Information System (INIS)
Domarco, Raquel Elisabeth; Walder, Julio Marcos Melges; Spoto, Marta Helena Fillet; Blumer, Lucimara; Matraia, Clarice
1996-01-01
In order to extend the shelf-life, nanica bananas (Musa AAA group, Caverdish subgroup), packed and not in polyethylene bags (o,03 mm thickness) were irradiated with 0.0, 0.2, 0.4, 0.6, 0.8 and 1.0 kGy. A cobalt-60 source, with an activity of 109.71 Gbq was used at a dose rate of 1,71 kGy/h. The experiment was developed on three different storage conditions: a) room temperature, without packaging, for 15 days; b) 17 deg C, without packaging, for 21 days; c) 17 deg C, with packaging, for 42 days. The chemical analysis were done on soluble solids, tirable acidity, pH and ratio. The bananas were subjected to sensorial evaluation by visual and flavor analysis. The results were analysed using SAS-ANOVA. Bananas irradiated with 0,4 kGy could be storage for 15 days on room temperatures without packaging. Dose of 0,6 kGy and 17 deg C temperature were enough to conservate bananas, without packaging for 21 days and with packaging for 42 days. (author)
9. Living with AIDS in Uganda : impacts on banana-farming households in two districts
NARCIS (Netherlands)
Karuhanga, M.
2008-01-01
The research was carried out among banana-farming households in the districts of Masaka and Kabarole in Uganda. A gendered livelihood approach was used. The research focused on the identification of critical factors that need to be taken into consideration in the development of relevant policies for
10. banana cultivar distribution in rwanda abstract résumé
African Journals Online (AJOL)
Gisubi (ABB), Gros Michel (AAA), and 'Kamaramasenge'. (AAB) were the most abundant. Farms with a higher proportion of Gisubi contained fewer other cultivars. Also, new cultivars were identified and these should be added to the National Banana germplasm collection. Key Words: Cultivars, diversity indexes, germplasm, ...
11. Effect of surface coating on ripening and early peel spotting in 'Sucrier' banana (Musa acuminata)
NARCIS (Netherlands)
Promyou, S.; Ketsa, S.; Doorn, van W.G.
2007-01-01
Sucrier¿ bananas (Musa acuminata, AA Group) show peel spotting when the peel is just about as yellow as green, which coincides with optimum eating quality. As consumers might relate the spotting to overripe fruit, early spotting is considered undesirable, especially for export markets. Fruit were
12. Using possibilities of some agricultural wastes in open-field banana cultivation
Directory of Open Access Journals (Sweden)
Mehmet ÖTEN
2016-06-01
Full Text Available Usage of farmyard manure is the one of the major factors to increase production cost in banana cultivation. Besides increasing the production costs, other disadvantages of farmyard manure are playing active role on carrying diseases and pests and also difficulty in obtaining. Due to the stated disadvantages, the use farmyard manure of banana farmers is decreasing. Therefore, we need alternative ways to increase the organic matter capacity of the soil. The effects of alternative applications to farmyard manure, namely banana waste and mushroom compost were investigated. The objective of the study was to evaluate effects of these applications on some morphological properties (plant height, plant circumference and number of leaves, yield (number of hands, number of fingers, bunch weight, finger weight and length and quality properties (flesh/skin ratio, total soluble solids matter, sugars etc. under open-field banana cultivation. The experiment was conducted in Kargıcak location of Alanya in randomized complete block design (RCBD with 3 replications. Experimental results revealed that using of farmyard manure and waste treatments positively affected the yield parameters like the number of hands and fingers, finger length, finger weight and bunch weight. On the other hand, treatments did not have a statistically significant effect on fruit quality parameters like soluble solids content, titratable acidity, pH and ash.
13. Involvement of WRKY Transcription Factors in Abscisic-Acid-Induced Cold Tolerance of Banana Fruit.
Science.gov (United States)
Luo, Dong-Lan; Ba, Liang-Jie; Shan, Wei; Kuang, Jian-Fei; Lu, Wang-Jin; Chen, Jian-Ye
2017-05-10
Phytohormone abscisic acid (ABA) and plant-specific WRKY transcription factors (TFs) have been implicated to play important roles in various stress responses. The involvement of WRKY TFs in ABA-mediated cold tolerance of economical fruits, such as banana fruit, however remains largely unknown. Here, we reported that ABA application could induce expressions of ABA biosynthesis-related genes MaNCED1 and MaNCED2, increase endogenous ABA contents, and thereby enhance cold tolerance in banana fruit. Four banana fruit WRKY TFs, designated as MaWRKY31, MaWRKY33, MaWRKY60, and MaWRKY71, were identified and characterized. All four of these MaWRKYs were nuclear-localized and displayed transactivation activities. Their expressions were induced by ABA treatment during cold storage. More importantly, the gel mobility shift assay and transient expression analysis revealed that MaWRKY31, MaWRKY33, MaWRKY60, and MaWRKY71 directly bound to the W-box elements in MaNCED1 and MaNCED2 promoters and activated their expressions. Taken together, our findings demonstrate that banana fruit WRKY TFs are involved in ABA-induced cold tolerance by, at least in part, increasing ABA levels via directly activating NECD expressions.
14. Market access and agricultural production : the case of banana production in Uganda
NARCIS (Netherlands)
Bagamba, F.
2007-01-01
Keywords: Smallholder poor farmers, market access, bananas, productivity, efficiency, labour demand, labour supply,Uganda.This study investigates the effects of factor and commodity markets on the
15. Effects of HIV/AIDS on the livelihood of banana-farming households in Central Kenya
NARCIS (Netherlands)
Nguthi, F.N.; Niehof, A.
2008-01-01
This paper explores the effects of HIV/AIDS on the livelihoods of banana-farming households in Maragua district, Central Kenya. It is based on the results of a field study carried out during 2004-2005. The study applied the sustainable livelihood approach, using both quantitative and qualitative
16. Structural properties and digestion of green banana flour as a functional ingredient in pasta.
Science.gov (United States)
Zheng, Zeqi; Stanley, Roger; Gidley, Michael J; Dhital, Sushil
2016-02-01
Gluten free pasta was made from raw banana flour in combination with vegetable gums and protein for comparison to pasta similarly made from wheat flour. After cooking, it was found that the banana flour pasta was less susceptible to alpha-amylase digestion compared to conventional wheat flour pasta. Release of glucose by alpha-amylase digestion followed first order kinetics with an initial rapid rate of digestion and a subsequent second slower phase. The structure of green banana pasta starch at the inner and outer pasta surfaces was observed under confocal laser scanning microscopy (CLSM) and the viscosities of the flour mixtures were measured by a Rapid Visco Analyser (RVA). The digestibility of banana flour pasta was found to be related, not only to the properties of the starch granules, but also to the protein network of the surrounding food matrix. The effects of gums and proteins on pasta formation and digestibility are discussed in the context of its potential use as a gluten free lower glycaemic alternative to conventional wheat based pastas.
17. Agrobacterium-mediated transformation of Mycosphaerella fijiensis, the devastating Black Sigatoka pathogen of bananas
NARCIS (Netherlands)
Díaz-Trujillo, C.; Adibon, H.; Kobayashi, K.; Zwiers, L.H.; Souza, M.T.; Kema, G.H.J.
2010-01-01
Mycosphaerella fijiensis, M. musicola en M. eumusae veroorzaken de Sigatoka-ziekte in banaan. Op dit moment is de toepassing van fungiciden de enige optie om deze ziekte te bestrijden. Het PRPB (Pesticide Reduction Program for Banana) investeert in de ontwikkeling van technieken voor de genotype- en
18. Evolutionary dynamics of mating-type loci of Mycosphaerella spp. occurring on banana
NARCIS (Netherlands)
Arzanlou, M.; Crous, P.W.; Zwiers, L.H.
2010-01-01
The devastating Sigatoka disease complex of banana is primarily caused by three closely related heterothallic fungi belonging to the genus Mycosphaerella: M. fijiensis, M. musicola, and M. eumusae. Previous phylogenetic work showing common ancestry led us to analyze the mating-type loci of these
19. Same Disease—different research strategies: Bananas and Black Sigatoka in Brazil
NARCIS (Netherlands)
Cordoba, D.M.; Jansen, K.
2014-01-01
Fungal disease epidemics have the potential to bring about drastic innovations. However, in the case of the Black Sigatoka (Mycosphaerella fijiensis) fungus in bananas, producers and international traders are still awaiting a breakthrough in crop protection research. Using the cases of Brazil and
20. Consumer Perceptions towards Introducing a Genetically Modified Banana (Musa spp.) in Uganda
NARCIS (Netherlands)
Kikulwe, E.M.; Wesseler, J.H.H.; Falck-Zepeda, J.
2010-01-01
The introduction of a genetically modified (GM) banana (Musa spp.) in Uganda is not without controversy. It is likely to generate a wide portfolio of concerns as the technology of genetic engineering is still in its early stages of development in Uganda. The purpose of this study is to show how
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5675565004348755, "perplexity": 15055.7988217363}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-34/segments/1534221217006.78/warc/CC-MAIN-20180820195652-20180820215652-00157.warc.gz"}
|
https://gilkalai.wordpress.com/2020/03/22/kelman-kindler-lifshitz-minzer-and-safra-towards-the-entropy-influence-conjecture/
|
Kelman, Kindler, Lifshitz, Minzer, and Safra: Towards the Entropy-Influence Conjecture
Let me briefly report on a remarkable new paper by Esty Kelman, Guy Kindler, Noam Lifshitz, Dor Minzer, and Muli Safra, Revisiting Bourgain-Kalai and Fourier Entropies. The paper describes substantial progress towards the Entropy-Influence conjecture, posed by Ehud Friedgut and me in 1996. (See this blog post from 2007.)
Abstract: The total influence of a function is a central notion in analysis of Boolean functions, and characterizing functions that have small total influence is one of the most fundamental questions associated with it. The KKL theorem and the Friedgut junta theorem give a strong characterization of such functions whenever the bound on the total influence is o(logn), however, both results become useless when the total influence of the function is ω(logn). The only case in which this logarithmic barrier has been broken for an interesting class of function was proved by Bourgain and Kalai, who focused on functions that are symmetric under large enough subgroups of $S_n$.
In this paper, we revisit the key ideas of the Bourgain-Kalai paper. We prove a new general inequality that upper bounds the correlation between a Boolean function f and a real-valued, low degree function g in terms of their norms, Fourier coefficients and total influences.
Some corollaries of this inequality are:
1. The Bourgain-Kalai result.
2. A slightly weaker version of the Fourier-Entropy Conjecture. More precisely, we prove that the Fourier entropy of the low-degree part of f is at most $O(I[f]log^2 I[f])$, where I[f] is the total influence of f. In particular, we conclude that the Fourier spectrum of a Boolean function is concentrated on at most $2O(I[f]log^2 I[f])$ Fourier coefficients.
3. Using well-known learning algorithms of sparse functions, the previous point implies that the class of functions with sub-logarithmic total influence (i.e. at most $O(\log n/(\log \log n)2))$ is learnable in polynomial time, using membership queries.
Our theorem on influence under symmetry. From my videotaped lecture on Jean Bourgain. See this post about Jean Bourgain.
A few remarks:
A) Given a Boolean function $f:\{-1,1\}^n\to \{-1,1\}$ let $f=\sum_{S \subset \{1,2,\dots ,n\}}\hat f(S)W_S$ be its Fourier-Walsh expansion. (Here $W_S(x_1,x_2,\dots ,x_n)=\prod_{i\in S}x_i$.) The total influence $I(f)$ of $f$ is described by
$I(f)=\sum {S \subset \{1,2,\dots ,n\}}\hat f^2(S)|S|$.
The spectral entropy $E(f)$ of $f$ is defined by
$E(f)=\sum_{S \subset \{1,2,\dots ,n\}}\hat f^2(S) \log (\hat f^2(S))$.
The entropy-influence conjecture (here called the Fourier-entropy conjecture) asserts that for some universal constant C
$I(f) \ge C E(f)$.
B) One interesting feature of the conjecture is that the RHS is invariant under arbitrary linear transformations of $\{-1,1\}^n$ (regarded as an $n$-dimensional vector space) while the LHS is invariant only to permutations of the variables.
C) My paper with Jean Bourgain, Influences of variables and threshold intervals under group symmetries, describes lower bounds on total influences for Boolean function invariant under certain groups of permutations (of the variables). Those results are stronger (but more restrictive) than what the Entropy/influence conjecture directly implies. (This was overlooked for a while.) The new paper gives (much hoped for, see below) simpler proofs and stronger results compared to those in my paper with Jean Bourgain.
D) Ryan O’Donnel wrote about Bourgain and some of his contributions to the analysis of Boolean functions:
“I spent close to 5 years understanding one 6-page paper of his (“On the distribution of the Fourier spectrum of Boolean functions”), and also close to 10 years understanding a 10-page paper of his (the k-SAT sharp threshold ‘appendix’). If anyone’s up for a challenge, I’m pretty sure no one on earth fully understands the second half of his paper “Influences of variables and threshold intervals under group symmetries” with Kalai (including Gil 🙂 )”
It looks that by now we have pretty good understanding and even some far-reaching progress regarding all three papers that Ryan mentioned. (It is unclear if we can hope now for an exponential spread of understanding for Bourgain’s proofs.)
This entry was posted in Combinatorics, Computer Science and Optimization, Open problems and tagged , , , , . Bookmark the permalink.
1 Response to Kelman, Kindler, Lifshitz, Minzer, and Safra: Towards the Entropy-Influence Conjecture
1. Gil Kalai says:
Here is a link to two videotaped lectures on the new proof by Dor Minzer https://simons.berkeley.edu/events/boolean-1
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 16, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9160208106040955, "perplexity": 1383.9249432150684}, "config": {"markdown_headings": false, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439740343.48/warc/CC-MAIN-20200814215931-20200815005931-00304.warc.gz"}
|
http://mathhelpforum.com/math-software/140806-m-file-failure.html
|
1. ## m file failure?
Hey:
Iv written this m-file, basically i want it to have a function, annd to plot it on a graph, then run another m-file within it to show the graph with small limits using its derivative, but when I run it, I don't get any results, i just get this:
f =
[]
Heres the m file:
function [ y ] = f( x )
% calculate f(x)=exp(x)
y=exp(x);
plot(x,f(x));
function [ y ] = fd( x )
% calculates fd(x)=exp(x)
y=exp(x);
plot(x,fd(x),'k')
pause
plot(x,(f(x+0.1)-f(x))/0.1,'r')
pause
plot(x,(f(x+0.01)-f(x))/0.01,'g')
pause
plot(x,(f(x+0.001)-f(x))/0.001,'b')
pause
plot(x,(f(x+0.0001)-f(x))/0.0001,'m')
end
end
cheerz
2. Originally Posted by ramdrop
Hey:
Iv written this m-file, basically i want it to have a function, annd to plot it on a graph, then run another m-file within it to show the graph with small limits using its derivative, but when I run it, I don't get any results, i just get this:
f =
[]
Heres the m file:
function [ y ] = f( x )
% calculate f(x)=exp(x)
y=exp(x);
plot(x,f(x));
function [ y ] = fd( x )
% calculates fd(x)=exp(x)
y=exp(x);
plot(x,fd(x),'k')
pause
plot(x,(f(x+0.1)-f(x))/0.1,'r')
pause
plot(x,(f(x+0.01)-f(x))/0.01,'g')
pause
plot(x,(f(x+0.001)-f(x))/0.001,'b')
pause
plot(x,(f(x+0.0001)-f(x))/0.0001,'m')
end
end
cheerz
This is a function definition file the first function in it is executed by invoking it, the second is only executed if invoked from the first (and it is not).
1. Please post the calling statement (what you type on at the command prompt or have in the script file)
2. Consider executing in the debugger.
3. Look at all the red or amber marks to the right in the editor, each of these indicates a syntactical mistake or a warning that the code may not be doing what you think, so check them.
4. You appear to be calling the top level function recursively, I doubt you intend this.
CB
3. Originally Posted by CaptainBlack
This is a function definition file the first function in it is executed by invoking it, the second is only executed if invoked from the first (and it is not).
1. Please post the calling statement (what you type on at the command prompt or have in the script file)
2. Consider executing in the debugger.
3. Look at all the red or amber marks to the right in the editor, each of these indicates a syntactical mistake or a warning that the code may not be doing what you think, so check them.
4. You appear to be calling the top level function recursively, I doubt you intend this.
CB
Basically (I think this is rigHT), What I want it to do is simple. To have a function defined, and then to (through a series of commands) run an m-file, so I think its called a Script m file or something like that.
1. I don't type anything in the command window, I just right click on it, and run it - that is obviously wrong though.
2. Not sure what you mean by this, but I will have a look.
3. There are no marks in the "m file maker" its a green square, so nothing is wrong. Well something obviously is.
4. Again, not sure what you mean by this, but Il take a second look at it.
4. Originally Posted by ramdrop
Basically (I think this is rigHT), What I want it to do is simple. To have a function defined, and then to (through a series of commands) run an m-file, so I think its called a Script m file or something like that.
1. I don't type anything in the command window, I just right click on it, and run it - that is obviously wrong though.
How does it find its input data?
2. Not sure what you mean by this, but I will have a look.
3. There are no marks in the "m file maker" its a green square, so nothing is wrong. Well something obviously is.
It is in the Matlab editor (answer for both of the above)
4. Again, not sure what you mean by this, but Il take a second look at it.
This is a function file not a script, only the top function is visible out side the file.
CB
5. Well, Iv had a major play with it at the moment, im trying to run the file first to see what happens. Except it just doesn't work,
I get in the command window:
??? Input argument "x" is undefined.
Error in ==> f at 3
y=x^2;
This clearly indicates to me that the variable, x is undefined and it will not work unless I define it.
Im unsure as to how to define it, I know of one command, "syms 'x'" but that doesn't seem to work,
Any ideas?
6. Originally Posted by ramdrop
Well, Iv had a major play with it at the moment, im trying to run the file first to see what happens. Except it just doesn't work,
I get in the command window:
This clearly indicates to me that the variable, x is undefined and it will not work unless I define it.
Im unsure as to how to define it, I know of one command, "syms 'x'" but that doesn't seem to work,
Any ideas?
Yes, but I have already posted them, have you read them and acted on them?
Please repost the content of the .m file its name and the data you are trying to pass to it and an explanation of how you are trying to invoke this function.
Also post a clear statement of what you think you are trying to do (the exact wording of the problem as set)
CB
7. function [ y ] = f( x )
% calculate f(x)=exp(x)
y=exp(x);
plot(x,f(x));
function [ y ] = fd( x )
% calculates fd(x)=exp(x)
y=exp(x);
plot(x,fd(x),'k')
pause
plot(x,(f(x+0.1)-f(x))/0.1,'r')
pause
plot(x,(f(x+0.01)-f(x))/0.01,'g')
pause
plot(x,(f(x+0.001)-f(x))/0.001,'b')
pause
plot(x,(f(x+0.0001)-f(x))/0.0001,'m')
end
end
That is the m file, what I want it to do, is for values of x, plot the function $\displaystyle e^x$ and show this on a graph.
Then I want it to (after I input some command I think) probably the actual plot line in the file, to plot the derivative of $\displaystyle e^x$ which is obviously the same, but with the small values towards the end of the m-file. We know the actual derivative, and its been plotted on a graph, the other "plots" get closer to the correct derivative. Thats what I want it to do
8. Originally Posted by ramdrop
That is the m file, what I want it to do, is for values of x, plot the function $\displaystyle e^x$ and show this on a graph.
Then I want it to (after I input some command I think) probably the actual plot line in the file, to plot the derivative of $\displaystyle e^x$ which is obviously the same, but with the small values towards the end of the m-file. We know the actual derivative, and its been plotted on a graph, the other "plots" get closer to the correct derivative. Thats what I want it to do
That is an m-file what happens when you run it?
Also what about the other questions I asked? Most importantly how are you calling this and what argument are you passing to it?
Also I have already pointed out the you are calling F recursively and you probably do not want to do that, do you want to do that? Have you done anything to fix that problem?
CB
9. It comes up with the error when I run it:
??? Input argument "x" is undefined.
Error in ==> f at 3
y=x^2
I have tried changing the functions, 1 being f and the other being g.
The file itself is called "f.m"
Im trying to write something in the command window, like f(5) or something and it comes with errors, I just need it to run basically
10. Originally Posted by ramdrop
It comes up with the error when I run it:
I have tried changing the functions, 1 being f and the other being g.
The file itself is called "f.m"
Im trying to write something in the command window, like f(5) or something and it comes with errors, I just need it to run basically
f requires a vector of points as input something like f([1,2,3,4,5]).
CB
11. I want it for all values of x.
Okay, so to write it, I would write in the m-file on the second line:
f = [-inf,inf] ?
12. Originally Posted by ramdrop
I want it for all values of x.
Okay, so to write it, I would write in the m-file on the second line:
f = [-inf,inf] ?
No you can't do such a plot! You are also not passing it a range to plot but a set of points.
It seems to me that your level of understanding of what you want and what Matlab does is poor. So if we are to make any progress with this you will first have to explain why you are doing what you are trying to do.
CB
13. My MATLAB knowledge is indeed poor because the lectures we have been given are just a sheet, the lecturer is never around to ask questions, usually these "sheets" are not really informative, they just work through examples, we're given them and then the lecturer wanders off for the majority of the class.
Im trying to learn MATLAB by myself because next year, I have a lot of MATLAB to do..
Okay, so this is EXACTLY what I want the m-file to do.
I want the m-file (for any function, x², exp(x)) defined as f, to be plotted on a graph, probably the best would be a 10 by 10. This is represented by:
function [ y ] = f( x )
% calculate f(x)=exp(x)
y=exp(x);
plot(x,f(x));
Then, I want it to differentiate the function f, and plot that graph. Then it "approximates" the derivative for values of x = 0.01, 0.001, 0.0001. This causes the "curve/line" of the graph to change and become more accurate.
So in basic terms, the smaller the value of x, the more accurate the derivative plot is. This is represented by the latter part of the m-file.
function [ y ] = fd( x )
% calculates fd(x)=exp(x)
y=exp(x);
plot(x,fd(x),'k')
pause
plot(x,(f(x+0.1)-f(x))/0.1,'r')
pause
plot(x,(f(x+0.01)-f(x))/0.01,'g')
pause
plot(x,(f(x+0.001)-f(x))/0.001,'b')
pause
plot(x,(f(x+0.0001)-f(x))/0.0001,'m')
end
en
14. Originally Posted by ramdrop
My MATLAB knowledge is indeed poor because the lectures we have been given are just a sheet, the lecturer is never around to ask questions, usually these "sheets" are not really informative, they just work through examples, we're given them and then the lecturer wanders off for the majority of the class.
Im trying to learn MATLAB by myself because next year, I have a lot of MATLAB to do..
Okay, so this is EXACTLY what I want the m-file to do.
I want the m-file (for any function, x², exp(x)) defined as f, to be plotted on a graph, probably the best would be a 10 by 10. This is represented by:
Then, I want it to differentiate the function f, and plot that graph. Then it "approximates" the derivative for values of x = 0.01, 0.001, 0.0001. This causes the "curve/line" of the graph to change and become more accurate.
So in basic terms, the smaller the value of x, the more accurate the derivative plot is. This is represented by the latter part of the m-file.
Lets start at the begining:
[code]
function [ y ] = f( x )
% calculate f(x)=exp(x)
y=exp(x);
plot(x,f(x));
[\code]
1. I have told you before that you have a recursive call to f in this function definition that you almost certainly don't intend.
Create a file f.m which contains:
Code:
function [y]=f(x)
y=exp(x);
Now call this from the console with:
Code:
>>x=[-10:0.5:10]
>>y=f(x)
>>plot(x,y);
Now report what happens.
CB
15. I got a very nice plot, thanks, so I guess I do the same for the next part of the file?
Page 1 of 2 12 Last
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8058874011039734, "perplexity": 770.1537611747076}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794865468.19/warc/CC-MAIN-20180523082914-20180523102914-00156.warc.gz"}
|
https://www.physicsforums.com/threads/simplifying-this-equation.600615/
|
# Homework Help: Simplifying this equation
1. Apr 26, 2012
### Cyclopse
1. The problem statement, all variables and given/known data
Is it possible to simplify this even further?
2. Relevant equations
This is a Secant-tangent equation.
3. The attempt at a solution
I cross multiplied to get 18=324radical 3 (AC)
and in the end i got 0 = (18radical3) (AC) which makes no sense
so is there a different way to simplify that?
2. Apr 26, 2012
### Staff: Mentor
Are you trying to solve for AC?
When you divide both sides by 18, you will be left with 1 on the left side, not 0.
If you are trying to find AC, divide both sides by 324√3.
3. Apr 26, 2012
### Cyclopse
Yes.
Here is the original question
Last edited: Apr 26, 2012
4. Apr 26, 2012
### Staff: Mentor
It doesn't seem to me that you have enough information.
How did you go from AB = 18 and AD = 18/√3 to 18 = 324√3 * AC?
5. Apr 26, 2012
### Cyclopse
I'm really terrible at maths...that's why I'm here.
6. Apr 26, 2012
### Staff: Mentor
Draw a radius from O to B, and call its length r. The radius OB is perpendicular to AB, so that ABO is a right triangle.
Let θ = the angle at A.
cos(θ) = 18/(r + 18/√3), and
sin(θ) = r/(r + 18/√3)
Now you have two equations in two unknowns, so you should be able to solve for both unknowns, although you only need r. Once you know r, it's pretty easy to get AC.
Last edited: Apr 26, 2012
7. Apr 26, 2012
### Staff: Mentor
Start out like Mark44 said by drawing a radius from O to B, and call its length r, but then apply the Pythagorean Theorem to the right triangle ABO to find r.
Chet
8. Apr 27, 2012
### Staff: Mentor
That's simpler than what I suggested. Simple is good!
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8088799118995667, "perplexity": 2557.5803387797855}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267156724.6/warc/CC-MAIN-20180921013907-20180921034307-00052.warc.gz"}
|
http://www.lastfm.es/user/erazorhead/library/music/65daysofstatic/_/When+We+Were+Younger+&+Better?setlang=es
|
# Colección
Música » 65daysofstatic »
## When We Were Younger & Better
26 scrobblings | Ir a la página del tema
Temas (26)
Tema Álbum Duración Fecha
When We Were Younger & Better 6:52 15 May 2012, 22:52
When We Were Younger & Better 6:52 16 Feb 2012, 17:04
When We Were Younger & Better 6:52 16 Sep 2011, 23:21
When We Were Younger & Better 6:52 16 Sep 2011, 17:25
When We Were Younger & Better 6:52 22 Jun 2011, 21:33
When We Were Younger & Better 6:52 18 May 2011, 18:44
When We Were Younger & Better 6:52 8 Abr 2011, 3:25
When We Were Younger & Better 6:52 21 Feb 2011, 21:48
When We Were Younger & Better 6:52 8 Ene 2011, 1:22
When We Were Younger & Better 6:52 29 Dic 2010, 6:59
When We Were Younger & Better 6:52 25 Nov 2010, 5:34
When We Were Younger & Better 6:52 24 Nov 2010, 20:28
When We Were Younger & Better 6:52 24 Nov 2010, 8:08
When We Were Younger & Better 6:52 24 Nov 2010, 2:28
When We Were Younger & Better 6:52 3 Jul 2010, 2:19
When We Were Younger & Better 6:52 19 May 2010, 22:03
When We Were Younger & Better 6:52 29 Ago 2008, 6:04
When We Were Younger & Better 6:52 12 Ago 2008, 5:49
When We Were Younger & Better 6:52 26 May 2008, 22:10
When We Were Younger & Better 6:52 23 May 2008, 7:34
When We Were Younger & Better 6:52 22 May 2008, 7:54
When We Were Younger & Better 6:52 21 May 2008, 8:49
When We Were Younger & Better 6:52 20 May 2008, 0:36
When We Were Younger & Better 6:52 19 May 2008, 20:16
When We Were Younger & Better 6:52 18 May 2008, 22:17
When We Were Younger & Better 6:52 11 May 2008, 5:10
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.96942538022995, "perplexity": 22761.959895562584}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-27/segments/1435375097354.86/warc/CC-MAIN-20150627031817-00007-ip-10-179-60-89.ec2.internal.warc.gz"}
|
https://www.physicsforums.com/threads/inequality-math-troubles.372222/
|
# Inequality math troubles
1. Jan 24, 2010
### Nikolas15
1. The problem statement, all variables and given/known data
Basically I was reviewing a proof of one of the inequality properties and there was a statement that a <= a , or in other words for ex. 7<=7. So my question is why is that, since 7 is really = to 7, at least I think so.
thanks.
2. Jan 24, 2010
### jgens
Re: inequality
The sign $\leq$ mean less than or equal to. So saying $7 \leq 7$ means that $7$ is less than or equal to $7$. Now, in mathematics (and pretty much everywhere else) the word "or" means that either one or the other holds (and depending on the context, potentially both). Clearly $7 < 7$ is absurd, but $7 = 7$ is true which means that one of the two clauses holds, thus $7 \leq 7$.
Similar Discussions: Inequality math troubles
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9659703373908997, "perplexity": 419.81904250669754}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948513611.22/warc/CC-MAIN-20171211144705-20171211164705-00192.warc.gz"}
|
https://gitter.im/serverless-offline/community?at=60425f8f457d6b4a94a5c121
|
## Where communities thrive
• Join over 1.5M+ people
• Join over 100K+ communities
• Free without limits
##### Activity
Dustin Goodman
@dustinsgoodman
No worries! Thank you :)
David Hérault
@dherault
Does it crash the same way using v6.8.0?
Dustin Goodman
@dustinsgoodman
Pulling it down now to test.
All better in v6.8.0! :D
Closed the above issue with the notes regarding resolution. Thanks for the fix!
tan-t
@tan-t
hi!
I'm using serverless-offline 6.5.0 in my project, and want to test my functions which don't have an HTTP events with invocation endpoint (via aws-cli).
I remember when it was 5.10.0, the serverless-offline plugin just gave me invocation endpoint for all the functions like below:
...
Serverless: Routes for helloworld:
Serverless: POST /{apiVersion}/functions/myservice-dev-helloworld/invocations
...
but with 6.5.0, it seems that the plugin isn't giving invocation endpoints unless the function has an HTTP event.
I don't find any breaking changes or migration guides related to this in v6,
but when I downgraded my serverless-offline to 5.10.0, it started to create endpoints, so I think it is a version problem.
does anyone know how to make it create invocation endpoint for all the functions with serverless-offline v6?
tan-t
@tan-t
sorry I've mistakenly thought that the plugin is not making endpoints but it turns out that it is making them silently. with SLS_DEBUG=* , I can see invocation endpoints and they are working nicely actually.
David Hérault
@dherault
:thumbsup:
Ross Coundon
@rcoundon
I have a serverless.yml with a collection of lambdas. There are currently some API Gateway triggered lambdas and one that is SQS triggered.
The environment is set at the provider level.
I'm running this using serverless-offline and for SQS I'm using the serverless-offline-sqs-esmq plugin.
The peculiar thing is, for the API Gateway triggered lambdas, all the process.env values are set correctly. However, when the SQS triggered lambda runs, none of the environment variables set in serverless.yml are available. Serverless is setting something as I can see that the IS_OFFLINE property is set, just none of my own specified ones.
There is no difference in config other than the event source for the two different types. I've also tried specifying the environment at the function level but this makes no difference.
Has anyone seen this or know what it could be?
Thanks
Gurkiran Singh
@Gurkiran-Singh
Hi,
I am new to Serverless.
I just want to know if there is any way around that we can create s3 buckets using serverless if else plugin.
What I am trying to achieve is that I want to create new s3 buckets as per Stage I gave in if else statement.
Fr example if script(serverless.yml) runs in production stage it should create different s3 bucket and if script runs in development stage then it should create different bucket according to if else statement.
Please let me know if there is any way around to achieve this.
Vishesh M
@visheshCuriousDev
Hi guys !! Is there no hot reload for serverless offline ?
I run "SLS_DEBUG=* sls offline start". It run in debug mode but any changes made to any of the handler files, it does not reload !!! Any idea ?
Jack Fazackerley
@JackFazackerley
Hi I'm running serverless-offline v6.8.0, serverless v2.16.1, Go v1.15.5 and for some reason i can only GET, any time i POST it's as if serverless-offline is blocking me? The connection just hangs and there are no lo gs. The command i'm running is: SLS_DEBUG=* sls --verbose offline start --useDocker. My serverless file looks like:
service: example
frameworkVersion: '2'
plugins:
- serverless-offline
provider:
name: aws
runtime: go1.x
stage: ${self:custom.stage} region:${self:custom.region}
custom:
stage: ${opt:stage, 'develop'} region:${file(env.yml):regions.${self:custom.stage}} package: individually: true exclude: - ./** functions: api: runtime: go1.x handler: bin/api events: - http: cors: origin: '*' path: api method: any package: include: - bin/api environment:${file(env.yml):environment.example}.
I'm unsure if i'm doing anything wrong or it's the versions?
Ahmed Bayoumy
@bayoumymac
Hello, I have been running into this weird error on the second request after running sls offline start
node[91757]: ../src/env-inl.h:1036:void node::Environment::AddCleanupHook(void (*)(void *), void *): Assertion (insertion_info.second) == (true)' failed.
1: 0x10009ce91 node::Abort() [/usr/local/bin/node]
2: 0x10009cd1a node::AppendExceptionLine(node::Environment*, v8::Local<v8::Value>, v8::Local<v8::Message>, node::ErrorHandlingMode) [/usr/local/bin/node]
3: 0x10001f6be node::Environment::AddCleanupHook(void (*)(void*), void*) [/usr/local/bin/node]
4: 0x100008b00 node::AddEnvironmentCleanupHook(v8::Isolate*, void (*)(void*), void*) [/usr/local/bin/node]
5: 0x107903206 WrappedRE2::Init() [/Users/Test/Documents/instagram-bot/serverless/node_modules/re2/build/Release/re2.node]
6: 0x107903f23 node_register_module_v88 [/Users/Test/Documents/instagram-bot/serverless/node_modules/re2/build/Release/re2.node]
7: 0x10007bf86 std::__1::__function::__func<node::binding::DLOpen(v8::FunctionCallbackInfo<v8::Value> const&)::$_0, std::__1::allocator<node::binding::DLOpen(v8::FunctionCallbackInfo<v8::Value> const&)::$_0>, bool (node::binding::DLib*)>::operator()(node::binding::DLib*&&) [/usr/local/bin/node]
9: 0x10021da46 v8::internal::FunctionCallbackArguments::Call(v8::internal::CallHandlerInfo) [/usr/local/bin/node]
10: 0x10021d0c9 v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/usr/local/bin/node]
11: 0x10021c8b0 v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/usr/local/bin/node]
12: 0x1007fd979 Builtins_CEntry_Return1_DontSaveFPRegs_ArgvOnStack_BuiltinExit [/usr/local/bin/node]
Abort trap: 6
that only occurs on the second request, the first request goes through just fine
cdmpicker
@cdmpicker
anyone got serverless offline working with rollup?
Zhani Baramidze
@jbaramidze
hi!,
I've installed serverless offline plugin, all works good, port 3000 got exposed and handles fine, but for some reason it does not seem to expose 3002 to do Lambda invocations by hand. shouldn't it start working automatically?
Andy Swanson
@awswanson
Hello, I have been running into this weird error on the second request after running sls offline start
@bayoumymac Were you able to resolve this issue? The exact same thing is happening to me.
Jack Howard
@JackHowa
is there a good way of mocking alb? I'm seeing urls that aren't matching what I would expect running serverless-offline?
[offline] Lambda Invocation Routes (for AWS SDK or AWS CLI):
* POST http://localhost:3002/2015-03-31/functions/x-api-sandbox-getBundles/invocations
Jack Howard
@JackHowa
^ this is related to issue dherault/serverless-offline#598 if any one is working on this. I saw someone try to bounty it I think
Jack Howard
@JackHowa
also is there a way to specify a different serverless.yml file to run? I haven't been able to find the equivalent of --config equivalent to serverless. thanks for your help!
looks like it's being tracked here by dherault/serverless-offline#1146
Rafael Franco
@rfoel
How about releasing the support for nodejs14.x?
Zac Tolley
@ztolley
Another Vote here for 14. Is there ANY traction on this?
mikemeerschaert
@mikemeerschaert
Is there a way to just run one function with serverless offline when I have many functions defined in my serverless.yml file?
e.g. given my serverless.yml file looks like this, how do I just run functionB. I tried sls offline --functionB but that builds and starts all the functions, and function A takes a very long time to build. Also, I intend to build many more functions in the future, but I want to be able to run a single functional locally while working on it without having to worry about the rest.
service:
name: backend
plugins:
- serverless-webpack
- serverless-offline
provider:
name: aws
region: us-west-1
runtime: nodejs12.x
functions:
functionA:
handler: handler.functionAHandler
events:
- http:
path: functionA
method: post
- http:
path: functionA
method: get
functionB:
handler: handler.functionBHandler
events:
- http:
path: functionB
method: post
- http:
path: functionB
method: get
Jake Wood
@jakemwood
+1 for Node v14, as well as a +1 for dherault/serverless-offline#1175. I'm happy to help with either of these initiatives.
Zac Tolley
@ztolley
At what point do we have to resort to created an official fork?
selected-pixel-jameson
@selected-pixel-jameson
Does anyone know if this repo is still being supported? It seems as if @dherault has not responded since October. I'm just concerned as this is a pretty big part of a lot of peoples development and the repo now hasn't seen a new release for almost 6 months, was monthly prior to 6.8.0. I've created this question, dherault/serverless-offline#1193, on the github repo. First and foremost I hope that @dherault is ok, but would like to know if the community needs to have a discussion as to where this project is going.
Mirko Buholzer
@buholzer
I have an issue when running serverless-offline with the JWT Authorization function. serverless-offline tries to compare my jwt token with jwtOptions.issuerUrl that points to a cloudformation function.
My jwtOptions from createJWTAuthScheme.js look like this:
{
authorizerName: 'cognitoAuthorizer',
name: 'cognitoAuthorizer',
issuerUrl: { 'Fn::Join': [ '', [Array] ] },
audience: [ { 'Fn::ImportValue': 'authUserPoolClientId' } ]
}
Any way I can ignore issuerUrl and audience?
Arslan Ali
@arslanalidev
Hi Guys,
I am facing this below error while I am trying to run the serverless offline.
Error ---------------------------------------------------
Require stack:
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\checkEngine.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\index.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\main.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\Serverless.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\scripts\serverless.js
C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\bin\serverless.js
at require (internal/modules/cjs/helpers.js:74:18)
at Object. (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\checkEngine.js:3:49)
at require (internal/modules/cjs/helpers.js:74:18)
at Object. (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\index.js:13:1)
at require (internal/modules/cjs/helpers.js:74:18)
at Object. (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless-offline\dist\main.js:11:18)
at require (internal/modules/cjs/helpers.js:74:18)
at requireServicePlugin (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:28:12)
at C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:164:20
at Array.map ()
at PluginManager.resolveServicePlugins (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\classes\PluginManager.js:161:8)
at Serverless.init (C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\lib\Serverless.js:171:30)
at async C:\Users\Arslan Titan\AppData\Roaming\npm\node_modules\serverless\scripts\serverless.js:235:7
Tim Stackhouse
@tstackhouse
How does one run the integration tests properly in order to contribute to the project?
I keep getting timeouts because it doesn't look like it's spinning up a local server
Dmitry Kireev
@AutomationD
Something common, prob ably, but Serverless command "offline start" not found. even after installing from "serverless-offline": "github:dherault/serverless-offline#master",
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2944914996623993, "perplexity": 8970.58322244568}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243991648.40/warc/CC-MAIN-20210514060536-20210514090536-00365.warc.gz"}
|
http://math.stackexchange.com/questions/146804/understanding-fundamental-principles-of-counting
|
# Understanding fundamental principles of counting.
There are two fundamental principles of counting; Fundamental principle of addition and fundamental principle of multiplication.
I often got confused applying them. I know that if there are two jobs, say m and n, such that they can be performed independently in $m$ and $n$ ways respectively, then either of the two jobs can be performed in $m+n$ ways and when two jobs are performed in succession, they can be performed in $m\times n$ ways.
My question is how to identify whether jobs are independent or in succession?
Is there any simple way to identify this? Are there any keywords?
Thanks for helping me.
-
@Arturo Dear sir Does independent refers to two different events? – srijan May 18 '12 at 19:49
"Independent" refers to the fact that a choice associated to one set of options does not affect your available choices for the second. For example, say you go to dinner, and want to have an entree and wine with it, and you can choose from among 5 different entrees and 4 different wines. If you are a stickler for wine, your choice of entree may restrict your choice of wine (if you order fish for your entree, you cannot order red wine; if you order meat for your entree, you cannot order white wine), so the choices are not independent. (cont) – Arturo Magidin May 18 '12 at 19:51
So in that case, the product and sum rule don't apply, because one choice affects the other. If the choice of entree does not affect your options of wine (if you don't care about choosing the "wrong wine" for your dish), then the two options are "independent", in that one decision for one of the items does not restrict or expand your available choices for the other. – Arturo Magidin May 18 '12 at 19:52
@srijan: You have to be careful about this keyword business. We do often use such keywords on tests, so that people who don't quite know what's going on but have studied have a reasonable chance to do OK. However, real world problems do not come with keywords carefully attached. – André Nicolas May 18 '12 at 20:23
@AndréNicolas Ok sir i will take care of that. I didn't mean that. – srijan May 18 '12 at 20:41
The distinction is not between "independently" and "in succession". The real distinction is whether you are doing both, or you are doing just one.
The number of ways to do one of the jobs is $m+n$ (either do the first job, which can be done in $m$ ways; or do the second job, which can be done in $n$ ways; total number of ways, $m+n$).
The number of ways of doing both is $mn$, because you have $m$ ways of doing the first job, and $n$ ways of doing the second job, and you have to do both.
Say you have $5$ pants and $3$ shirts. If you are giving one piece of clothing away, then you have $5+3$ ways of deciding which piece to give away. But if you are deciding what to wear, you need to pick a pair of pants, and a shirt. That gives you $5\times 3$ possible combinations. You can make the choices simultaneously (they don't have to be "in succession"). The "independence" clause is jut that choice of shirt/pants should not restrict your choice of pants/shirt (so you don't have a plaid shirt and a pair of striped pants that cannot be worn together...). What you choose for job one does not affect what you can choose for job two.
So you have to think about what you are doing. I'm not sure if there are "keywords" that you should be looking at.
I hadn't ever realized that there was this relationship between "and/exclusive or" and numeric multiplication/addition, even though I knew that, for Boolean algebras, taking "and" and "exclusive or" as $\otimes$ and $\oplus$ makes the algebra a ring... – Thomas Andrews May 18 '12 at 20:32
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6750279068946838, "perplexity": 558.7037122624762}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1409535921869.7/warc/CC-MAIN-20140901014521-00340-ip-10-180-136-8.ec2.internal.warc.gz"}
|
https://deepai.org/publication/the-quadratic-wasserstein-metric-for-inverse-data-matching
|
# The quadratic Wasserstein metric for inverse data matching
This work characterizes, analytically and numerically, two major effects of the quadratic Wasserstein (W_2) distance as the measure of data discrepancy in computational solutions of inverse problems. First, we show, in the infinite-dimensional setup, that the W_2 distance has a smoothing effect on the inversion process, making it robust against high-frequency noise in the data but leading to a reduced resolution for the reconstructed objects at a given noise level. Second, we demonstrate that for some finite-dimensional problems, the W_2 distance leads to optimization problems that have better convexity than the classical L^2 and Ḣ^-1 distances, making it a more preferred distance to use when solving such inverse matching problems.
## Authors
• 1 publication
• 28 publications
• 6 publications
• ### Sliced Gromov-Wasserstein
Recently used in various machine learning contexts, the Gromov-Wasserste...
05/24/2019 ∙ by Titouan Vayer, et al. ∙ 10
• ### A Smoothed Dual Approach for Variational Wasserstein Problems
Variational problems that involve Wasserstein distances have been recent...
03/09/2015 ∙ by Marco Cuturi, et al. ∙ 0
• ### Sensor Clusterization in D-optimal Design in Infinite Dimensional Bayesian Inverse Problems
We investigate the problem of sensor clusterization in optimal experimen...
07/23/2020 ∙ by Yair Daon, et al. ∙ 0
• ### On the well-posedness of Bayesian inverse problems
02/26/2019 ∙ by Jonas Latz, et al. ∙ 0
• ### A Framework for Wasserstein-1-Type Metrics
We propose a unifying framework for generalising the Wasserstein-1 metri...
01/08/2017 ∙ by Bernhard Schmitzer, et al. ∙ 0
• ### The Unbalanced Gromov Wasserstein Distance: Conic Formulation and Relaxation
Comparing metric measure spaces (i.e. a metric space endowed with a prob...
09/09/2020 ∙ by Thibault Séjourné, et al. ∙ 5
• ### Stability of Gibbs Posteriors from the Wasserstein Loss for Bayesian Full Waveform Inversion
Recently, the Wasserstein loss function has been proven to be effective ...
04/07/2020 ∙ by Matthew M. Dunlop, et al. ∙ 0
##### This week in AI
Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.
## 1 Introduction
This paper is concerned with inverse problems where we intend to match certain given data to data predicted by a (discrete or continuous) mathematical model, often called the forward model. To set up the problem, we denote by a function () the input of the mathematical model that we are interested in reconstructing from a given datum . We denote by the forward operator that maps the unknown quantity to the datum , that is
f(m)=g, (1)
where the operator is assumed to be nonlinear in general. We denote by the linearization of the operator at the background . With a bit of abuse of notation, we write to denote a linear inverse problem of the form (1) where . The space of functions where we take our unknown object , denoted by , and datum , denoted by , as well as the exact form of the forward operator , will be given later when we study specific problems.
Inverse problems for (1) are mostly solved computationally due to the lack of analytic inversion formulas. Numerical methods often reformulate the problem as a data matching problem where one takes the solution as the function that minimizes the data discrepancy, measured in a metric , between the model prediction and the measured datum . That is,
m∗=argminm∈MΦ(m),with,Φ(m):=12D2(f(m),g). (2)
The most popular metric used in the past to measure the prediction-data discrepancy is the metric due to its mathematical and computational simplicity. Moreover, it is often the case that a regularization term is added to the mismatch functional to impose extra prior knowledge on the unknown (besides of the fact that it is an element of ) to be reconstructed.
In recent years, the quadratic Wasserstein metric [1, 27, 32] is proposed as an alternative for the metric in solving such inverse data matching problems [6, 7, 13, 18, 20, 19, 22, 23, 29, 34]. Numerical experiments suggest that the quadratic Wasserstein metric has attractive properties for some inverse data matching problems that the classical metric does not have [14, 35]. The objective of this work is trying to understand mathematically these numerical observations reported in the literature. More precisely, we attempt to characterize the numerical inversion of (1) based on the quadratic Wasserstein metric and compare that with the inversion based on the classical metric.
In the rest of the paper, we first briefly review some background materials on the quadratic Wasserstein metric and its connection to inverse data match problems in Section 2. We then study in Section 3 the Fourier domain behavior of the solutions to (1) in the asymptotic regime of the Wasserstein metric: the regime where the model prediction and the datum are sufficiently close. We show that in the asymptotic regime, the Wasserstein inverse solution tends to be smoother than the based inverse solution. We then show in Section 4 that this smoothing effect of the Wasserstein metric also exists in the non-asymptotic regime, but in a less explicit way. In Section 5, we demonstrate, in perhaps overly simplified settings, some advantages of the Wasserstein metric over the metric in solving some important inverse matching problems: inverse transportation, back-projection from (possibly partial) data, and deconvolution of highly concentrated sources. Numerical simulations are shown in Section 6 to demonstrate the main findings of our study. Concluding remarks are offered in Section 7.
## 2 Background and problem setup
Let and
be two probability densities on
that have the same total mass. The square of the quadratic Wasserstein distance between and , denoted as , is defined as
W22(f,g):=infT∈T∫Rd|x−T(x)|2f(x)dx, (3)
where is the set of measure-preserving maps from to . The map that achieves the infimum is called the optimal transport map between and . In the context of (1), the probability density depends on the unknown function . Therefore, can be viewed as a mismatch functional of for solving the inverse problem. Since the quadratic Wasserstein distance is only defined between probability measures of the same total mass, one has to normalize and and turn them into probability densities when applying them to solve inverse matching problems where and
cannot be interpreted as nonnegative probability density functions. This introduces new issues that need to be addressed
[15].
It is well-known by now that the quadratic Wasserstein distance between and is connected to a weighted distance between them; see [32, Section 7.6] and [21, 25]. For any , let be the space of functions
Hs(Rd):={m(x):∥m∥2Hs(Rd):=∫Rd⟨ξ⟩2s|ˆm(ξ)|2dξ<+∞}
where
denotes the Fourier transform of
and . When , is the usual Hilbert space of functions with square integrable derivatives, and . The space with is understood as the dual of . We also introduce the space , , with the (semi-) norm defined through the relation
∥m∥2Hs(Rd)=∥m∥2L2(Rd)+∥m∥2˙Hs(Rd).
The space is defined as the dual of via the norm
∥m∥˙H−s:=sup{|⟨w,m⟩|:∥w∥˙Hs≤1}. (4)
It was shown [32, Section 7.6] that asymptotically is equivalent to , where the subscript indicates that the space is defined with respect to the reference probability measure . To be precise, if is the probability measure and is an infinitesimal perturbation that has zero total mass, then
W2(μ,μ+dπ)=∥dπ∥˙H−1(dμ)+o(dπ). (5)
This fact allows one to show that, for two positive probability measures and with densities and that are sufficiently regular, we have the following non-asymptotic equivalence between and :
c1∥μ−ν∥˙H−1(dμ)≤W2(μ,ν)≤c2∥μ−ν∥˙H−1(dμ), (6)
for some constants and . The second inequality is generic with [25, Theorem 1] while the first inequality, proved in [21, Proposition 2.8] and [25, Theorem 5] independently, requires further that and be bounded from above.
In the rest of this paper, we study numerical solutions to the inverse data matching problem for (1) under three different mismatch functionals:
ΦHs(m)≡12∥f(m)−g∥2Hs:=12∫Rd⟨ξ⟩2s|ˆf(m)(ξ)−ˆg(ξ)|2dξ, (7)
(8)
where , denotes the convolution operation, and
ΦW2(m)≡12W22(f(m),g):=12infT∈T∫Rd|x−T(x)|2f(m(x))dx. (9)
Our main goal is to analyze the differences between the Fourier contents of the inverse matching results, a main motivation for us to choose the Fourier domain definition of the norms. These norms allow us to systematically study: (i) the differences between the (i.e. the special case of of ) and the , with a positive or negative , inversion results; (ii) the differences between and inversion results caused by the weight ; and (iii) the similarities and differences between and inversion results. This is our roadmap toward better understandings of the differences between -based and -based inverse data matching.
###### Remark 2.1.
Note that since the norm is only a shift away from the corresponding norm in the Fourier representation, by replacing with , we do not introduce extra mismatch functionals for those (semi-) norms. We will, however, discuss inversions when we study the corresponding inversions.
###### Remark 2.2.
In the definition of the objective function, we take the weight function such that is consistent with the linearization of [32].
We refer interested readers to [32, 21, 25] for technical discussions on the results in (5) and (6) (under more general settings than what we present here) that connect with . For our purpose, these results say that: (i) in the asymptotic regime when two signals and , both being probability density functions, are sufficiently close to each other, their distance can be well approximated by their distance; and (ii) if , then and vice versa, that is, the exact matching solutions to the model (1), if exists, are global minimizers to both and . However, let us emphasize that the non-asymptotic equivalence in (6) does NOT imply that the functional and (if we define one) have exactly the same optimization landscape. In fact, numerical evidences show that the two functionals have different optimization landscapes that are both quite different from that of the mismatch functional ; see for instance Section (5) for analytical and numerical evidences.
## 3 Frequency responses in asymptotic regime
We first study the Fourier-domain behavior of the solutions to (1) obtained through the minimizing the functionals we introduced in the previous section. At the solution, and are sufficiently close to each other. Therefore their distance can be replaced with their distance according to (5). In the leading order, the solution is simply the solution in this regime.
### 3.1 Hs-based inverse matching for linear problems
Let us start with a linear inverse matching problem given by the model:
Am=gδ, (10)
where denotes the datum in (1) polluted by an additive noise introduced in the measuring process. The subscript is used to denote the size (in appropriate norms to be specified soon) of the noise, that is, the size of . Besides, we assume that is still in the range of the operator . When the model is viewed as the linearization of the nonlinear model (1), should be regarded as the perturbation of the background . The model perturbation is also often denoted as . We assume that the linear operator is diagonal in the Fourier domain, that is, it has the symbol,
ˆA(ξ)∼⟨ξ⟩−α, (11)
for some . This assumption is to make some of the calculations more concise but is not essential as we will comment on later. When the exponent , the operator is “smoothing”, in the sense that it maps a given to an output with better regularity than itself. The inverse matching problem of solving for in (10), on the other hand, is ill-conditioned (so would be the corresponding nonlinear inverse problem if is regarded as the linearization of ). The size of , to some extent, can describe the degree of ill-conditionedness of the inverse matching problem.
We assume a priori that for some . Therefore, could be viewed as an operator . We now look at the inversion of the problem under the () framework.
We seek the solution of the inverse problem as the minimizer of the functional , defined as in (7) with and replaced with . We verify that the Fréchet derivative of at in the direction is given by
Φ′Hs(m)[δm]=∫RdˆA∗(ξ){⟨ξ⟩2s¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯¯[ˆA(ξ)ˆm(ξ)−ˆgδ(ξ)]}ˆδm(ξ)dξ,
where we used to denote the adjoint of the operator . The minimizer of is located at the place where its Fréchet derivative vanishes. Therefore the minimizer solves the following (modified) normal equation at frequency :
ˆA∗(ξ){⟨ξ⟩2sˆA}ˆm=ˆA∗(ξ){⟨ξ⟩2sˆgδ(ξ)}. (12)
The solution at frequency is therefore
ˆm(ξ)=(ˆA∗(ξ)(⟨ξ⟩2sˆA))−1ˆA∗(ξ)(⟨ξ⟩2sˆgδ(ξ)).
We can then perform an inverse Fourier transform to find the solution in the physical space. The result is
m=(A∗PA)−1A∗Pgδ,P:=(I−Δ)s/2, (13)
where the operator is defined through the relation
(I−Δ)s/2m:=F−1(⟨ξ⟩sˆm),
being the inverse Fourier transform, being the usual Laplacian operator, and being the identity operator.
#### Key observations.
Let us first remark that the calculations above can be carried out in the same way if the norm is replaced with the norm. The only changes are that should be replaced with and the operator in has to be replaced with .
When , assuming that , the above solution reduces to the classical least-square solution . Moreover, when is invertible (so will be ), the solution can be simplified to , using the fact that , which is simply the true solution to the original problem (10). Therefore, in the same manner, as the classical least-square method, the least-square method based on the norm does not change the solution to the original inverse problem when it is uniquely solvable. This is, however, not the case for the inversion in general. For instance, inversion only matches the derivative of the predicted data to the measured data.
When , is a differential operator. Applying to the datum amplifies high-frequency components of the datum. When , is a (smoothing) integral operator. Applying to the datum suppresses high-frequency components of the datum. Even though the presence of the operator in will un-do the effect of on the datum in a perfect world (when is invertible, and all calculations are done with arbitrary precision), when operated under a given accuracy, inversion with is less sensitive to high-frequency noise in the data while inversion with is more sensitive to high-frequency noise in the data, compared to the case of (that is, the classical least-square inversion). Therefore, inversion with can be seen as a “preconditioned” (by the operator ) least-square inversion.
### 3.2 Resolution analysis of linear inverse matching
We now perform a simple resolution analysis, following the calculations in [2], on the inverse matching result for the linear model (10).
###### Theorem 3.1.
Let be given as in (11) and an approximation to defined through its symbol:
ˆRc(ξ)∼{⟨ξ⟩α,|ξ|<ξc0,|ξ|>ξc.
Let be the norm of the additive noise in . Then the reconstruction error , with obtained as the minimizer of , is bounded by
∥m−mcδ∥L2≲∥m∥α−sα+β−sHβδβα+β−s, (14)
when we select
⟨ξc⟩−1∼(δ∥m∥−1Hβ)1α+β−s. (15)
###### Proof.
Following classical results in [12], it is straightforward to verify that the difference between the true solution and the approximated noisy solution is
∥m−mcδ∥L2=∥m−Rcgδ∥L2=∥m−Rcg+Rcg−Rcgδ∥L2=∥(I−RcA)m+Rc(g−gδ)∥L2≤∥(I−RcA)m∥L2+∥Rc(g−gδ)∥L2. (16)
We then verify that operators and have the following norms respectively
∥Rc∥Hs↦L2∼⟨ξc⟩α−sand∥(I−RcA)∥Hβ↦L2∼⟨ξc⟩−β.
This allows us to conclude that
∥m−mcδ∥L2≤∥Rc∥Hs↦L2δ+∥(I−RcA)∥Hβ↦L2∥m∥Hβ≲⟨ξc⟩α−sδ+⟨ξc⟩−β∥m∥Hβ.
We can now select , i.e. the relation given in (15), to minimize the error of the reconstruction, which gives the bound in (14). ∎
#### Optimal resolution.
One main message carried by the theorem above is that reconstruction based on the mismatch has a spatial resolution
ε:=⟨ξc⟩−1∼δ1α+β−s,
under the conditions in the theorem. At a fixed noise level , for fixed and , the optimal resolution of the inverse matching result degenerates when gets smaller. The case of corresponds to the usual reconstruction in the framework. The optimal resolution one could get in this case is decided by . When (), the best resolution one could get is better than the case in a perfect world. When , the reconstructions in the framework provides an optimal resolution that is worse than the case. In other words, the reconstructions based on the negative norms appear to be smoother than optimal reconstructions in this case. See Section 6 for numerical examples that illustrate this resolution analysis.
However, we should emphasize that the above simple calculation only provides the best-case scenarios. It does not tell us exactly how to achieve the best results in a general setup (when the symbol of
, i.e. the singular value decomposition of
in the discrete case, is not precisely known). Nevertheless, the guiding principle of the analysis is well demonstrated: least-square with a stronger (than the ) norm yield higher resolution reconstructions while least-square with a weaker norm yield lower (again compared to the case) resolution reconstructions in the best case.
### 3.3 Hs(dμ)-based inverse matching for linear problems
Inverse matching with the weighted norm can be analyzed in the same manner to study the impact of the weight on the inverse matching result. The solution to (10) in this case is sought as the minimizer of the functional defined in (8) with and . This means that the weight in our definition of the objective function.
Following the same calculation as in the previous subsection, we find that the minimizer of the functional solves the following normal equation at frequency :
(17)
where is the adjoint of the operator defined through the relation .
We first observe that the right-hand side of (17) and that of (12) are different. In (12), the -th Fourier mode of the datum is amplified or attenuated, depending on the sign of , by a factor of . While in (17), this mode is further convoluted with other modes of the datum after the amplification/attenuation. The convolution induces mixing between different modes of the datum. Therefore, inverse matching with the weighted norm cannot be done mode by mode as what we did for the unweighted norm, even when we assume that the forward operator is diagonal. However, main effect of the norm on the inversion, the smoothing/sharpening effect introduced by the factor (half of which come from the factor in front of while the other half come from the factor in ), are the same in both the unweighted and the weighted norms.
The inverse matching solution, in this case, written in physical space, is:
m=(A∗PgA)−1A∗Pggδ,Pg:=(I−Δ)s/2ω(I−Δ)s/2. (18)
We can again compare this with the unweighted solution in (13). The only difference is the introduction of the inhomogeneity, which depends on the datum , in the preconditioning operator by replacing it with . When (), and are (local) differential operators. Roughly speaking, compared to , emphasizes places where is small, be reminded that , or the -th order derivative of is large. At those locations, amplifies the same modes of the datum more than does. When , and are non-local operators. The impact of is more global (as we have seen in the previous paragraph in the Fourier domain). It is hard to precisely characterize the impact of without knowing its form explicitly. However, we can still see, for instance, from (17), that the smoother is, the smoother the inverse matching result will be (since has fast decay and the convolution will further smooth out ). If is very rough, say that it behaves like Gaussian noise, then decays very slow. The convolution, in this case, will not smooth out as much as in the previous case. The main effect of on the inverse matching result in this case mainly comes from the norm, not the weight.
###### Remark 3.2.
Thanks to the asymptotic equivalence between and in (5), the smoothing effect we observe in this section for the inverse matching (and therefore inverse matching since is only different from
on the zeroth moment in the Fourier domain) is also what we observe in the
inverse matching. This observation will be demonstrated in more detail in our numerical simulations in Section 6.
### 3.4 Iterative solution of nonlinear inverse matching
The simple analysis in the previous sections based on the linearized quadratic Wasserstein metric, i.e., a weighted norm, on the inverse matching of linear model (10) does not translate directly to the case of inverse matching with the fully nonlinear model (1). Nevertheless, the analysis does provide us some insights.
Let us consider an iterative matching algorithm for the nonlinear problem, starting with a given initial guess , characterized by the following iteration:
mk+1=mk+ℓkζk, k≥0, (19)
where is a chosen descent direction of the objective functional at iteration , and is the step length at this iteration. For simplicity, let us take the steepest descent method where the descent direction is taken as the negative gradient direction. Following the calculations in the previous section, we verify that the Fréchet derivative of at the current iteration in the direction is given by
(20)
assuming that the forward model is Fréchet differentiable at the with derivative . This leads to the following descent direction chosen by a gradient descent method:
ζk=−(f′(mk)∗Pgf′(mk))−1f′(mk)∗Pg(f(mk)−gδ), (21)
Let us compare this with the descent direction resulted from the least-square functional:
ζk=(f′(mk)∗f′(mk))−1f′(mk)∗(f(mk)−g). (22)
We see that the iterative process of the inverse matching can be viewed as a preconditioned version of the corresponding iteration. The preconditioning operator, , depends on the datum but is independent of the iteration. When the iteration is stopped after a finite step, the effect we observed for linear problems, that is, the smoothing effect of in the case of or its de-smoothing effect in the case of , is carried to the solution of nonlinear problems.
#### Wasserstein smoothing in the asymptotic regime.
To summarize, when the model predictions and the measured data are sufficiently close to each other, inverse matching with the quadratic Wasserstein metric, or equivalently the metric, can be viewed as a preconditioned -based inverse matching. The preconditioning operator is roughly the inverse Laplacian operator with a coefficient given by the datum. The optimal resolution of the inversion result from the Wasserstein metric, with data at a given noise level is roughly of the order ( being the order of the operator at the optimal solution and ) instead of as given in the case. The shape of the datum distorts the Wasserstein matching result slightly from the inverse matching result with a regular (semi-) norm.
## 4 Wasserstein iterations in non-asymptotic regime
As we have seen from the previous sections, in the asymptotic regime, the smoothing effect of the quadratic Wasserstein metric in solving inverse matching problems can be characterized relatively precise thanks to the equivalence between and given in (5). The demonstrated smoothing effect makes -based inverse matching very robust against high-frequency noise in the measured data. This phenomenon has been reported in the numerical results published in recent years [6, 13, 14, 34, 35] and is one of the main reasons that is considered as a good alternative for -based matching methods. In this section, we argue that the smoothing effect of can also be observed in the non-asymptotic regime, that is, a regime where signals and are sufficiently far away from each other. The smoothing effect in the non-asymptotic regime implies that the landscape of the objective functional is smoother than that of the classical objective functional.
To see the smoothing effect of in non-asymptotic regime, we analyze the inverse matching procedure described by the iterative scheme (19) for the objective functional , defined in (9). For the sake of being technically correct, we assume that the data we are comparing in this section are sufficiently regular. More precisely, we assume that and for some . We also assume that the map () is Fréchet differentiable at any admissible and denote by the derivative in direction . We can then write down the Fréchet derivative of at the current iteration in the direction following the differentiability result of with respect to [32]. It is,
(23)
where denotes the optimal transport map at iteration (that is, for ), and denotes the derivative of with respect to (not ) in the direction .
Following the optimal transport theorem of Brenier [32], the optimal transport map at the current iteration , , is given as where is the unique (up to a constant) convex solution to the Monge-Ampère equation:
det(D2u(x))=f(mk(x))/g(∇u(x)), u being convex. (24)
Here is the Hessian operator defined through the Hessian matrix (with the notation ).
Let be the Fréchet derivative of at in the direction , we then verify that solves the following second-order elliptic equation to the leading order:
∑ijaij∂2φ∂xi∂xj+∑jbj∂φ∂xj=δf, (25)
where while depend on the dimension. When , () and (). When , we have
aij=g(Tk(x))⎧⎪ ⎪ ⎪ ⎪⎨⎪ ⎪ ⎪ ⎪⎩∂2u∂xk∂xi∂2u∂xj∂xk−∂2u∂xj∂xi∂2u∂xk∂xk,i≠j≠k,∂2u∂xk′∂xk′∂2u∂xk∂xk−∂2u∂xk′∂xk∂2u∂xk∂xk′,i=j≠k≠k′.
Let be the solution to the (adjoint) equation:
∑ijaij∂2ψ∂xi∂xj−∑jbj∂ψ∂xj=−∇⋅((x−Tk(x))f(x)). (26)
It is then straightforward to verify, following standard adjoint state calculations [33], that update direction can be written as
ζk(x)=f′∗(mk)[|x−Tk(x)|22+ψ(x)], (27)
where denotes the adjoint of the operator .
We first observe that unlike in the classical case where is applied directly to the residual , that is, , the descent direction here depends on the model prediction and the datum only implicitly through the transfer map and its derivative with respect to . Only in the asymptotic regime of being very close to can we make the connection between and the normalized residual. This is where the approximation to comes from.
From Caffarelli’s regularity theory (c.g. [32, Theorem 4.14]), which states that when and we have that the Monge-Ampère solution , we see that is at least . Therefore the solution to the adjoint problem, , is in
by standard theory for elliptic partial differential equations when the problem is not degenerate and in
if it is degenerate. Therefore, is one derivative smoother than and (and therefore the residual). This is exactly what the preconditioning operator (with ) did to the residual in the asymptotic regime, for instance, as shown in (13). This shows that inverse matching has smoothing effect even in the non-asymptotic regime.
In one-dimensional case, we can see the smoothing effect more explicitly since we are allowed to construct the optimal map explicitly in this case. Let and be the cumulative density functions for and respectively. The optimal transportation theorem in one-dimensional setting (c.g. [32, Theorem 2.18]) then says that the optimal transportation map from to is given by . This allows us to verify that, the gradient of at in direction , given in (23), is simplified to:
Φ′W2(mk)[δm]=∫R((x−Tk(x))22+pk(+∞)−pk(x))f′(mk)[δm]dx=∫Rf′∗(mk)[(x−Tk(x))22−pk(+∞)+pk(x)]δm(x)dx (28)
where the function is defined as . Therefore the descent direction (27) simplifies to
ζk(x)=f′∗(mk)[(x−Tk(x))22−pk(+∞)+pk(x)]. (29)
It is clear from (29) that the Fréchet derivative of at iteration depends only on the anti-derivatives of , and , through and . Therefore, it is smoother than the Fréchet derivative of in general, whether or not the signals and are close to each other. This shows why the smoothing effect of exists also in non-asymptotic regime.
To provide some numerical evidences, we show in Figure 1 and Figure 2 some gradients of the and objective functions, with respect to the absorption coefficient (i.e. ), for the inverse diffusion problem we study in Section 6.2, in one- and two-dimensional domains and respectively. The synthetic data, generated by applying the forward operator to the true absorption coefficient and then adding multiplicative random noise, contains roughly of random noise. We intentionally choose initial guesses to be relatively far away from the true coefficient so that the model prediction is far from the data to be matched. We are not interested in a direct quantitative comparison between the gradient of the Wasserstein objective function and that of the objective function since we do not have a good basis for the comparison. However, it is evident from these numerical results that the gradient of the Wasserstein functional is smoother, or contains fewer frequencies to be more precise, compared to the corresponding case.
## 5 Wasserstein inverse matching for transportation and convolution
Its robustness against high-frequency noise in data, resulted from its smoothing effect we demonstrated in the previous two sections, is not the only reason why is thought as better than for many inverse data matching problems. We show in this section another advantage of the distance in studying inverse matching problems: its convexity with respect to translation and dilation of signals.
### 5.1 W2 convexity with respect to affine transformations
For a given probability density function with finite moments and , we define:
f(m(x)):=1√|Σ|ϕ(Σ−12(x−λ¯x)). (30)
where with symmetric and positive-definite, and . This is simply a translation (by ) and dilation (by ) of the function . We verify that .
Let be generated from with symmetric and positive-definite. Then we check that the optimal transport map from to is given by . In other words, the function is a convex solution to the Monge-Ampère equation (24) with this pair. This observation allows us to find that,
W22(f,g)=|λ¯x−λg¯xg|2+2(λ¯x−λg¯xg)t(Σ1/2−Σ1/2g)M1+∫Rd|(Σ1/2−Σ1/2g)x|2ϕ(x)dx. (31)
This calculation shows that is convex with respect to and for rather general probability density function .
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9822043776512146, "perplexity": 476.0434944116934}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153474.19/warc/CC-MAIN-20210727170836-20210727200836-00134.warc.gz"}
|
https://www.physicsforums.com/threads/relativity-without-the-aether-pseudoscience.90552/
|
# Relativity without the aether: pseudoscience?
1. Sep 24, 2005
### Aether
Special relativity (SR) SR and Lorentz ether theory (LET) are empirically equivalent systems for interpreting local Lorentz symmetry. These two theories are equally valid, but it is not possible (so far) to demonstrate the truth or falseness of the postulates of either theory over the other by experimentation. Still, a superstition persists in the minds of many that somehow "SR is true, and LET is false". Why isn't "relativity without the aether" fairly described by the term "pseudoscience"?
pseudoscience - Refers to any body of knowledge or practice which purports to be scientific or supported by science but which is judged by the mainstream scientific community to fail to comply with the scientific method.
scientific method n - The principles and empirical processes of discovery and demonstration considered characteristic of or necessary for scientific investigation, generally involving the observation of phenomena, the formulation of a hypothesis concerning the phenomena, experimentation to demonstrate the truth or falseness of the hypothesis, and a conclusion that validates or modifies the hypothesis.
su·per·sti·tion n An irrational belief that an object, action, or circumstance not logically related to a course of events influences its outcome.
1) A belief, practice, or rite irrationally maintained by ignorance of the laws of nature or by faith in magic or chance.
2) A fearful or abject state of mind resulting from such ignorance or irrationality.
3) Idolatry.
2. Sep 24, 2005
### Perspicacious
Special relativity is physics according to the definition of physics whereas adherence to the aether is a religious belief that doesn't generate any physics.
3. Sep 24, 2005
### Aether
Lorentz symmetry, $\eta_{\mu \nu}$, is coordinate independent physics. Adherence to either one of SR, $\Lambda_\nu^\mu$, or LET, $T_\nu^\mu$, to the exclusion of the other is, it seems to me, a religious superstition that doesn't generate any physics (so far).
Last edited: Sep 24, 2005
4. Sep 24, 2005
### Perspicacious
Postulating Lorentz symmetry generates a lot of physics. Taking as an axiom the existence of an aether only produces narrow-mindedness and noise on newsgroups and physics forums. If an axiom doesn't generate any quantifiable predictions, it's worthless and needs to be thrown out.
5. Sep 24, 2005
### Aether
I agree, but what I am in doubt about is that one can truly appreciate Lorentz symmetry without the combined perspectives of SR & LET. Can you actually visualize Lorentz symmetry from both of these perspectives yourself?
SR and LET are empirically equivalent. Name one quantifiable prediction that is generated by postulating that the speed of light is a constant in all inertial frames that is not also generated by postulating an aether.
I'm not saying that LET should replace SR (yet). I'm saying that either one without the other is a potentially misleading representation of Lorentz symmetry.
Last edited: Sep 24, 2005
6. Sep 24, 2005
### Tom Mattson
Staff Emeritus
Because relativity is testable, and it passes every test to which it is subjected.
As for the fact that there are other theories that are empirically equivalent to SR: this situation is not unique to SR. There are also the pilot waves of Bohmian mechanics, which is empirically equivalent to QM. Should QM be considered pseudoscience? Of course not, the question is ridiculous. And it's just as ridiculous for SR.
Honestly Aether, how many threads do you need to push your agenda?
7. Sep 24, 2005
### benjamincarson
8. Sep 24, 2005
### Perspicacious
The Santa Reindeer Postulate
So what! I can create an infinite number of new theories empirically equivalent to SR.
Permit me to illustrate.
Start with SR and create a new theory SR* by adjoining to SR the postulate of Santa Claus and flying reindeer. Realize that SR and SR* are empirically equivalent if we wisely stipulate that Santa Claus and flying reindeer are undetectable. Show the physics of this and how it integrates so naturally with the Santa Reindeer postulate. Estimate the speed and distances covered by Santa Claus on Christmas night. Post a lot of hoopla about it and declare that the physics is unassailable. Also note that the theory has been peer-reviewed and is published on the Fermi National Accelerator Laboratory website.
http://www.fnal.gov/pub/ferminews/santa/
9. Sep 24, 2005
### Aether
I'm not trying to push an agenda. You have my blessing to kill this thread if you don't think that the question is a valid one. If there was a sub-forum to discuss "Lorentz symmetry" that might be a better place to raise this question, but I really am looking for answers.
10. Sep 24, 2005
### Staff: Mentor
The fact that they are equivalent, but that one involves assumptions for which there is no evidence and the other doesn't is precisely what makes one science and the other not.
It's the invisible purple elephant postulate: I can postulate that there is an invisible purple elephant in my garage (caveat: I do not have a garage). But that postulate would not affect how we understand gravity, so it would be useless to include it in our theory of gravity.
Aether theorists do, of course, hope that eventually evidence is found that makes SR and LET not empirically equivalent, but until such evidence is found, the postulate is just a superfluous invisible purple elephant. It's a piggyback theory.
It can also not be denied that the box in which the Aether could possibly be found in has been shrinking as physics advances. It is not unreasonable to operate on the assumption that the box is empty until physicists perform an experiment that does not fit with predictions.
edit: dang, Perspicacious beat me to it with his "Santa and the flying reindeer" postulate. However, that postulate is compatible with my invisible purple elephant theory. So how 'bout we co-author a paper about a combined "Relativity and Santa and the Flying Reindeer and the Invisible Purple Elephant Theory of Gravity"?
Last edited: Sep 24, 2005
11. Sep 24, 2005
### Aether
I think that both theories involve such an assumption: SR assumes that the one-way speed of light is isotropic, but this can not be proven by experiment; LET assumes absolute simultaneity, but this hasn't been proven either (yet). Why would an impartial observer, not from our culture, prefer either one of these theories (today) over the other?
Some people are under the false impression that the speed of light postulate from SR has been "proven by experiment", but that's not true. How can a preference for SR as opposed to SR+LET be a benign choice when it leads so readily to such a misunderstanding?
The box for violations of the rotation and boost invariance components of Lorentz symmetry is shrinking, but that still does not predict which theory (SR or LET) an impartial observer would choose (I would hope that they would either choose SR+LET, or better yet a completely coordinate independent approach). Can you visualize Lorentz symmetry from the LET perspective as well as the SR perspective? It's like starting with both eyes open, and closing your left eye, and that's one perspective; then open your left eye and close your right eye, and that's the other perspective. They are both equally valid. I'm just saying, now open both eyes.
That's fine, but it would apply equally well to SR alone as it does to LET alone if you were an impartial observer. SR+LET has an advantage over either one alone.
12. Sep 24, 2005
### Perspicacious
Forget about Einstein's tortured derivation. There's no reason to base relativity on Einstein's original presentation. There's no need to fixate on the clumsiness of anyone's baby steps.
Special relativity is best understood without the clutter of unnecessary assumptions. Having to assume anything about the one-way speed of light is absolutely unnecessary. The best derivation of special relativity available anywhere to date in terms of sheer elegance, physical intuitiveness and mathematical simplicity is given here:
http://www.everythingimportant.org/relativity/
http://scitation.aip.org/getabs/servlet/GetabsServlet?prog=normal&id=AJPIAS000043000005000434000001
http://arxiv.org/PS_cache/physics/pdf/0302/0302045.pdf
Only the mathematically inept believe that Einstein discovered an admirable derivation of special relativity.
13. Sep 24, 2005
### Staff: Mentor
One thing I've never understood is why people say that the 1-way speed of light cannot be verified by experiment. It should be simple: place two clocks a set distance apart and synchronize them (there are lots of ways to do this, but the simplest would probably be to use a 3rd clock halfway between them to set them). Then fire a laser from one to the other and measure how long it takes to get there.
It has been my understanding that physicists who accept Relativity consider such an experiment to be unnecessary, while aether theorists simply choose not to do it - perhaps because they don't want to see the result.
14. Sep 24, 2005
### JesseM
If a given aether theory makes precisely the same predictions as SR, then it is not really an alternate "theory" at all, it's more akin to one of the various "interpretations" of quantum mechanics. This post from sci.physics.relativity has some good reasons for rejecting such an interpretation on the basis of "elegance" and Occam's razor:
Last edited: Sep 24, 2005
15. Sep 24, 2005
### Aether
The experiment really is just that simple, and anyone can do it, but exactly how you synchronize the clocks is what will always determine the outcome. If we synchronize our clocks so that the speed of light is the same in both (all) directions, then that is Einstein synchronization, relative simultaneity, and SR. However, we are also perfectly free to synchronize them to maintain absolute simultaneity with an "ether" frame, and we will then measure a generally different speed of light going in every direction, and that is LET. What we can not do by experiment(yet) is to say that one synchronization scheme is better than the other.
Such an experiment is unnecessary if you accept relativity on philosophical grounds, but it is not appropriate to then forget exactly how it was that you ever came to accept relativity in the first place and then go on to claim that you have done an experiment and that as a result of that experiment we are all compelled to accept relativity. Aether theorists haven't found a locally preferred frame yet, and that maintains their theory at parity (so far) with special relativity as far as experiments go.
16. Sep 24, 2005
### Perspicacious
No. Actually, they like that method because of the arbitrariness of it. They will quickly point out an explicit assumption in your approach. You're going to end up assuming that light speed from A to B is equal to light speed from B to A. It's better to do an ultraslow clock transport instead.
17. Sep 24, 2005
### Tom Mattson
Staff Emeritus
It's not considered unecessary at all. In fact it's been done!
I think that the most dramatic example is the following:
Alvaeger F.J.M. Farley, J. Kjellman and I Wallin, Physics Letters 12, 260 (1964).
Alvaeger et al measured the speed of gamma rays from decaying pions moving near light speed. The speed of the gamma rays was found to be 'c', modulo a small experimental error.
18. Sep 24, 2005
### JesseM
When you say "yet", are you suggesting you see no reason why some future phenomenon might not respect local Lorentz-invariance? If so, you are underscoring point #6 from the sci.physics.relativity post by Tom Roberts I quoted in my last post:
19. Sep 24, 2005
### Hurkyl
Staff Emeritus
"True" and "false" aren't really (internally) applicable to science. What is true is that each of the postulates of SR are empirically verifiable, whereas the same cannot be said of LET.
No it doesn't.
To even begin to say something like "the one-way speed of light is isotropic", it requires one to specify a coordinate chart.
Coordinare charts are nonphysical choices.
You're thinking about the fact that, among all possible rectilinear coordinate charts we could use, SR chooses to define the orthogonal ones as the inertial reference frames.
This choice is exactly analogous to the fact that we generally like, when studying 3-space, for our x, y, and z axes to be perpendicular.
Special Relativity is generally done in these orthogonal, linear coordinate charts for exactly the same reason that coordinate geometry is generally done with perpendicular axes.
Another good reason to use the orthogonal, linear coordinate charts is that such things can be defined experimentally. (Of course, so can other types of coordinate charts)
Contrast to the choice of charts used by aether theories which invoke some principle of absolute simultaneity which cannot be experimentally determined.
(Since I'm talking about "orthogonal" in the above, that means I'm using some sort of "metric". Of course, I'm using the "metric" of Minowski 4-space, because that's the one that appears in all the physical laws)
20. Sep 24, 2005
### Aether
Here's the first two lines from a paper from Kostelecky & Mewes for example: http://www.citebase.org/cgi-bin/citations?id=oai:arXiv.org:hep-ph/0111026 "Lorentz violation is a promising candidate signal for Planck-scale physics. For instance, it could arise in string theory and is a basic feature of noncommutative field theories...". So, when I say "yet" I simply mean that I am aware of many physicists who expect to find violations eventually.
I know that there are some good points in Tom Robert's articles, and I have read them in the past, but he's talking about a whole infinity of ether theories and I'm just talking about one; LET. So, kindly extract the point from the article that you think applies here.
He (and you by quoting it) seems to think that local Lorentz invariance is synonymous with SR to the exclusion of LET:
"Note that all phenomena discovered since 1905 do indeed exhibit the local Lorentz invariance of SR -- what is happenstance in ether theory was directly predicted by SR."
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8343714475631714, "perplexity": 1225.2770185838683}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988721555.36/warc/CC-MAIN-20161020183841-00347-ip-10-171-6-4.ec2.internal.warc.gz"}
|
https://safetylinks.net/training/safety-management/incident-investigation/
|
Incident Investigator Training
## Course Introduction
Incidents don’t just happen. They are caused! Therefore the overall objective of this course is to present an analytical approach to incident investigations so you can discover the causes of incidents and prevent reoccurrence.
## Course Details
This interactive session provides participants with the knowledge and tools they need to conduct and document formal Accident/ Incident investigations on their sites. This is an advanced course designed for seasoned safety professionals.
## Topics Include
• Section 1: Introduction to incidents
• Section 2: Incident Reporting
• Section 3: Sources of Failure
• Section 4: Primary Response
• Section 5: Investigation process
• Section 6: Analytical techniques (Including Causal Factors Analysis, Change Analysis, Barrier Analysis, Tree analysis, and Alternative Techniques and Computer-Based Simulation)
• Section 7: Recommendations
• Section 8: Feedback and the Presentation of Reports
• Section 9: Monitoring
• Section 10: The six step incident investigation procedure
• Section 11: Case Studies
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8448070287704468, "perplexity": 9372.231793774505}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704833804.93/warc/CC-MAIN-20210127214413-20210128004413-00267.warc.gz"}
|
https://bitbucket.org/pypy/pypy/src/0f1e91da6cb2/pypy/doc/cli-backend.rst?at=default
|
Full commit
# The CLI backend
The goal of GenCLI is to compile RPython programs to the CLI virtual machine.
## Target environment and language
The target of GenCLI is the Common Language Infrastructure environment as defined by the Standard Ecma 335.
While in an ideal world we might suppose GenCLI to run fine with every implementation conforming to that standard, we know the world we live in is far from ideal, so extra efforts can be needed to maintain compatibility with more than one implementation.
At the moment of writing the two most popular implementations of the standard are supported: Microsoft Common Language Runtime (CLR) and Mono.
Then we have to choose how to generate the real executables. There are two main alternatives: generating source files in some high level language (such as C#) or generating assembly level code in Intermediate Language (IL).
The IL approach is much faster during the code generation phase, because it doesn't need to call a compiler. By contrast the high level approach has two main advantages:
• the code generation part could be easier because the target language supports high level control structures such as structured loops;
• the generated executables take advantage of compiler's optimizations.
In reality the first point is not an advantage in the PyPy context, because the flow graph we start from is quite low level and Python loops are already expressed in terms of branches (i.e., gotos).
About the compiler optimizations we must remember that the flow graph we receive from earlier stages is already optimized: PyPy implements a number of optimizations such a constant propagation and dead code removal, so it's not obvious if the compiler could do more.
Moreover by emitting IL instruction we are not constrained to rely on compiler choices but can directly choose how to map CLI opcodes: since the backend often know more than the compiler about the context, we might expect to produce more efficient code by selecting the most appropriate instruction; e.g., we can check for arithmetic overflow only when strictly necessary.
The last but not least reason for choosing the low level approach is flexibility in how to get an executable starting from the IL code we generate:
• write IL code to a file, then call the ilasm assembler;
• directly generate code on the fly by accessing the facilities exposed by the System.Reflection.Emit API.
## Handling platform differences
Since our goal is to support both Microsoft CLR we have to handle the differences between the twos; in particular the main differences are in the name of the helper tools we need to call:
Tool CLR Mono
IL assembler ilasm ilasm2
C# compiler csc gmcs
Runtime ... mono
The code that handles these differences is located in the sdk.py module: it defines an abstract class which exposes some methods returning the name of the helpers and one subclass for each of the two supported platforms.
Since Microsoft ilasm is not capable of compiling the PyPy standard interpreter due to its size, on Windows machines we also look for an existing Mono installation: if present, we use CLR for everything except the assembling phase, for which we use Mono's ilasm2.
## Targeting the CLI Virtual Machine
In order to write a CLI backend we have to take a number of decisions. First, we have to choose the typesystem to use: given that CLI natively supports primitives like classes and instances, ootypesystem is the most natural choice.
Once the typesystem has been chosen there is a number of steps we have to do for completing the backend:
• map ootypesystem's types to CLI Common Type System's types;
• map ootypesystem's low level operation to CLI instructions;
• map Python exceptions to CLI exceptions;
• write a code generator that translates a flow graph into a list of CLI instructions;
• write a class generator that translates ootypesystem classes into CLI classes.
### Mapping primitive types
The rtyper give us a flow graph annotated with types belonging to ootypesystem: in order to produce CLI code we need to translate these types into their Common Type System equivalents.
For numeric types the conversion is straightforward, since there is a one-to-one mapping between the two typesystems, so that e.g. Float maps to float64.
For character types the choice is more difficult: RPython has two distinct types for plain ASCII and Unicode characters (named UniChar), while .NET only supports Unicode with the char type. There are at least two ways to map plain Char to CTS:
• map UniChar to char, thus maintaining the original distinction between the two types: this has the advantage of being a one-to-one translation, but has the disadvantage that RPython strings will not be recognized as .NET strings, since they only would be sequences of bytes;
• map both char, so that Python strings will be treated as strings also by .NET: in this case there could be problems with existing Python modules that use strings as sequences of byte, such as the built-in struct module, so we need to pay special attention.
We think that mapping Python strings to .NET strings is fundamental, so we chose the second option.
### Mapping built-in types
As we saw in section ootypesystem defines a set of types that take advantage of built-in types offered by the platform.
For the sake of simplicity we decided to write wrappers around .NET classes in order to match the signatures required by pypylib.dll:
ootype CLI
String System.String
StringBuilder System.Text.StringBuilder
List System.Collections.Generic.List<T>
Dict System.Collections.Generic.Dictionary<K, V>
CustomDict pypy.runtime.Dict
DictItemsIterator pypy.runtime.DictItemsIterator
Wrappers exploit inheritance for wrapping the original classes, so, for example, pypy.runtime.List<T> is a subclass of System.Collections.Generic.List<T> that provides methods whose names match those found in the _GENERIC_METHODS of ootype.List
The only exception to this rule is the String class, which is not wrapped since in .NET we can not subclass System.String. Instead, we provide a bunch of static methods in pypylib.dll that implement the methods declared by ootype.String._GENERIC_METHODS, then we call them by explicitly passing the string object in the argument list.
### Mapping instructions
PyPy's low level operations are expressed in Static Single Information (SSI) form, such as this:
v2 = int_add(v0, v1)
By contrast the CLI virtual machine is stack based, which means the each operation pops its arguments from the top of the stacks and pushes its result there. The most straightforward way to translate SSI operations into stack based operations is to explicitly load the arguments and store the result into the appropriate places:
LOAD v0
STORE v2
The code produced works correctly but has some inefficiency issues that can be addressed during the optimization phase.
The CLI Virtual Machine is fairly expressive, so the conversion between PyPy's low level operations and CLI instruction is relatively simple: many operations maps directly to the corresponding instruction, e.g int_add and sub.
By contrast some instructions do not have a direct correspondent and have to be rendered as a sequence of CLI instructions: this is the case of the "less-equal" and "greater-equal" family of instructions, that are rendered as "greater" or "less" followed by a boolean "not", respectively.
Finally, there are some instructions that cannot be rendered directly without increasing the complexity of the code generator, such as int_abs (which returns the absolute value of its argument). These operations are translated by calling some helper function written in C#.
The code that implements the mapping is in the modules opcodes.py.
### Mapping exceptions
Both RPython and CLI have their own set of exception classes: some of these are pretty similar; e.g., we have OverflowError, ZeroDivisionError and IndexError on the first side and OverflowException, DivideByZeroException and IndexOutOfRangeException on the other side.
The first attempt was to map RPython classes to their corresponding CLI ones: this worked for simple cases, but it would have triggered subtle bugs in more complex ones, because the two exception hierarchies don't completely overlap.
At the moment we've chosen to build an RPython exception hierarchy completely independent from the CLI one, but this means that we can't rely on exceptions raised by built-in operations. The currently implemented solution is to do an exception translation on-the-fly.
As an example consider the RPython int_add_ovf operation, that sums two integers and raises an OverflowError exception in case of overflow. For implementing it we can use the built-in add.ovf CLI instruction that raises System.OverflowException when the result overflows, catch that exception and throw a new one:
.try
{
ldarg 'x_0'
ldarg 'y_0'
stloc 'v1'
leave __check_block_2
}
catch [mscorlib]System.OverflowException
{
newobj instance void class OverflowError::.ctor()
throw
}
### Translating flow graphs
As we saw previously in PyPy function and method bodies are represented by flow graphs that we need to translate CLI IL code. Flow graphs are expressed in a format that is very suitable for being translated to low level code, so that phase is quite straightforward, though the code is a bit involved because we need to take care of three different types of blocks.
The code doing this work is located in the Function.render method in the file function.py.
First of all it searches for variable names and types used by each block; once they are collected it emits a .local IL statement used for indicating the virtual machine the number and type of local variables used.
Then it sequentially renders all blocks in the graph, starting from the start block; special care is taken for the return block which is always rendered at last to meet CLI requirements.
Each block starts with an unique label that is used for jumping across, followed by the low level instructions the block is composed of; finally there is some code that jumps to the appropriate next block.
Conditional and unconditional jumps are rendered with their corresponding IL instructions: brtrue, brfalse.
Blocks that needs to catch exceptions use the native facilities offered by the CLI virtual machine: the entire block is surrounded by a .try statement followed by as many catch as needed: each catching sub-block then branches to the appropriate block:
# RPython
try:
# block0
...
except ValueError:
# block1
...
except TypeError:
# block2
...
// IL
block0:
.try {
...
leave block3
}
catch ValueError {
...
leave block1
}
catch TypeError {
...
leave block2
}
block1:
...
br block3
block2:
...
br block3
block3:
...
There is also an experimental feature that makes GenCLI to use its own exception handling mechanism instead of relying on the .NET one. Surprisingly enough, benchmarks are about 40% faster with our own exception handling machinery.
### Translating classes
As we saw previously, the semantic of ootypesystem classes is very similar to the .NET one, so the translation is mostly straightforward.
The related code is located in the module class_.py. Rendered classes are composed of four parts:
• fields;
• user defined methods;
• default constructor;
• the ToString method, mainly for testing purposes
Since ootype implicitly assumes all method calls to be late bound, as an optimization before rendering the classes we search for methods that are not overridden in subclasses, and declare as "virtual" only the one that needs to.
The constructor does nothing more than calling the base class constructor and initializing class fields to their default value.
Inheritance is straightforward too, as it is natively supported by CLI. The only noticeable thing is that we map ootypesystem's ROOT class to the CLI equivalent System.Object.
### The Runtime Environment
The runtime environment is a collection of helper classes and functions used and referenced by many of the GenCLI submodules. It is written in C#, compiled to a DLL (Dynamic Link Library), then linked to generated code at compile-time.
The DLL is called pypylib and is composed of three parts:
• a set of helper functions used to implements complex RPython low-level instructions such as runtimenew and ooparse_int;
• a set of helper classes wrapping built-in types
• a set of helpers used by the test framework
The first two parts are contained in the pypy.runtime namespace, while the third is in the pypy.test one.
## Testing GenCLI
As the rest of PyPy, GenCLI is a test-driven project: there is at least one unit test for almost each single feature of the backend. This development methodology allowed us to early discover many subtle bugs and to do some big refactoring of the code with the confidence not to break anything.
The core of the testing framework is in the module pypy.translator.cli.test.runtest; one of the most important function of this module is compile_function(): it takes a Python function, compiles it to CLI and returns a Python object that runs the just created executable when called.
This way we can test GenCLI generated code just as if it were a simple Python function; we can also directly run the generated executable, whose default name is main.exe, from a shell: the function parameters are passed as command line arguments, and the return value is printed on the standard output:
# Python source: foo.py
from pypy.translator.cli.test.runtest import compile_function
def foo(x, y):
return x+y, x*y
f = compile_function(foo, [int, int])
assert f(3, 4) == (7, 12)
# shell
\$ mono main.exe 3 4
(7, 12)
GenCLI supports only few RPython types as parameters: int, r_uint, r_longlong, r_ulonglong, bool, float and one-length strings (i.e., chars). By contrast, most types are fine for being returned: these include all primitive types, list, tuples and instances.
## Installing Python for .NET on Linux
With the CLI backend, you can access .NET libraries from RPython; programs using .NET libraries will always run when translated, but you might also want to test them on top of CPython.
To do so, you can install Python for .NET. Unfortunately, it does not work out of the box under Linux.
To make it work, download and unpack the source package of Python for .NET; the only version tested with PyPy is the 1.0-rc2, but it might work also with others. Then, you need to create a file named Python.Runtime.dll.config at the root of the unpacked archive; put the following lines inside the file (assuming you are using Python 2.7):
<configuration>
<dllmap dll="python27" target="libpython2.7.so.1.0" os="!windows"/>
</configuration>
The installation should be complete now. To run Python for .NET, simply type mono python.exe.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2784358263015747, "perplexity": 2870.9435270628014}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1429246636213.71/warc/CC-MAIN-20150417045716-00149-ip-10-235-10-82.ec2.internal.warc.gz"}
|
http://tex.stackexchange.com/questions/62185/how-to-use-the-translator-package-with-beamer
|
How to use the translator package with beamer?
I'd like to have my headers like Theorem, Definition etc. in a beamer presentation translated to Polish. After skimming through the last pages of beamer user guide, I tried \usepackage[polish]{translator}, but it resulted in "Option clash for package translator". Any ideas what I can do?
-
You need to put your language option amongst the class options where both babel and the translator package (implicitly loaded by beamer) can see it:
\documentclass[polish]{beamer}
\usepackage{babel}
\begin{document}
\begin{frame}
\begin{theorem}[Theorem]
\end{theorem}
\end{frame}
\end{document}
An alternative for your implicit question "How do I pass options to a package that I don't load myself" is to use \PassOptionsToPackage:
\PassOptionsToPackage{polish}{translator}
\documentclass{beamer}
\usepackage[polish]{babel}
-
Thanks! And what if I don't want to use babel? – mbork Jul 4 '12 at 19:00
OK, the \PassOptionsToPackage worked without babel. Incidentally, there was another error: there's a bug in the Polish dictionary for translator; I'll email the maintainer. – mbork Jul 4 '12 at 19:03
And just [polish]{beamer} works even better. Thanks again! – mbork Jul 4 '12 at 19:04
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6187015771865845, "perplexity": 5082.419700405138}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-32/segments/1438042989018.40/warc/CC-MAIN-20150728002309-00033-ip-10-236-191-2.ec2.internal.warc.gz"}
|
https://me.gateoverflow.in/1461/gate2019-me-2-1
|
In matrix equation $[A] \{X\}=\{R\}$,
$[A] = \begin{bmatrix} 4 & 8 & 4 \\ 8 & 16 & -4 \\ 4 & -4 & 15 \end{bmatrix} \{X\} = \begin{Bmatrix} 2 \\ 1 \\ 4 \end{Bmatrix} \text{ and} \{ R \} = \begin{Bmatrix} 32 \\ 16 \\ 64 \end{Bmatrix}$
One of the eigen values of matrix $[A]$ is
1. $4$
2. $8$
3. $15$
4. $16$
recategorized
Let $A$ be a square matrix of order $n$ and $\lambda$ is one of its eigenvalues. Let $X$ be an eigenvector associated to eigenvalue $\lambda$ then we must have equation
$AX = \lambda X$ --------- Equation $(1)$
In Question, It is given that $AX=R$ ----------Equation $(2)$
Now, From Equation $(1)$ and $(2),$ $\lambda X = R$
So, $\lambda \begin{Bmatrix} 2\\ 1\\ 4 \end{Bmatrix} = \begin{Bmatrix} 32\\ 16\\64 \end{Bmatrix}$
$\Rightarrow$ $\lambda \begin{Bmatrix} 2\\ 1\\ 4 \end{Bmatrix} =16 \begin{Bmatrix} 2\\ 1\\4 \end{Bmatrix}$
$\Rightarrow$ $\lambda = 16$
So, Answer should be $(D)$
by (390 points) 1 4
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9731176495552063, "perplexity": 289.77464313389277}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986660323.32/warc/CC-MAIN-20191015205352-20191015232852-00187.warc.gz"}
|
https://dash.harvard.edu/handle/1/7470490?show=full
|
# Radio Frequency Spectroscopy of a Strongly Imbalanced Feshbach-Resonant Fermi Gas
dc.contributor.author Veillette, Martin dc.contributor.author Moon, Eun Gook dc.contributor.author Lamacraft, Austen dc.contributor.author Radzihovsky, Leo dc.contributor.author Sachdev, Subir dc.contributor.author Sheehy, D. E. dc.date.accessioned 2012-01-10T18:05:37Z dc.date.issued 2008 dc.identifier.citation Veillette, Martin, Eun Gook Moon, Austen Lamacraft, Leo Radzihovsky, Subir Sachdev, and D. E. Sheehy. 2008. Radio frequency spectroscopy of a strongly imbalanced Feshbach-resonant Fermi gas. Physical Review A 78(3): 033614. en_US dc.identifier.issn 1094-1622 en_US dc.identifier.uri http://nrs.harvard.edu/urn-3:HUL.InstRepos:7470490 dc.description.abstract A sufficiently large species imbalance (polarization) in a two-component Feshbach resonant Fermi gas is known to drive the system into its normal state. We show that the resulting strongly-interacting state is a conventional Fermi liquid, that is, however, strongly renormalized by pairing fluctuations. Using a controlled 1/N expansion, we calculate the properties of this state with a particular emphasis on the atomic spectral function, the momentum distribution functions displaying the Migdal discontinuity, and the radio frequency (RF) spectrum. We discuss the latter in the light of the recent experiments of Schunck et al. (Science 316, 867 (2007)) on such a resonant Fermi gas, and show that the observations are consistent with a conventional, but strongly renormalized Fermi-liquid picture. en_US dc.description.sponsorship Physics en_US dc.language.iso en_US en_US dc.publisher American Physical Society en_US dc.relation.isversionof doi:10.1103/PhysRevA.78.033614 en_US dc.relation.hasversion http://arxiv.org/abs/0803.2517 en_US dash.license OAP dc.title Radio Frequency Spectroscopy of a Strongly Imbalanced Feshbach-Resonant Fermi Gas en_US dc.type Journal Article en_US dc.description.version Author's Original en_US dc.relation.journal Physical Review A en_US dash.depositing.author Sachdev, Subir dc.date.available 2012-01-10T18:05:37Z
## Files in this item
Files Size Format View
0803.2517v2.pdf 318.8Kb PDF View/Open
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.916330099105835, "perplexity": 23740.06938723618}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886105455.37/warc/CC-MAIN-20170819143637-20170819163637-00493.warc.gz"}
|
https://www.khanacademy.org/math/recreational-math/math-warmup
|
# Math warmups
Introducing key concepts using physical analogies
Community Questions
All content in “Math warmups”
## Expectation warmup
The 'problem of points' is a classic problem Fermat and Pascal famously debated in the 17th century. Their solution to this problem formed the basis of modern day probability theory. Now it's your turn to relive this challenge!
## Random sampling warmup
Introduction to random sampling (also known as the weak law of large numbers)
## Independent events warmup
Introduction to independent events and frequency analysis using histograms.
## Distribution warmup
Introduction to probability distributions, center, spread, and overall shape. In this warmup we will discover the binomial distribution!
## Arithmetic warmups
Arithmetic warmups
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9402281045913696, "perplexity": 4329.435505110813}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398447773.21/warc/CC-MAIN-20151124205407-00309-ip-10-71-132-137.ec2.internal.warc.gz"}
|
http://cgi.10yearsafter.info/dddbpf/OUTPUT.php?ID=1039
|
ҏW@LN^[ꗗ@@@@Pathfinder LN^[f[^x[X
^O ̐X LN^[ SEND ɂđP vC[ Avasnia ŏIXV 2017/12/29 21:08 NX/LN^[x W[ M_ @ ̋ VFAbNX 푰 @i TCY ^ j N 24 g '"@cm ̏d lb.@kg ̐F ̐F ̐F @
\͖ \͒l \Cl ꎞI\͒l ꎞICl ؗ 16 +3 @ @ q 18 +4 @ @ ϋv 12 +1 @ @ m 12 +1 @ @ f 12 +1 @ @ 6 -2 @ @
go qbg|g ől 36(10+18+4+4) _[W @ /݂̃qbg|Cg 1d10 v_[W @ CjVA`uCl 4 = +4 + @ v yqzCl ̑Cl
ړx n 30ft./6sq. ft./0sq. {ړx hp ft./ ft. 20ft. ft. s/@ j o @i
Z\ NXZ\ Z\ Z\Cl \Cl N NXZ\ Zɂè ̑Cl qЈr -1 = yz-2 + @ + @ + @ + 1 qBr 13 = yqz+4 + 4 + +3 + 0 + 2 qyƁr 13 = yqz+4 + 4 + +3 + 0 + 2 qӒr 1 = ymz+1 + @ + @ + @ + @ qRr 4 = yqz+4 + @ + @ + 0 + @ q|\r() -2 = yz-2 + @ + @ + @ + @ q|\r() -2 = yz-2 + @ + @ + @ + @ qwr* @ = ymz+1 + @ + @ + @ + @ qr -2 = yz-2 + @ + @ + @ + @ qwr* 5 = ymz+1 + 1 + +3 + @ + @ qE\r() 1 = yz+1 + @ + @ + @ + @ qE\r() 1 = yz+1 + @ + @ + @ + @ q^ӊŔjr 1 = yz+1 + @ + @ + @ + @ qjr 7 = yz+3 + 1 + +3 + 0 + @ qr() 1 = ymz+1 + @ + @ + @ + @ qr() 1 = ymz+1 + @ + @ + @ + @ qr() 1 = ymz+1 + @ + @ + @ + @ qr 8 = yz+1 + 4 + +3 + @ + @ qu͉r* @ = yqz+4 + @ + @ + 0 + @ qEopr 4 = yqz+4 + @ + @ + 0 + @ qmor 8 = yz+1 + 4 + +3 + @ + @ qmFMr* @ = ymz+1 + @ + @ + @ + @ qmFHwr* @ = ymz+1 + @ + @ + @ + @ qmFEr* @ = ymz+1 + @ + @ + @ + @ qmFRr* 8 = ymz+1 + 4 + +3 + @ + @ qmF@r* @ = ymz+1 + @ + @ + @ + @ qmF_wr* @ = ymz+1 + @ + @ + @ + @ qmFݼޮݒTr* 5 = ymz+1 + 1 + +3 + @ + @ qmFnr* @ = ymz+1 + @ + @ + @ + @ qmFnr* 5 = ymz+1 + 1 + +3 + @ + @ qmFjr* @ = ymz+1 + @ + @ + @ + @ qÁr 1 = yz+1 + @ + @ + @ + @ q̑Ɓr* @ = yqz+4 + @ + @ + 0 + @ qor 10 = yz+3 + 4 + +3 + 0 + @ qgr -2 = yz-2 + @ + @ + @ + @ q͂r -2 = yz-2 + @ + @ + @ + @ qsr 4 = yqz+4 + @ + @ + 0 + @ qϑr -2 = yz-2 + @ + @ + @ + @ q@ugpr* @ = yz-2 + @ + @ + @ + @ *K̂݁ByzyqzZɂ锻yieBKp
`b A[}[NX 18 =PO+ 5 + @ + 3 + @ + @ + @ + @ v Z{[iX {[iX yqzCl TCYCl O{[iX {[iX ̑{[iX
ڐG A[}[NX 13 A[}[NX @ C
Z[BOEX[ v {Z[ \Cl @ɂCl ̑Cl ꎞICl 挒yϋv́z 6 = 4 + +1 + 1 + @ + @ yq́z 9 = 4 + +4 + 1 + @ + @ ӎuyf́z 3 = 1 + +1 + 1 + @ + @
@@@@@{U{[iX@@@@@ 4 @@R@@ @
Z{[iX 7 = 4 + +3 + @ v {U{[iX yzCl TCYCl
Zhl 21 = 4 + +3 + +4 + @ +10 v {U{[iX yzCl yqzCl TCYCl
U1 U{[iX NeBJ d R|WbgEO{E 9 x3 3 ^Cv ˒ ̑Ee _[W h 110ft 100gp+2000gp+300gp=2400gp 1d8+3+1
U2 U{[iX NeBJ d R|WbgEO{Eߋ 10 x3 @ ^Cv ˒ ̑Ee _[W @ 30ft @ 1d8+3+1+1
U3 U{[iX NeBJ d @ @ @ @ ^Cv ˒ ̑Ee _[W @ @ @ @
U4 U{[iX NeBJ d @ @ @ @ ^Cv ˒ ̑Ee _[W @ @ @ @
U5 U{[iX NeBJ d @ @ @ @ ^Cv ˒ ̑Ee _[W @ @ @ @
U6 U{[iX NeBJ d @ @ @ @ ^Cv ˒ ̑Ee _[W @ @ @ @
ZEhACe AC{[iX ^Cv ACւ́yqz{[iX yieB psm d Ȇ ~XVc+1 5 yZ 6 @ 10 10 2100gp N[NEIEWX^X+1 @ @ @ @ 1 1000gp @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ v 5 @ 6 0 10 11 @
i ACe d A[(100)5gp 15 W[p(9gp) w܁AgїpQAxg|[`AΑłƑłAS̐[AgѐHꎮA[vAi~1jAۑHi5jA 19 \vf|[`(5gp) 2 h:OXgC_[(25`[W)375gp ACE[Eg[`75gp ^dʂ̍v(d݊) 50.66
(cp)
@
(sp)
@
(gp)
33
(pp)
@
yd 76 Ɏグőd 230 d 153 nʂ玝グődQ 460 dd 230 ődT 1150
Z sˌt ߐڐ퓬ɎQĂGɑāAʏ|4̃yieBɁAuˌ蓊łB sߋˌt 30tB[gȓ̓GΏۂƂuU̍U[ƃ_[WE[Ɂ{1̃{[iXB sˁt uőS͍UہA1̒ljU邱ƂłBȀꍇׂĂ̍Ú|2̃yieBB sv́t N͈ȉ̔Ɂ{4̃{[iXFv_[WɒRۂ́qjrA𑱂邽߂́yϋv́zAsRɂv_[WȂ߂́yϋv́zA~߂ۂ́yϋv́zAQ⊉ɂv_[WȂ߂́yϋv́zAMCCv_[WȂ߂̊挒Z[AсAɂ_[WɒR邽߂̊挒Z[B \ y푰z ͂ޔ Ɩh̏KnF SĂ̒PƌRp yZAZAi^[EV[h͏j yW[z ǐ 쐶Ƃ̋ W[̏WF2 퓬X^CZF|p sߋˌt sv́t ӂȒn`FX n`ƂJ o_ ̃x @ @
OBeBE{E 1ANV 1/x A[1TCYޑ傫̂悤Ƀ_[W^B 2d6ɂ yLN^[z k}㏸鑖ҁl cAN͔ђˊ댯ȍɋ삯AȂȂNVԂ̂ɏ\ȎRĂB̏ꏊł炾B vFN͕toێ邽߂Wv邽߂́qyƁrɁ{2̓{[iXB܂qyƁr͌ÑNXZ\ƂȂB krҁl N͑fꂽŐBāAbĂ炤߂ɋЂ\͂ɑiȂȂȂBŃqЈrɁ{1̓{[iXBāqЈr͏ɃNXZ\ƂĈB
ݒȂ F@iAʌAX
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9772534966468811, "perplexity": 19540.58729005647}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084889617.56/warc/CC-MAIN-20180120122736-20180120142736-00444.warc.gz"}
|
https://cabal.readthedocs.io/en/3.6/cabal-commands.html
|
# 6. cabal-install Commands¶
We now give an in-depth description of all the commands, describing the arguments and flags they accept.
## 6.1. cabal v2-configure¶
cabal v2-configure takes a set of arguments and writes a cabal.project.local file based on the flags passed to this command. cabal v2-configure FLAGS; cabal v2-build is roughly equivalent to cabal v2-build FLAGS, except that with v2-configure the flags are persisted to all subsequent calls to v2-build.
cabal v2-configure is intended to be a convenient way to write out a cabal.project.local for simple configurations; e.g., cabal v2-configure -w ghc-7.8 would ensure that all subsequent builds with cabal v2-build are performed with the compiler ghc-7.8. For more complex configuration, we recommend writing the cabal.project.local file directly (or placing it in cabal.project!)
cabal v2-configure inherits options from Cabal. semantics:
• Any flag accepted by ./Setup configure.
• Any flag accepted by cabal configure beyond ./Setup configure, namely --cabal-lib-version, --constraint, --preference and --solver.
• Any flag accepted by cabal install beyond ./Setup configure.
• Any flag accepted by ./Setup haddock.
The options of all of these flags apply only to local packages in a project; this behavior is different than that of cabal install, which applies flags to every package that would be built. The motivation for this is to avoid an innocuous addition to the flags of a package resulting in a rebuild of every package in the store (which might need to happen if a flag actually applied to every transitive dependency). To apply options to an external package, use a package stanza in a cabal.project file.
## 6.2. cabal v2-update¶
cabal v2-update updates the state of the package index. If the project contains multiple remote package repositories it will update the index of all of them (e.g. when using overlays).
Some examples:
$cabal v2-update # update all remote repos$ cabal v2-update head.hackage # update only head.hackage
## 6.3. cabal v2-build¶
cabal v2-build takes a set of targets and builds them. It automatically handles building and installing any dependencies of these targets.
A target can take any of the following forms:
• A package target: package, which specifies that all enabled components of a package to be built. By default, test suites and benchmarks are not enabled, unless they are explicitly requested (e.g., via --enable-tests.)
• A component target: [package:][ctype:]component, which specifies a specific component (e.g., a library, executable, test suite or benchmark) to be built.
• All packages: all, which specifies all packages within the project.
• Components of a particular type: package:ctypes, all:ctypes: which specifies all components of the given type. Where valid ctypes are:
• libs, libraries,
• flibs, foreign-libraries,
• exes, executables,
• tests,
• benches, benchmarks.
• A module target: [package:][ctype:]module, which specifies that the component of which the given module is a part of will be built.
• A filepath target: [package:][ctype:]filepath, which specifies that the component of which the given filepath is a part of will be built.
In component targets, package: and ctype: (valid component types are lib, flib, exe, test and bench) can be used to disambiguate when multiple packages define the same component, or the same component name is used in a package (e.g., a package foo defines both an executable and library named foo). We always prefer interpreting a target as a package name rather than as a component name.
Some example targets:
$cabal v2-build lib:foo-pkg # build the library named foo-pkg$ cabal v2-build foo-pkg:foo-tests # build foo-tests in foo-pkg
$cabal v2-build src/Lib.s # build the library component to # which "src/Lib.hs" belongs$ cabal v2-build app/Main.hs # build the executable component of
# "app/Main.hs"
$cabal v2-build Lib # build the library component to # which the module "Lib" belongs Beyond a list of targets, cabal v2-build accepts all the flags that cabal v2-configure takes. Most of these flags are only taken into consideration when building local packages; however, some flags may cause extra store packages to be built (for example, --enable-profiling will automatically make sure profiling libraries for all transitive dependencies are built and installed.) In addition cabal v2-build accepts these flags: • --only-configure: When given we will forego performing a full build and abort after running the configure phase of each target package. ## 6.4. cabal v2-repl¶ cabal v2-repl TARGET loads all of the modules of the target into GHCi as interpreted bytecode. In addition to cabal v2-build’s flags, it takes an additional --repl-options flag. To avoid ghci specific flags from triggering unneeded global rebuilds these flags are now stripped from the internal configuration. As a result --ghc-options will no longer (reliably) work to pass flags to ghci (or other repls). Instead, you should use the new --repl-options flag to specify these options to the invoked repl. (This flag also works on cabal repl and Setup repl on sufficiently new versions of Cabal.) Currently, it is not supported to pass multiple targets to v2-repl (v2-repl will just successively open a separate GHCi session for each target.) It also provides a way to experiment with libraries without needing to download them manually or to install them globally. This command opens a REPL with the current default target loaded, and a version of the vector package matching that specification exposed. $ cabal v2-repl --build-depends "vector >= 0.12 && < 0.13"
Both of these commands do the same thing as the above, but only exposes base, vector, and the vector package’s transitive dependencies even if the user is in a project context.
$cabal v2-repl --ignore-project --build-depends "vector >= 0.12 && < 0.13"$ cabal v2-repl --project='' --build-depends "vector >= 0.12 && < 0.13"
This command would add vector, but not (for example) primitive, because it only includes the packages specified on the command line (and base, which cannot be excluded for technical reasons).
$cabal v2-repl --build-depends vector --no-transitive-deps ## 6.5. cabal v2-run¶ cabal v2-run [TARGET [ARGS]] runs the executable specified by the target, which can be a component, a package or can be left blank, as long as it can uniquely identify an executable within the project. Tests and benchmarks are also treated as executables. See the v2-build section for the target syntax. Except in the case of the empty target, the strings after it will be passed to the executable as arguments. If one of the arguments starts with - it will be interpreted as a cabal flag, so if you need to pass flags to the executable you have to separate them with --. $ cabal v2-run target -- -a -bcd --argument
v2-run also supports running script files that use a certain format. With a script that looks like:
#!/usr/bin/env cabal
{- cabal:
build-depends: base ^>= 4.11
, shelly ^>= 1.8.1
-}
main :: IO ()
main = do
...
It can either be executed like any other script, using cabal as an interpreter, or through this command:
$cabal v2-run script.hs$ cabal v2-run script.hs -- --arg1 # args are passed like this
## 6.6. cabal v2-freeze¶
cabal v2-freeze writes out a freeze file which records all of the versions and flags that are picked by the solver under the current index and flags. Default name of this file is cabal.project.freeze but in combination with a --project-file=my.project flag (see project-file) the name will be my.project.freeze. A freeze file has the same syntax as cabal.project and looks something like this:
constraints: HTTP ==4000.3.3,
HTTP +warp-tests -warn-as-error -network23 +network-uri -mtl1 -conduit10,
QuickCheck ==2.9.1,
QuickCheck +templatehaskell,
-- etc...
For end-user executables, it is recommended that you distribute the cabal.project.freeze file in your source repository so that all users see a consistent set of dependencies. For libraries, this is not recommended: users often need to build against different versions of libraries than what you developed against.
## 6.7. cabal v2-bench¶
cabal v2-bench [TARGETS] [OPTIONS] runs the specified benchmarks (all the benchmarks in the current package by default), first ensuring they are up to date.
## 6.8. cabal v2-test¶
cabal v2-test [TARGETS] [OPTIONS] runs the specified test suites (all the test suites in the current package by default), first ensuring they are up to date.
## 6.9. cabal v2-haddock¶
cabal v2-haddock [FLAGS] [TARGET] builds Haddock documentation for the specified packages within the project.
If a target is not a library haddock-benchmarks, haddock-executables, haddock-internal, haddock-tests will be implied as necessary.
## 6.10. cabal v2-exec¶
cabal v2-exec [FLAGS] [--] COMMAND [--] [ARGS] runs the specified command using the project’s environment. That is, passing the right flags to compiler invocations and bringing the project’s executables into scope.
## 6.11. cabal v2-install¶
cabal v2-install [FLAGS] PACKAGES builds the specified packages and symlinks/copies their executables in installdir (usually ~/.cabal/bin).
For example this command will build the latest cabal-install and symlink its cabal executable:
$cabal v2-install cabal-install In addition, it’s possible to use cabal v2-install to install components of a local project. For example, with an up-to-date Git clone of the Cabal repository, this command will build cabal-install HEAD and symlink the cabal executable: $ cabal v2-install exe:cabal
Where symlinking is not possible (eg. on some Windows versions) the copy method is used by default. You can specify the install method by using --install-method flag:
$cabal v2-install exe:cabal --install-method=copy --installdir=$HOME/bin
Note that copied executables are not self-contained, since they might use data-files from the store.
### 6.11.1. Adding libraries to GHC package environments¶
It is also possible to “install” libraries using the --lib flag. For example, this command will build the latest Cabal library and install it:
$cabal v2-install --lib Cabal This works by managing GHC package environment files. By default, it is writing to the global environment in ~/.ghc/$ARCH-$OS-$GHCVER/environments/default. v2-install provides the --package-env flag to control which of these environments is modified.
This command will modify the environment file in the current directory:
$cabal v2-install --lib Cabal --package-env . This command will modify the environment file in the ~/foo directory: $ cabal v2-install --lib Cabal --package-env foo/
Do note that the results of the previous two commands will be overwritten by the use of other v2-style commands, so it is not recommended to use them inside a project directory.
This command will modify the environment in the local.env file in the current directory:
$cabal v2-install --lib Cabal --package-env local.env This command will modify the myenv named global environment: $ cabal v2-install --lib Cabal --package-env myenv
If you wish to create a named environment file in the current directory where the name does not contain an extension, you must reference it as ./myenv.
You can learn more about how to use these environments in this section of the GHC manual.
## 6.12. cabal v2-clean¶
cabal v2-clean [FLAGS] cleans up the temporary files and build artifacts stored in the dist-newstyle folder.
By default, it removes the entire folder, but it can also spare the configuration and caches if the --save-config option is given, in which case it only removes the build artefacts (.hi, .o along with any other temporary files generated by the compiler, along with the build output).
## 6.13. cabal v2-sdist¶
cabal v2-sdist [FLAGS] [TARGETS] takes the crucial files needed to build TARGETS and puts them into an archive format ready for upload to Hackage. These archives are stable and two archives of the same format built from the same source will hash to the same value.
cabal v2-sdist takes the following flags:
• -l, --list-only: Rather than creating an archive, lists files that would be included. Output is to stdout by default. The file paths are relative to the project’s root directory.
• -o, --output-directory: Sets the output dir, if a non-default one is desired. The default is dist-newstyle/sdist/. --output-directory - will send output to stdout unless multiple archives are being created.
• --null-sep: Only used with --list-only. Separates filenames with a NUL byte instead of newlines.
v2-sdist is inherently incompatible with sdist hooks (which were removed in Cabal-3.0), not due to implementation but due to fundamental core invariants (same source code should result in the same tarball, byte for byte) that must be satisfied for it to function correctly in the larger v2-build ecosystem. autogen-modules is able to replace uses of the hooks to add generated modules, along with the custom publishing of Haddock documentation to Hackage.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5391758680343628, "perplexity": 8673.687089431296}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320303356.40/warc/CC-MAIN-20220121101528-20220121131528-00651.warc.gz"}
|
http://kitchingroup.cheme.cmu.edu/blog/2015/04/24/Commenting-in-org-files/
|
## Commenting in org-files
| categories: org | tags: | View Comments
There was an interesting discussion on the org-mode mail list about putting comments in org files. Eric Fraga suggested using inline tasks, and customizing the export of them so they make a footnote, or use the todonotes package (suggested by Marcin Borkowski). Here is Eric's export. A big advantage of this is integration with the Agenda, so you can see what there is todo in your document.
(setq org-inlinetask-export-templates
'((latex "%s\\footnote{%s\\\\ %s}\\marginpar{\\fbox{\\thefootnote}}"
'((unless
(eq todo "")
(format "\\fbox{\\textsc{%s%s}}" todo priority))
Eric Abrahamsen suggested an idea to use a link syntax. I like the idea a lot, so here we develop some ideas. A link has two parts, the path, and description. A simple comment would just be a simple link, probably in double square brackets so you can have spaces in your comment. COMMENT It might be feasible to use COMMENT the description to "mark text" that the comment refers to. The remaining question is what functionality should our link have when you click on it, and how to export it. For functionality, a click will show the comment in the minibuffer and offer to delete it. For export, for now we will make it export with todonotes in LaTeX, and as a red COMMENT with a tooltip in html. To use this, you need to have the LaTeX package todonotes included in your org file.
(org-add-link-type
"comment"
(let ((elm (org-element-context))
(use-dialog-box nil))
(when (y-or-n-p "Delete comment? ")
(setf (buffer-substring
(org-element-property :begin elm)
(org-element-property :end elm))
(cond
((org-element-property :contents-begin elm)
(buffer-substring
(org-element-property :contents-begin elm)
(org-element-property :contents-end elm)))
(t
""))))))
(lambda (keyword desc format)
(cond
((eq format 'html)
(format "<font color=\"red\"><abbr title=\"%s\" color=\"red\">COMMENT</abbr></font> %s" keyword (or desc "")))
((eq format 'latex)
(format "\\todo{%s}{%s}" keyword (or desc ""))))))
It would be convenient to have a quick function for adding a comment to some highlighted text.
(defun add-comment (begin end)
(interactive "r")
(if (region-active-p)
(let ((selected-text (buffer-substring begin end)))
(setf (buffer-substring begin end)
(format "[[comment:%s][%s]]"
(insert (format "[[comment:%s]]" (read-input "Comment: ")))))
Test 1: COMMENT
COMMENT Test 2
That is it. I could see a few other enhancements that might be very useful, e.g. a command to list all the comments, remove all the comments, etc… I am pretty satisfied with this for now though.
## 1 An updated approach to comments.
Rainer asked about making some comments inline. It would be nice if a single link syntax could accommodate both styles of comments. I previously developed an approach to extend links with attributes (http://kitchingroup.cheme.cmu.edu/blog/2015/02/05/Extending-the-org-mode-link-syntax-with-attributes/ ), which I will reuse here for that purpose. The idea is to add an ":inline" attribute to change the export behavior. We only modify the LaTeX export here.
(org-add-link-type
"comment"
(let ((elm (org-element-context))
(use-dialog-box nil))
(when (y-or-n-p "Delete comment? ")
(setf (buffer-substring
(org-element-property :begin elm)
(org-element-property :end elm))
(cond
((org-element-property :contents-begin elm)
(buffer-substring
(org-element-property :contents-begin elm)
(org-element-property :contents-end elm)))
(t
""))))))
;; format function
(lambda (path description format)
(let* ((data (read (concat "(" path ")")))
(plist (cdr data)))
(cond
((eq format 'html)
(format "<font color=\"red\"><abbr title=\"%s\" color=\"red\">COMMENT</abbr></font> %s" path (or description "")))
((eq format 'latex)
(format "\\todo%s{%s}%s"
(if (-contains? data :inline) "[inline]" "")
(mapconcat (lambda (s)
(format "%s" s))
(-remove-item :inline data) " ")
(if description (format "{%s}" description) "")))))))
Here are some examples of the syntax:
[[comment: :inline the rest of your text]]
[[comment:Some text you want to highlight]]
[[comment:Some text you want to highlight :inline]]
It doesn't matter where the :inline attribute is added. This seems to work pretty well.
We can modify our convenience function to allow us to use a prefix arg to make the comment inline. Here is one way to do it.
(defun add-comment (begin end &optional arg)
"Comment the region. With a prefix ARG, make the comment inline."
(interactive (list (region-beginning)
(region-end)
current-prefix-arg))
(let ((inline (if arg ":inline " "")))
(if (region-active-p)
(let ((selected-text (buffer-substring begin end)))
(setf (buffer-substring begin end)
(format
"[[comment:%s%s][%s]]"
inline
(insert (format
"[[comment:%s%s]]"
inline
add-comment
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3974528908729553, "perplexity": 14995.038746423834}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583879117.74/warc/CC-MAIN-20190123003356-20190123025356-00529.warc.gz"}
|
https://www.clutchprep.com/chemistry/practice-problems/77006/calculate-the-molarity-of-the-two-solutions-a-0-300-mol-of-naoh-in-1-80-l-of-sol
|
🤓 Based on our data, we think this question is relevant for Professor Hillyard's class at CSU CHICO.
# Solution: Calculate the molarity of the two solutions. a) 0.300 mol of NaOH in 1.80 L of solution. b) 14.7 g of NaCI in 719 mL of solution.
###### Problem
Calculate the molarity of the two solutions.
a) 0.300 mol of NaOH in 1.80 L of solution.
b) 14.7 g of NaCI in 719 mL of solution.
View Complete Written Solution
|
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9243021011352539, "perplexity": 8772.243897028486}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585371826355.84/warc/CC-MAIN-20200408233313-20200409023813-00260.warc.gz"}
|
https://www.quantopian.com/posts/pipeline-preview-overview-of-pipeline-content-easy-to-add-to-your-backtest
|
Pipeline Preview: Overview of pipeline content easy to add to your backtest (log preview any series or dataframe)
It's good to know what values you're working with. Try this to keep an eye on them.
Provides an overview of pipeline output (or any series or dataframe) one time at the beginning of a backtest (or anywhere).
Or set to run every day or conditionally.
It is simpler than it looks, just set its main argument to match your pipeline output (or any series or dataframe), instead of context.out currently.
To limit logging to only selected columns, set their names in the 'fields' list arguement like the line currently commented out.
If you have long column names, for vertical alignment, you might find it useful to simply shorten them.
Also this shows starting capital and start/stop dates like this:
2014-10-29 06:31 log_data:117 INFO \$10,000,000 2014-10-29 to 2018-03-28
Code
Updated 2019-10-24
def log_data(context, z, num, fields=None):
'''
if not len(z):
log.info('Empty pipe')
return
try: context.log_data_done
except:
context.log_data_done = 1
# Options
log_nan_only = 0 # Only log if nans are present.
show_sectors = 0 # If sectors, see them or not.
show_sorted_details = 1 # [num] high & low securities sorted, each column.
padmax = 6 # num characters for each field, starting point.
def out(lines): # log data lines of output efficiently
buffer_len = 1024 # each group
chunk = ':'
for line in lines:
if line is None or not len(line):
continue # skip if empty string for example
if len(chunk) + len(line) < buffer_len:
# Add to chunk if will still be under buffer_len
chunk += '\n{}'.format(line)
else: # Or log chunk and start over with new line.
log.info(chunk)
chunk = ':\n{}'.format(line)
if len(chunk) > 2: # if anything remaining
log.info(chunk)
if 'dict' in str(type(z)):
log.info('Not set up to handle a dictionary, only dataframe & series, bailing out of log_data()')
return
elif 'MultiIndex' in str(type(z.index)):
log.info('Found MultiIndex, not set up to handle it, bailing out of log_data()')
return
# Change index to just symbols for readability, meanwhile, right-aligned
z = z.rename(index=dict(zip(z.index.tolist(), [i.symbol.rjust(6) for i in z.index.tolist()])))
# Series ......
if 'Series' in str(type(z)): # is Series, not DataFrame
nan_count = len(z[z != z])
nan_count = 'NaNs {}/{}'.format(nan_count, len(z)) if nan_count else ''
if (log_nan_only and nan_count) or not log_nan_only:
log.info('{}{}{} {}'.format(
nan_count
))
log.info('Low\n{}' .format(z.sort_values(ascending=False).tail(num)))
return
# DataFrame ......
content_min_max = [ ['','min','mean','max',''] ] ; content = []
for col in z.columns:
try: z[col].max()
except: continue # skip non-numeric
if col == 'sector' and not show_sectors: continue
nan_count = len(z[col][z[col] != z[col]])
nan_count = 'NaNs {}/{}'.format(nan_count, len(z)) if nan_count else ''
if len(str(z[col].max())) > 8 and 'float' in str(z[col].dtype):
z[col] = z[col].round(6) # Reduce number of decimal places for floating point values
if 'float' in str(z[col].dtype): mean_ = str(round(z[col].mean(), 6))
elif 'int' in str(z[col].dtype): mean_ = str(round(z[col].mean(), 1))
content_min_max.append([col, str(z[col] .min()), mean_, str(z[col] .max()), nan_count])
if log_nan_only and nan_count or not log_nan_only:
log.info('Rows: {} Columns: {}'.format(z.shape[0], z.shape[1]))
if len(z.columns) == 1: content.append('Rows: {}'.format(z.shape[0]))
paddings = [6 for i in range(4)]
for lst in content_min_max: # set max lengths
i = 0
for val in lst[:4]: # value in each sub-list
i += 1
content.append(('{}{}{}{}{}'.format(
''
)))
for lst in content_min_max[1:]: # populate content using max lengths
content.append(('{}{}{}{} {}'.format(
lst[4],
)))
out(content)
if not show_sorted_details: return
if len(z.columns) == 1: return # skip detail if only 1 column
if fields == None: details = z.columns
content = []
for detail in details:
if detail == 'sector' and not show_sectors: continue
lo = z[details].sort_values(by=detail, ascending=False).tail(num)
content.append(('_ _ _ {} _ _ _' .format(detail)))
content.append(('{} highs ...\n{}'.format(detail, str(hi))))
content.append(('{} lows ...\n{}'.format(detail, str(lo))))
if log_nan_only and not len(lo[lo[detail] != lo[detail]]):
continue # skip if no nans
out(content)
Output
(partial output just for illustration, here, everything was being ranked)
2016-03-24 05:45 log_pipe:306 INFO Rows: 400 Columns: 7
min mean max
alpha 358.5 1364.53625 2376.0
bbr 2.0 419.02 838.0
dir 1.0 210.855 419.0
fcf 2.0 209.8625 419.0
rev 1.0 209.5075 419.0
rsi 0.5 105.06375 209.5
yld 1.0 210.2275 419.0
2016-03-24 05:45 log_pipe:321 INFO _ _ _ alpha _ _ _
... alpha highs
alpha bbr dir fcf rev rsi yld
AMBC 2376.0 804.0 266.0 371.0 360.0 156.0 419.0
SALE 2268.5 738.0 332.0 373.0 365.0 181.5 279.0
BPT 2265.5 724.0 415.0 350.0 166.0 194.5 416.0
RTRX 2261.0 802.0 377.0 149.0 340.0 181.0 412.0
... alpha lows
alpha bbr dir fcf rev rsi yld
ARNA 472.5 102.0 14.0 32.0 243.0 39.5 42.0
HES 395.0 128.0 137.0 48.0 3.0 29.0 50.0
AMD 359.0 16.0 36.0 49.0 162.0 60.0 36.0
DVN 358.5 90.0 145.0 78.0 14.0 29.5 2.0
4 responses
Updated the code above. Improved output.
@Blue
As always...super useful...I'm going to try it out!
alan
Thank you!
Thanks for the kind words, a reason to keep caring.
Updated 2/2019 for easier reading, symbols instead of security id's in the first column and with right align. Limited number of decimal points for floats
While this would usually be called in before_trading_start (bts), and just once on start, there can be times when it might be useful anywhere else in an algo, conditionally. For example maybe the portfolio just had a jump or dip. Can reset context.log_data_done = 0and make the same call as the one above in bts and diff them in case of any clues.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.29312124848365784, "perplexity": 21061.720500791493}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738723.55/warc/CC-MAIN-20200810235513-20200811025513-00496.warc.gz"}
|
https://physics.stackexchange.com/questions/433748/real-world-intuitive-explanation-of-jerk
|
# Real world intuitive explanation of Jerk
If $$a(t)$$ denotes the acceleration of an object, then $$a^\prime(t)$$ represents the jerk.
I'm looking for an intuitive explanation of this phenomena. I'm hoping the following anecdote provides the intuition of how one might experience / explain jerk.
Suppose you're driving your car, and you approach a stoplight. The car then has a non-constant change in acceleration, $$a'(t).$$ Now, the moment the car comes to a full stop, we experience the car rocking back and forth. This is because the acceleration $$a(t)$$ goes to zero, but the change in acceleration at that time $$a'(t)$$ is . . .
Now, I would attempt to finish the explanation, but I lack the requisite language. My intuition is that $$a'(t)$$ is large, infinite, or is best approximated by an impulse function as $$a(t)$$ gets close to $$0$$. I'm not assuming that $$a'(t)$$ is differentiable, nor am I assuming it is even continuous.
Is this the right intuition? Would one of you be willing to provide another anecdote that one might experience on a day to day basis?
• An electric shuttle bus at an airport is a great demonstration of jerk and all bus drivers there are jerk experts. – safesphere Oct 10 '18 at 19:37
$$m\ddot x = F,$$ then, $$m\dddot x = \dot F.$$
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 11, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.48839545249938965, "perplexity": 200.0919112135}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655906934.51/warc/CC-MAIN-20200710082212-20200710112212-00092.warc.gz"}
|
http://mathoverflow.net/questions/91595/pseudoanosov-mapping-torus-and-length-of-curves?sort=newest
|
# Pseudoanosov mapping torus and length of curves.
Let $M_{\phi}$ be a hyperbolic mapping torus coming from a pseudo-Anosov map $\phi$ in a surface $S$. Is there any way to estimate the length of the geodesic representing a given curve in the surface in terms of the map $\phi$? That is, knowing something like the stable and unstable foliations for the map or something equivalent, can you estimate the length of a given curve? Any references for something like this are really appreciated.
For example, if you take a mapping torus $M_{\phi}$, drill one simple nontrivial curve $\alpha$ in the surface and re-glue by $\sigma^n$, a large Dehn twist about $\alpha$, you are going to get a hyperbolic mapping torus $M_{\phi\sigma^{n}}$. In this manifold $\alpha$ is going to be very short.
Another example, if you take a map $\psi = \phi\sigma^n$ where $\phi$ is pseudo-anosov in all of $S$ and $\sigma$ is a pseudo-Anosov just in a subsurface $X \subset S$, I think the curves in the complement of $X$ have to be very small for $n$ large, right?
-
Regarding the last paragraph - Suppose $\phi$ is pA in all of $S$, and $\sigma$ is pA in $X$, a strict subsurface. The lengths of the curves of $\partial X$ go to zero in $M_{\phi \sigma^n}$, as $n$ goes to infinity. But for curves in the complement of $X$, yet not in the boundary, their length does not go to zero. – Sam Nead Mar 19 '12 at 11:34
(I edited your post to make the notation in the last paragraph match the previous paragraphs.) – Sam Nead Mar 19 '12 at 11:44
Thanks for that and for your answer. Do you know any example where the curves in the complement of $X$ don't have short length? – shurtados Mar 25 '12 at 19:23
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9445899128913879, "perplexity": 179.34677465465182}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416931007056.0/warc/CC-MAIN-20141125155647-00232-ip-10-235-23-156.ec2.internal.warc.gz"}
|
https://worldwidescience.org/topicpages/c/change+ipcc+fourth.html
|
#### Sample records for change ipcc fourth
1. Global climate change: An introduction and results from the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4)
OpenAIRE
Seth, Anji
2007-01-01
This presentation gives summary of the results of the Intergovernmental Panel on Climate Change (IPCC) Working Group I (WG1) Fourth Assessment Report (AR4): The physical science basis for climate change. It begins with a history of the theory of global climate change, followed by the important concepts surrounding global climate change: the greenhouse effect and carbon cycle and how the climate has changed throughout the earth's history. It then discusses the IPCC's assessment reports, focusi...
2. Climate change science - beyond IPCC
International Nuclear Information System (INIS)
Full text: Full text: The main conclusions of the IPCC Working Group I assessment of the physical science of climate change, from the Fourth IPCC Assessment, will be presented, along with the evidence supporting these conclusions. These conclusions include: Global atmospheric concentrations of carbon dioxide, methane and nitrous oxide have increased markedly as a result of human activities since 1750 and now far exceed pre-industrial values determined from ice cores spanning many thousands of years. The global increases in carbon dioxide concentration are due primarily to fossil fuel use and land use change, while those of methane and nitrous oxide are primarily due to agriculture; The understanding of anthropogenic warming and cooling influences on climate has improved since the Third Assessment Report, leading to very high confidence that the global average net effect of human activities since 1750 has been one of warming, with a radiative forcing of +1.6 [+0.6 to +2.4] Wm-2; Warming of the climate system is unequivocal, as is now evident from observations of increases in global average air and ocean temperatures, widespread melting of snow and ice, and rising global average sea level; At continental, regional and ocean basin scales, numerous long-term changes in climate have been observed. These include changes in arctic temperatures and ice, widespread changes in precipitation amounts, ocean salinity, wind patterns and aspects of extreme weather including droughts, heavy precipitation, heat waves and the intensity of tropical cyclones. Palaeo-climatic information supports the interpretation that the warmth of the last half-century is unusual in at least the previous 1,300 years; Most of the observed increase in global average temperatures since the mid-20th century is very likely due to the observed increase in anthropogenic greenhouse gas concentrations; Discernible human influences now extend to other aspects of climate, including ocean warming, continental
3. Radiative forcing by well-mixed greenhouse gases: Estimates from climate models in the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4)
Science.gov (United States)
Collins, W. D.; Ramaswamy, V.; Schwarzkopf, M. D.; Sun, Y.; Portmann, R. W.; Fu, Q.; Casanova, S. E. B.; Dufresne, J.-L.; Fillmore, D. W.; Forster, P. M. D.; Galin, V. Y.; Gohar, L. K.; Ingram, W. J.; Kratz, D. P.; Lefebvre, M.-P.; Li, J.; Marquet, P.; Oinas, V.; Tsushima, Y.; Uchiyama, T.; Zhong, W. Y.
2006-07-01
The radiative effects from increased concentrations of well-mixed greenhouse gases (WMGHGs) represent the most significant and best understood anthropogenic forcing of the climate system. The most comprehensive tools for simulating past and future climates influenced by WMGHGs are fully coupled atmosphere-ocean general circulation models (AOGCMs). Because of the importance of WMGHGs as forcing agents it is essential that AOGCMs compute the radiative forcing by these gases as accurately as possible. We present the results of a radiative transfer model intercomparison between the forcings computed by the radiative parameterizations of AOGCMs and by benchmark line-by-line (LBL) codes. The comparison is focused on forcing by CO2, CH4, N2O, CFC-11, CFC-12, and the increased H2O expected in warmer climates. The models included in the intercomparison include several LBL codes and most of the global models submitted to the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4). In general, the LBL models are in excellent agreement with each other. However, in many cases, there are substantial discrepancies among the AOGCMs and between the AOGCMs and LBL codes. In some cases this is because the AOGCMs neglect particular absorbers, in particular the near-infrared effects of CH4 and N2O, while in others it is due to the methods for modeling the radiative processes. The biases in the AOGCM forcings are generally largest at the surface level. We quantify these differences and discuss the implications for interpreting variations in forcing and response across the multimodel ensemble of AOGCM simulations assembled for the IPCC AR4.
4. Summary for Policymakers IPCC Fourth Assessment Report, WorkingGroup III
Energy Technology Data Exchange (ETDEWEB)
Barker, Terry; Bashmakov, Igor; Bernstein, Lenny; Bogner,Jean; Bosch, Peter; Dave, Rutu; Davidson, Ogunlade; Fisher, Brian; Grubb,Michael; Gupta, Sujata; Halsnaes, Kirsten; Heij, Bertjan; Kahn Ribeiro,Suzana; Kobayashi, Shigeki; Levine, Mark; Martino, Daniel; MaseraCerutti, Omar; Metz, Bert; Meyer, Leo; Nabuurs, Gert-Jan; Najam, Adil; Nakicenovic, Nebojsa; Rogner, Hans Holger; Roy, Joyashree; Sathaye,Jayant; Schock, Robert; Shukla, Priyaradshi; Sims, Ralph; Smith, Pete; Swart, Rob; Tirpak, Dennis; Urge-Vorsatz, Diana; Zhou, Dadi
2007-04-30
A. Introduction 1. The Working Group III contribution to theIPCC Fourth Assessment Report (AR4) focuses on new literature on thescientific, technological, environmental, economic and social aspects ofmitigation of climate change, published since the IPCC Third AssessmentReport (TAR) and the Special Reports on COB2B Capture and Storage (SRCCS)and on Safeguarding the Ozone Layer and the Global Climate System (SROC).The following summary is organised into six sections after thisintroduction: - Greenhouse gas (GHG) emission trends, - Mitigation in theshort and medium term, across different economic sectors (until 2030), -Mitigation in the long-term (beyond 2030), - Policies, measures andinstruments to mitigate climate change, - Sustainable development andclimate change mitigation, - Gaps in knowledge. References to thecorresponding chapter sections are indicated at each paragraph in squarebrackets. An explanation of terms, acronyms and chemical symbols used inthis SPM can be found in the glossary to the main report.
5. IPCC underestimates the Sun’s role in climate change
NARCIS (Netherlands)
B. van Geel; P.A. Ziegler
2013-01-01
For the understanding of current and future climate change it is a basic pre requisite to properly understand the mechanisms, which controlled climate change after the Last Ice Age. According to the IPCC 5th assessment report (in prep.) the Sun has not been a major driver of climate change during th
6. EVIDENCE OF ARCTIC AND ANTARCTIC CHANGES AND THEIR REGULATION OF GLOBAL CLIMATE CHANGE (FURTHER FINDINGS SINCE THE FOURTH IPCC ASSESSMENT REPORT RELEASED)%南极和北极地区变化对全球气候变化的指示和调控作用——第四次IPCC评估报告以来一些新认知
Institute of Scientific and Technical Information of China (English)
陈立奇
2013-01-01
Changes in the climate of the Arctic and the Antarctic have been of great concern to international scientific and social communities since the Fourth Assessment Report of the Intergovernmental Panel on Climate Change (IPCC) was released in 2007. Since then, many new findings have been reported from observations and research carried out in the Arctic and Antarctic during the 4th International Polar Year (IPY). There is evidence that rapid changes in the Arctic and Antarctic induced by global warming are occurring in a quantitative and qualitative sense, and Arctic and Antarctic regional changes could be used indicators of global climate change. Declining Arctic sea ice could impact on winter snowfall in much of the Northern Hemisphere, with colder winters and more snow. Projections suggest that summertime Arctic sea ice will disappear by 2040. By 2050, the Antarctic ozone hole will have recovered to the level of the early 1980s, when the production of freon was completely prohibited. With weakening the shielding effect of the ozone hole to the global warming, it will become warmer in Antarctica and East Antarctica, leading to melting of ice sheets and retreating sea ice. Sea level rise will be a serious issue. As sea surface temperature rises the air-sea exchange of CO2 will be enhanced and surface water will take up more C02. This will lead to ocean acidification with important effects on ecological systems and food chains.%政府间气候变化专门委员会(IPCC) 2007年发布了第四次评估报告,全球气候变化问题再次成为一个重要的国际科学和政治议题.2007年以来,通过实施第四次国际极地年行动所获得的成果进一步证明,全球变暖所诱发极区出现的快速变化正在经历由量到质的转变,表明两极变化对全球气候变化起着一种指示和调控作用.一些研究指出:北冰洋会在2040年前后出现夏季无海冰并将引起北半球大范围的持续暴雪的寒冷冬季发生;2050
7. Evaluation of an IPCC climate report. An analysis of conclusions on the possible regional consequences of climate change
International Nuclear Information System (INIS)
This report contains the results of a study of the reliability of the regional chapters (H9-16) of the contribution of Working Group II to the Fourth Climate Report of the IPCC (the sub report on consequences, adaptation and vulnerability). Moreover an assessment was made of the possible consequences of errors for the conclusions in the high level summaries of that report. The Netherlands Environmental Assessment Agency did not detect any errors that may undermine the main conclusions of the scientific UN Climate Panel IPCC of 2007 on the possible future consequences of climate change. However, some of the substantiations of the conclusions lack clarity. To prevent lack of clarity and inaccuracies the IPCC needs to invest more in quality checks.
8. IPCC Climate Change 2013: Mitigation of Climate Change - Key Findings and Lessons Learned
Science.gov (United States)
Sokona, Youba
2014-05-01
The Working Group III contribution to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, Mitigation of Climate Change, examines the results of scientific research about mitigation, with special attention on how knowledge has evolved since the Fourth Assessment Report published in 2007. Throughout, the focus is on the implications of its findings for policy, without being prescriptive about the particular policies that governments and other important participants in the policy process should adopt. The report begins with a framing of important concepts and methods that help to contextualize the findings presented throughout the assessment. The valuation of risks and uncertainties, ethical concepts and the context of sustainable development and equity are among the guiding principles for the assessment of mitigation strategies. The report highlights past trends in stocks and flows of greenhouse gases and the factors that drive emissions at global, regional, and sectoral scales including economic growth, technology or population changes. It provides analyses of the technological, economic and institutional requirements of long-term mitigation scenarios and details on mitigation measures and policies that are applied in different economic sectors and human settlements. It then discusses interactions of mitigation policies and different policy instrument types at national, regional and global governance levels and between economic sectors, The Working Group III report comprises 16 chapters and in assembling this assessment authors were guided by the principles of the IPCC mandate: to be explicit about mitigation options, to be explicit about their costs and about their risks and opportunities vis-à-vis other development priorities, and to be explicit about the underlying criteria, concepts, and methods for evaluating alternative policies.
9. Changes in climate extremes and their impacts on the natural physical environment: An overview of the IPCC SREX report
Science.gov (United States)
Seneviratne, S. I.; Nicholls, N.; Easterling, D.; Goodess, C. M.; Kanae, S.; Kossin, J.; Luo, Y.; Marengo, J.; McInnes, K.; Rahimi, M.; Reichstein, M.; Sorteberg, A.; Vera, C.; Zhang, X.
2012-04-01
In April 2009, the Intergovernmental Panel on Climate Change (IPCC) decided to prepare a new special report with involvement of the UN International Strategy for Disaster Reduction (ISDR) on the topic "Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation" (SREX, http://ipcc-wg2.gov/SREX/). This special report reviews the scientific literature on past and projected changes in weather and climate extremes, and the relevance of such changes to disaster risk reduction and climate change adaptation. The SREX Summary for Policymakers was approved at an IPCC Plenary session on November 14-18, 2011, and the full report is planned for release in February 2012. This presentation will provide an overview on the structure and contents of the SREX, focusing on Chapter 3: "Changes in climate extremes and their impacts on the natural physical environment" [1]. It will in particular present the main findings of the chapter, including differences between the SREX's conclusions and those of the IPCC Fourth Assessment of 2007, and the implications of this new assessment for disaster risk reduction. Finally, aspects relevant to impacts on the biogeochemical cycles will also be addressed. [1] Seneviratne, S.I., N. Nicholls, D. Easterling, C.M. Goodess, S. Kanae, J. Kossin, Y. Luo, J. Marengo, K. McInnes, M. Rahimi, M. Reichstein, A. Sorteberg, C. Vera, and X. Zhang, 2012: Changes in climate extremes and their impacts on the natural physical environment. In: Intergovernmental Panel on Climate Change Special Report on Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation [Field, C. B., Barros, V., Stocker, T.F., Qin, D., Dokken, D., Ebi, K.L., Mastrandrea, M. D., Mach, K. J., Plattner, G.-K., Allen, S. K., Tignor, M. and P. M. Midgley (eds.)]. Cambridge University Press, Cambridge, United Kingdom and New York, NY, USA
10. Are the Projections of Future Climate Change Reliable in the IPCC Reports?
Institute of Scientific and Technical Information of China (English)
Zongci Zhao
2011-01-01
@@ As we know,the projections of future climate change including impacts and strategies in the IPCC Assessment Reports were based on global climate models with scenarios on various human activities.Global climate model simulations provide key inputs for climate change assessments. In this study,the main objective is to analyze if the projections of fu-ture climate change by global climate models are reli-able.Several workshops have been held on this issue,such as the IPCC expert meeting on assessing and combining multi-model climate projections in January of 2010 (presided by the co-chairs of the IPCC WGI and WGII AR5),and the workshop of the combined global climate model group held by NCAR in June of 2010.
11. Climate Change on Twitter: Topics, Communities and Conversations about the 2013 IPCC Working Group 1 Report
Science.gov (United States)
Pearce, Warren; Holmberg, Kim; Hellsten, Iina; Nerlich, Brigitte
2014-01-01
In September 2013 the Intergovernmental Panel on Climate Change published its Working Group 1 report, the first comprehensive assessment of physical climate science in six years, constituting a critical event in the societal debate about climate change. This paper analyses the nature of this debate in one public forum: Twitter. Using statistical methods, tweets were analyzed to discover the hashtags used when people tweeted about the IPCC report, and how Twitter users formed communities around their conversational connections. In short, the paper presents the topics and tweeters at this particular moment in the climate debate. The most used hashtags related to themes of science, geographical location and social issues connected to climate change. Particularly noteworthy were tweets connected to Australian politics, US politics, geoengineering and fracking. Three communities of Twitter users were identified. Researcher coding of Twitter users showed how these varied according to geographical location and whether users were supportive, unsupportive or neutral in their tweets about the IPCC. Overall, users were most likely to converse with users holding similar views. However, qualitative analysis suggested the emergence of a community of Twitter users, predominantly based in the UK, where greater interaction between contrasting views took place. This analysis also illustrated the presence of a campaign by the non-governmental organization Avaaz, aimed at increasing media coverage of the IPCC report. PMID:24718388
12. Climate change on Twitter: topics, communities and conversations about the 2013 IPCC Working Group 1 report.
Directory of Open Access Journals (Sweden)
Warren Pearce
Full Text Available In September 2013 the Intergovernmental Panel on Climate Change published its Working Group 1 report, the first comprehensive assessment of physical climate science in six years, constituting a critical event in the societal debate about climate change. This paper analyses the nature of this debate in one public forum: Twitter. Using statistical methods, tweets were analyzed to discover the hashtags used when people tweeted about the IPCC report, and how Twitter users formed communities around their conversational connections. In short, the paper presents the topics and tweeters at this particular moment in the climate debate. The most used hashtags related to themes of science, geographical location and social issues connected to climate change. Particularly noteworthy were tweets connected to Australian politics, US politics, geoengineering and fracking. Three communities of Twitter users were identified. Researcher coding of Twitter users showed how these varied according to geographical location and whether users were supportive, unsupportive or neutral in their tweets about the IPCC. Overall, users were most likely to converse with users holding similar views. However, qualitative analysis suggested the emergence of a community of Twitter users, predominantly based in the UK, where greater interaction between contrasting views took place. This analysis also illustrated the presence of a campaign by the non-governmental organization Avaaz, aimed at increasing media coverage of the IPCC report.
13. Climate change on Twitter: topics, communities and conversations about the 2013 IPCC Working Group 1 report.
Science.gov (United States)
Pearce, Warren; Holmberg, Kim; Hellsten, Iina; Nerlich, Brigitte
2014-01-01
In September 2013 the Intergovernmental Panel on Climate Change published its Working Group 1 report, the first comprehensive assessment of physical climate science in six years, constituting a critical event in the societal debate about climate change. This paper analyses the nature of this debate in one public forum: Twitter. Using statistical methods, tweets were analyzed to discover the hashtags used when people tweeted about the IPCC report, and how Twitter users formed communities around their conversational connections. In short, the paper presents the topics and tweeters at this particular moment in the climate debate. The most used hashtags related to themes of science, geographical location and social issues connected to climate change. Particularly noteworthy were tweets connected to Australian politics, US politics, geoengineering and fracking. Three communities of Twitter users were identified. Researcher coding of Twitter users showed how these varied according to geographical location and whether users were supportive, unsupportive or neutral in their tweets about the IPCC. Overall, users were most likely to converse with users holding similar views. However, qualitative analysis suggested the emergence of a community of Twitter users, predominantly based in the UK, where greater interaction between contrasting views took place. This analysis also illustrated the presence of a campaign by the non-governmental organization Avaaz, aimed at increasing media coverage of the IPCC report. PMID:24718388
14. The impact of SciDAC on US climate change research and the IPCC AR4
International Nuclear Information System (INIS)
SciDAC has invested heavily in climate change research. We offer a candid opinion as to the impact of the DOE laboratories' SciDAC projects on the upcoming Fourth Assessment Report of the Intergovernmental Panel on Climate Change
15. The Intergovernmental Panel on Climate Change (IPCC) and scientific consensus. How scientists come to say what they say about climate change
Energy Technology Data Exchange (ETDEWEB)
Alfsen, Knut H.; Skodvin, Tora
1998-12-01
This document reviews the background, organization and operation of the Intergovernmental Panel on Climate Change (IPCC). It gives some background on climate change in the past and finally discusses what IPCC says about the likely future impact of anthropogenic emission of greenhouse gases. 14 refs., 8 figs.
16. Mitigation of climate change: back to IPCC's fifth report
International Nuclear Information System (INIS)
This article provides an overview of current knowledge on climate change mitigation, based on the Intergovernmental Panel on Climate Change (IPCC) Working Group III fifth assessment report. The report emphasizes how little room for manoeuvre there is to meet the target of a global mean surface temperature increase below 2 deg. C, if ambitious policies to reduce greenhouse gases are not implemented by 2020. It also assesses sectoral potentials for emissions reductions and addresses emerging questions, in particular regarding the financing of decarbonization pathways. The report finally highlights the need for integrated policies to take advantage of co-benefits of climate policies (health, energy security, etc.), the evaluation of which is becoming more systematic. (authors)
17. Sea level change under IPCC-A2 scenario in Bohai, Yellow, and East China Seas
Directory of Open Access Journals (Sweden)
Chang-lin CHEN
2015-10-01
Full Text Available Because of the environmental and socioeconomic impacts of anthropogenic sea level rise (SLR, it is very important to understand the processes leading to past and present SLRs towards more reliable future SLR projections. A regional ocean general circulation model (ROGCM, with a grid refinement in the Bohai, Yellow, and East China Seas (BYECSs, was set up to project SLR induced by the ocean dynamic change in the 21st century. The model does not consider the contributions from ice sheets and glacier melting. Data of all forcing terms required in the model came from the simulation of the Community Climate System Model version 3.0 (CCSM3 under the International Panel on Climate Change (IPCC-A2 scenario. Simulation results show that at the end of the 21st century, the sea level in the BYECSs will rise about 0.12 to 0.20 m. The SLR in the BYECSs during the 21st century is mainly caused by the ocean mass redistribution due to the ocean dynamic change of the Pacific Ocean, which means that water in the Pacific Ocean tends to move to the continental shelves of the BYECSs, although the local steric sea level change is another factor.
18. Sea level change under IPCC-A2 scenario in Bohai, Yellow, and East China Seas
Institute of Scientific and Technical Information of China (English)
Chang-lin CHEN; Jun-cheng ZUO; Mei-xiang CHEN; Zhi-gang GAO; C-K SHUM
2014-01-01
Because of the environmental and socioeconomic impacts of anthropogenic sea level rise (SLR), it is very important to understand the processes leading to past and present SLRs towards more reliable future SLR projections. A regional ocean general circulation model (ROGCM), with a grid refinement in the Bohai, Yellow, and East China Seas (BYECSs), was set up to project SLR induced by the ocean dynamic change in the 21st century. The model does not consider the contributions from ice sheets and glacier melting. Data of all forcing terms required in the model came from the simulation of the Community Climate System Model version 3.0 (CCSM3) under the International Panel on Climate Change (IPCC)-A2 scenario. Simulation results show that at the end of the 21st century, the sea level in the BYECSs will rise about 0.12 to 0.20 m. The SLR in the BYECSs during the 21st century is mainly caused by the ocean mass redistribution due to the ocean dynamic change of the Pacific Ocean, which means that water in the Pacific Ocean tends to move to the continental shelves of the BYECSs, although the local steric sea level change is another factor.
19. Global Warming Induced Changes in Rainfall Characteristics in IPCC AR5 Models
Science.gov (United States)
Lau, William K. M.; Wu, Jenny, H.-T.; Kim, Kyu-Myong
2012-01-01
Changes in rainfall characteristic induced by global warming are examined from outputs of IPCC AR5 models. Different scenarios of climate warming including a high emissions scenario (RCP 8.5), a medium mitigation scenario (RCP 4.5), and 1% per year CO2 increase are compared to 20th century simulations (historical). Results show that even though the spatial distribution of monthly rainfall anomalies vary greatly among models, the ensemble mean from a sizable sample (about 10) of AR5 models show a robust signal attributable to GHG warming featuring a shift in the global rainfall probability distribution function (PDF) with significant increase (>100%) in very heavy rain, reduction (10-20% ) in moderate rain and increase in light to very light rains. Changes in extreme rainfall as a function of seasons and latitudes are also examined, and are similar to the non-seasonal stratified data, but with more specific spatial dependence. These results are consistent from TRMM and GPCP rainfall observations suggesting that extreme rainfall events are occurring more frequently with wet areas getting wetter and dry-area-getting drier in a GHG induced warmer climate.
20. Climate change. Important findings from the 4. fact finding report of the intergovernmental commission on climate change of the United Nations (IPCC); Klimaaenderung. Wichtige Erkentnisse aus dem 4. Sachstandsbericht des Zwischenstaatlichen Ausschusses fuer Klimaaenderungen der Vereinten Nationen (IPCC)
Energy Technology Data Exchange (ETDEWEB)
Maeder, Claudia
2009-12-19
The Report covers the following topics: 1. anthropogenic climate change - since when do we know about it? 2. IPCC - the intergovernmental commission for climate change. 3. Assignable causes for climate change: changes of incoming solar radiation, changes of the reflected solar radiation, change of the heat radiation lost into space, aerosols, internal variability of the climate system. 4. Historical climate changes in long periods. 5. Development of the greenhouse gases in the atmosphere. 6. Observed climate changes. 7. Projections of future climate changes. 8. Consequences of climate change: consequences of the actual temperature increase, possible future consequences, freshwater resources and their management, ecosystems, agricultural production, coastal regions and low lying areas.
1. Mesh climate change data of Japan ver. 2 for climate change impact assessments under IPCC SRES A1B and A2
International Nuclear Information System (INIS)
The Intergovernmental Panel on Climate Change (IPCC) published the Fourth Assessment Report (AR4) in 2007 and stated that recent climate change and variation are induced by increases in the atmospheric greenhouse gases (GHG) concentration due to anthropogenic activities. The report includes the results of impact assessments on a wide range of sectors. These assessments have been conducted based on future climate projections, which refer to aspects of the future climate evaluated by Atmosphere-Ocean Coupled General Circulation Models (CGCMs). The projection data used in the AR4 are archived under the Program for Climate Model Diagnosis and Intercomparison (PCMDI) promoted by the U.S. Department of Energy. We interpolated the projection data around Japan and constructed a dataset entitled the 'Mesh climate change data of Japan Ver. 2' for the climate change impact study. Nine projections performed by seven models under the A1B and A2 of the Special Report on Emissions Scenarios (SRES) were implemented for the dataset. They consist of mesh data with a size of 7.5 min in longitude and 5.0 min in latitude, i.e. approximately 10 X 10 km (45 sec in longitude and 30 sec in latitude, approximately 1 x 1 km, for one high-resolution model). The dataset includes five climatic elements, i.e. the daily mean, maximum, and minimum surface air temperatures, daily total precipitation, and daily accumulated shortwave radiation for three periods, 1981-2000, 2046-2065, and 2081-2100. This article describes the details concerning the construction and characteristics of the data
2. Methods for interfacing IPCC climate change scenarios with higher resolution watershed management models in the Ethiopian Blue Nile Basin
Science.gov (United States)
Easton, Z. M.; MacAlister, C.; Fuka, D. R.
2013-12-01
As much as 90% of the Nile River flow that reaches Egypt originates in the Highlands of the Ethiopian Blue Nile Basin. This imbalance in water availability poses a threat to water security in the region, and could be severely impacted by projected climate change. This analysis coupled hydrodynamic/watershed models with the Intergovernmental Panel on Climate Change (IPCC) AR4 climate change scenarios to assess the potential impact on water resources and sediment dynamics. Specific AR4 scenarios include the A1B, B1, B2 and COMMIT, which were used to force the baseline hydrodynamic models calibrated against 1979-2011 streamflow for 20 sub-watersheds in the Tana and Beles basins. Transfer functions were developed to distribute the model parameters from the calibrated sub-watersheds to un-gauged portions of the basins based on a similarity index of hydrologic response units. We analyzed the scenario in two manners: first we ran all of the seven individual Global Circulation Model results in the IPCC AR4 report though our watershed models to asses the potential spread of climate change predictions; then we assessed the mean value produced for each IPCC AR4 scenario to better estimate convergence. Results indicate that the Tana basin is expected to experience an increase in mean annual flow. The Beles basin is predicted to experience a small decrease in mean annual flow. Sediment concentrations in the Tana basin increase proportionally more than the flow increase. Interestingly, and perhaps counter to what might be expected for a decrease in flow in the Beles basin, sediment concentrations increase.
3. IPCC's Climate Communication
DEFF Research Database (Denmark)
Almlund, Pernille
The work of IPCC is an important work and contribution to the global discussion and global challenge of climate change. But this work is primarily based on computer modelling, natural science, economic science and as a new perspective a stronger focus on the risk perspective than in earlier IPCC...... reports. This paper is based on a wonder of why the IPCC’s analysis and reports are not, to a higher degree, based on social science and human science. Are these scientific perspectives with many different approaches not important to this global political awareness of climate change? Especially now when...... all the IPCC’s assessment report have concluded that climate changes are human made and recently stated that 97 % of all climate researchers agree in that conclusion. Due to the theoretical work of Michel Callon, Lascoumes and Barthe (2011) and their ANT perspective, climate change can be observed...
4. IPCC and other assessments as vehicles for integrating natural and social science research to address human dimensions of climate change
Science.gov (United States)
Field, C. B.
2012-12-01
IPCC and other assessments address both natural and social science aspects of climate change, but this approach has historically involved relatively little integration across the two sets of disciplines. In a framing that is only slightly oversimplified, past relationships were mostly sequential. From a physical climate perspective, human behavior was a boundary condition setting the trajectory of atmospheric forcing. And from an impacts perspective, changes in the physical climate set the stage upon which humans experienced impacts and made decisions about adaptation and mitigation. Integrated assessment models have been the main locus of research on questions about bi-directional coupling, where the trajectory of the physical climate influences GHG balance related to the need for agricultural land as well as GHG emissions from other activities. In the IPCC AR4 (2007), feedbacks from the natural carbon cycle to climate were a focus, but with little discussion of the potentially important feedbacks from climate-carbon interactions in the human domain. Detailed research and modeling in this area are still in the relatively early stages. For the future, IPCC and other assessments potentially provide a vehicle for new insights about the interaction of natural and social science dimensions of climate change. Several aspects could be interesting. Some of these relate to the decisions that modulate GHG emissions. For example, how does scientific understanding of climate change influence people's interest in mitigation and adaptation? How does it influence their willingness to pay? How are these modulated by regional and global geopolitics? Other potentially interesting aspects relate to interactions between mitigation and adaptation. For example, how does local experience of climate change alter the balance of focus on adaptation and mitigation? Still others relate to the nature of impacts and the role of sustainable development. With an aggress sustainable development
5. Noaa contributions to the 1995 IPCC assessments: A summary of the current and future activities of the intergovernmental panel on climate change
International Nuclear Information System (INIS)
Contents: introduction -the IPCC and NOAA; NOAA contributions to the IPCC 1995 science assessment; NOAA contributions to the IPCC 1995 impacts, adaptation, and mitigation assessment; NOAA contributions to the 1995 IPCC economics and greenhouse-gas scenario assessment
6. Global observed long-term changes in temperature and precipitation extremes: A review of progress and limitations in IPCC assessments and beyond
Directory of Open Access Journals (Sweden)
Lisa V. Alexander
2016-03-01
Full Text Available The Intergovernmental Panel on Climate Change (IPCC first attempted a global assessment of long-term changes in temperature and precipitation extremes in its Third Assessment Report in 2001. While data quality and coverage were limited, the report still concluded that heavy precipitation events had increased and that there had been, very likely, a reduction in the frequency of extreme low temperatures and increases in the frequency of extreme high temperatures. That overall assessment had changed little by the time of the IPCC Special Report on Extremes (SREX in 2012 and the IPCC Fifth Assessment Report (AR5 in 2013, but firmer statements could be added and more regional detail was possible. Despite some substantial progress throughout the IPCC Assessments in terms of temperature and precipitation extremes analyses, there remain major gaps particularly regarding data quality and availability, our ability to monitor these events consistently and our ability to apply the complex statistical methods required. Therefore this article focuses on the substantial progress that has taken place in the last decade, in addition to reviewing the new progress since IPCC AR5 while also addressing the challenges that still lie ahead.
7. The Fourth Way of Technology and Change
Science.gov (United States)
Shirley, Dennis
2011-01-01
Recent social policy reforms have sought to overcome the limitations of "First Way" strategies emphasizing the welfare state and "Second Way" approaches advocating markets. Scholars and policymakers instead have begun to explore optimal synthesis of the public and private sector in a new "Third Way" of leadership and change. According to one line…
8. Trends in marine climate change research in the Nordic region since the first IPCC report
DEFF Research Database (Denmark)
Pedersen, Martin Wæver; Kokkalis, Alexandros; Bardarson, H.;
2016-01-01
Oceans are exposed to anthropogenic climate change shifting marine systems toward potential instabilities. The physical, biological and social implications of such shifts can be assessed within individual scientific disciplines, but can only be fully understood by combining knowledge and expertise...... evaluated the development of climate change related marine science by quantifying trends in number of publications, disciplinarity, and scientific focus of 1362 research articles published between 1990 and 2011. Our analysis showed a faster increase in publications within climate change related marine...... science than in general marine science indicating a growing prioritisation of research with a climate change focus. The composition of scientific disciplines producing climate change related publications, which initially was dominated by physical sciences, shifted toward a distribution with almost even...
9. Assessing impact of climate change on season length in Karnataka for IPCC SRES scenarios
Indian Academy of Sciences (India)
Aavudai Anandhi
2010-08-01
Changes in seasons and season length are an indicator, as well as an effect, of climate change. Seasonal change profoundly affects the balance of life in ecosystems and impacts essential human activities such as agriculture and irrigation. This study investigates the uncertainty of season length in Karnataka state, India, due to the choice of scenarios, season type and number of seasons. Based on the type of season, the monthly sequences of variables (predictors) were selected from datasets of NCEP and Canadian General Circulation Model (CGCM3). Seasonal stratifications were carried out on the selected predictors using K-means clustering technique. The results of cluster analysis revealed increase in average, wet season length in A2, A1B and B1 scenarios towards the end of 21st century. The increase in season length was higher for A2 scenario whereas it was the least for B1 scenario. COMMIT scenario did not show any change in season length. However, no change in average warm and cold season length was observed across the four scenarios considered. The number of seasons was increased from 2 to 5. The results of the analysis revealed that no distinct cluster could be obtained when the number of seasons was increased beyond three.
10. The social costs of climate change. The IPCC second assessment report and beyond
International Nuclear Information System (INIS)
Climate change is expected to have far-reaching impacts. Earlier studies have estimated an aggregated monetised damage equivalent to 1.5 to 2.0 % of World GDP (for 2 x CO2). According to these estimates, the OECD would face losses equivalent to 1.0 to 1.5 % of GDP, and developing countries 2.0 to 9.0 %. While these figures are preliminary and highly uncertain, recent findings have not, as yet, changed the general picture. As is shown in this paper, estimates that are fully corrected for differences in purchasing power parity do not significantly differ from the initial figures. Newer studies increasingly emphasise adaptation, variability, extreme events, other (non-climate change) stress factors, and the need for integrated assessment of damages. Incorporating these factors has lead to increased differences in estimated impacts between different regions and sectors. Estimates of market impacts in developed countries tended to fall, while non-market impacts have become more important. Marginal damages are more interesting from a policy point of view. Earlier estimates range from about $5 to$125 per tonne of carbon, with most estimates at the lower end of this range. These figures are based on power functions in the level of climate change. The rate of change may be equally important, as are the speed of adaptation, restoration and value adjustment. Furthermore, future vulnerability to climate change will differ from current vulnerability: market impacts could fall (relatively) with economic growth while non-market impacts may rise. 6 tabs., 65 refs
11. Trends in marine climate change research in the Nordic region since the first IPCC report
NARCIS (Netherlands)
Pedersen, M.W.; Kokkalis, A.; Bardarson, H.; Bonanomi, S.; Boonstra, W.J.; Butler, W.E.; Diekert, F.K.; Fouzai, N.; Holma, M.; Holt, R.E.; Kvile, K.; Nieminen, E.; Ottosen, K.M.; Richter, A.P.; Rogers, L.A.; Romagnoni, G.; Snickars, M.; Tornroos, A.; Weigel, B.; Whittington, J.D.; Yletyinen, J.; Ferreira, A.S.A.
2016-01-01
Oceans are exposed to anthropogenic climate change shifting marine systems toward potential instabilities. The physical, biological and social implications of such shifts can be assessed within individual scientific disciplines, but can only be fully understood by combining knowledge and expertise a
12. Climate Change Draws World Attention: The 2007 Nobel Peace Award Goes to Gore and IPCC
Science.gov (United States)
Bisland, Beverly Milner; Ahmad, Iftikhar
2008-01-01
In the fall of 2007, the Nobel Committee awarded their Peace Prize to the Intergovernmental Panel on Climate Change (a scientific intergovernmental body set up by the World Meteorological Organization and by the United Nations Environment Program) and to former Vice-President Al Gore, Jr. The committee praised the United Nations panel for creating…
13. IPCC Climate Change 2013: Impacts, Adaptation and Vulnerability: Key findings and lessons learned
Science.gov (United States)
Giorgi, Filippo; Field, Christopher; Barros, Vicente
2014-05-01
The Working Group II contribution to the Fifth Assessment Report of the Intergivernmental Panel on Climate Change, Impacts, Adaptation and Vulnerability, will be completed and approved in March 2014. It includes two parts, Part A covering Global and Sectoral Aspects, and Part B, covering Regional Aspects. The WGII report spans a very broad range of topics which are approached in a strong interdisciplinary context. It highlights how observed impacts of climate change are now widespread and consequential, particularly for natural systems, and can be observed on all continents and across the oceans. Vulnerability to climate change depends on interactions with non-climatic stressors and inequalities, resulting in highly differential risks associated with climate change. It is also found that adaptation is already occurring across scales and is embedded in many planning processes. Continued sustained warming thrughout the 21st century will exacerbate risks and vulnerabilities across multiple sectors, such as freshwater resources, terrestrial and inland water systems, coastal and marine systems, food production, human health, security and livelihood. The report stresses how risks and vulnerabilities need to be assessed within a multi-stressor and regionally specific context, and can be reduced and managed by adopting climate-resilient pathwyas combining suitable adaptation and mitigation options with synergies and tradeoffs occurring both within and across regions. The Working group II report includes a large number of Chapters (30) and contributors (310 including authors and review editors), with expertise in a broad range of disciplines, from the physical science to the impact and socio-economic sciences. The communication across chapters and disciplines has been a challenge, and will continue to be one as the Global Change problem will increasingly require a fully integrated and holistic approach. Note that text on this abstract is not approved at the time its
14. Simulation of the future change of East Asian monsoon climate using the IPCC SRES A2 and B2 scenarios
Institute of Scientific and Technical Information of China (English)
2003-01-01
In this paper, we applied the newest emission scenarios of the sulfur and greenhouse gases, i.e. Intergovernmental Panel on Climate Change (IPCC) Special Report on Emission Scenarios (SRES) A2 and B2 scenarios, to investigating the change of the East Asian climate in the last three decades of the 21st century with an atmosphere-ocean coupled general circulation model. The global warming enlarges the land-sea thermal contrast and, hence, enhances (reduces) the East Asian summer (winter) monsoon circulation. The precipitation from the Yangtze and Huaihe river valley to North China increases significantly. In particular, the strong rainfall increase over North China implies that the East Asian rainy area would expand northward. In addition, from the southeastern coastal area to North China, the rainfall would increase significantly in September, implying that the rainy period of the East Asian monsoon would be prolonged about one month. In July, August and September, the interannual variability of the precipitation enhances evidently over North China, meaning a risk of flooding in the future.
15. Why We Don't Collaborate in Response to Climate Change: The Knowledge Deficit, Co-Production, and the Future of the IPCC.
Science.gov (United States)
Cook, B. R.; Overpeck, J. T.
2014-12-01
Scientific knowledge production is based on recognizing and filling knowledge deficits or 'gaps' in understanding, but for climate adaptation and mitigation, the applicability of this approach is questionable. The Intergovernmental Panel on Climate Change (IPCC) mandate is an example of this type of 'gap filling,' in which the elimination of uncertainties is presumed to enable rational decision making for individuals and rational governance for societies. Presumed knowledge deficits, though, are unsuited to controversial problems with social, cultural, and economic dimensions; likewise, communication to educate is an ineffective means of inciting behavioural change. An alternative is needed, particularly given the economic, social, and political scale that action on climate change requires. We review the 'deficit-education framing' and show how it maintains a wedge between those affected and those whose knowledge is required. We then review co-production to show how natural and social scientists, as well as the IPCC, might more effectively proceed.
16. Emissions from tropical hydropower and the IPCC
International Nuclear Information System (INIS)
Highlights: • Tropical dams emit greenhouse gases, which are undercounted in IPCC guidelines. • IPCC comparisons with other energy sources undercount hydroelectric emissions. • GHG inventories must fully count emissions as a basis for negotiating national quotas. • The IPCC needs to reassess emissions from dams independent of the hydropower industry. - Abstract: Tropical hydroelectric emissions are undercounted in national inventories of greenhouse gases under the United Nations Framework Convention on Climate Change (UNFCCC), giving them a role in undermining the effectiveness of as-yet undecided emission limits. These emissions are also largely left out of the Intergovernmental Panel on Climate Change (IPCC) Special Report on Renewable Energy Sources and Climate Change Mitigation, and have been excluded from a revision of the IPCC guidelines on wetlands. The role of hydroelectric dams in emissions inventories and in mitigation has been systematically ignored
17. Renewable energy sources and climate change mitigation. Special report of the Intergovernmental Panel on Climate Change (IPCC)
Energy Technology Data Exchange (ETDEWEB)
Edenhofer, O. (Potsdam Institute for Climate Impact Research (PIK), Potsdam (Germany)); Pichs Madruga, R. (Centro de Investigaciones de la Economia Mundial (CIEM), Hanoi (Viet Nam)); Sokona, Y. (African Climate Policy Centre, United Nations Economic Commission for Africa, Addis Ababa (Ethiopia)) (and others)
2012-07-01
Climate change is one of the great challenges of the 21st century. Its most severe impacts may still be avoided if efforts are made to transform current energy systems. Renewable energy sources have a large potential to displace emissions of greenhouse gases from the combustion of fossil fuels and thereby to mitigate climate change. If implemented properly, renewable energy sources can contribute to social and economic development, to energy access, to a secure and sustainable energy supply, and to a reduction of negative impacts of energy provision on the environment and human health. This Special Report on Renewable Energy Sources and Climate Change Mitigation (SRREN) impartially assesses the scientific literature on the potential role of renewable energy in the mitigation of climate change for policymakers, the private sector, academic researchers and civil society. It covers six renewable energy sources - bioenergy, direct solar energy, geothermal energy, hydropower, ocean energy and wind energy - as well as their integration into present and future energy systems. It considers the environmental and social consequences associated with the deployment of these technologies, and presents strategies to overcome technical as well as non-technical obstacles to their application and diffusion. The authors also compare the levelized cost of energy from renewable energy sources to recent non-renewable energy costs. (Author)
18. Fifth IPCC Assessment Report Now Out
Science.gov (United States)
Kundzewicz, Zbigniew W.
2014-01-01
The Fifth Assessment Report (AR5) of the Intergovernmental Panel on Climate Change (IPCC) is now available. It provides policymakers with an assessment of information on climate change, its impacts and possible response options (adaptation and mitigation). Summaries for policymakers of three reports of IPCC working groups and of the Synthesis Report have now been approved by IPCC plenaries. This present paper reports on the most essential findings in AR5. It briefly informs on the contents of reports of all IPCC working groups. It discusses the physical science findings, therein observed changes (ubiquitous warming, shrinking cryosphere, sea level rise, changes in precipitation and extremes, and biogeochemical cycles). It deals with the drivers of climate change, progress in climate system understanding (evaluation of climate models, quantification of climate system responses), and projections for the future. It reviews impacts, adaptation and vulnerability, including observed changes, key risks, key reasons for concern, sectors and systems, and managing risks and building resilience. Finally, mitigation of climate change is discussed, including greenhouse gas emissions in the past, present and future, and mitigation in sectors. It is hoped that the present article will encourage the readership of this journal to dive into the AR5 report that provides a wealth of useful information.
19. Linguistic analysis of IPCC summaries for policymakers and associated coverage
Science.gov (United States)
Barkemeyer, Ralf; Dessai, Suraje; Monge-Sanz, Beatriz; Renzi, Barbara Gabriella; Napolitano, Giulio
2016-03-01
The Intergovernmental Panel on Climate Change (IPCC) Summary for Policymakers (SPM) is the most widely read section of IPCC reports and the main springboard for the communication of its assessment reports. Previous studies have shown that communicating IPCC findings to a variety of scientific and non-scientific audiences presents significant challenges to both the IPCC and the mass media. Here, we employ widely established sentiment analysis tools and readability metrics to explore the extent to which information published by the IPCC differs from the presentation of respective findings in the popular and scientific media between 1990 and 2014. IPCC SPMs clearly stand out in terms of low readability, which has remained relatively constant despite the IPCC’s efforts to consolidate and readjust its communications policy. In contrast, scientific and quality newspaper coverage has become increasingly readable and emotive. Our findings reveal easy gains that could be achieved in making SPMs more accessible for non-scientific audiences.
20. Regulating Knowledge Monopolies: The Case of the IPCC
OpenAIRE
Tol, R.S.J.
2010-01-01
The Intergovernmental Panel on Climate Change has a monopoly on the provision of climate policy advice at the international level and a strong market position in national policy advice. This may have been the intention of the founders of the IPCC. I argue that the IPCC has a natural monopoly, as a new entrant would have to invest time and effort over a longer period to perhaps match the reputation, trust, goodwill, and network of the IPCC. The IPCC is a not-for-profit organization, and it is ...
1. Hydrologic impacts of climate change on the Nile River basin: Implications of the 2007 IPCC climate scenarios
NARCIS (Netherlands)
Beyene, T.; Lettenmaier, D.P.; Kabat, P.
2010-01-01
We assess the potential impacts of climate change on the hydrology and water resources of the Nile River basin using a macroscale hydrology model. Model inputs are bias corrected and spatially downscaled 21st Century simulations from 11 General Circulation Models (GCMs) and two global emissions scen
2. Simulating Soil Organic Carbon Stock Changes in Agro-ecosystems using CQESTR, DayCent, and IPCC Tier 1 Methods
Science.gov (United States)
Models are often used to quantify how land use change and management impact soil organic carbon (SOC) stocks because it is often not feasible to use direct measuring methods. Because models are simplifications of reality, it is essential to compare model outputs with measured values to evaluate mode...
3. Changes of Atmospheric Water Balance over China under the IPCC SRES A1B Scenario Based on RegCM3 Simulations
Institute of Scientific and Technical Information of China (English)
SUN Bo; JIANG Da-Bang
2012-01-01
Simulations of the Regional Climate Model Version 3 (RegCM3) under the Intergovernmental Panel on Climate Change (IPCC) Special Report on Emissions Scenarios (SRES) A1B scenario were employed to investigate possible decadal changes and long-term trends of annual mean atmospheric water balance components over China in the 21st century with reference to the period of 1981-2000. An evaluation showed that RegCM3 can reasonably reproduce annual evapotranspiration, precipitation, and water vapor transport over China, with a better performance for March–June. It was found that the water vapor exchange between the land surface and atmosphere would be significantly intensified in Northwest China by the mid-to late-21st century and that the region would possibly shift to a wetter or drought-mitigated state under global warming. Conversely, the water vapor exchange evidently weakened over the Tibetan Plateau and South-west China by the mid-to late-21st century. In addition, there appears to be a drier state for Northeast China and the middle and lower reaches of the Yangtze River valley by the mid-to late-21st century, with slight mitigation by the end compared with the mid-21st century. The westerly and southwesterly water vapor transport over China generally presents an increasing trend, with increasing diver-gence over the Tibetan Plateau and Northeast China, corresponding to a loss of atmospheric water vapor by water vapor transport.
4. Uncertainty quantification of US Southwest climate from IPCC projections.
Energy Technology Data Exchange (ETDEWEB)
Boslough, Mark Bruce Elrick
2011-01-01
The Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4) made extensive use of coordinated simulations by 18 international modeling groups using a variety of coupled general circulation models (GCMs) with different numerics, algorithms, resolutions, physics models, and parameterizations. These simulations span the 20th century and provide forecasts for various carbon emissions scenarios in the 21st century. All the output from this panoply of models is made available to researchers on an archive maintained by the Program for Climate Model Diagnosis and Intercomparison (PCMDI) at LLNL. I have downloaded this data and completed the first steps toward a statistical analysis of these ensembles for the US Southwest. This constitutes the final report for a late start LDRD project. Complete analysis will be the subject of a forthcoming report.
5. Applying IPCC Representative Concentration Pathway (RCP) land-use projections in a regional assessment of land-use change in the conterminous United States.
Science.gov (United States)
Sherba, J.; Sleeter, B. M.
2015-12-01
The Intergovernmental Panel on Climate Change (IPCC) Representative Concentration Pathways (RCPs) include global land-use change projections for four global emissions scenarios. These projections are potentially useful for driving regional-scale models needed for informing land-use and management interactions. Here, we applied global gridded RCP land-use projections within a regional-scale state-and-transition simulation model (STSM) projecting land-use change in the conterminous United States. First, we cross-walked RCP land-use transition classes to land-use classes more relevant for modeling at the regional scale. Coarse grid RCP land-use transition values were then downscaled to EPA Level III ecoregion boundaries using historical land-use transition data from the USGS Land Cover Trends (LCT) dataset. Downscaled transitions were aggregated to the ecoregion level. Ecoregions were chosen because they represent areas with consistent land-use patterns that have proven useful for studying land-use and management interactions. Ecoregion-level RCP projections were applied in a state-and-transition simulation model (STSM) projecting land-use change between 2005 and 2100 at the 1-km scale. Resulting RCP-based STSM projections were compared to STSM projections created using scenario projections from the Special Report on Emissions Scenarios (SRES) and the USGS LCT dataset. While most land-use trajectories appear plausible, some transitions such as forest harvest are unreasonable in the context of historical land-use patterns and the socio-economic drivers of change outlined for each scenario. This effort provides a method for using the RCP land-use projections in a wide range of regional scale models. However, further investigation is needed into the performance of RCP land-use projections at the regional scale.
6. Steric Sea Level Change in Twentieth Century Historical Climate Simulation and IPCC-RCP8.5 Scenario Projection: A Comparison of Two Versions of FGOALS Model
Institute of Scientific and Technical Information of China (English)
DONG Lu; ZHOU Tianjun
2013-01-01
To reveal the steric sea level change in 20th century historical climate simulations and future climate change projections under the IPCC's Representative Concentration Pathway 8.5 (RCP8.5) scenario,the results of two versions of LASG/IAP's Flexible Global Ocean-Atmosphere-Land System model (FGOALS) are analyzed.Both models reasonably reproduce the mean dynamic sea level features,with a spatial pattern correlation coefficient of 0.97 with the observation.Characteristics of steric sea level changes in the 20th century historical climate simulations and RCP8.5 scenario projections are investigated.The results show that,in the 20th century,negative trends covered most parts of the global ocean.Under the RCP8.5 scenario,global-averaged steric sea level exhibits a pronounced rising trend throughout the 21st century and the general rising trend appears in most parts of the global ocean.The magnitude of the changes in the 21st century is much larger than that in the 20th century.By the year 2100,the global-averaged steric sea level anomaly is 18 cm and 10 cm relative to the year 1850 in the second spectral version of FGOALS (FGOALS-s2) and the second grid-point version of FGOALS (FGOALS-g2),respectively.The separate contribution of the thermosteric and halosteric components from various ocean layers is further evaluated.In the 20th century,the steric sea level changes in FGOALS-s2 (FGOALS-g2) are largely attributed to the thermosteric (halosteric) component relative to the pre-industrial control run.In contrast,in the 21st century,the thermosteric component,mainly from the upper 1000 m,dominates the steric sea level change in both models under the RCP8.5 scenario.In addition,the steric sea level change in the marginal sea of China is attributed to the thermosteric component.
7. 2007 status of climate change: Mitigation of Climate Change. Contribution of Working Group III to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change. Summary for Policy-makers
International Nuclear Information System (INIS)
The Working Group III contribution to the IPCC Fourth Assessment Report (AR4) focuses on new literature on the scientific, technological, environmental, economic and social aspects of mitigation of climate change, published since the IPCC Third Assessment Report (TAR) and the Special Reports on CO2 Capture and Storage (SRCCS) and on Safeguarding the Ozone Layer and the Global Climate System (SROC).The main aim of this summary report is to assess options for mitigating climate change. Several aspects link climate change with development issues. This report explores these links in detail, and illustrates where climate change and sustainable development are mutually reinforcing. Economic development needs, resource endowments and mitigative and adaptive capacities differ across regions. There is no one-size-fits-all approach to the climate change problem, and solutions need to be regionally differentiated to reflect different socio-economic conditions and, to a lesser extent, geographical differences. Although this report has a global focus, an attempt is made to differentiate the assessment of scientific and technical findings for the various regions. Given that mitigation options vary significantly between economic sectors, it was decided to use the economic sectors to organize the material on short- to medium-term mitigation options. Contrary to what was done in the Third Assessment Report, all relevant aspects of sectoral mitigation options, such as technology, cost, policies etc., are discussed together, to provide the user with a comprehensive discussion of the sectoral mitigation options. The report is organised into six sections after the introduction: - Greenhouse gas (GHG) emission trends; - Mitigation in the short and medium term, across different economic sectors (until 2030); - Mitigation in the long-term (beyond 2030); - Policies, measures and instruments to mitigate climate change; - Sustainable development and climate change mitigation; - Gaps in
8. Mitigation of global greenhouse gas emissions from waste: conclusions and strategies from the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report. Working Group III (Mitigation).
Science.gov (United States)
Bogner, Jean; Pipatti, Riitta; Hashimoto, Seiji; Diaz, Cristobal; Mareckova, Katarina; Diaz, Luis; Kjeldsen, Peter; Monni, Suvi; Faaij, Andre; Gao, Qingxian; Zhang, Tianzhu; Ahmed, Mohammed Abdelrafie; Sutamihardja, R T M; Gregory, Robert
2008-02-01
Greenhouse gas (GHG) emissions from post-consumer waste and wastewater are a small contributor (about 3%) to total global anthropogenic GHG emissions. Emissions for 2004-2005 totalled 1.4 Gt CO2-eq year(-1) relative to total emissions from all sectors of 49 Gt CO2-eq year(-1) [including carbon dioxide (CO2), methane (CH4), nitrous oxide (N2O), and F-gases normalized according to their 100-year global warming potentials (GWP)]. The CH4 from landfills and wastewater collectively accounted for about 90% of waste sector emissions, or about 18% of global anthropogenic methane emissions (which were about 14% of the global total in 2004). Wastewater N2O and CO2 from the incineration of waste containing fossil carbon (plastics; synthetic textiles) are minor sources. Due to the wide range of mature technologies that can mitigate GHG emissions from waste and provide public health, environmental protection, and sustainable development co-benefits, existing waste management practices can provide effective mitigation of GHG emissions from this sector. Current mitigation technologies include landfill gas recovery, improved landfill practices, and engineered wastewater management. In addition, significant GHG generation is avoided through controlled composting, state-of-the-art incineration, and expanded sanitation coverage. Reduced waste generation and the exploitation of energy from waste (landfill gas, incineration, anaerobic digester biogas) produce an indirect reduction of GHG emissions through the conservation of raw materials, improved energy and resource efficiency, and fossil fuel avoidance. Flexible strategies and financial incentives can expand waste management options to achieve GHG mitigation goals; local technology decisions are influenced by a variety of factors such as waste quantity and characteristics, cost and financing issues, infrastructure requirements including available land area, collection and transport considerations, and regulatory constraints. Existing studies on mitigation potentials and costs for the waste sector tend to focus on landfill CH4 as the baseline. The commercial recovery of landfill CH4 as a source of renewable energy has been practised at full scale since 1975 and currently exceeds 105 Mt CO2-eq year(-1). Although landfill CH4 emissions from developed countries have been largely stabilized, emissions from developing countries are increasing as more controlled (anaerobic) landfilling practices are implemented; these emissions could be reduced by accelerating the introduction of engineered gas recovery, increasing rates of waste minimization and recycling, and implementing alternative waste management strategies provided they are affordable, effective, and sustainable. Aided by Kyoto mechanisms such as the Clean Development Mechanism (CDM) and Joint Implementation (JI), the total global economic mitigation potential for reducing waste sector emissions in 2030 is estimated to be > 1000 Mt CO2-eq (or 70% of estimated emissions) at costs below 100 US$t(-1) CO2-eq year(-1). An estimated 20-30% of projected emissions for 2030 can be reduced at negative cost and 30-50% at costs 130 Mt waste year(-1) incinerated at more than 600 plants. Current uncertainties with respect to emissions and mitigation potentials could be reduced by more consistent national definitions, coordinated international data collection, standardized data analysis, field validation of models, and consistent application of life-cycle assessment tools inclusive of fossil fuel offsets. 9. Mitigation of global greenhouse gas emissions from waste: conclusions and strategies from the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report DEFF Research Database (Denmark) Bogner, J.P.; Pipatti, R.; Hashimoto, S.; 2008-01-01 protection, and sustainable development co-benefits, existing waste management practices can provide effective mitigation of GHG emissions from this sector. Current mitigation technologies include landfill gas recovery, improved landfill practices, and engineered wastewater management. In addition...... through the conservation of raw materials, improved energy and resource efficiency, and fossil fuel avoidance. Flexible strategies and financial incentives can expand waste management options to achieve GHG mitigation goals; local technology decisions are influenced by a variety of factors such as waste...... quantity and characteristics, cost and financing issues, infrastructure requirements including available land area, collection and transport considerations, and regulatory constraints. Existing studies on mitigation potentials and costs for the waste sector tend to focus on landfill CH4 as the baseline... 10. Mitigation of global greenhouse gas emissions from waste: conclusions and strategies from the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report. Working Group III (Mitigation). Science.gov (United States) Bogner, Jean; Pipatti, Riitta; Hashimoto, Seiji; Diaz, Cristobal; Mareckova, Katarina; Diaz, Luis; Kjeldsen, Peter; Monni, Suvi; Faaij, Andre; Gao, Qingxian; Zhang, Tianzhu; Ahmed, Mohammed Abdelrafie; Sutamihardja, R T M; Gregory, Robert 2008-02-01 Greenhouse gas (GHG) emissions from post-consumer waste and wastewater are a small contributor (about 3%) to total global anthropogenic GHG emissions. Emissions for 2004-2005 totalled 1.4 Gt CO2-eq year(-1) relative to total emissions from all sectors of 49 Gt CO2-eq year(-1) [including carbon dioxide (CO2), methane (CH4), nitrous oxide (N2O), and F-gases normalized according to their 100-year global warming potentials (GWP)]. The CH4 from landfills and wastewater collectively accounted for about 90% of waste sector emissions, or about 18% of global anthropogenic methane emissions (which were about 14% of the global total in 2004). Wastewater N2O and CO2 from the incineration of waste containing fossil carbon (plastics; synthetic textiles) are minor sources. Due to the wide range of mature technologies that can mitigate GHG emissions from waste and provide public health, environmental protection, and sustainable development co-benefits, existing waste management practices can provide effective mitigation of GHG emissions from this sector. Current mitigation technologies include landfill gas recovery, improved landfill practices, and engineered wastewater management. In addition, significant GHG generation is avoided through controlled composting, state-of-the-art incineration, and expanded sanitation coverage. Reduced waste generation and the exploitation of energy from waste (landfill gas, incineration, anaerobic digester biogas) produce an indirect reduction of GHG emissions through the conservation of raw materials, improved energy and resource efficiency, and fossil fuel avoidance. Flexible strategies and financial incentives can expand waste management options to achieve GHG mitigation goals; local technology decisions are influenced by a variety of factors such as waste quantity and characteristics, cost and financing issues, infrastructure requirements including available land area, collection and transport considerations, and regulatory constraints. Existing studies on mitigation potentials and costs for the waste sector tend to focus on landfill CH4 as the baseline. The commercial recovery of landfill CH4 as a source of renewable energy has been practised at full scale since 1975 and currently exceeds 105 Mt CO2-eq year(-1). Although landfill CH4 emissions from developed countries have been largely stabilized, emissions from developing countries are increasing as more controlled (anaerobic) landfilling practices are implemented; these emissions could be reduced by accelerating the introduction of engineered gas recovery, increasing rates of waste minimization and recycling, and implementing alternative waste management strategies provided they are affordable, effective, and sustainable. Aided by Kyoto mechanisms such as the Clean Development Mechanism (CDM) and Joint Implementation (JI), the total global economic mitigation potential for reducing waste sector emissions in 2030 is estimated to be > 1000 Mt CO2-eq (or 70% of estimated emissions) at costs below 100 US$ t(-1) CO2-eq year(-1). An estimated 20-30% of projected emissions for 2030 can be reduced at negative cost and 30-50% at costs 130 Mt waste year(-1) incinerated at more than 600 plants. Current uncertainties with respect to emissions and mitigation potentials could be reduced by more consistent national definitions, coordinated international data collection, standardized data analysis, field validation of models, and consistent application of life-cycle assessment tools inclusive of fossil fuel offsets. PMID:18338699
11. Projection of future climate change conditions using IPCC simulations, neural networks and Bayesian statistics. Part 2: Precipitation mean state and seasonal cycle in South America
Energy Technology Data Exchange (ETDEWEB)
Boulanger, Jean-Philippe [LODYC, UMR CNRS/IRD/UPMC, Tour 45-55/Etage 4/Case 100, UPMC, Paris Cedex 05 (France); University of Buenos Aires, Departamento de Ciencias de la Atmosfera y los Oceanos, Facultad de Ciencias Exactas y Naturales, Buenos Aires (Argentina); Martinez, Fernando; Segura, Enrique C. [University of Buenos Aires, Departamento de Computacion, Facultad de Ciencias Exactas y Naturales, Buenos Aires (Argentina)
2007-02-15
Evaluating the response of climate to greenhouse gas forcing is a major objective of the climate community, and the use of large ensemble of simulations is considered as a significant step toward that goal. The present paper thus discusses a new methodology based on neural network to mix ensemble of climate model simulations. Our analysis consists of one simulation of seven Atmosphere-Ocean Global Climate Models, which participated in the IPCC Project and provided at least one simulation for the twentieth century (20c3m) and one simulation for each of three SRES scenarios: A2, A1B and B1. Our statistical method based on neural networks and Bayesian statistics computes a transfer function between models and observations. Such a transfer function was then used to project future conditions and to derive what we would call the optimal ensemble combination for twenty-first century climate change projections. Our approach is therefore based on one statement and one hypothesis. The statement is that an optimal ensemble projection should be built by giving larger weights to models, which have more skill in representing present climate conditions. The hypothesis is that our method based on neural network is actually weighting the models that way. While the statement is actually an open question, which answer may vary according to the region or climate signal under study, our results demonstrate that the neural network approach indeed allows to weighting models according to their skills. As such, our method is an improvement of existing Bayesian methods developed to mix ensembles of simulations. However, the general low skill of climate models in simulating precipitation mean climatology implies that the final projection maps (whatever the method used to compute them) may significantly change in the future as models improve. Therefore, the projection results for late twenty-first century conditions are presented as possible projections based on the &apos
12. Emulating IPCC AR4 atmosphere-ocean and carbon cycle models for projecting global-mean, hemispheric and land/ocean temperatures: MAGICC 6.0
OpenAIRE
Meinshausen, M.; RAPER S.c.b.; Wigley, T. M. L.
2008-01-01
Current scientific knowledge on the future response of the climate system to human-induced perturbations is comprehensively captured by various model intercomparison efforts. In the preparation of the Fourth Assessment Report (AR4) of the Intergovernmental Panel on Climate Change (IPCC), intercomparisons were organized for atmosphere-ocean general circulation models (AOGCMs) and carbon cycle models, named "CMIP3" and "C4MIP", respectively. Despite their tremendous value fo...
13. Climate Change and its Impacts on Water Resources and Management of Tarbela Reservoir under IPCC Climate Change Scenarios in Upper Indus Basin, Pakistan
Science.gov (United States)
Khan, Firdos; Pilz, Jürgen
2014-05-01
Water resources play a vital role in agriculture, energy, industry, households and ecological balance. The main source of water to rivers is the Himalaya-Karakorum-Hindukush (HKH) glaciers and rainfall in Upper Indus Basin (UIB). There is high uncertainty in the availability of water in the rivers due to the variability of the monsoon, Western Disturbances, prolonged droughts and melting of glaciers in the HKH region. Therefore, proper management of water resources is undeniably important. Due to the growing population, urbanization and increased industrialization, the situation is likely to get worse. For the assessment of possible climate change, maximum temperature, minimum temperature and precipitation were investigated and evidence was found in favor of climate change in the region. Due to large differences between historical meteorological data and Regional Climate Model (RCM) simulated data, different statistical techniques were used for bias correction in temperature and precipitation. The hydrological model was calibrated for the period of 1995-2004 and validated for the period of 1990-1994 with almost 90 % efficiencies. After the application of bias correction techniques output of RCM, Providing Regional Climate for Impact Studies (PRECIS) were used as input data to the hydrological model to produce inflow projections at Tarbela reservoir on Indus River. For climate change assessment, the results show that the above mentioned variables have greater increasing trend under A2 scenario compared to B2 scenario. The projections of inflow to Tarbela reservoir show that overall 59.42 % and 34.27 % inflow increasing to Tarbela Reservoir during 2040-2069 under A2 and B2 scenarios will occur, respectively. Highest inflow and comparatively more shortage of water is noted in the 2020s under A2 scenario. Finally, the impacts of changing climate are investigated on the operation of the Tarbela reservoir. The results show that there will be shortage of water in some
14. Social Media in the Changing Ecology of News: The Fourth and Fifth Estate in Britain
OpenAIRE
Dutton, William H.; Grant Blank; Nic Newman
2012-01-01
This paper provides a case study of the changing patterns of news production and consumption in the UK that are being shaped by the Internet and related social media. Theoretically, this focus addresses concern over whether the Internet is undermining the Fourth Estate role of the press in liberal democratic societies. The case study draws from multiple methods, including survey research of individuals in Britain from 2003-2011, analysis of log files of journalistic sites, and interviews with...
15. Social Media in the Changing Ecology of News: The Fourth and Fifth Estate in Britain
Directory of Open Access Journals (Sweden)
William H. Dutton
2012-07-01
Full Text Available This paper provides a case study of the changing patterns of news production and consumption in the UK that are being shaped by the Internet and related social media. Theoretically, this focus addresses concern over whether the Internet is undermining the Fourth Estate role of the press in liberal democratic societies. The case study draws from multiple methods, including survey research of individuals in Britain from 2003-2011, analysis of log files of journalistic sites, and interviews with journalists. Survey research shows a step-jump in the use of online news since 2003 but a levelling off since 2009. However, the apparent stability in news consumption masks the growing role of social network sites. The analyses show that the Fourth Estate—the institutional news media—is using social media to enhance their role in news production and dissemination. However, networked individuals have used social media to source and distribute their own information in ways that achieve a growing independence from the Fourth Estate journalism. As more information moves online and individuals become routinely linked to the Internet, an emerging Fifth Estate, built on the activities of networked individuals sourcing and distributing their own information, is developing a synergy with the Fourth Estate as each builds on and responds to the other in this new news ecology. Comparative data suggests that this phenomenon is likely to characterize the developing news ecology in other liberal democratic societies as well, but more comparative research is required to establish the validity of this model.
16. Global Water Cycle Agreement in the Climate Models Assessed in the IPCC AR4
Science.gov (United States)
Waliser, D.; Seo, K. -W.; Schubert, S.; Njoku, E.
2007-01-01
This study examines the fidelity of the global water cycle in the climate model simulations assessed in the IPCC Fourth Assessment Report. The results demonstrate good model agreement in quantities that have had a robust global observational basis and that are physically unambiguous. The worst agreement occurs for quantities that have both poor observational constraints and whose model representations can be physically ambiguous. In addition, components involving water vapor (frozen water) typically exhibit the best (worst) agreement, and fluxes typically exhibit better agreement than reservoirs. These results are discussed in relation to the importance of obtaining accurate model representation of the water cycle and its role in climate change. Recommendations are also given for facilitating the needed model improvements.
17. Communicating uncertainty in the IPCC's greenhouse gas emissions scenarios
NARCIS (Netherlands)
Schenk, Niels J.; Lensink, Sander M.
2007-01-01
The issue of climate change required the development of the Special Report on Emission Scenarios (SRES) by the IPCC. The complexity of the subject and the unique science-policy relation resulted in confusion and discussions appeared in popular media like The Economist. This paper reviews scenario li
18. Mudanças na circulação atmosférica sobre a América do Sul para cenários futuros de clima projetados pelos modelos globais do IPCC AR4 Changes in the atmospheric circulation pattern over South America in future climate scenarios derived from the IPCC AR4 model climate simulations
Directory of Open Access Journals (Sweden)
María C Valverde
2010-03-01
19. Certainties and probabilities of the IPCC
International Nuclear Information System (INIS)
Based on an analysis of information about the climate evolution, simulations of a global warming and the snow coverage monitoring of Meteo-France, the IPCC presented its certainties and probabilities concerning the greenhouse effect. (A.L.B.)
20. The impact of climate change on summer maize phenology in the northwest plain of Shandong province under the IPCC SRES A1B scenario
International Nuclear Information System (INIS)
Climate change will affect agricultural production. Combining a climate model and a crop growth model furnishes a good approach for analyzing this effect quantitatively. The purpose of this study is to analyze the effect of climate change on summer maize phenology in northwest Shandong province under the A1B climate scenario using a regional climate model and the CERES-Maize growth model. The results showed that the temperature would increase significantly during the maize growth season in the study region, that the increased temperature would shorten the maize growth stage and result in a potential yield loss using the current cultivar, and that it is critical to breed a heat-resistant and late-maturing cultivar to maintain the yield
1. Climate change: The necessary, the possible and the desirable Earth League climate statement on the implications for climate policy from the 5th IPCC Assessment
Science.gov (United States)
Rockström, Johan; Brasseur, Guy; Hoskins, Brian; Lucht, Wolfgang; Schellnhuber, John; Kabat, Pavel; Nakicenovic, Nebojsa; Gong, Peng; Schlosser, Peter; Máñez Costa, Maria; Humble, April; Eyre, Nick; Gleick, Peter; James, Rachel; Lucena, Andre; Masera, Omar; Moench, Marcus; Schaeffer, Roberto; Seitzinger, Sybil; van der Leeuw, Sander; Ward, Bob; Stern, Nicholas; Hurrell, James; Srivastava, Leena; Morgan, Jennifer; Nobre, Carlos; Sokona, Youba; Cremades, Roger; Roth, Ellinor; Liverman, Diana; Arnott, James
2014-12-01
The development of human civilisations has occurred at a time of stable climate. This climate stability is now threatened by human activity. The rising global climate risk occurs at a decisive moment for world development. World nations are currently discussing a global development agenda consequent to the Millennium Development Goals (MDGs), which ends in 2015. It is increasingly possible to envisage a world where absolute poverty is largely eradicated within one generation and where ambitious goals on universal access and equal opportunities for dignified lives are adopted. These grand aspirations for a world population approaching or even exceeding nine billion in 2050 is threatened by substantial global environmental risks and by rising inequality. Research shows that development gains, in both rich and poor nations, can be undermined by social, economic and ecological problems caused by human-induced global environmental change. Climate risks, and associated changes in marine and terrestrial ecosystems that regulate the resilience of the climate system, are at the forefront of these global risks. We, as citizens with a strong engagement in Earth system science and socio-ecological dynamics, share the vision of a more equitable and prosperous future for the world, yet we also see threats to this future from shifts in climate and environmental processes. Without collaborative action now, our shared Earth system may not be able to sustainably support a large proportion of humanity in the decades ahead.
2. The Change of North China Climate in Transient Simulations Using the IPCC SRES A2 and B2 Scenarios with a Coupled Atmosphere-Ocean General Circulation Model
Institute of Scientific and Technical Information of China (English)
BUHE Cholaw(布和朝鲁); Ulrich CUBASCH; LIN Yonghui(林永辉); JI Liren(纪立人)
2003-01-01
This paper applies the newest emission scenarios of the sulfur and greenhouse gases, namely IPCCSRES A2 and B2 scenarios, to investigate the change of the North China climate with an atmosphere-oceancoupled general circulation nodel. In the last three decades of the 21st century, the global warming enlargesthe land-sea thermal contrast, and hence, causes the East Asian summer (winter) monsoon circulation tobe strengthened (weakened). The rainfall seasonality strengthens and the summer precipitation increasessignificantly in North China. It is suggested that the East Asian rainy area would expand northward toNorth China in the last three decades of the 21st century. In addition, the North China precipitationwould increase significantly in September. In July, August, and September, the interannual variability ofthe precipitation enlarges evidently over North China, implying a risk of flooding in the future.
3. Fourth annual progress report for Canada's Climate Change Voluntary Challenge and Registry program
International Nuclear Information System (INIS)
Examples of how greenhouse gas issues are being integrated into management processes within Suncor Energy Inc. are described in this fourth annual progress report to the Climate Change Voluntary Challenge and Registry Program. The report covers Suncor's three operating businesses - oil sands and conventional oil exploration and production in Western Canada, and refining and marketing operation in Ontario. Oil sands was the largest source of greenhouse emissions, accounting for 2/3 of the total. Carbon dioxide emissions accounted for 93 per cent of total emissions. This report addresses three areas of change: one of these is Project Millennium in the oil sands division, which is a major expansion project planned for efficiency improvements. As a result of the project, total greenhouse gas emissions will increase to 9.3 million tonnes by the year 2002, in terms of operating efficiency, emissions per unit of production will continue to decline from 0.54 tonnes ECO2 in 1990 to 0.44 tonnes ECO2 in 2002, a reduction of 18 per cent. Another change is that target reductions in the Kyoto Protocol will supersede informal Canadian commitments for the year 2000, if the protocol is ratified. Thirdly, Suncor's greenhouse gas emission forecast has been extended to the year 2002 to demonstrate the impact of Project Millennium and to clarify the changes during the transition period relative to previous forecasts. New initiatives to be undertaken during 1998-2002 include heat recovery in new upgrader units, recycling diluent used in bitumen extraction without cooling, recovery of gas presently going to the flare system, installation of a 200,000 barrel hot water surge tank, addition of a third turbogenerator, and various projects to generate more electrical power internally. tabs., figs
4. Observing, studying, and managing for change-Proceedings of the Fourth Interagency Conference on Research in the Watersheds
Science.gov (United States)
Medley, Nicolas; Patterson, Glenn; Parker, Melanie J.
2011-01-01
These proceedings contain the abstracts, manuscripts, and posters of presentations given at the Fourth Interagency Conference on Research in the Watersheds-Observing, Studying, and Managing for Change, held at the Westmark Hotel in Fairbanks, Alaska, September 26-30, 2011. The conference was jointly hosted by the Bureau of Land Management and the National Park Service.
5. Possible future climates. The IPCC-scenarios simulated by dialogue
Energy Technology Data Exchange (ETDEWEB)
Hoekstra, J. [KEMA-KES, Arnheim (Netherlands)
1995-12-31
Global warming is an environmental problem that increasingly attracts the attention of governments, (inter)national organizations and the general public. Policymakers that want to attack this problem need to understand the causes and effects of all related aspects. For this reason integrated assessment tools are developed that allow policymakers to analyze and evaluate climate change scenarios. Dialogue is such an integrated assessment tool. This article presents the results of Dialogue when the socio-economic parameters of the six well-known IPCC-scenarios, IS92a-f (IPCC 1992) are taken as a point of departure. Using as input, variables as population growth and the energy intensity of an economy, Dialogue goes through a chain of processes and finally determines climatic changes in temperature and precipitation
6. Climate effects of land use changes and anthropogenic impact on surface radiation
OpenAIRE
2009-01-01
The fourth assessment report on climate change (AR4) was released in 2007 and the Intergovernmental Panel of Climate Change (IPCC) derive an increase of 0.74 ± 0.18°C in the 100 year global mean surface temperature linear trend between 1906 – 2005. IPCC state further that “there is very high confidence that the global average net effect of human activities since 1750 has been one of warming” (IPCC, 2007). The observed global warming has occurred during the same period as a considerable increa...
7. Maynard Participation in Alaska Forum on the Environment Panel Discussion on Increasing Input to the US National Climate Assessment (NCA) and the Intergovernmental Panel on Climate Change (IPCC) Processes from Alaska, with Emphasis on Indigenous Peoples Issues
Science.gov (United States)
Maynard, Nancy G.
2012-01-01
Dr. Nancy Maynard was invited by the Alaska Forum on the Environment to participate in a Panel Discussion to discuss (1) background about what the US NCA and International IPCC assessments are, (2) the impact the assessments have on policy-making, (3) the process for participation in both assessments, (4) how we can increase participation by Indigenous Peoples such as Native Americans and Alaska Natives, (5) How we can increase historical and current impacts input from Native communities through stories, oral history, "grey" literature, etc. The session will be chaired by Dr. Bull Bennett, a cochair of the US NCA's chapter on "Native and Tribal Lands and Resources" and Dr. Maynard is the other co-chair of that chapter and they will discuss the latest activities under the NCA process relevant to Native Americans and Alaska Natives. Dr. Maynard is also a Lead Author of the "Polar Regions" chapter of the IPCC WG2 (5th Assessment) and she will describes some of the latest approaches by the IPCC to entrain more Indigenous peoples into the IPCC process.
8. Who did write the report of the IPCC Group 3?
International Nuclear Information System (INIS)
After having briefly recalled the role of each of the groups of the IPCC, this brief document recalls the role of Group 3 (to define strategies aimed at climate change mitigation) and indicates the number of authors of different nationalities for the report published in April 2014. He analyses how these nationalities are distributed for some specific and important chapters. He outlines and denounces the low representation of France among these authors, and notices that some contents are in contradiction with the French energy policy. Somehow consequently, he also criticizes the way this report considers nuclear energy (mining risks, operational risks, civil-military relationship, nuclear waste control)
9. Greenhouse gas emissions from managed peat soils: is the IPCC reporting guidance realistic?
Directory of Open Access Journals (Sweden)
J. Couwenberg
2011-03-01
Full Text Available Drainage of peatlands leads to the decomposition of peat, resulting in substantial losses of carbon and nitrogen to the atmosphere. The conservation and restoration of peatlands can provide a major contribution to the mitigation of climate change. Improvements to guidance and capacity for reporting of greenhouse gas emissions from peatlands will be valuable in the context of the current negotiations towards a post-2012 climate agreement. This article evaluates IPCC approaches to greenhouse gas emissions from managed organic (peat soils and presents a summary table comparing IPCC default values with best estimates based on recent literature. Inconsistencies are pointed out with regard to the IPCC definitions of organic soils and climate zones. The 2006 IPCC Guidelines use a definition of organic soil that is not totally consistent with FAO definitions, use climate zones that are not fully compatible, present default CO2 values that are substantially (often an order of magnitude too low, and present a default N2O value for tropical cropland that is also an order of magnitude too low. An update of IPCC default values is desirable. The IPCC Emission Factor Database offers a platform for establishing more accurate emission factors, but so far contains little information about emissions from peat soils.
10. Report of the Joint IPCC WG 2 and 3 expert meeting on the integration of adaptation, mitigation and sustainable development into the 4. IPCC assessment report
International Nuclear Information System (INIS)
The objectives for this meeting at Reunion Island were: - To feed new views from outside the climate change literature into the assessment of Working Group II (WG II) and WG III concerning the strongly interrelated area of adaptation, mitigation and sustainable development. - to dove-tail zero-order draft texts of WG II and WG III (by the authors) with a view to ensuring that the treatment of Adaptation and Mitigation (AM) and Sustainable Development (SD) issues in both assessments is: 'Consistent, Complementary, Concise and Complete' ('4 Cs'). Furthermore, it was decided that the deliverable should be: - Recommendations for the writing team of WG II fourth Assessment Report (AR4) for incorporation of AM and SD issues in their First Order Draft (following their 2. Lead Author meeting in Cairns, 14-17 March 2005); - Recommendations for the writing team of WG III for incorporation in their Zero-order Draft (ZOD, to be completed 11 March 2005) The programme of the meeting was developed by the TSUs of WG II and III under the responsibility of the co-chairs of WG II and III. Day 1 the programme was devoted to a series of key note speakers, covering both potential user views as well as relevant new perspectives on the handling of AM and SD issues. These areas have not been fully addressed in the IPCC assessment work to date. The invited experts elaborated on 'new science areas' or 'new literatures' that inform parts of the AR4. The morning programme of Day 1 also contained an opening session featuring several ministers of Environment of neighbouring Small Island States, a representative of the European Parliament, and government officials from both the French Republic and Reunion Island. Day 2 and 3 were used for working sessions between authors on the integration of adaptation, mitigation and sustainable development into the contributions of Working Groups II and III of the AR4. The full programme is attached to the document. The meeting brought together more than forty
11. Report of the Joint IPCC WG 2 and 3 expert meeting on the integration of adaptation, mitigation and sustainable development into the 4. IPCC assessment report
Energy Technology Data Exchange (ETDEWEB)
NONE
2005-07-01
The objectives for this meeting at Reunion Island were: - To feed new views from outside the climate change literature into the assessment of Working Group II (WG II) and WG III concerning the strongly interrelated area of adaptation, mitigation and sustainable development. - to dove-tail zero-order draft texts of WG II and WG III (by the authors) with a view to ensuring that the treatment of Adaptation and Mitigation (AM) and Sustainable Development (SD) issues in both assessments is: 'Consistent, Complementary, Concise and Complete' ('4 Cs'). Furthermore, it was decided that the deliverable should be: - Recommendations for the writing team of WG II fourth Assessment Report (AR4) for incorporation of AM and SD issues in their First Order Draft (following their 2. Lead Author meeting in Cairns, 14-17 March 2005); - Recommendations for the writing team of WG III for incorporation in their Zero-order Draft (ZOD, to be completed 11 March 2005) The programme of the meeting was developed by the TSUs of WG II and III under the responsibility of the co-chairs of WG II and III. Day 1 the programme was devoted to a series of key note speakers, covering both potential user views as well as relevant new perspectives on the handling of AM and SD issues. These areas have not been fully addressed in the IPCC assessment work to date. The invited experts elaborated on 'new science areas' or 'new literatures' that inform parts of the AR4. The morning programme of Day 1 also contained an opening session featuring several ministers of Environment of neighbouring Small Island States, a representative of the European Parliament, and government officials from both the French Republic and Reunion Island. Day 2 and 3 were used for working sessions between authors on the integration of adaptation, mitigation and sustainable development into the contributions of Working Groups II and III of the AR4. The full programme is attached to the document. The
12. IPCC. 4. climate assessment report, 2007; GIEC. 4. rapport d'evaluation du climat, 2007
Energy Technology Data Exchange (ETDEWEB)
NONE
2007-07-01
The mission and challenge of the Intergovernmental panel on climate change (IPCC, GIEC in French) is to evaluate, synthesize and make available the sum of scientific and economic information of the complex domain of climatic change, and in addition to make the results of these works accepted by government representatives of 192 states. This document makes a brief synthesis in three parts of the 4. assessment report of the IPCC: 1 - physical scientific bases of climatic change: characteristic of the phenomenon, greenhouse gas emissions trend, already observed effects, forecasts of climate models; 2 - impacts, adaptations and vulnerabilities of climatic change: types of future impacts, impacts per sector, regional impacts, limits of ecosystems adaptation; 3 - mitigation of climatic changes: past emissions and future trends, possible mitigation actions and cost, possible political levers for emissions abatement. A last part introduces the French researchers involved in IPCC's works. (J.S.)
13. 2007 status of climate change: Mitigation of Climate Change. Contribution of Working Group III to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change. Summary for Policy-makers; Bilan 2007 des changements climatiques: l'attenuation des changements climatiques. Contribution du Groupe de travail 3 au quatrieme rapport d'evaluation du Groupe d'Experts Intergouvernemental sur l'Evolution du Climat (GIEC). Resume a l'attention des decideurs
Energy Technology Data Exchange (ETDEWEB)
Barker, T.; Bashmakov, I.; Bernstein, L.; Bogner, J.; Bosch, P.; Dave, R.; Davidson, O.; Fisher, B.; Grubb, M.; Gupta, S.; Halsnaes, K.; Heij, B.; Kahn Ribeiro, S.; Kobayashi, S.; Levine, M.; Martino, D.; Masera Cerutti, O.; Metz, B.; Meyer, L.; Nabuurs, G.J.; Najam, A.; Nakicenovic, N.; Holger Rogner, H.; Roy, J.; Sathaye, J.; Schock, R.; Shukla, P.; Sims, R.; Smith, P.; Swart, R.; Tirpak, D.; Urge-Vorsatz, D.; Dadi, Z
2007-07-01
The Working Group III contribution to the IPCC Fourth Assessment Report (AR4) focuses on new literature on the scientific, technological, environmental, economic and social aspects of mitigation of climate change, published since the IPCC Third Assessment Report (TAR) and the Special Reports on CO{sub 2} Capture and Storage (SRCCS) and on Safeguarding the Ozone Layer and the Global Climate System (SROC).The main aim of this summary report is to assess options for mitigating climate change. Several aspects link climate change with development issues. This report explores these links in detail, and illustrates where climate change and sustainable development are mutually reinforcing. Economic development needs, resource endowments and mitigative and adaptive capacities differ across regions. There is no one-size-fits-all approach to the climate change problem, and solutions need to be regionally differentiated to reflect different socio-economic conditions and, to a lesser extent, geographical differences. Although this report has a global focus, an attempt is made to differentiate the assessment of scientific and technical findings for the various regions. Given that mitigation options vary significantly between economic sectors, it was decided to use the economic sectors to organize the material on short- to medium-term mitigation options. Contrary to what was done in the Third Assessment Report, all relevant aspects of sectoral mitigation options, such as technology, cost, policies etc., are discussed together, to provide the user with a comprehensive discussion of the sectoral mitigation options. The report is organised into six sections after the introduction: - Greenhouse gas (GHG) emission trends; - Mitigation in the short and medium term, across different economic sectors (until 2030); - Mitigation in the long-term (beyond 2030); - Policies, measures and instruments to mitigate climate change; - Sustainable development and climate change mitigation; - Gaps in
14. Downscaling the Impacts of Large-Scale LUCC on Surface Temperature along with IPCC RCPs: A Global Perspective
Directory of Open Access Journals (Sweden)
Xiangzheng Deng
2014-04-01
Full Text Available This study focuses on the potential impacts of large-scale land use and land cover changes (LUCC on surface temperature from a global perspective. As important types of LUCC, urbanization, deforestation, cultivated land reclamation, and grassland degradation have effects on the climate, the potential changes of the surface temperature caused by these four types of large-scale LUCC from 2010 to 2050 are downscaled, and this issue analyzed worldwide along with Representative Concentration Pathways (RCPs of the Intergovernmental Panel on Climate Change (IPCC. The first case study presents some evidence of the effects of future urbanization on surface temperature in the Northeast megalopolis of the United States of America (USA. In order to understand the potential climatological variability caused by future forest deforestation and vulnerability, we chose Brazilian Amazon region as the second case study. The third selected region in India as a typical region of cultivated land reclamation where the possible climatic impacts are explored. In the fourth case study, we simulate the surface temperature changes caused by future grassland degradation in Mongolia. Results show that the temperature in built-up area would increase obviously throughout the four land types. In addition, the effects of all four large-scale LUCC on monthly average temperature change would vary from month to month with obviously spatial heterogeneity.
15. The emerging fourth sector
OpenAIRE
Friis, Andreas
2009-01-01
The Fourth Sector is a new phenomenon related to dual social and financial value creation, until now not clearly defined with crisp sectoral boundaries and an operationalized definition. The phenomenon is getting increased attention in media, conferences, business schools and by organizations all over the world, and this study intends to explain the macroenvironmental changes that led to the rise of the fourth sector, describe new trends of social value creation in the private ...
16. Evaluating the use of verbal probability expressions to communicate likelihood information in IPCC reports
Science.gov (United States)
2014-05-01
The Intergovernmental Panel on Climate Change (IPCC) prescribes that the communication of risk and uncertainty information pertaining to scientific reports, model predictions etc. be communicated with a set of 7 likelihood expressions. These range from "Extremely likely" (intended to communicate a likelihood of greater than 99%) through "As likely as not" (33-66%) to "Extremely unlikely" (less than 1%). Psychological research has investigated the degree to which these expressions are interpreted as intended by the IPCC, both within and across cultures. I will present a selection of this research and demonstrate some problems associated with communicating likelihoods in this way, as well as suggesting some potential improvements.
17. The IPCC in an age of social media
Science.gov (United States)
Hickman, Leo
2015-04-01
How should the IPCC communicate its findings, not just to policymakers, but to a wider audience? In today's online environment, readers demand an open and transparent interaction, but the responses must be both rapid and authoritative. As the IPCC debates its future, it must be bold in engaging with social media.
18. Exploring the impact of the IPCC Assessment Reports on science
NARCIS (Netherlands)
Vasileiadou, E.; Heimeriks, G.J.; Petersen, A.C.
2011-01-01
Even though critique to IPCC is certainly not new, the climate controversies of 2009 and 2010 brought this critique again to the fore in public media. The paper contributes to this ongoing debate, and investigates empirically the impact of the four Assessment Reports of the IPCC on scientific public
19. Effects of Conceptual Change Texts and Laboratory Experiments on Fourth Grade Students' Understanding of Matter and Change Concepts
Science.gov (United States)
Durmus, Jale; Bayraktar, Sule
2010-01-01
The purpose of this study was to investigate whether conceptual change texts and laboratory experiments are effective in overcoming misconceptions and whether the concepts were acquired permanently when these methods were utilized. In this study, we addressed some topics from the "Matter and Change" unit in science and technology class of…
20. A method for evaluating climate change adaptation strategies for small-scale farmers using survey, experimental and modeled data
NARCIS (Netherlands)
Claessens, L.F.G.; Antle, J.M.; Stoorvogel, J.J.; Valdivia, R.O.; Thornton, P.K.; Herrero, M.
2012-01-01
Sub-Saharan Africa (SSA) is predicted to experience considerable negative impacts of climate change. The IPCC Fourth Assessment emphasizes that adaptation strategies are essential. Addressing adaptation in thecontext of small-scale, semi-subsistence agriculture raises special challenges. High data d
1. Cloud Radiative Forcing in Asian Monsoon Region Simulated by IPCC AR4 AMIP Models
Institute of Scientific and Technical Information of China (English)
LI Jiandong; LIU Yimin; WU Guoxiong
2009-01-01
This study examines cloud radiative forcing (CRF) in the Asian monsoon region (0°-50°N,60°-150°E)simulated by Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4) AMIP models.During boreal winter,no model realistically reproduces the larger long-wave cloud radiative forcing (LWCF) over the Tibet Plateau (TP) and only a couple of models reasonably capture the larger short-wave CRF (SWCF) to the east of the TP.During boreal summer,there are larger biases for central location and intensity of simulated CRF in active convective regions.The CRF biases are closely related to the rainfall biases in the models.Quantitative analysis further indicates that the correlation between simulated CRF and observations are not high,and that the biases and diversity in SWCF are larger than that in LWCF.The annual cycle of simulated CRF over East Asia (0°-50°N,100°-145°E) is also examined.Though many models capture the basic annual cycle in tropics,strong LWCF and SWCF to the east of the TP beginning in early spring are underestimated by most models.As a whole,GFDL-CM2.1,MPI-ECHAM5,UKMO-HadGAM1,and MIROC3.2 (medres) perform well for CRF simulation in the Asian monsoon region,and the multi-model ensemble (MME) has improved results over the individual simulations. It is suggested that strengthening the physical parameterizations involved over the TP,and improving cumulus convection processes and model experiment design are crucial to CRF simulation in the Asian monsoon region.
2. Changes in Test Scores Distribution for Students of the Fourth Grade in Brazil: A Relative Distribution Analysis for the Years 1997-2005
Science.gov (United States)
Rodrigues, Clarissa Guimaraes; Rios-Neto, Eduardo Luiz Goncalves; de Xavier Pinto, Cristine Campos
2013-01-01
In Brazil, the mean of math test scores for students of the fourth grade declined by approximately 0.2 standard deviation in the late 1990s. However, the potential changes in the distribution of scores have never been addressed. It is unclear if the decline was caused by deterioration in student performance levels at the upper and/or lower tails…
3. A multistage crucible of revision and approval shapes IPCC policymaker summaries.
Science.gov (United States)
Mach, Katharine J; Freeman, Patrick T; Mastrandrea, Michael D; Field, Christopher B
2016-08-01
Intergovernmental Panel on Climate Change (IPCC) member governments approve each report's summary for policymakers (SPM) by consensus, discussing and agreeing on each sentence in a plenary session with scientist authors. A defining feature of IPCC assessment, the governmental approval process builds joint ownership of current knowledge by scientists and governments. The resulting SPM revisions have been extensively discussed in anecdotes, interviews, and perspectives, but they have not been comprehensively analyzed. We provide an in-depth evaluation of IPCC SPM revisions, establishing an evidential basis for understanding their nature. Revisions associated with governmental review and approval generally expand SPMs, with SPM text growing by 17 to 53% across recent assessment reports. Cases of high political sensitivity and failure to reach consensus are notable exceptions, resulting in SPM contractions. In contrast to recent claims, we find that IPCC SPMs are as readable, for multiple metrics of reading ease, as other professionally edited assessment summaries. Across reading-ease metrics, some SPMs become more readable through governmental review and approval, whereas others do not. In an SPM examined through the entire revision process, most revisions associated with governmental review and approval occurred before the start of the government-approval plenary session. These author revisions emphasize clarity, scientific rigor, and explanation. In contrast, the subsequent plenary revisions place greater emphasis especially on policy relevance, comprehensiveness of examples, and nuances of expert judgment. Overall, the value added by the IPCC process emerges in a multistage crucible of revision and approval, as individuals together navigate complex science-policy terrain. PMID:27532046
4. Patterns of authorship in the IPCC Working Group III report
Science.gov (United States)
Corbera, Esteve; Calvet-Mir, Laura; Hughes, Hannah; Paterson, Matthew
2016-01-01
The Intergovernmental Panel on Climate Change (IPCC) has completed its Fifth Assessment Report (AR5). Here, we explore the social scientific networks informing Working Group III (WGIII) assessment of mitigation for the AR5. Identifying authors’ institutional pathways, we highlight the persistence and extent of North-South inequalities in the authorship of the report, revealing the dominance of US and UK institutions as training sites for WGIII authors. Examining patterns of co-authorship between WGIII authors, we identify the unevenness in co-authoring relations, with a small number of authors co-writing regularly and indicative of an epistemic community’s influence over the IPCC’s definition of mitigation. These co-authoring networks follow regional patterns, with significant EU-BRICS collaboration and authors from the US relatively insular. From a disciplinary perspective, economists, engineers, physicists and natural scientists remain central to the process, with insignificant participation of scholars from the humanities. The shared training and career paths made apparent through our analysis suggest that the idea that broader geographic participation may lead to a wider range of viewpoints and cultural understandings of climate change mitigation may not be as sound as previously thought.
5. Reconstruction of the boundary between climate science and politics: the IPCC in the Japanese mass media, 1988-2007.
Science.gov (United States)
Asayama, Shinichiro; Ishii, Atsushi
2014-02-01
The Intergovernmental Panel on Climate Change (IPCC) plays a significant role in bridging the boundary between climate science and politics. Media coverage is crucial for understanding how climate science is communicated and embedded in society. This study analyzes the discursive construction of the IPCC in three Japanese newspapers from 1988 to 2007 in terms of the science-politics boundary. The results show media discourses engaged in boundary-work which rhetorically separated science and politics, and constructed the iconic image of the IPCC as a pure scientific authority. In the linkages between the global and national arenas of climate change, the media "domesticate" the issue, translating the global nature of climate change into a discourse that suits the national context. We argue that the Japanese media's boundary-work is part of the media domestication that reconstructed the boundary between climate science and politics reflecting the Japanese context.
6. Climate modelling: IPCC gazes into the future
Science.gov (United States)
Raper, Sarah
2012-04-01
In 2013, the Intergovernmental Panel on Climate Change will report on the next set of future greenhouse-gas emission scenarios, offering a rational alternative pathway for avoiding dangerous climate change.
7. Global Income Distribution and Poverty: Implications from the IPCC SRES Scenarios
OpenAIRE
2010-01-01
The Special Report on Emissions Scenarios (SRES) has been widely used to analyze climate change impacts, vulnerability and adaptation. The storylines behind these scenarios outline alternative development pathways, which have been the base for climate research and other studies at global, regional and country level. Based on the global income distribution and poverty module (GlobPov), we assess the implication of the IPCC SRES scenarios on global poverty and inequality. We find that global po...
8. Mutations in the fourth EGF-like domain affect thrombomodulin-induced changes in the active site of thrombin
OpenAIRE
Koeppe, Julia R.; Beach, Muneera A.; Baerga-Ortiz, Abel; Jordan Kerns, S.; Komives, Elizabeth A.
2008-01-01
A number of alanine and more conservative mutants of residues in the fourth domain of thrombomodulin (TM) were prepared and assayed for protein C activation and for thrombin binding. Several of the alanine mutations appeared to cause misfolding or structural defects as assessed by poor expression and/or NMR HSQC experiments, while more conservative mutations at the same site appeared to fold correctly and retain activity. Several of the conservative mutants bound more weakly to thrombin despi...
9. Climate in Peril. A popular guide to the latest IPCC reports
Energy Technology Data Exchange (ETDEWEB)
2009-07-01
In 2007 the Intergovernmental Panel on Climate Change (IPCC) shared the Nobel Peace Prize with former US Vice President Al Gore for their work to provide policy makers and the general public around the world with the best possible science base for understanding and combating the increasing threat from climate change. But as the messages from the scientists are becoming increasingly explicit, the gap between the need for action they project and the climate policy the world leaders put in place is steadily increasing. One illustration is the trend in emissions of greenhouse gases. According to the IPCC global emissions would need to peak between 2000 and 2015 in order to limit the global temperature increase to between 2 and 2.4 degrees C compared to pre-industrial times. In 2007, when ideally the emissions should have peaked, the world instead experienced a new record in annual emission increase. For each day we fail to twist development towards a low-carbon society, the damage to the world's ecosystems become more severe, and the costs of mitigation and adaptation increases. The main purpose of this short guide is to help bridging the gap between science and policy and to increase public awareness about the urgency of action to combat climate change and its impacts. This booklet is intended for those who do not have the time - and may not have the scientific expertise - to read the entire Synthesis Report from the IPCC. (Author)
10. A review of uncertainty visualization within the IPCC reports
Science.gov (United States)
Nocke, Thomas; Reusser, Dominik; Wrobel, Markus
2015-04-01
Results derived from climate model simulations confront non-expert users with a variety of uncertainties. This gives rise to the challenge that the scientific information must be communicated such that it can be easily understood, however, the complexity of the science behind is still incorporated. With respect to the assessment reports of the IPCC, the situation is even more complicated, because heterogeneous sources and multiple types of uncertainties need to be compiled together. Within this work, we systematically (1) analyzed the visual representation of uncertainties in the IPCC AR4 and AR5 reports, and (2) executed a questionnaire to evaluate how different user groups such as decision-makers and teachers understand these uncertainty visualizations. Within the first step, we classified visual uncertainty metaphors for spatial, temporal and abstract representations. As a result, we clearly identified a high complexity of the IPCC visualizations compared to standard presentation graphics, sometimes even integrating two or more uncertainty classes / measures together with the "certain" (mean) information. Further we identified complex written uncertainty explanations within image captions even within the "summary reports for policy makers". In the second step, based on these observations, we designed a questionnaire to investigate how non-climate experts understand these visual representations of uncertainties, how visual uncertainty coding might hinder the perception of the "non-uncertain" data, and if alternatives for certain IPCC visualizations exist. Within the talk/poster, we will present first results from this questionnaire. Summarizing, we identified a clear trend towards complex images within the latest IPCC reports, with a tendency to incorporate as much as possible information into the visual representations, resulting in proprietary, non-standard graphic representations that are not necessarily easy to comprehend on one glimpse. We conclude that
11. Tropical Intraseasonal Variability in 14 IPCC AR4 Climate Models Part I: Convective Signals
Energy Technology Data Exchange (ETDEWEB)
Lin, J; Kiladis, G N; Mapes, B E; Weickmann, K M; Sperber, K R; Lin, W; Wheeler, M; Schubert, S D; Genio, A D; Donner, L J; Emori, S; Gueremy, J; Hourdin, F; Rasch, P J; Roeckner, E; Scinocca, J F
2005-05-06
This study evaluates the tropical intraseasonal variability, especially the fidelity of Madden-Julian Oscillation (MJO) simulations, in 14 coupled general circulation models (GCMs) participating in the Inter-governmental Panel on Climate Change (IPCC) Fourth Assessment Report (AR4). Eight years of daily precipitation from each model's 20th century climate simulation are analyzed and compared with daily satellite retrieved precipitation. Space-time spectral analysis is used to obtain the variance and phase speed of dominant convectively coupled equatorial waves, including the MJO, Kelvin, equatorial Rossby (ER), mixed Rossby-gravity (MRG), and eastward inertio-gravity (EIG) and westward inertio-gravity (WIG) waves. The variance and propagation of the MJO, defined as the eastward wavenumbers 1-6, 30-70 day mode, are examined in detail. The results show that current state-of-the-art GCMs still have significant problems and display a wide range of skill in simulating the tropical intraseasonal variability. The total intraseasonal (2-128 day) variance of precipitation is too weak in most of the models. About half of the models have signals of convectively coupled equatorial waves, with Kelvin and MRG-EIG waves especially prominent. However, the variances are generally too weak for all wave modes except the EIG wave, and the phase speeds are generally too fast, being scaled to excessively deep equivalent depths. An interesting result is that this scaling is consistent within a given model across modes, in that both the symmetric and antisymmetric modes scale similarly to a certain equivalent depth. Excessively deep equivalent depths suggest that these models may not have a large enough reduction in their ''effective static stability'' due to diabatic heating. The MJO variance approaches the observed value in only two of the 14 models, but is less than half of the observed value in the other 12 models. The ratio between the eastward MJO variance
12. Methods for Assessing Uncertainties in Climate Change, Impacts and Responses (Invited)
Science.gov (United States)
Manning, M. R.; Swart, R.
2009-12-01
Assessing the scientific uncertainties or confidence levels for the many different aspects of climate change is particularly important because of the seriousness of potential impacts and the magnitude of economic and political responses that are needed to mitigate climate change effectively. This has made the treatment of uncertainty and confidence a key feature in the assessments carried out by the Intergovernmental Panel on Climate Change (IPCC). Because climate change is very much a cross-disciplinary area of science, adequately dealing with uncertainties requires recognition of their wide range and different perspectives on assessing and communicating those uncertainties. The structural differences that exist across disciplines are often embedded deeply in the corresponding literature that is used as the basis for an IPCC assessment. The assessment of climate change science by the IPCC has from its outset tried to report the levels of confidence and uncertainty in the degree of understanding in both the underlying multi-disciplinary science and in projections for future climate. The growing recognition of the seriousness of this led to the formation of a detailed approach for consistent treatment of uncertainties in the IPCC’s Third Assessment Report (TAR) [Moss and Schneider, 2000]. However, in completing the TAR there remained some systematic differences between the disciplines raising concerns about the level of consistency. So further consideration of a systematic approach to uncertainties was undertaken for the Fourth Assessment Report (AR4). The basis for the approach used in the AR4 was developed at an expert meeting of scientists representing many different disciplines. This led to the introduction of a broader way of addressing uncertainties in the AR4 [Manning et al., 2004] which was further refined by lengthy discussions among many IPCC Lead Authors, for over a year, resulting in a short summary of a standard approach to be followed for that
13. Contribution of the working group 2 to the fourth evaluation report of the inter government expert group on the climatic change. Evaluation 2007 of the climatic changes: impacts, adaptation and vulnerability
International Nuclear Information System (INIS)
This document exposes the results of the fourth evaluation report of the working group II of the inter government experts group on the climatic change. This evaluation presents the today scientific understanding of the climatic change impacts on the humans and their adaptation ability and vulnerability. It is based on the GIEC evaluations and new knowledge added since the third evaluation report. (A.L.B.)
14. Run-based multi-model interannual variability assessment of precipitation and temperature over Pakistan using two IPCC AR4-based AOGCMs
Science.gov (United States)
Asmat, U.; Athar, H.
2015-09-01
The interannual variability of precipitation and temperature is derived from all runs of the Intergovernmental Panel on Climate Change (IPCC) fourth Assessment Report (AR4)-based two Atmospheric Oceanic General Circulation Model (AOGCM) simulations, over Pakistan, on an annual basis. The models are the CM2.0 and CM2.1 versions of Geophysical Fluid Dynamics Laboratory (GFDL)-based AOGCM. Simulations for a recent 22-year period (1979-2000) are validated using Climate Research Unit (CRU) and NCEP/NCAR datasets over Pakistan, for the first time. The study area of Pakistan is divided into three regions: all Pakistan, northern Pakistan, and southern Pakistan. Bias, root mean square error, one sigma standard deviation, and coefficient of variance are used as validation metrics. For all Pakistan and northern Pakistan, all three runs of GFDL-CM2.0 perform better under the above metrics, both for precipitation and temperature (except for one sigma standard deviation and coefficient of variance), whereas for southern Pakistan, third run of GFDL-CM2.1 perform better expect for the root mean square error for temperature. A mean and variance-based bias correction is applied to bias in modeled precipitation and temperature variables. This resulted in a reduced bias, except for the months of June, July, and August, when the reduction in bias is relatively lower.
15. Federated Quality Control Procedure for CMIP5 / IPCC-AR5 Data
Science.gov (United States)
Stockhause, M.; Höck, H.; Kurtz, M.; Lautenschlager, M.; Toussaint, F.
2012-04-01
The International Panel on Climate Change (IPCC) aims to advance the knowledge of climate change and climate variability. The results collected within the Climate Model Intercomparison Project No. 5 (CMIP5) are intended to underlie the coming fifth assessment report (IPCC-AR5). In comparison with the CMIP3 (IPCC-AR4) three main improvements have been implemented in the data infrastructure: Decentral data storage on local data nodes of the Earth System Grid (ESG, http://pcmdi3.llnl.gov/esgcet) and replication of the most important data (relevant for IPCC-AR5) among the three primary CMIP5 archive centers, PCMDI (Program for Climate Model Diagnosis and Intercomparison), BADC (British Atmospheric Data Centre), and WDCC (World Data Center for Climate) at DKRZ. Detailed descriptions of numerical climate models and the simulations using the CIM (Common Information Model) developed by METAFOR (http://metaforclimate.eu). Data curation was improved by introducing a versioning concept and a quality assessment process providing a uniform identification of datasets as well as a persistent identifier DOI (Digital Object Identifier) for data citation in scientific publications (http://cmip5qc.wdc-climate.de). The quality control (QC) concept was developed on the background of the existing federated data infrastructure of the ESG and the external metadata source. For this reason as well as for sharing the work load of the quality checks a federated / distributed quality control procedure was developed, consisting of: a QC repository for QC result and information storage and exchange within the QC process, a QC checker tool, and a QC service package to support QC repository storage, QC result analyses, QC information access for QC managers as well as for data users. For CMIP5 the quality control procedure consists of three quality levels. With increasing quality level the checks are performed more centralized: QC level 1 Separate technical QC checks on data (CMOR2, ESG conformance
16. South Asian Summer Monsoon and Its Relationship with ENSO in the IPCC AR4 Simulations
Energy Technology Data Exchange (ETDEWEB)
Annamalai, H; Hamilton, K; Sperber, K R
2005-09-07
In this paper we use the extensive integrations produced for the IPCC Fourth Assessment Report (AR4) to examine the relationship between ENSO and the monsoon at interannual and decadal timescales. We begin with an analysis of the monsoon simulation in the 20th century integrations. Six of the 18 models were found to have a reasonably realistic representation of monsoon precipitation climatology. For each of these six models SST and anomalous precipitation evolution along the equatorial Pacific during El Nino events display considerable differences when compared to observations. Out of these six models only four (GFDL{_}CM{_}2.0, GFDL{_}CM{_}2.1, MRI, and MPI{_}ECHAM5) exhibit a robust ENSO-monsoon contemporaneous teleconnection, including the known inverse relationship between ENSO and rainfall variations over India. Lagged correlations between the all-India rainfall (AIR) index and Nino3.4 SST reveal that three models represent the timing of the teleconnection, including the spring predictability barrier which is manifested as the transition from positive to negative correlations prior to the monsoon onset. Furthermore, only one of these three models (GFDL{_}CM{_}2.1) captures the observed phase lag with the strongest anticorrelation of SST peaking 2-3 months after the summer monsoon, which is partially attributable to the intensity of simulated El Nino itself. We find that the models that best capture the ENSO-monsoon teleconnection are those that correctly simulate the timing and location of SST and diabatic heating anomalies in the equatorial Pacific, and the associated changes to the equatorial Walker Circulation during El Nino events. The strength of the AIR-Nino3.4 SST correlation in the model runs waxes and wanes to some degree on decadal timescales. The overall magnitude and timescale for this decadal modulation in most of the models is similar to that seen in observations. However, there is little consistency in the phase among the realizations
17. ON THE PAUSED WARMING CONTROVERSY BASED ON IPCC AR5 AND BEYOND
Directory of Open Access Journals (Sweden)
MIKA J.
2014-03-01
Full Text Available The paused warming since ca. 2002 (maybe, 1998 is not satisfactorily reflected by the IPCC WGI (2013 Report. The aim of the present study is to collect, present and discuss the key arguments of the issue, selected strictly from this valuable Report. Our study tackles three aspects: (i Symptoms of pausing, including atmospheric changes, near-surface oceans, cryosphere and geographical differences. (ii Possible reasons of the paused warming, including external forcing factors, playing rather minor role, and the enhanced ocean heat uptake. Though missing warming is 0.2 K/decade compared to the model expectations, the whole climate system integrates continuously increasing amount of heat, 95 % of which is locked in the oceans. (iii Consequences of the pausing for the three main branches of the IPCC activity. For climate science, correct simulation of the enhanced heat uptake is a challenge. Since characteristic time scale of most adaptation measures is 1-2 decades, or shorter, near-term projections may not drive adaptation until climate models become able meet this challenge. On the other hand, pausing warming does not question the need for mitigation, since it is physically unlikely, that oceans can uptake endless amount of heat. Vertical temperature gradients of the upper ocean layers already show stagnation.
18. Geoethics: IPCC disgraced by violation of observational facts and physical laws in their sea level scenario
Science.gov (United States)
Mörner, Nils-Axel
2014-05-01
the order of 0.4o C. The improved ARGO measurements starting 2004 give virtually no change, however. The physically possible amount of expansion decreases, of course, with the decreasing water columns towards the coasts, and at the coasts it is zero (±0.0 mm). The redistribution of water masses in response to the Earth's rotation, surface current beat, wind stress, air pressure, etc. is an important factor. It gives local to regional changes, cancelled out on the global scale, however. From a geoethical point of view, it is of course quite blameworthy that IPCC excels in spreading these horror scenarios of a rapid, even accelerating, sea level rise. Besides, modern understanding of the planetary-solar-terrestrial interaction shows that we are now on our way into grand solar minimum with severely colder climate - that is just the opposite to IPCC's talk about an accelerating warming. In science we should debate - but we should not dictate (as IPCC insist upon), and it is here the perspectives of geoethics comes into the picture.
19. IPCC workshop on impacts of ocean acidification on marine biology and ecosystems. Workshop report
Energy Technology Data Exchange (ETDEWEB)
Field, C.B.; Barros, V.; Stocker, T.F.; Dahe, Q.; Mach, K.J.; Plattner, G.-K.; Mastrandrea, M.D.; Tignor, M.; Ebi, K.L.
2011-09-15
Understanding the effects of increasing atmospheric CO{sub 2} concentrations on ocean chemistry, commonly termed ocean acidification, as well as associated impacts on marine biology and ecosystems, is an important component of scientific knowledge about global change. The Fifth Assessment Report (AR5) of the Intergovernmental Panel on Climate Change (IPCC) will include comprehensive coverage of ocean acidification and its impacts, including potential feedbacks to the climate system. To support ongoing AR5 assessment efforts, Working Group II and Working Group I (WGII and WGI) of the IPCC held a joint Workshop on Impacts of Ocean Acidification on Marine Biology and Ecosystems in Okinawa, Japan, from 17 to 19 January 2011. The workshop convened experts from the scientific community, including WGII and WGI AR5 authors and review editors, to synthesise scientific understanding of changes in ocean chemistry due to increased CO{sub 2} and of impacts of this changing chemistry on marine organisms, ecosystems, and ecosystem services. This workshop report summarises the scientific content and perspectives presented and discussed during the workshop. It provides syntheses of these perspectives for the workshop's core topics: (i) the changing chemistry of the oceans, (ii) impacts of ocean acidification for individual organisms, and (iii) scaling up responses from individual organisms to ecosystems. It also presents summaries of workshop discussions of key cross-cutting themes, ranging from detection and attribution of ocean acidification and its impacts to understanding ocean acidification in the context of other stressors on marine systems. Additionally, the workshop report includes extended abstracts for keynote and poster presentations at the workshop. (Author)
20. Evaluation of the IPCC Models (AR4 and AR5) in the Precipitation Simulation in the Northeast of Brazil
Science.gov (United States)
Alves, José; Vasconcelos Junior, Francisco; Chaves, Rosane; Silva, Emerson; Servain, Jacques; Costa, Alexandre; Sombra, Sérgio; Barbosa, Augusto; Dos Santos, Antonio
2016-04-01
With the simulations of the models used in the latest reports from the Intergovernmental Panel on Climate Change (IPCC), comparative studies are necessary between observations and the so-called historical run (C20) and future projections of the AR4 (A2) and AR5 (RCP8.5) experiments, in order to assess whether the AR5 models had a better performance in the representation of physical processes. This article compares the sensitivity of IPCC models (AR4 and AR5) in representing the anuall average and seasonal rainfall variation (summer and autumn) in three regions of the Northeast of Brazil between 1979 and 2000, using the CMAP - CPC (Merged Analysis of Precipitation) data as reference. The projections made by these models for the period 2040-2070 were also analyzed.
1. Evaluation of the IPCC Models (AR4 and AR5 in the Precipitation Simulation in the Northeast of Brazil
Directory of Open Access Journals (Sweden)
José Brabo Alves
2016-04-01
Full Text Available With the simulations of the models used in the latest reports from the Intergovernmental Panel on Climate Change (IPCC, comparative studies are necessary between observations and the so-called historical run (C20 and future projections of the AR4 (A2 and AR5 (RCP8.5 experiments, in order to assess whether the AR5 models had a better performance in the representation of physical processes. This article compares the sensitivity of IPCC models (AR4 and AR5 in representing the anuall average and seasonal rainfall variation (summer and autumn in three regions of the Northeast of Brazil between 1979 and 2000, using the CMAP - CPC (Merged Analysis of Precipitation data as reference. The projections made by these models for the period 2040-2070 were also analyzed.
2. The fourth dimension
CERN Document Server
Rucker, Rudy
2014-01-01
""This is an invigorating book, a short but spirited slalom for the mind."" - Timothy Ferris, The New York Times Book Review ""Highly readable. One is reminded of the breadth and depth of Hofstadter's Gödel, Escher, Bach."" - Science""Anyone with even a minimal interest in mathematics and fantasy will find The Fourth Dimension informative and mind-dazzling... [Rucker] plunges into spaces above three with a zest and energy that is breathtaking."" - Martin Gardner ""Those who think the fourth dimension is nothing but time should be encouraged to read The Fourth Dimension, along with anyone else
3. Solar cycle length hypothesis appears to support the IPCC on global warming
DEFF Research Database (Denmark)
Laut, Peter; Gundermann, Jesper
1999-01-01
warming from the enhanced concentrations of greenhouse gases. The "solar hypothesis" claims that solar activity causes a significant component of the global mean temperature to vary in phase opposite to the filtered solar cycle lengths. In an earlier paper we have demonstrated that for data covering...... as the contributions from man-made greenhouse gases and sulphate aerosols by using an upwelling diffusion-energy balance model similar to the model of Wigley and Raper employed in the Second Assessment Report of The Intergovernmental Panel on Climate Change. It turns out that the agreement of the filtered solar cycle...... lengths with the "corrected" temperature anomalies is substantially better than with the historical anomalies. Therefore our findings support a total reversal of the common assumption that a verification of the solar hypothesis would challenge the IPCC assessment of man-made global warming....
4. A model of enteric fermentation in dairy cows to estimate methane emission for the Dutch National Inventory Report using the IPCC Tier 3 approach
NARCIS (Netherlands)
Bannink, A.; Schijndel, van M.W.; Dijkstra, J.
2011-01-01
The protocol for the National Inventory of agricultural greenhouse gas emissions in The Netherlands includes a dynamic and mechanistic model of animal digestion and fermentation as an Intergovernmental Panel on Climate Change (IPCC) Tier 3 approach to estimate enteric CH4 emission by dairy cows. The
5. The land-use projections and resulting emissions in the IPCC SRES scenarios as simulated by the IMAGE 2.2 model
NARCIS (Netherlands)
Strengers, B.; Leemans, R.; Eickhout, B.; Vries, de B.; Bouwman, L.
2004-01-01
The Intergovernmental Panel on Climate Change (IPCC) developed a new series of emission scenarios (SRES). Six global models were used to develop SRES but most focused primarily on energy and industry related emissions. Land-use emissions were only covered by three models, where IMAGE included the mo
6. 2007 status of climate changes: synthesis report. Summary for policy-makers; Bilan 2007 des changements climatiques: rapport de synthese. Resume a l'intention des decideurs
Energy Technology Data Exchange (ETDEWEB)
NONE
2007-07-01
This Synthesis Report is based on the assessment carried out by the three Working Groups of the Intergovernmental Panel on Climate Change (IPCC). It provides an integrated view of climate change as the final part of the IPCC's Fourth Assessment Report (AR4). Topic 1 summarises observed changes in climate and their effects on natural and human systems, regardless of their causes, while topic 2 assesses the causes of the observed changes. Topic 3 presents projections of future climate change and related impacts under different scenarios. Topic 4 discusses adaptation and mitigation options over the next few decades and their interactions with sustainable development. Topic 5 assesses the relationship between adaptation and mitigation on a more conceptual basis and takes a longer-term perspective. Topic 6 summarises the major robust findings and remaining key uncertainties in this assessment.
7. Regional Climate and Streamflow Projections in North America Under IPCC CMIP5 Scenarios
Science.gov (United States)
Chang, H. I.; Castro, C. L.; Troch, P. A. A.; Mukherjee, R.
2014-12-01
The Colorado River system is the predominant source of water supply for the Southwest U.S. and is already fully allocated, making the region's environmental and economic health particularly sensitive to annual and multi-year streamflow variability. Observed streamflow declines in the Colorado Basin in recent years are likely due to synergistic combination of anthropogenic global warming and natural climate variability, which are creating an overall warmer and more extreme climate. IPCC assessment reports have projected warmer and drier conditions in arid to semi-arid regions (e.g. Solomon et al. 2007). The NAM-related precipitation contributes to substantial Colorado streamflows. Recent climate change studies for the Southwest U.S. region project a dire future, with chronic drought, and substantially reduced Colorado River flows. These regional effects reflect the general observation that climate is being more extreme globally, with areas climatologically favored to be wet getting wetter and areas favored to be dry getting drier (Wang et al. 2012). Multi-scale downscaling modeling experiments are designed using recent IPCC AR5 global climate projections, which incorporate regional climate and hydrologic modeling components. The Weather Research and Forecasting model (WRF) has been selected as the main regional modeling tool; the Variable Infiltration Capacity model (VIC) will be used to generate streamflow projections for the Colorado River Basin. The WRF domain is set up to follow the CORDEX-North America guideline with 25km grid spacing, and VIC model is individually calibrated for upper and lower Colorado River basins in 1/8° resolution. The multi-scale climate and hydrology study aims to characterize how the combination of climate change and natural climate variability is changing cool and warm season precipitation. Further, to preserve the downscaled RCM sensitivity and maintain a reasonable climatology mean based on observed record, a new bias correction
8. Floods in Mekong Delta Under Sea-Level Rise Projections By IPCC AR5
Science.gov (United States)
Takagi, H.
2014-12-01
One of the mightiest rivers in the planet, the Mekong ranks 10th amongst the world's great rivers on the basis of mean annual flow at the mouth. It flows southwards over a distance of approximately 4,800 km from its source to the sea, through six different countries: China, Myanmar, Lao PDR, Thailand, Cambodia and Vietnam. This great basin has been considered to be one of the most sensitive areas in the world to climate change. The present paper investigates fluvial flood hazards in urban areas in the Mekong Delta to inundation due to seasonal flooding, a phenomenon which is likely to be exacerbated by future sea-level rise. Unlike past researches which mainly focus on flooding due to river discharge from upstream or heavy precipitation, the present paper scrutinizes the influence of ocean tides. The research reveals that ocean tides predominantly determine water elevation even in an upstream location such as Can Tho City, 80 km inland from the river mouth, and that the river flow causes tidal damping and effectively reduces the energy of the incoming tides. This tidal damping is especially pronounced during the rainy season. Analysis based on the water levels monitored by the Mekong River Commission reveals that the ground near the riverbank of Can Tho had experienced inundation for a total of 215 hours between July 2009 and June 2010 (2.5% of the time over a one year period). It is also shown that inundation reached up to a maximum height of 47 cm above the roads of Can Tho downtown in this one-year period. Assuming two scenarios of sea-level rise of 25 cm in the middle of the 21st century and 60 cm in the end of the century, all based on the Fifth Assessment Report of Intergovernmental Panel on Climate Change (IPCC AR5) projections, it was found that the duration of inundation will be prolonged from the present percentage of 2.5% to 7.5% and 24% of the year, respectively. It is important to note that while at present this flooding is seasonal and limited, in the
9. River discharge and flood inundation over the Amazon based on IPCC AR5 scenarios
Science.gov (United States)
Paiva, Rodrigo; Sorribas, Mino; Jones, Charles; Carvalho, Leila; Melack, John; Bravo, Juan Martin; Beighley, Edward
2015-04-01
Climate change and related effects over the hydrologic regime of the Amazon River basin could have major impacts over human and ecological communities, including issues with transportation, flood vulnerability, fisheries and hydropower generation. We examined future changes in discharge and floodplain inundation within the Amazon River basin. We used the hydrological model MGB-IPH (Modelo de Grandes Bacias - Instituto de Pesquisas Hidráulicas) coupled with a 1D river hydrodynamic model simulating water storage over the floodplains. The model was forced using satellite based precipitation from the TRMM 3B42 dataset, and it had a good performance when validated against discharge and stage measurements as well as remotely sensed data, including radar altimetry-based water levels, gravity anomaly-based terrestrial water storage and flood inundation extent. Future scenarios of precipitation and other relevant climatic variables for the 2070 to 2100 time period were taken from five coupled atmosphere-ocean general circulation models (AOGCMs) from IPCC's Fifth Assessment Report (AR5) Coupled Model Intercomparison Project Phase 5 (CMIP5). The climate models were chosen based on their ability to represent the main aspects of recent (1970 to 2000) Amazon climate. A quantile-quantile bias removal procedure was applied to climate model precipitation to mitigate unreliable predictions. The hydrologic model was then forced using past observed climate data altered by delta change factors based on the past and future climate models aiming to estimate projected discharge and floodplain inundation in climate change scenario at several control points in the basin. The climate projections present large uncertainty, especially the precipitation rate, and predictions using different climate models do not agree on the sign of changes on total Amazon flood extent or discharge along the main stem of the Amazon River. However, analyses of results at different regions indicate an increase
10. Fourth-rank cosmology
International Nuclear Information System (INIS)
Some cosmological implications of the recently proposed fourth-rank theory of gravitation are studied. The model exhibits the possibility of being free from the horizon and flatness problems at the price of introducing a negative pressure. The field equations we obtain are compatible with kobs=0 and Ωobstclas approx. 1020tPlanck approx. 10-23s. When interpreted at the light of General Relativity the treatment is shown to be almost equivalent to that of the standard model of cosmology combined with the inflationary scenario. Hence, an interpretation of the negative pressure hypothesis is provided. (author). 8 refs
11. Dynamic EROI Assessment of the IPCC 21st Century Electricity Production Scenario
Directory of Open Access Journals (Sweden)
Charles Neumeyer
2016-04-01
Full Text Available The Energy Return on Investment (EROI is an important measure of the energy gain of an electrical power generating facility that is typically evaluated based on the life cycle energy balance of a single facility. The EROI concept can be extended to cover a collection of facilities that comprise a complete power system and used to assess the expansion and evolution of a power system as it transitions from one portfolio mix of technologies to another over time. In this study we develop a dynamic EROI model that simulates the evolution of a power system and we perform an EROI simulation of one of the electricity production scenarios developed under the auspices of the Intergovernmental Panel on Climate Change (IPCC covering the global supply of electricity in the 21st century. Our analytic tool provides the means for evaluation of dynamic EROI based on arbitrary time-dependent demand scenarios by modeling the required expansion of power generation, including the plowback needed for new construction and to replace facilities as they are retired. The results provide insight into the level of installed and delivered power, above and beyond basic consumer demand, that is required to support construction during expansion, as well as the supplementary power that may be required if plowback constraints are imposed. In addition, sensitivity to EROI parameters, and the impact of energy storage efficiency are addressed.
12. A Coupled Model Study on the Intensification of the Asian Summer Monsoon in IPCC SRES Scenarios
Institute of Scientific and Technical Information of China (English)
2005-01-01
The Asian summer monsoon is an important part of the climate system. Investigating the response of the Asian summer monsoon to changing concentrations of greenhouse gases and aerosols will be meaningful to understand and predict climate variability and climate change not only in Asia but also globally. In order to diagnose the impacts of future anthropogenic emissions on monsoon climates, a coupled general circulation model of the atmosphere and the ocean has been used at the Max-Planck-Institute for Meteorology. In addition to carbon dioxide, the major well mixed greenhouse gases such as methane, nitrous oxide, several chlorofluorocarbons, and CFC substitute gases are prescribed as a function of time. The sulfur cycle is simulated interactively, and both the direct aerosol effect and the indirect cloud albedo effect are considered.Furthermore, changes in tropospheric ozone have been pre-calculated with a chemical transport model and prescribed as a function of time and space in the climate simulations. Concentrations of greenhouse gases and anthropogenic emissions of sulfur dioxide are prescribed according to observations (1860-1990) and projected into the future (1990-2100) according to the Scenarios A2 and B2 in Special Report on Emissions Scenarios (SRES, Nakicenovic et al., 2000) developed by the Intergovernmental Panel on Climate Change (IPCC). It is found that the Indian summer monsoon is enhanced in the scenarios in terms of both mean precipitation and interannual variability. An increase in precipitation is simulated for northern China but a decrease for the southern part. Furthermore, the simulated future increase in monsoon variability seems to be linked to enhanced ENSO variability towards the end of the scenario integrations.
13. A Canonical Response in Rainfall Characteristics to Global Warming: Projections by IPCC CMIP5 Models
Science.gov (United States)
Lau, William K. M.; Wu, H. T.; Kim, K. M.
2012-01-01
Changes in rainfall characteristics induced by global warming are examined based on probability distribution function (PDF) analysis, from outputs of 14 IPCC (Intergovernmental Panel on Climate Change), CMIP (5th Coupled Model Intercomparison Project) models under various scenarios of increased CO2 emissions. Results show that collectively CMIP5 models project a robust and consistent global and regional rainfall response to CO2 warming. Globally, the models show a 1-3% increase in rainfall per degree rise in temperature, with a canonical response featuring large increase (100-250 %) in frequency of occurrence of very heavy rain, a reduction (5-10%) of moderate rain, and an increase (10-15%) of light rain events. Regionally, even though details vary among models, a majority of the models (>10 out of 14) project a consistent large scale response with more heavy rain events in climatologically wet regions, most pronounced in the Pacific ITCZ and the Asian monsoon. Moderate rain events are found to decrease over extensive regions of the subtropical and extratropical oceans, but increases over the extratropical land regions, and the Southern Oceans. The spatial distribution of light rain resembles that of moderate rain, but mostly with opposite polarity. The majority of the models also show increase in the number of dry events (absence or only trace amount of rain) over subtropical and tropical land regions in both hemispheres. These results suggest that rainfall characteristics are changing and that increased extreme rainfall events and droughts occurrences are connected, as a consequent of a global adjustment of the large scale circulation to global warming.
14. Searching for the fourth family quarks through anomalous decays
International Nuclear Information System (INIS)
The flavor democracy hypothesis predicts the existence of the fourth standard model family. Because of the high masses of the fourth family quarks, their anomalous decays could be dominant if certain criteria are met. This will drastically change the search strategy at hadron colliders. We show that the fourth standard model family down quarks with masses up to 400-450 GeV can be observed (or excluded) via anomalous decays by Tevatron.
15. Sustainability assessment of nuclear power: Discourse analysis of IAEA and IPCC frameworks
International Nuclear Information System (INIS)
Highlights: • Sustainability assessments (SAs) are methodologically precarious. • Discourse analysis reveals how the meaning of sustainability is constructed in SAs. • Discourse analysis is applied on the SAs of nuclear power of IAEA and IPCC. • For IAEA ‘sustainable’ equals ‘complying with best international practices’. • The IAEA framework largely inspires IPCC Fifth Assessment Report. - Abstract: Sustainability assessments (SAs) are methodologically precarious. Value-based judgments inevitably play a role in setting the scope of the SA, selecting assessment criteria and indicators, collecting adequate data, and developing and using models of considered systems. Discourse analysis can reveal how the meaning and operationalization of sustainability is constructed in and through SAs. Our discourse-analytical approach investigates how sustainability is channeled from ‘manifest image’ (broad but shallow), to ‘vision’, to ‘policy targets’ (specific and practical). This approach is applied on the SA frameworks used by IAEA and IPCC to assess the sustainability of the nuclear power option. The essentially problematic conclusion is that both SA frameworks are constructed in order to obtain answers that do not conflict with prior commitments adopted by the two institutes. For IAEA ‘sustainable’ equals ‘complying with best international practices and standards’. IPCC wrestles with its mission as a provider of “policy-relevant and yet policy-neutral, never policy-prescriptive” knowledge to decision-makers. IPCC avoids the assessment of different visions on the role of nuclear power in a low-carbon energy future, and skips most literature critical of nuclear power. The IAEA framework largely inspires IPCC AR5
16. Drawing Boundaries: Boundary Arrangements of the IPCC Working Groups
OpenAIRE
van Eck, Christel
2016-01-01
The present research investigates how the IPCC’s Working Groups safeguard their scientific character while communicating with policymakers. Due to the different nature of Working Groups’ assessments, all Working Groups make different boundary arrangements of how science is defined; what is considered as relevant knowledge; and what the division of labor is amongst Working Groups. The results show that science is a context-specific activity in a constantly changing landscape, which in turn aff...
17. Time Series of Aerosol Column Optical Depth at the Barrow, Alaska, ARM Climate Research Facility for 2008 Fourth Quarter 2009 ARM and Climate Change Prediction Program Metric Report
Energy Technology Data Exchange (ETDEWEB)
C Flynn; AS Koontz; JH Mather
2009-09-01
The uncertainties in current estimates of anthropogenic radiative forcing are dominated by the effects of aerosols, both in relation to the direct absorption and scattering of radiation by aerosols and also with respect to aerosol-related changes in cloud formation, longevity, and microphysics (See Figure 1; Intergovernmental Panel on Climate Change, Assessment Report 4, 2008). Moreover, the Arctic region in particular is especially sensitive to changes in climate with the magnitude of temperature changes (both observed and predicted) being several times larger than global averages (Kaufman et al. 2009). Recent studies confirm that aerosol-cloud interactions in the arctic generate climatologically significant radiative effects equivalent in magnitude to that of green house gases (Lubin and Vogelmann 2006, 2007). The aerosol optical depth is the most immediate representation of the aerosol direct effect and is also important for consideration of aerosol-cloud interactions, and thus this quantity is essential for studies of aerosol radiative forcing.
18. Future coal production outlooks in the IPCC Emission Scenarios: Are they plausible?
Energy Technology Data Exchange (ETDEWEB)
Hoeoek, Mikael
2010-10-15
Anthropogenic climate change caused by CO{sub 2} emissions is strongly and fundamentally linked to the future energy production. The Special Report on Emission Scenarios (SRES) from 2000 contains 40 scenarios for future fossil fuel production and is used by the IPCC to assess future climate change. Coal, with its 26% share of world energy, is a major source of greenhouse gas emissions and commonly seen as a key contributor to anthropogenic climate change. SRES contains a wide array of different coal production outlooks, ranging from a complete coal phase-out by 2100 to a roughly tenfold increase from present world production levels. Scenarios with high levels of global warming also have high expectations on future fossil fuel production. The assumptions on resource availability are in SRES based on Rogner's assessment of world hydrocarbon resources from 1997, where it is stated that 'the sheer size of the fossil resource base makes fossil sources an energy supply option for many centuries to come'. Regarding the future coal production it is simply assumed to be dependent on economics, accessibility, and environmental acceptance. It is also generally assumed that coal is abundant, and will thus take a dominating part in the future energy system. Depletion, geographical location and geological parameters are not given much influence in the scenario storylines. This study quantifies what the coal production projection in SRES would imply in reality. SRES is riddled with future production projections that would put unreasonable expectation on just a few countries or regions. Is it reasonable to expect that China, among the world's largest coal reserve and resource holder and producer, would increase their production by a factor of 8 over the next 90 years, as implied by certain scenarios? Can massive increases in global coal output really be justified from historical trends or will reality rule out some production outlooks as implausible? The
19. Fourth Light at Paranal!
Science.gov (United States)
2000-09-01
VLT YEPUN Joins ANTU, KUEYEN and MELIPAL It was a historical moment last night (September 3 - 4, 2000) in the VLT Control Room at the Paranal Observatory , after nearly 15 years of hard work. Finally, four teams of astronomers and engineers were sitting at the terminals - and each team with access to an 8.2-m telescope! From now on, the powerful "Paranal Quartet" will be observing night after night, with a combined mirror surface of more than 210 m 2. And beginning next year, some of them will be linked to form part of the unique VLT Interferometer with unparalleled sensitivity and image sharpness. YEPUN "First Light" Early in the evening, the fourth 8.2-m Unit Telescope, YEPUN , was pointed to the sky for the first time and successfully achieved "First Light". Following a few technical exposures, a series of "first light" photos was made of several astronomical objects with the VLT Test Camera. This instrument was also used for the three previous "First Light" events for ANTU ( May 1998 ), KUEYEN ( March 1999 ) and MELIPAL ( January 2000 ). These images served to evaluate provisionally the performance of the new telescope, mainly in terms of mechanical and optical quality. The ESO staff were very pleased with the results and pronounced YEPUN fit for the subsequent commissioning phase. When the name YEPUN was first given to the fourth VLT Unit Telescope, it was supposed to mean "Sirius" in the Mapuche language. However, doubts have since arisen about this translation and a detailed investigation now indicates that the correct meaning is "Venus" (as the Evening Star). For a detailed explanation, please consult the essay On the Meaning of "YEPUN" , now available at the ESO website. The first images At 21:39 hrs local time (01:39 UT), YEPUN was turned to point in the direction of a dense Milky Way field, near the border between the constellations Sagitta (The Arrow) and Aquila (The Eagle). A guide star was acquired and the active optics system quickly optimized the
20. Investigation of the climate change within Moscow metropolitan area
Science.gov (United States)
Varentsov, Mikhail; Trusilova, Kristina; Konstantinov, Pavel; Samsonov, Timofey
2014-05-01
As the urbanization continues worldwide more than half of the Earth's population live in the cities (U.N., 2010). Therefore the vulnerability of the urban environment - the living space for millions of people - to the climate change has to be investigated. It is well known that urban features strongly influence the atmospheric boundary layer and determine the microclimatic features of the local environment, such as urban heat island (UHI). Available temperature observations in cities are, however, influenced by the natural climate variations, human-induced climate warming (IPCC, 2007) and in the same time by the growth and structural modification of the urban areas. The relationship between these three factors and their roles in climate changes in the cities are very important for the climatic forecast and requires better understanding. In this study, we made analysis of the air temperature change and urban heat island evolution within Moscow urban area during decades 1970-2010, while this urban area had undergone intensive growth and building modification allowing the population of Moscow to increase from 7 to 12 million people. Analysis was based on the data from several meteorological stations in Moscow region and Moscow city, including meteorological observatory of Lomonosov Moscow State University. Differences in climate change between urban and rural stations, changes of the power and shape of urban heat island and their relationships with changes of building height and density were investigated. Collected data and obtained results are currently to be used for the validation of the regional climate model COSMO-CLM with the purpose to use this model for further more detailed climate research and forecasts for Moscow metropolitan area. References: 1. U.N. (2010), World Urbanization Prospects. The 2009 Revision.Rep., 1-47 pp, United Nations. Department of Economic and Social Affairs. Population Division., New York. 2. IPCC (2007), IPCC Fourth Assessment Report
1. On a fourth order superlinear elliptic problem
Directory of Open Access Journals (Sweden)
M. Ramos
2001-01-01
Full Text Available We prove the existence of a nonzero solution for the fourth order elliptic equation $$Delta^2u= mu u +a(xg(u$$ with boundary conditions $u=Delta u=0$. Here, $mu$ is a real parameter, $g$ is superlinear both at zero and infinity and $a(x$ changes sign in $Omega$. The proof uses a variational argument based on the argument by Bahri-Lions cite{BL}.
2. The climate crisis: An introductory guide to climate change
Science.gov (United States)
Trenberth, Kevin E.
2011-06-01
Human-induced climate change, sometimes called “global warming,” has, unfortunately, become a “hot” topic, embroiled in controversy, misinformation, and claims and counterclaims. It should not be this way, because there are many scientific facts that provide solid information on which to base policy. There is a very strong observational, theoretical, and modeling base in physical science that underpins current understanding of what has happened to Earth's climate and why and what the prospects are for the future under certain assumptions. Moreover, these changes have impacts, which are apt to grow, on the environment and human society. To avoid or reduce these impacts and the economic and human effects of undesirable future climate change requires actions that are strongly opposed by those with vested interests in the status quo, some of whom have funded misinformation campaigns that have successfully confused the public and some politicians, leading to paralysis in political action. Without mitigation of climate change, one would suppose that at least society would plan sensibly for the changes already happening and projected, but such future adaptation plans are also largely in limbo. The implication is that we will suffer the consequences. All of these aspects are addressed in this informative and attractive book, which is written for a fairly general but technically informed audience. The book is strongly based upon the 2007 Fourth Assessment Report (AR4) of the Intergovernmental Panel on Climate Change (IPCC) and therefore has a solid scientific basis. Many figures, graphs, and maps come from the three IPCC working group reports, although the captions often do not explain the detail shown. Given that the IPCC reports totaled nearly 3000 pages, to distill the complex material down to 249 pages is no mean task, and the authors have succeeded quite well.
3. “管理极端气候事件和灾害风险特别报告”对我国的启示%The Implications on China' s Disaster Prevention and Mitigation from IPCC Special Report on "Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation"
Institute of Scientific and Technical Information of China (English)
刘冰; 薛澜
2012-01-01
政府间气候变化专门委员会(IPCC)最新发布的"管理极端气候事件和灾害风险促进气候变化适应特别报告决策者摘要(SREX)"从"极端气候事件+脆弱性+暴露程度"的角度剖析了灾害风险的根源,综合考虑了气候、环境、社会经济条件等因素,提出了管理灾害风险和适应气候变化的各种政策选项,对于我国把风险管理纳入应对气候变化行动的整体框架提供了重要的科学依据。本文基于特别报告的主要结论,结合我国防灾减灾工作的实际情况,提出加快社会经济发展、实现社会系统重构、发挥政策协同效应是今后防灾减灾工作的重要着力点。%The newly issued Intergovernmental Panel on Climate Change (IPCC) Special Report: Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation (SREX) analyzes the root of disaster risks from three perspectives of "extreme climate events, vulnerability and exposure". It considers factors such as climatic, environmental and socio - economic conditions, and puts forward various policy options for disaster risk management and climate change adaptation. This report provides scientific evidence for integrating risk management into the action framework of addressing climate change in China. This article provides a synthesis of the main conclusions of the report. Based on the actual situation of China' s disaster prevention and reduction efforts, the article makes several policy recommendations, including 1 ) speeding up socio - economic development, 2) promoting social and institutional innovation and transformation, and 3 ) developing policy synergy across different policy arenas related to disaster prevention and reduction.
4. Evaluation of the effect of accounting method, IPCC v. LCA, on grass-based and confinement dairy systems' greenhouse gas emissions.
Science.gov (United States)
O'Brien, D; Shalloo, L; Patton, J; Buckley, F; Grainger, C; Wallace, M
2012-09-01
Life cycle assessment (LCA) and the Intergovernmental Panel on Climate Change (IPCC) guideline methodology, which are the principal greenhouse gas (GHG) quantification methods, were evaluated in this study using a dairy farm GHG model. The model was applied to estimate GHG emissions from two contrasting dairy systems: a seasonal calving pasture-based dairy farm and a total confinement dairy system. Data used to quantify emissions from these systems originated from a research study carried out over a 1-year period in Ireland. The genetic merit of cows modelled was similar for both systems. Total mixed ration was fed in the Confinement system, whereas grazed grass was mainly fed in the grass-based system. GHG emissions from these systems were quantified per unit of product and area. The results of both methods showed that the dairy system that emitted the lowest GHG emissions per unit area did not necessarily emit the lowest GHG emissions possible for a given level of product. Consequently, a recommendation from this study is that GHG emissions be evaluated per unit of product given the growing affluent human population and increasing demand for dairy products. The IPCC and LCA methods ranked dairy systems' GHG emissions differently. For instance, the IPCC method quantified that the Confinement system reduced GHG emissions per unit of product by 8% compared with the grass-based system, but the LCA approach calculated that the Confinement system increased emissions by 16% when off-farm emissions associated with primary dairy production were included. Thus, GHG emissions should be quantified using approaches that quantify the total GHG emissions associated with the production system, so as to determine whether the dairy system was causing emissions displacement. The IPCC and LCA methods were also used in this study to simulate, through a dairy farm GHG model, what effect management changes within both production systems have on GHG emissions. The findings suggest that
5. The handshake between the integrated assessment and climate modeling communities: IPCC AR5 as a catalyst for improved networking, collaboration and communication
Science.gov (United States)
Hibbard, K. A.; Meehl, G.; Edmonds, J.; Nakićenović, N.; Lamarque, J.; Rose, S.; van Vuuren, D.; Moss, R.; Hurtt, G. C.
2009-12-01
In the wake of AR4, it was clear that there needed to be a much more proactive modeling strategy for the climate modeling community. In 2006 a series of conversations were initiated within the global environmental change communities about how to approach a modeling strategy for the next IPCC assessment (AR5). At that time, the idea of developing both a near- and long- term experimental design that addressed questions around carbon cycle diagnostics, Earth system predictability as well as inertia of the climate system were introduced. From these discussions, a joint meeting was held in Aspen, Colorado (Aspen Global Change Institute) that included both the climate and integrated assessment modeling communities as well as representatives from the impacts, adaptation and vulnerability (IAV) research groups. Many activities have been spawned from this initial set of conversations, including the development, for the first time ever, Representative Concentration Pathways (RCPs) which are a set of four radiative forcing pathways, of which three include, for the first time ever in IPCC history, mitigation for stabilization of greenhouse gasses and radiative forcings (mitigation scenarios include two stabilization and one overshoot and decline), a land use/land cover and emissions harmonization between integrated assessment and climate modeling communities as well as the near-term prediction experiment designed to address model skill in predictability, downscaling, extreme events, air quality, etc.. The consensus development of the RCPs provided an initial motivation for closer collaboration between the ESM, IAM and IAV communities towards the development of New Scenarios that were driven by the scientific communities and not the IPCC. These collaborations are leading to greater mutual understanding of modeling paradigms and strengthening ties with integrated assessment, climate and emissions inventory and modeling communities.
6. Projecting the Global Distribution of the Emerging Amphibian Fungal Pathogen, Batrachochytrium dendrobatidis, Based on IPCC Climate Futures
Science.gov (United States)
Olson, Deanna H.; Blaustein, Andrew R.
2016-01-01
Projected changes in climate conditions are emerging as significant risk factors to numerous species, affecting habitat conditions and community interactions. Projections suggest species range shifts in response to climate change modifying environmental suitability and is supported by observational evidence. Both pathogens and their hosts can shift ranges with climate change. We consider how climate change may influence the distribution of the emerging infectious amphibian chytrid fungus, Batrachochytrium dendrobatidis (Bd), a pathogen associated with worldwide amphibian population losses. Using an expanded global Bd database and a novel modeling approach, we examined a broad set of climate metrics to model the Bd-climate niche globally and regionally, then project how climate change may influence Bd distributions. Previous research showed that Bd distribution is dependent on climatic variables, in particular temperature. We trained a machine-learning model (random forest) with the most comprehensive global compilation of Bd sampling records (~5,000 site-level records, mid-2014 summary), including 13 climatic variables. We projected future Bd environmental suitability under IPCC scenarios. The learning model was trained with combined worldwide data (non-region specific) and also separately per region (region-specific). One goal of our study was to estimate of how Bd spatial risks may change under climate change based on the best available data. Our models supported differences in Bd-climate relationships among geographic regions. We projected that Bd ranges will shift into higher latitudes and altitudes due to increased environmental suitability in those regions under predicted climate change. Specifically, our model showed a broad expansion of areas environmentally suitable for establishment of Bd on amphibian hosts in the temperate zones of the Northern Hemisphere. Our projections are useful for the development of monitoring designs in these areas, especially for
7. Calculation of Co2 emissions from the italian energy system; Calcolo delle emissioni di CO2 dal settore energetico italiano. 1990-2000. Metodo di riferimento IPCC
Energy Technology Data Exchange (ETDEWEB)
Contaldi, M. [Agenzia Nazionale per la Protezione dell' Ambiente, Rome (Italy); La Motta, S. [ENEA, Funzione Centrale Studi, Centro Ricerche Casaccia, Rome (Italy)
2001-07-01
The calculation of Co2 emissions from the Italian energy system is the object of this work. The inventory method used is the Reference Approach from the Intergovernmental Panel for Climate Change (IPCC, 1996 revised Guidelines for National Greenhouse Gas Inventories) and the energy consumption data are taken from the Italian Energy Balance edited by the Ministry of Industry. The years analysed are those from 1990 to 2000. [Italian] Lo scopo di questo lavoro e' quello di contabilizzare le emissioni di CO2 provenienti dal settore energetico per fonte di utilizzo dell'energia, a partire direttamente dal Bilancio Energetico Nazionale (Bilancio Energetico Nazionale, BEN, a cura del Ministero Industria, Commercio ed Artigianato) ed applicando all'Italia la metodologia di riferimento per il calcolo delle emissioni della CO2 elaborata dall'Intergovernmental Panel for Climate Change (IPCC, 1996 revised Guidelines for National Greenhouse Gas Inventories). Gli anni presi in considerazione in queto lavoro sono quelli dal 1990 al 2000.
8. Dominant frames in legacy and social media coverage of the IPCC Fifth Assessment Report
Science.gov (United States)
O'Neill, Saffron; Williams, Hywel T. P.; Kurz, Tim; Wiersma, Bouke; Boykoff, Maxwell
2015-04-01
The media are powerful agents that translate information across the science-policy interface, framing it for audiences. Yet frames are never neutral: they define an issue, identify causes, make moral judgements and shape proposed solutions. Here, we show how the IPCC Fifth Assessment Report (AR5) was framed in UK and US broadcast and print coverage, and on Twitter. Coverage of IPCC Working Group I (WGI) was contested and politicized, employing the Settled Science, Uncertain Science, Political or Ideological Struggle and Role of Science’ frames. WGII coverage commonly used Disaster or Security. More diverse frames were employed for WGII and WGIII, including Economics and Morality and Ethics. Framing also varied by media institution: for example, the BBC used Uncertain Science, whereas Channel 4 did not. Coverage varied by working group, with WGIII gaining far less coverage than WGI or WGII. We suggest that media coverage and framing of AR5 was influenced by its sequential three-part structure and by the availability of accessible narratives and visuals. We recommend that these communication lessons be applied to future climate science reports.
9. Problems with the North American Monsoon in CMIP/IPCC GCM Precipitation
Science.gov (United States)
Schiffer, N. J.; Nesbitt, S. W.
2011-12-01
Successful water management in the Desert Southwest and surrounding areas hinges on anticipating the timing and distribution of precipitation. IPCC AR4 models predict a more arid climate, more extreme precipitation events, and an earlier peak in springtime streamflow in the North American Monsoon region as the area warms. This study aims to assess the summertime skill with which general circulation models (GCMs) simulate precipitation and related dynamics over this region, a necessary precursor to reliable hydroclimate projections. Thirty-year climatologies of several GCMs in the third and fifth Climate Model Intercomparison Projects (CMIP) are statistically evaluated against each other and observed climatology for their skill in representing the location, timing, variability, character, and large-scale forcing of precipitation over the southwestern United States and northwestern Mexico. The results of this study will lend greater credence to more detailed, higher resolution studies, based on the CMIP and IPCC models, of the region's future hydrology. Our ultimate goal is to provide guidance such that decision-makers can plan future water management with more confidence.
10. THE FOURTH STATE OF WATER
Directory of Open Access Journals (Sweden)
E.V. Savich
2013-04-01
Full Text Available The fourth state of water is aqueous vapor which, under power action of magnetic-field strength at Curie point at the moment of magnetic phase transition, acquires properties of ferroelectric plasma. The latter is the basis of origination of such natural phenomena as a thunderstorm cloud and a ball lightning capable of releasing, under certain conditions, lightning strokes and, during explosion, plenty of thermal energy, which is typical of a low-temperature reactor.
11. The Fourth Way in Finland
Directory of Open Access Journals (Sweden)
Vesa Iitti
2008-01-01
Full Text Available This article focuses on the general history of the Fourth Way in Finland. The Fourth Way, or simply ‘the Work’, began as a Greco-Armenian man named Georges Ivanovich Gurdjieff (1866?–1949 gathered groups of pupils in St Petersburg and Moscow in 1912. To these groups, Gurdjieff started to teach what he had learned and synthesized between ca 1896 and 1912 during his travels on spiritual search of Egypt, Crete, Sumeria, Assyria, the Holy Land, Mecca, Ethiopia, Sudan, India, Afghanistan, the northern valleys of Siberia, and Tibet. Neither Gurdjieff nor any of his disciples called themselves a church, a sect, or anything alike, but referred to themselves simply as ‘the Work’, or as ‘the Fourth Way’. The name ‘the Fourth Way’ originates in a Gurdjieffian view that there are essentially three traditional ways of spiritual work: those of a monk, a fakir, and a yogi. These ways do not literally refer to the activities of a monk, a fakir, and a yogi, but to similar types of spiritual work emphasizing exercise of emotion, body, or mind. Gurdjieff’s teaching is a blend of various influences that include Sufism, orthodox Christianity, Buddhism, Kabbalah, and general elements of various occult teachings of both the East and the West. Gurdjieff’s teaching is a blend of various influences that include Sufism, orthodox Christianity, Buddhism, Kabbalah, and general elements of various occult teachings of both the East and the West. It is a unique combination of cosmology, psychology, theory of evolution, and overall theory and practise aiming to help individuals in their efforts towards what is called ‘self-remembering’.
12. The fourth generation in supergravity
Science.gov (United States)
Enqvist, K.; Nanopoulos, D. V.; Zwirner, F.
1985-12-01
We consider model-independent constraints on the fourth-generation fermion masses and the magnitude of the D-term contribution to the scalar masses. We find that the ratio of vacuum expectation values is limited to the range 1/5 ~ 150 GeV. A general feature of the four-generation models is thus a heavy spectrum of sparticles. On leave from International School for Advanced Studies, Trieste, and Istituto Nazionale di Fisica Nucleare, Padua, Italy.
13. The Fourth Microlensing Planet Revisited
CERN Document Server
Yock, Philip
2015-01-01
The fourth microlensing planet, otherwise known as OGLE-2005-BLG-169Lb, was discovered by a collaboration of US, NZ, Polish and UK astronomers in 2005-2006. Recently the results were confirmed by the Hubble Space Telescope and by the Keck Observatory. OGLE-2005-BLG-169Lb is the first microlensing planet to receive such confirmation. Its discovery and confirmation are described here in an historical context.
14. The Copyright Book: A Practical Guide. Fourth Edition.
Science.gov (United States)
Strong, William S.
In response to important changes in copyright law as the United States accommodates itself to the Berne Convention and develops means to take account of new technologies, this guide puts these changes in a form and context that will make sense to persons who are concerned about their rights under the law. New material in the fourth edition of this…
15. Climate change in China and China’s policies and actions for addressing climate change
Directory of Open Access Journals (Sweden)
Luo Y.
2010-12-01
Full Text Available Since the first assessment report (FAR of Inter-Governmental Panel on Climate Change (IPCC in 1990, the international scientific community has made substantial progresses in climate change sciences. Changes in components of climate system, including the atmosphere, oceans and cryosphere, indicate that global warming is unequivocal. Instrumental records demonstrate that the global mean temperature has a significant increasing trend during the 20th century and in the latest 50 years the warming become faster. In the meantime, the global sea level has a strong increasing trend, as well as the snow coverage of Northern Hemisphere showed an obvious downward trend. Moreover, the global warming plays a key role in significantly affecting the climate system and social-economy on both global and regional scales, such as sea level rise, melting of mountain glaciers and ice sheets, desertification, deforestation, increase of weather extremes (typhoon, hurricane and rainstorm and so on. The state of the art understanding of IPCC Fourth Assessment Report (AR4 was most of the observed increase in global average temperatures since the mid-20th century is very likely due to the observed increase in the concentrations of anthropogenic greenhouse gases. Climate change issues, as a grave challenge to the sustainable development of the human society, have received ever greater attention from the international community. Deeply cognizant of the complexity and extensive influence of these issues and fully aware of the arduousness and urgency of the task of addressing climate change, the Chinese government is determined to address climate change in the process of pursuing sustainable development. The facts of climate change in China and its impacts, and China’s policies and actions for addressing climate change are introduced in this paper.
16. Search for the fourth standard model family
International Nuclear Information System (INIS)
Existence of the fourth family follows from the basics of the standard model (SM) and the actual mass spectrum of the third family fermions. We discuss possible manifestations of the fourth SM family at existing and future colliders. The LHC and Tevatron potentials to discover the fourth SM family have been compared. The scenario with dominance of the anomalous decay modes of the fourth-family quarks has been considered in detail.
17. The Super-Fast Chemistry Mechanism for IPCC AR5 Simulations with CCSM
Science.gov (United States)
Cameron-Smith, P. J.; Prather, M. J.; Lamarque, J.; Hess, P. G.; Connell, P. S.; Bergmann, D. J.; Vitt, F. M.
2009-12-01
When atmospheric chemistry is included in long climate and Earth system simulations, the completeness of the chemical mechanism must be balanced with the computational cost, and the scientific interests of the atmospheric chemist must be tempered by the chemical needs of other components of the climate model (e.g, greenhouse gas concentrations for radiative heating, and deposition rates for biosphere interactions). We have implemented a super-fast chemical mechanism for use in IPCC AR5 simulations with the CCSM climate model. The mechanism uses just 17 species to calculate ozone, OH, and sulfate, with the OH providing the means to back-out the lifetime of methane and other species of interest. The recent addition of isoprene offers significant improvements to the sensitivity of the super-fast mechanism to emission perturbations when compared to our state-of-the-art full chemistry mechanism that uses 90 species. An analysis of the impact of different tropopause definitions will also be presented.
18. Geoethics disgraced by Copernicus in its desperate act of covering up for the IPCC
Science.gov (United States)
Mörner, Nils-Axel
2014-05-01
All my life I have put observational facts in the centre. This is what Leonardo da Vinci called "to read the book of mother Earth". For a geologist this is very natural because this is the core of the geological profession. The IPCC, on the other hand, put all credits on models and scenarios. One of its founders and its first president, Bert Bolin, at a debate in the Geophysical Society of Sweden in 2001, admitted (quotation): "you can order whatever scenario you wish". With this as a background, we can put a recent action by Copernicus at a test with respect to geoethics and normal decency. The idea that the planetary motions affect and control the solar variability is old, but in the stage of an unproven hypothesis. In recent years major advancements have occurred and in 2013, it seemed that time was ripe for a major, multi-authored, reinvestigation. Therefore, a Special Issue of Pattern Recognition in Physics was devoted to: "Pattern in solar variability, their planetary origin and terrestrial impacts". The volume includes 12 separate research papers and General Conclusions, co-authored by 19 prominent scientists. Indeed, they agreed that the driving factor of solar variability must emerge from the planetary beat on the Sun, and by that its emission of luminosity and Solar Wind both factors of which affect the Earth-Moon system. This may be held as a benchmark event in our understanding of the planetary-solar-terrestrial interaction. Furthermore, they noted two implications of this: partly that the old hypothesis was now lifted to a firm theory, maybe even a new paradigm, and partly that we are on our way into a new grand solar minimum which "sheds serious doubts on the issue of a continued, even accelerated, warming as claimed by the IPCC". "We were alarmed by the second implication", Martin Rasmussen, VD of Copernicus, stated, and took the unbelievable decision immediately to close down the entire journal. This happened on January 17 without any discussion
19. The fourth dimension simply explained
CERN Document Server
Manning, Henry P
2005-01-01
To remove the contents of an egg without puncturing its shell or to drink the liquor in a bottle without removing the cork is clearly unthinkable - or is it? Understanding the world of Einstein and curved space requires a logical conception of the fourth dimension.This readable, informative volume provides an excellent introduction to that world, with 22 essays that employ a minimum of mathematics. Originally written for a contest sponsored by Scientific American, these essays are so well reasoned and lucidly written that they were judged to merit publication in book form. Their easily unders
20. Stratospheric Temperature Changes and Ozone Recovery in the 21st Century
Institute of Scientific and Technical Information of China (English)
HU Yongyun; XIA Yan; GAO Mei; LU Daren
2009-01-01
Increasing greenhouse gases and likely ozone recovery will be the two most important factors influencing changes in stratospheric temperatures in the 21st century. The radiative effect of increasing greenhouse gases will cause cooling in the stratosphere, while ozone recovery will lead to stratospheric warming. To investigate how stratospheric temperatures change under the two opposite forcings in the 21st century, we use observed ozone and reanalysis data as well as simulation results from four coupled oceanic and atmospheric general circulation models (GISS-ER, GFDL-CM20, NCAR-CCSM3, and UKMO-HadCM3) used in the IPCC (Intergovernment Panel for Climate Change) Fourth Assessment Report (AR4). Observational analysis shows that total column ozone and lower stratospheric temperatures all show increasing in the past 10 years, while middle stratospheric temperatures demonstrate cooling. IPCC AR4 simulations show that greenhouse forcing alone will lead to stratospheric cooling. However, with forcing of both increasing greenhouse gases and ozone recovery, the middle stratosphere will be cooled, while the lower stratosphere will be warmed. Warming magnitudes vary from one model to another. UKMO-HadCM3 generates relatively strong warming for all three greenhouse scenarios, and warming extends to 40 hPa. GFDL-CM20 and NCAR-CCSM3 produce weak warming, and warming mainly exists at lower levels, below about 60 hPa. In addition, we also discuss the effect of temperature changes on ozone recovery.
1. Simulation evaluation and future prediction of the IPCC-AR4 GCMs on the extreme temperatures in China
Institute of Scientific and Technical Information of China (English)
WANG Ji; JIANG Zhihong; SONG Jie; LOU Dejun
2008-01-01
On the basis of the temperature observations during 1961-2000 in China, seven coupled general circulation models' (GCMs) extreme temperature products are evaluated supplied by the Intergovemmental Panel on Climate Change's 4th Assessment Report (IPCC-AR4). The extreme temperature indices in use are frost days (FD), growing season length (GSL), extreme tempera-ture range (ETR), warm nights (TN90), and heat wave duration index (HWDI). Results indicate that all the seven models are capable of simulating spatial and temporal variations in temperature characteristics, and their ensemble acts more reliable than any single one. Among the seven models, GFDL-CM2.0 and MIROC3.2 performances are much better. Besides, most of the mod-els are able to present linear trends of the same positive/negative signs as the observations but for weaker intensities. The simula-tion effects are different on a nationwide basis, with 110°N as the division, east (west) of which the effects are better (worse) and the poorer over the Qinghai-Tibetan Plateau in China. The predictions for the 21st century on emissions scenarios show that except decreases in the FD and ETR, other indices display significant increasing trend, especially for the indices of HWDI and TN90, which represent the notable extreme climate. This indicates that the temperature-related climate is moving towards the ex-treme. In the late 21st century, the GSL and TN90 (HWDI) increase most notably in southwest China (the Qinghai--Tibetan Plateau), and the FD decrease most remarkably in the Qinghai-Tibetan Plateau, northwest and northeast of China. Apart from South China, the yearly change range of the extreme temperature is reduced in most of China.
2. Contribution of the working group 2 to the fourth evaluation report of the inter government expert group on the climatic change. Evaluation 2007 of the climatic changes: impacts, adaptation and vulnerability; Contribution du Groupe de travail 2 au quatrieme rapport d'evaluation du Groupe d'expert intergouvernemental sur l'evolution du climat. Bilan 2007 des changements climatiques: impacts, adaptation et vulnerabilite
Energy Technology Data Exchange (ETDEWEB)
NONE
2007-07-01
This document exposes the results of the fourth evaluation report of the working group II of the inter government experts group on the climatic change. This evaluation presents the today scientific understanding of the climatic change impacts on the humans and their adaptation ability and vulnerability. It is based on the GIEC evaluations and new knowledge added since the third evaluation report. (A.L.B.)
3. Future climate in world regions: an intercomparison of model-based projections for the new IPCC emissions scenarios
Energy Technology Data Exchange (ETDEWEB)
Ruosteenoja, K.; Carter, T.R.; Jylhae, K.; Tuomenvirta, H.
2003-07-01
Projections of changes in seasonal surface air temperature and precipitation for three 30-year periods during the 21st century in 32 sub-continental scale regions are presented. This information may offer useful guidance on the selection of climate scenarios for regional impact studies. The climate changes have been simulated by seven coupled atmosphere-ocean general circulation models (AOGCMs), the greenhouse gas and aerosol forcing being inferred from the SRES emission scenarios A1F1, A2, B1 and B2. For a majority of the AOGCMs, simulations have only been conducted for scenarios A2 and B2. Projections for other scenarios were then extrapolated from the available runs applying a pattern-scaling technique. In tests, this method proved to be fairly accurate, the correlation between the AOGCM-simulated and the corresponding pattern-scaled response to the A2 scenario for the end of the 21st century being generally {approx} 0.97 - 0.99 for temperature and {approx} 0.9 or higher for precipitation. Projected changes of temperature and precipitation are presented in the form of 384 scatter diagrams. The model-simulated temperature changes were almost invariably statistically significant, i.e., they fell clearly outside the natural multi-decadal variability derived from 1000-year unforced coupled AOGCM simulations. For precipitation, fewer modelled changes were statistically significant, especially in the earliest projection period 2010-2039. Differences in the projections given by various models were substantial, of the same order of magnitude by the end of the century as differences among the responses to separate forcing scenarios. Nevertheless, the surface air temperature increased in all regions and seasons. For precipitation, changes with both sign occurred, but an increase of regional precipitation was more common than a decrease. All models simulate higher precipitation at high latitudes and enhanced summer monsoon precipitation for Southern and Eastern Asia. There
4. Simulations of the 100-hPa South Asian High and Precipitation over East Asia with IPCC Coupled GCMs
Institute of Scientific and Technical Information of China (English)
ZHOU Ningfang; YU Yongqiang; QIAN Yongfu
2006-01-01
The South Asian High (SAH) and precipitation over East Asia simulated by 11 coupled GCMs associated with the forthcoming Intergovernmental Panel on Climate Change's (IPCC) 4th Assessment Report are evaluated. The seasonal behavior of the SAH is presented for each model. Analyses of the results show that all models are able to reproduce the seasonal cycle of the SAH. Locations of the SAH center are also basically reproduced by these models. All models underestimate the intensity and the extension of coverage in summer. The anomalous SAH can be divided into east and west modes according to its longitu dinal position in summer on the interannual timescale, and the composite anomalies of the observed precipitation for these two modes tend to have opposite signs over East Asia. However, only several coupled GCMs can simulate the relationship between rainfall and SAH similar to the observed one, which may be associated with the bias in simulation of the subtropical anticyclone over the West Pacific (SAWP) at 500 hPa. In fact, it is found that any coupled GCM, that can reproduce the reasonable summer mean state of SAWP and the southward (northward) withdrawal (extension) for the east (west) mode of SAH as compared to the observed, will also simulate similar rainfall anomaly patterns for the east and west SAH modes over East Asia. Further analysis indicates that the observed variations in the SAH, SAWP and rainfall are closely related to the sea surface temperature (SST) over the equatorial tropical Pacific. Particularly, some models cannot simulate the SAWP extending northward in the west mode and withdrawing southward in the east mode, which may be related to weak major El Ni(n)o or La Ni(n)a events.The abilities of the coupled GCMs to simulate the SAWP and ENSO events are associated partly with their ability to reproduce the observed relationship between SAH and the rainfall anomaly over East Asia.
5. Methane production as a result from rumen fermentation in cattle calculated by using the IPCC-GPG tier 2 method
NARCIS (Netherlands)
Smink W; Pellikaan WF; Kolk LJ van der; Hoek KW van der; LVM
2005-01-01
Following the IPCC GPG Guidelines the enteric fermentation emission is calculated as a percentage of the gross energy intake in cattle feedstuffs. The energy consumption for maintenance, activity, growth, lactation and pregnancy is described in detail. Combining this with the average rations for cat
6. A Search for the Fourth SM Family Quarks through Anomalous Decays
CERN Document Server
Sahin, M; Turkoz, S
2010-01-01
Existence of the fourth family follows from the basics of the Standard Model. Because of the high masses of the fourth family quarks, their anomalous decays could be dominant. This will drastically change the search strategy at hadron colliders. We show that the fourth SM family down quarks with masses up to 400-450 GeV can be observed (or excluded) via anomalous decays by Tevatron before the LHC.
7. Gravitational lensing in fourth order gravity
CERN Document Server
Capozziello, S; Troisi, A
2006-01-01
Gravitational lensing is investigated in the weak field limit of fourth order gravity in which the Lagrangian of the gravitational field is modified by replacing the Ricci scalar curvature R with an analytical expression $f(R)$. Considering the case of a pointlike lens, we study the behaviour of the deflection angle in the case of power law Lagrangians, i.e. with f(R) = f_0 R^n. In order to investigate possible detectable signatures, the position of the Einstein ring and the solutions of the lens equation are evaluated considering the change with respect to the standard case. Effects on the amplification of the images and the Paczynski curve in microlensing experiments are also estimated.
8. The Health Effects of Climate Change in the WHO European Region
Directory of Open Access Journals (Sweden)
Tanja Wolf
2015-11-01
Full Text Available The evidence of observed health effects as well as projections of future health risks from climate variability and climate change is growing. This article summarizes new knowledge on these health risks generated since the IPCC fourth assessment report (AR4 was published in 2007, with a specific focus on the 53 countries comprising the WHO European Region. Many studies on the effects of weather, climate variability, and climate change on health in the European Region have been published since 2007, increasing the level of certainty with regard to already known health threats. Exposures to temperature extremes, floods, storms, and wildfires have effects on cardiovascular and respiratory health. Climate- and weather-related health risks from worsening food and water safety and security, poor air quality, and ultraviolet radiation exposure as well as increasing allergic diseases, vector- and rodent-borne diseases, and other climate-sensitive health outcomes also warrant attention and policy action to protect human health.
9. PERSPECTIVE: Climate change: seeking balance in media reports
Science.gov (United States)
Huntingford, Chris; Fowler, David
2008-06-01
10. Energetics of IPCC4AR Climate Models: Energy Balance and Meridional Enthalpy Transports
CERN Document Server
Lucarini, Valerio
2009-01-01
We consider the climate simulations performed using pre-industrial and SRESA1B scenarios and analyse the outputs of the state-of-the-art models included in IPCC4AR. For control simulations, large energy biases are present for several models both when global climate budgets and when energy budgets of the atmospheric, oceanic, and land subdomains are considered. The energy biases depend on the imperfect closure of the energy cycle in the fluid components of the climate system and on issues in the treatment of phase transitions and heat fluxes over land. Additionally, the consequence of a positive global energy bias, which is what most models feature, is the underestimation of the thermodynamic emission temperature of the planet and of the globally averaged surface temperature. This may help explaining the cold bias of climate models. Models agree on the representation of meridional enthalpy transports in terms of location of the peaks of the total and atmospheric transports, whereas quantitative disagreements o...
11. Towards developing IPCC methane ‘emission factors’ for peatlands (organic soils
Directory of Open Access Journals (Sweden)
J. Couwenberg
2012-03-01
Full Text Available (1 Huge reductions of carbon dioxide (CO2 and nitrous oxide (N2O effluxes can be attained by rewetting drained peatlands, but this will increase methane (CH4 effluxes.(2 The scientific data base for methane effluxes from peatlands is much larger than that for CO2 or N2O. Once anoxic conditions are provided, the availability of fresh plant material is the major factor in methane production. Old (recalcitrant peat plays only a subordinate role in gas efflux.(3 The annual mean water level is a surprisingly good indicator for methane effluxes, but at high water levels the cover of aerenchymous shunts (gas conductive plant tissue becomes a better proxy. Ideally, both water level and cover of aerenchymous shunts should be assessed to arrive at robust estimates of methane effluxes.(4 The available data provide sufficient guidance for arriving at moderately accurate (Tier 1 estimates consistent with IPCC methodologies. For more accurate estimation (higher tier approaches, vegetation provides a promising basis for development of more detailed efflux factors. Vegetation is a good proxy for mean water levels and can provide - with extra attention to aerenchymous shunts - a robust proxy for accurate and spatially explicit estimates of methane effluxes over large areas.
12. Communicating Uncertainties on Climate Change
Science.gov (United States)
Planton, S.
2009-09-01
The term of uncertainty in common language is confusing since it is related in one of its most usual sense to what cannot be known in advance or what is subject to doubt. Its definition in mathematics is unambiguous but not widely shared. It is thus difficult to communicate on this notion through media to a wide public. From its scientific basis to the impact assessment, climate change issue is subject to a large number of sources of uncertainties. In this case, the definition of the term is close to its mathematical sense, but the diversity of disciplines involved in the analysis process implies a great diversity of approaches of the notion. Faced to this diversity of approaches, the issue of communicating uncertainties on climate change is thus a great challenge. It is also complicated by the diversity of the targets of the communication on climate change, from stakeholders and policy makers to a wide public. We will present the process chosen by the IPCC in order to communicate uncertainties in its assessment reports taking the example of the guidance note to lead authors of the fourth assessment report. Concerning the communication of uncertainties to a wide public, we will give some examples aiming at illustrating how to avoid the above-mentioned ambiguity when dealing with this kind of communication.
13. Disaggregated N2O emission factors in China based on cropping parameters create a robust approach to the IPCC Tier 2 methodology
Science.gov (United States)
Shepherd, Anita; Yan, Xiaoyuan; Nayak, Dali; Newbold, Jamie; Moran, Dominic; Dhanoa, Mewa Singh; Goulding, Keith; Smith, Pete; Cardenas, Laura M.
2015-12-01
China accounts for a third of global nitrogen fertilizer consumption. Under an International Panel on Climate Change (IPCC) Tier 2 assessment, emission factors (EFs) are developed for the major crop types using country-specific data. IPCC advises a separate calculation for the direct nitrous oxide (N2O) emissions of rice cultivation from that of cropland and the consideration of the water regime used for irrigation. In this paper we combine these requirements in two independent analyses, using different data quality acceptance thresholds, to determine the influential parameters on emissions with which to disaggregate and create N2O EFs. Across China, the N2O EF for lowland horticulture was slightly higher (between 0.74% and 1.26% of fertilizer applied) than that for upland crops (values ranging between 0.40% and 1.54%), and significantly higher than for rice (values ranging between 0.29% and 0.66% on temporarily drained soils, and between 0.15% and 0.37% on un-drained soils). Higher EFs for rice were associated with longer periods of drained soil and the use of compound fertilizer; lower emissions were associated with the use of urea or acid soils. Higher EFs for upland crops were associated with clay soil, compound fertilizer or maize crops; lower EFs were associated with sandy soil and the use of urea. Variation in emissions for lowland vegetable crops was closely associated with crop type. The two independent analyses in this study produced consistent disaggregated N2O EFs for rice and mixed crops, showing that the use of influential cropping parameters can produce robust EFs for China.
14. Fourth international radiopharmaceutical dosimetry symposium
International Nuclear Information System (INIS)
15. Fourth order spatial derivative gravity
Science.gov (United States)
Bemfica, F. S.; Gomes, M.
2011-10-01
In this work, we study a modified theory of gravity that contains up to fourth order spatial derivatives as a model for the Hořava-Lifshitz gravity. The propagator is evaluated and, as a result, one extra pole is obtained, corresponding to a spin-2 nonrelativistic massless particle, an extra term which jeopardizes renormalizability, besides the unexpected general relativity unmodified propagator. Then unitarity is proved at the tree level, where the general relativity pole has been shown to have no dynamics, remaining only the 2 degrees of freedom of the new pole. Next, the nonrelativistic effective potential is determined from a scattering process of two identical massive gravitationally interacting bosons. In this limit, Newton’s potential is obtained, together with a Darwin-like term that comes from the extra nonpole term in the propagator. Regarding renormalizability, this extra term may be harmful by power counting, but it can be eliminated by adjusting the free parameters of the model. This adjustment is in accord with the detailed balance condition suggested in the literature and shows that the way in which extra spatial derivative terms are added is of fundamental importance.
16. Scenarios of climate change
International Nuclear Information System (INIS)
This article provides an overview of current and prospected climate changes, their causes and implied threats, and of a possible route to keep the changes within a tolerable level. The global mean temperature has up to 2005 risen by almost 0.8 C deg., and the change expected by 2100 is as large as glacial-interglacial changes in the past, which were commonly spread out over 10 000 years. As is well known, the principle actor is man-made CO2, which, together with other anthropogenic gases, enhances the atmosphere's greenhouse effect. The only man-made cooling agent appears to be atmospheric aerosols. Atmospheric CO2 has now reached levels unprecedented during the past several million years. Principal threats are a greatly reduced biodiversity (species extinction), changes in the atmospheric precipitation pattern, more frequent weather extremes, and not the least, sea level rise. The expected precipitation pattern will enhance water scarcity in and around regions that suffer from water shortage already, affecting many countries. Sea level rise will act on a longer time scale. It is expected to amount to more than 50 cm by 2100, and over the coming centuries the potential rise is of the order of 10 m. A global-mean temperature increase of 2 C deg. is often quoted as a safe limit, beyond which irreversible effects must be expected. To achieve that limit, a major, rapid, and coordinated international effort will be needed. Up to the year 2050, the man-made CO2 releases must be reduced by at least 50%. This must be accompanied by a complete overhaul of the global energy supply toward depending increasingly on the Sun's supply of energy, both directly and in converted form, such as wind energy. Much of the information and insight available today has been generated by the Intergovernmental Panel on Climate Change (IPCC), in particular its Fourth Assessment Report of 2007, which greatly advanced both public attention and political action. (author)
17. Methane production as a result from rumen fermentation in cattle calculated by using the IPCC-GPG tier 2 method; Methaanproductie als gevolg van pensfermentatie bij rundvee berekend middels de IPCC-GPG Tier 2 methode
Energy Technology Data Exchange (ETDEWEB)
Smink, W.; Pellikaan, W.F.; Van der Kolk, L.J. [Feed Innovation Services, Aarle-Rixtel (Netherlands); Van der Hoek, K.W. [Rijksinstituut voor Volksgezondheid en Milieu RIVM, Bilthoven (Netherlands)
2004-07-01
Following the IPCC GPG Guidelines the enteric fermentation emission is calculated as a percentage of the gross energy intake in cattle feedstuffs. The energy consumption for maintenance, activity, growth, lactation and pregnancy is described in detail. Combining this with the average rations for cattle the methane emissions are calculated. The report presents for the period 1990-2002 for all cattle categories the methane emission resulting from enteric fermentation. [Dutch] Volgens de IPCC-GPG methode wordt de methaanemissie berekend als percentage van de door het dier opgenomen bruto energie met het voer. Het rapport beschrijft in detail de benodigde energie voor onderhoud, activiteit, groei, lactatie en dracht. Met behulp van de rantsoenen van het Nederlandse rundvee wordt vervolgens de bruto energie opname berekend. Het rapport presenteert voor de volledige periode 1990-2002 voor alle rundveecategorieen de methaanemissie als gevolg van pensfermentatie.
18. Systems Prototyping with Fourth Generation Tools.
Science.gov (United States)
Sholtys, Phyllis
1983-01-01
The development of information systems using an engineering approach that uses both traditional programing techniques and fourth generation software tools is described. Fourth generation applications tools are used to quickly develop a prototype system that is revised as the user clarifies requirements. (MLW)
19. Detection and Attribution of Anthropogenic Climate Change Impacts
Science.gov (United States)
Rosenzweig, Cynthia; Neofotis, Peter
2013-01-01
Human-influenced climate change is an observed phenomenon affecting physical and biological systems across the globe. The majority of observed impacts are related to temperature changes and are located in the northern high- and midlatitudes. However, new evidence is emerging that demonstrates that impacts are related to precipitation changes as well as temperature, and that climate change is impacting systems and sectors beyond the Northern Hemisphere. In this paper, we highlight some of this new evidence-focusing on regions and sectors that the Intergovernmental Panel on Climate Change Fourth Assessment Report (IPCC AR4) noted as under-represented-in the context of observed climate change impacts, direct and indirect drivers of change (including carbon dioxide itself), and methods of detection. We also present methods and studies attributing observed impacts to anthropogenic forcing. We argue that the expansion of methods of detection (in terms of a broader array of climate variables and data sources, inclusion of the major modes of climate variability, and incorporation of other drivers of change) is key to discerning the climate sensitivities of sectors and systems in regions where the impacts of climate change currently remain elusive. Attributing such changes to human forcing of the climate system, where possible, is important for development of effective mitigation and adaptation. Current challenges in documenting adaptation and the role of indigenous knowledge in detection and attribution are described.
20. Future projection of East China Sea temperature by dynamic downscaling of the IPCC_AR4 CCSM3 model result
Institute of Scientific and Technical Information of China (English)
YU Xiaolin; WANGF Fan; TANG Xiaohui
2012-01-01
Future temperature distributions of the marginal Chinese seas are studied by dynamic downscaling of global CCSM3 IPCC_AR4 scenario runs.Different forcing fields from 2080-2099 Special Report on Emissions Scenarios(SRES)B1,A1,and A2 to 1980-1999 20C3M are averaged and superimposed on CORE2 and SODA2.2.4 data to force high-resolution regional future simulations using the Regional Ocean Modeling System(ROMS).Volume transport increments in downscaling simulation support the CCSM3result that with a weakening subtropical gyre circulation,the Kuroshio Current in the East China Sea(ECS)is possibly strengthened under the global wanning scheme.This mostly relates to local wind change,whereby the summer monsoon is strengthened and winter monsoon weakened.Future temperature fluxes and their seasonal variations are larger than in the CCSM3 result.Downscaling 100 years' temperature increments are comparable to the CCSM3,with a minimum in B1 scenario of 1.2-2.0℃ and a maximum in A2 scenario of 2.5-4.5℃.More detailed temperature distributions are shown in the downscaling simulation.Larger increments are in the Bohai Sea and middle Yellow Sea,and smaller increments near the southeast coast of China,west coast of Korea,and southern ECS.There is a reduction of advective heat north of Taiwan Island and west of Tsushima in summer,and along the southern part of the Yellow Sea warm current in winter.There is enhancement of advective heat in the northern Yellow Sea in winter,related to the delicate temperature increment distribution.At 50 meter depth,the Yellow Sea cold water mass is destroyed.Our simulations suggest that in the formation season of the cold water mass,regional temperature is higher in the future and the water remains at the bottom until next summer.In summer,the mixed layer is deeper,making it much easier for the strengthened surface heat flux to penetrate to the bottom of this water.
1. Economic and policy issues in climate change
International Nuclear Information System (INIS)
Global climate change has emerged as one of today's most challenging and controversial policy issues. In this significant new contribution, a roster of premier scholars examines economic and social aspects of that far-reaching phenomenon. Although the 1997 summit in Kyoto focused world attention on climate, it was just one step in an ongoing process. Research by the U.N.'s Intergovernmental Panel on Climate Change (IPCC) has been ongoing since 1988. An extensive IPCC Working Group report published in 1995 examined the economic and social aspects of climate change. In this new volume, eminent economists assess that IPCC report and address the questions that emerge. William Nordhaus's introduction establishes the context for this book. It provides basic scientific background, reviews the IPCC's activities, and explains the genesis of the project
2. Testing an astronomically based decadal-scale empirical harmonic climate model versus the IPCC (2007) general circulation climate models
Science.gov (United States)
Scafetta, Nicola
2012-05-01
We compare the performance of a recently proposed empirical climate model based on astronomical harmonics against all CMIP3 available general circulation climate models (GCM) used by the IPCC (2007) to interpret the 20th century global surface temperature. The proposed astronomical empirical climate model assumes that the climate is resonating with, or synchronized to a set of natural harmonics that, in previous works (Scafetta, 2010b, 2011b), have been associated to the solar system planetary motion, which is mostly determined by Jupiter and Saturn. We show that the GCMs fail to reproduce the major decadal and multidecadal oscillations found in the global surface temperature record from 1850 to 2011. On the contrary, the proposed harmonic model (which herein uses cycles with 9.1, 10-10.5, 20-21, 60-62 year periods) is found to well reconstruct the observed climate oscillations from 1850 to 2011, and it is shown to be able to forecast the climate oscillations from 1950 to 2011 using the data covering the period 1850-1950, and vice versa. The 9.1-year cycle is shown to be likely related to a decadal Soli/Lunar tidal oscillation, while the 10-10.5, 20-21 and 60-62 year cycles are synchronous to solar and heliospheric planetary oscillations. We show that the IPCC GCM's claim that all warming observed from 1970 to 2000 has been anthropogenically induced is erroneous because of the GCM failure in reconstructing the quasi 20-year and 60-year climatic cycles. Finally, we show how the presence of these large natural cycles can be used to correct the IPCC projected anthropogenic warming trend for the 21st century. By combining this corrected trend with the natural cycles, we show that the temperature may not significantly increase during the next 30 years mostly because of the negative phase of the 60-year cycle. If multisecular natural cycles (which according to some authors have significantly contributed to the observed 1700-2010 warming and may contribute to an
3. Fourth-order mutual coherence function in oceanic turbulence.
Science.gov (United States)
Baykal, Yahya
2016-04-10
We have recently expressed the structure constant of atmospheric turbulence in terms of the oceanic turbulence parameters, which are the ratio of temperature to salinity contributions to the refractive index spectrum, rate of dissipation of kinetic energy per unit mass of fluid, rate of dissipation of the mean-squared temperature, wavelength, Kolmogorov microscale, and link length. In this paper, utilizing this recently found structure constant and the fourth-order mutual coherence function of atmospheric turbulence, we present the fourth-order mutual coherence function to be used in oceanic turbulence evaluations. Thus, the found fourth-order mutual coherence function of oceanic turbulence is evaluated for the special case of a point source located at the transmitter origin and at a single receiver point. The variations of this special case of the fourth-order mutual coherence function of oceanic turbulence against the changes in the ratio of temperature to salinity contributions to the refractive index spectrum, the rate of dissipation of kinetic energy per unit mass of fluid, the rate of dissipation of the mean-squared temperature, the wavelength, and the Kolmogorov microscale at various link lengths are presented. PMID:27139862
4. Predictive Understanding of Seasonal Hydrological Dynamics under Climate and Land Use-Land Cover Change
Science.gov (United States)
Batra, N.; Yang, Y. E.; Choi, H. I.; Kumar, P.; Cai, X.; Fraiture, C. D.
2008-12-01
Water has always been and will continue to be an important factor in agricultural production and any alteration in the seasonal distribution of water availability due to climate and land use-land cover change (LULCC) will significantly impact the future production. To achieve the ecologic, economic and social objectives of sustainability, physical understanding of the linkages between climatic changes, LULCC and hydrological response is required. Aided by satellite data, modeling and understanding of the interactions between physical processes of the climate system and society, more reliable regional LULCC and climate change projections are now available. However, resulting quantitative projection of changes on the regional scale hydrological components at the seasonal time scale are sparse. This study attempts to quantify the seasonal hydrological response due to projected LULCC and climate change scenario of Intergovernmental Panel on Climate Change (IPCC) in different hydro-climatic regions of the world. The Common Land Model (CLM) is used for global assessment of future hydrologic response with the development of a consistent global GIS based database for the surface boundary conditions and meteorological forcing of the model. Future climate change projections are derived from the IPCC Fourth Assessment Report: Working Group I - The Physical Science Basis. The study is performed over nine river basins selected from Asia, Africa and North America to present the broad climatic and landscape controls on the seasonal hydrological dynamics. Future changes in water availability are quite evident from our results based upon changes in the volume and seasonality of precipitation, runoff and evapotranspiration. Severe water scarcity is projected to occur in certain seasons which may not be known through annual estimates. Knowledge of the projected seasonal hydrological response can be effectively used for developing adaptive management strategies for the sustainability
5. The fourth state of water
OpenAIRE
E.V. Savich
2013-01-01
Четвертое состояние воды – это водяной пар под энергетическим воздействием напряженности магнитного поля в точке Кюри, в момент магнитного фазового перехода обретает свойство сегнетоэлектрической плазмы, которая есть основа образований таких природных явлений, как грозовое облако и шаровая молния, способных при определенных условиях выделять молнийные разряды и при взрыве большое количество тепловой энергии, что характерно для низкотемпературного реактора. The fourth state of water is aque...
6. Heading for the fourth nuclear age
International Nuclear Information System (INIS)
The author examines the evolution of the global nuclear order since the advent of nuclear weapons in 1945 to present by breaking down the sixty-plus years of nuclear history into three analytically distinct 'ages', each lasting roughly twenty years. By doing so, the author traces back the roots of the current nuclear predicament to some early seeds of trouble which have gradually grown more profound. He attributes much of the unraveling of the nuclear order to: - Certain inherent weaknesses in the original NPT formula; - Changes in the global distribution of power since the codification of the nuclear order in the 1960's; - The dissemination of nuclear weapon technology; and - Complacency and subsequent disillusionment with the nuclear order since the early 1990's. The paper further analyzes what could precipitate a new nuclear age around 2010. The author argues that such a 'fourth nuclear age' would likely be characterized by either a nuclear anarchy, which he believes has become the default option, or a more benign nuclear order manifested by lower numbers of weapons and stringent controls and restrictions on remaining nuclear arsenals and activities. He concludes by considering the more pressing requirements for regaining nuclear stability
7. Handle Fireworks with Care on The Fourth
Science.gov (United States)
... Fourth Take steps to avoid injury while celebrating independence To use the sharing features on this page, ... a university news release. And Dr. Jay McCollum, director of emergency services at the university's Eye Hospital, ...
8. Fourth Moments and Independent Component Analysis
OpenAIRE
Miettinen, Jari; Taskinen, Sara; Nordhausen, Klaus; Oja, Hannu
2014-01-01
In independent component analysis it is assumed that the components of the observed random vector are linear combinations of latent independent random variables, and the aim is then to find an estimate for a transformation matrix back to these independent components. In the engineering literature, there are several traditional estimation procedures based on the use of fourth moments, such as FOBI (fourth order blind identification), JADE (joint approximate diagonalization of eigenmatrices), a...
9. Al Gore' Inconvenient Truth, where? The IPCC report provides the answer; Al Gore' Inconvenient Truth, waar? Het IPCC rapport geeft het antwoord
Energy Technology Data Exchange (ETDEWEB)
Zeiler, W. [Kropman, Rijswijk (Netherlands)
2007-09-15
The climate is changing. Extreme weather conditions occur with greater frequency. In the Netherlands, for example, there have been heat waves, the drought of 2003, exceptional rainstorms and, in 2004, the wettest August in 150 years. The populations of developing countries and vulnerable areas such as the Netherlands will bear the consequences of climate change. The impact on the natural environment will also be severe. Some flora and fauna will extinct while others will proliferate. The outcome will be a strong decline in biodiversity. [Dutch] Het klimaat verandert. De gemiddelde temperatuur stijgt. Er komen vaker en vaker extreme weersomstandigheden voor. Denk in Nederland aan de hittegolven en de droogte van de zomer 2003 en aan de vele regenbuien in augustus 2004, de natste augustus maand in 150 jaar. Vooral de mensen in ontwikkelingslanden en de mensen in kwetsbare gebieden zoals Nededand zullen het meest worden getroffen door de gevolgen van klimaatverandering. Ook de natuur zal ingrijpend veranderen. Bepaalde plant- en diersoorten zullen uitsterven andere zullen gaan overwoekeren. De diversiteit zal sterk afnemen.
10. Predicted risk of cobalt deficiency in grazing sheep from a geochemical survey; communicating uncertainty with the IPCC verbal scale.
Science.gov (United States)
Lark, R. M.; Ander, E. L.; Cave, M. R.; Knights, K. V.; Glennon, M. M.; Scanlon, R. P.
2014-05-01
Deficiency or excess of certain trace elements in the soil causes problems for agriculture, including disorders of grazing ruminants. Farmers and their advisors in Ireland use index values for the concentration of total soil cobalt and manganese to identify where grazing sheep are at risk of cobalt deficiency. We used cokriging with topsoil data from a regional geochemical survey across six counties of Ireland to form local cokriging predictions of cobalt and manganese concentrations with an attendant distribution which reflects the joint uncertainty of these predictions. From this distribution we then computed conditional probabilities for different combinations of cobalt and manganese index values, and so for the corresponding inferred risk to sheep of cobalt deficiency and the appropriateness of different management interventions. The challenge is to communicate these results effectively to an audience comprising, inter alia, farmers, agronomists and veterinarians. Numerical probabilities are not generally well-understood by non-specialists. For this reason we presented our results as maps using a verbal scale to communicate the probability that a deficiency is indicated by local soil conditions, or that a particular intervention is indicated. In the light of recent research on the effectiveness of the verbal scale used by the Intergovernmental Panel on Climate Change to communicate probabilistic information we reported the geostatistical predictions as follows. First, we use the basic IPCC scale with intensifiers, but we also indicate the corresponding probabilities (as percentages) as recommended by Budescu et al. (2009). Second, we make it clear that the source of uncertainty in these predictions is the spatial variability of soil Co and Mn. The outcome under consideration is therefore that a particular soil management scenario would be indicated if the soil properties were known without error, possible uncertainty about the implications of particular soil
11. The Naturalness of the Fourth SM Family
CERN Document Server
Sultansoy, S
2009-01-01
The necessity of the fourth family follows from the SM basics. According to flavor democracy the Dirac masses of the fourth SM family fermions are almost equal with preferable value 450 GeV, which corresponds to common (for all fundamental fermions) Yukawa coupling equal to SU(2) gauge coupling gW. In principle, one expect u4 a little bit lighter than d4, while nu4 could be essentially lighter than l4 due to Majorana mass terms for right-handed components of neutrinos. Obviously, the fourth family quarks will be copiously produced at the LHC. However, the first indication of the fourth SM family may be provided by early Higgs boson observation due to almost an order enhancement of the gluon fusion to Higgs cross-section. For the same reason the Tevatron still has a chance to observe the Higgs boson before the LHC. Concerning the fourth family leptons, in general, best place will be NLC/CLIC. However, for some mass regions and MNS matrix elements double discovery of both the nu4 and H could be possible at the ...
12. Fourth Tennessee water resources symposium
International Nuclear Information System (INIS)
The annual Tennessee Water Resources Symposium was initiated in 1988 as a means to bring together people with common interests in the state's important water-related resources at a technical, professional level. Initially the symposium was sponsored by the American Institute of Hydrology and called the Hydrology Symposium, but the Tennessee Section of the American Water Resources Association (AWRA) has taken on the primary coordination role for the symposium over the last two years and the symposium name was changed in 1990 to water resources to emphasize a more inter-disciplinary theme. This year's symposium carries on the successful tradition of the last three years. Our goal is to promote communication and cooperation among Tennessee's water resources professionals: scientists, engineers, and researchers from federal, state, academic, and private institutions and organizations who have interests and responsibilities for the state's water resources. For these conference proceedings, individual papers are processed separately for the Energy Data Base
13. Fourth International Seminar and Fourth National Workshop 'Use and development of Health Related Industrial Isotope Products'
International Nuclear Information System (INIS)
The analysis of the main subjects was tackled during the Fourth International Seminar and Fourth National Workshop 'Use and development of Health Related Industrial Isotope Products' which took place on December 7th, 8th and 9th 2010, devoted to the XV Anniversary of the Isotope Centre is showed in the paper.(author)
14. Evaluation of cloud fraction and its radiative effect simulated by IPCC AR4 global models against ARM surface observations
Directory of Open Access Journals (Sweden)
Y. Qian
2012-02-01
Full Text Available Cloud Fraction (CF is the dominant modulator of radiative fluxes. In this study, we evaluate CF simulated in the IPCC AR4 GCMs against ARM long-term ground-based measurements, with a focus on the vertical structure, total amount of cloud and its effect on cloud shortwave transmissivity. Comparisons are performed for three climate regimes as represented by the Department of Energy Atmospheric Radiation Measurement (ARM sites: Southern Great Plains (SGP, Manus, Papua New Guinea and North Slope of Alaska (NSA. Our intercomparisons of three independent measurements of CF or sky-cover reveal that the relative differences are usually less than 10% (5% for multi-year monthly (annual mean values, while daily differences are quite significant. The total sky imager (TSI produces smaller total cloud fraction (TCF compared to a radar/lidar dataset for highly cloudy days (CF > 0.8, but produces a larger TCF value than the radar/lidar for less cloudy conditions (CF < 0.3. The compensating errors in lower and higher CF days result in small biases of TCF between the vertically pointing radar/lidar dataset and the hemispheric TSI measurements as multi-year data is averaged. The unique radar/lidar CF measurements enable us to evaluate seasonal variation of cloud vertical structures in the GCMs.
Both inter-model deviation and model bias against observation are investigated in this study. Another unique aspect of this study is that we use simultaneous measurements of CF and surface radiative fluxes to diagnose potential discrepancies among the GCMs in representing other cloud optical properties than TCF. The results show that the model-observation and inter-model deviations have similar magnitudes for the TCF and the normalized cloud effect, and these deviations are larger than those in surface downward solar radiation and cloud transmissivity. This implies that other dimensions of cloud in addition to cloud amount, such as cloud optical thickness and
15. A Fourth Chiral Generation And Susy Breaking
CERN Document Server
Wingerter, Akin
2011-01-01
We revisit four generations within the context of supersymmetry. We compute the perturbativity limits for the fourth generation Yukawa couplings and show that if the masses of the fourth generation lie within reasonable limits of their present experimental lower bounds, it is possible to have perturbativity only up to scales around 1000 TeV, i.e. the current experimental bounds and perturbative unification are mutually exclusive. Such low scales are ideally suited to incorporate gauge mediated supersymmetry breaking, where the mediation scale can be as low as 10-20 TeV. The minimal messenger model, however, is highly constrained. Lack of electroweak symmetry breaking rules out a large part of the parameter space, and in the remaining part, the fourth generation stau is tachyonic.
16. IPCC确定的几种未来大气CO2浓度水平对人为CO2排放的限制%The Limit to Anthropogenic Carbon Emissions under IPCC Scenarios
Institute of Scientific and Technical Information of China (English)
金心; 石广玉
2001-01-01
A three-dimensional ocean carbon cycle model and a simple terrestrial biosphere model are used to simulate anthropogenic CO2 uptake by the ocean and terrestrial biosphere under the IPCC (Intergovernmental Panel on Climate Change) scenarios for future atmospheric Pco2 levels. We estimate anthropogenic carbon emissions required to stabilize future atmospheric CO2 at various levels ranging from 350×10-6 to 750×10-6. All the stabilization scenarios require a substantial future reduction in emissions, but the amount to be reduction for each of the scenarios is quite difference.%用三维海洋碳循环模式和一个简单的陆地生物圈模式计算了IPCC(政府间气候变化委员会)未来大气CO2情景中海洋和生物圈的吸收,并结合土地变化的资料得出燃料的排放值。结果表明:尽管在所有的构想下,为了使大气中CO2浓度达到稳定必须减少排放,但对应不同的IPCC未来大气CO2情景,对人为CO2排放的限制是很不相同的。
17. Twenty first century climate change as simulated by European climate models
International Nuclear Information System (INIS)
Full text: Climate change simulation results for seven European state-of-the-art climate models, participating in the European research project ENSEMBLES (ENSEMBLE-based Predictions of Climate Changes and their Impacts), will be presented. Models from Norway, France, Germany, Denmark, and Great Britain, representing a sub-ensemble of the models contributing to the Intergovernmental Panel on Climate Change Fourth Assessment Report (IPCC AR4), are included. Climate simulations are conducted with all the models for present-day climate and for future climate under the SRES A1B, A2, and B1 scenarios. The design of the simulations follows the guidelines of the IPCC AR4. The 21st century projections are compared to the corresponding present-day simulations. The ensemble mean global mean near surface temperature rise for the year 2099 compared to the 1961-1990 period amounts to 3.2Kforthe A1B scenario, to 4.1 K for the A2 scenario, and to 2.1 K for the B1 scenario. The spatial patterns of temperature change are robust among the contributing models with the largest temperature increase over the Arctic in boreal winter, stronger warming overland than over ocean, and little warming over the southern oceans. The ensemble mean globally averaged precipitation increases for the three scenarios (5.6%, 5.7%, and 3.8% for scenarios A1B, A2, and B1, respectively). The precipitation signals of the different models display a larger spread than the temperature signals. In general, precipitation increases in the Intertropical Convergence Zone and the mid- to high latitudes (most pronounced during the hemispheric winter) and decreases in the subtropics. Sea-level pressure decreases over the polar regions in all models and all scenarios, which is mainly compensated by a pressure increase in the subtropical highs. These changes imply an intensification of the Southern and Northern Annular Modes
18. "Researching" with Third- and Fourth-Graders.
Science.gov (United States)
Liston, Barbara
1970-01-01
In order to instill in children the skills which will be basic to their school experience, words implying a process (such as "hemp,""parasite," and "vanilla") may be "researched" by third and fourth graders through the use of a dictionary, an encyclopedia, a supplementary book on the subject, and an interview with an adult. The child makes a…
19. Sex Differences in Cognitive Abilities. Fourth Edition
Science.gov (United States)
Halpern, Diane F.
2011-01-01
The fourth edition of "Sex Differences in Cognitive Abilities" critically examines the breadth of research on this complex and controversial topic, with the principal aim of helping the reader to understand where sex differences are found--and where they are not. Since the publication of the third edition, there have been many exciting and…
20. Fourth-Generation Computer Languages: An Overview.
Science.gov (United States)
Ricks, John
1988-01-01
Points out that mainframe computer users today can make their requirements known to the computer in simple English. Provides a listing of fourth generation computer language advantages over third generation languages. Summarizes a program to streamline faculty records on a mainframe computer. (MVL)
1. Recreational Reading: Choices of Fourth Graders.
Science.gov (United States)
Boulware, Beverly J.; Foley, Christy L.
1998-01-01
Compares readability levels of the recreational reading books selected by fourth graders. Finds that the students chose books from their school libraries and read (or chose not to read) books on all their reading levels. Concludes that students read books related to their interests, regardless of the book's readability or the student's reading…
2. Fourth Dutch Process Security Control Event
NARCIS (Netherlands)
Luiijf, H.A.M.; Zielstra, A.
2010-01-01
On December 1st, 2009, the fourth Dutch Process Control Security Event took place in Baarn, The Netherlands. The security event with the title ‘Manage IT!’ was organised by the Dutch National Infrastructure against Cybercrime (NICC). Mid of November, a group of over thirty people participated in the
3. Communicating uncertainty: lessons learned and suggestions for climate change assessment
International Nuclear Information System (INIS)
Assessments of climate change face the task of making information about uncertainty accessible and useful to decision-makers. The literature in behavior economics provides many examples of how people make decisions under conditions of uncertainty relying on inappropriate heuristics, leading to inconsistent and counterproductive choices. Modern risk communication practices recommend a number of methods to overcome these hurdles, which have been recommended for the Intergovernmental Panel on Climate Change (IPCC) assessment reports. This paper evaluates the success of the most recent IPCC approach to uncertainty communication, based on a controlled survey of climate change experts. Evaluating the results from the survey, and from a similar survey recently conducted among university students, the paper suggests that the most recent IPCC approach leaves open the possibility for biased and inconsistent responses to the information. The paper concludes by suggesting ways to improve the approach for future IPCC assessment reports. (authors)
4. Effects of IPCC SRES* emissions scenarios on river runoff: a global perspective
Directory of Open Access Journals (Sweden)
N. W. Arnell
2003-01-01
Full Text Available This paper describes an assessment of the implications of future climate change for river runoff across the entire world, using six climate models which have been driven by the SRES emissions scenarios. Streamflow is simulated at a spatial resolution of 0.5°x0.5° using a macro-scale hydrological model, and summed to produce total runoff for almost 1200 catchments. The effects of climate change have been compared with the effects of natural multi-decadal climatic variability, as determined from a long unforced climate simulation using HadCM3. By the 2020s, change in runoff due to climate change in approximately a third of the catchments is less than that due to natural variability but, by the 2080s, this falls to between 10 and 30%. The climate models produce broadly similar changes in runoff, with increases in high latitudes, east Africa and south and east Asia, and decreases in southern and eastern Europe, western Russia, north Africa and the Middle East, central and southern Africa, much of North America, most of South America, and south and east Asia. The pattern of change in runoff is largely determined by simulated change in precipitation, offset by a general increase in evaporation. There is little difference in the pattern of change between different emissions scenarios (for a given model, and only by the 2080s is there evidence that the magnitudes of change in runoff vary, with emissions scenario A1FI producing the greatest change and B1 the smallest. The inter-annual variability in runoff increases in most catchments due to climate change — even though the inter-annual variability in precipitation is not changed — and the frequency of flow below the current 10-year return period minimum annual runoff increases by a factor of three in Europe and southern Africa and of two across North America. Across most of the world climate change does not alter the timing of flows through the year but, in the marginal zone between cool and
5. The Effects of Background Music in the Classroom on the Productivity, Motivation, and Behavior of Fourth Grade Students
Science.gov (United States)
White, Kevin N.
2007-01-01
Many students in a fourth grade classroom at Logan Elementary School are expressing numerous types of negative behaviors, are not motivated to learn, and do not stay on-task. In an effort to change these students, an action research study was conducted that implemented background music in the classroom. There were ten fourth grade students who…
6. Forestal measures against climate change. Review and status after the Fourth Conference of the Parties of the Climate Convention; Skogtiltak mot klimaendringer. Oversikt og status etter fjerde partskonferanse til Klimakonvensjonen
Energy Technology Data Exchange (ETDEWEB)
Naess, Lars Otto
1999-08-01
The Kyoto Protocol of December 1997 opens up the possibility that forestal measures can be used to meet parts of the commitments of the industrialized countries to achieve a net reduction of emission of climate gases. The present report summarizes the issues involved in forestal measures that will mitigate global climate changes. The emphasis is on forestal measures in the climate negotiations and technical carbon binding potential. There is also a brief review of economic, environmental and social aspects. The next decades will be crucial to the many of the worlds forests. The forests contain a large part of the biological diversity. Above all this is true of tropical forests. But untouched areas in tempered and boreal areas are also experiencing various types of threats, including the effects of a possible global heating. It is a main conclusion that, in spite of many complex challenges, climate measures in the forests may play a constructive role both in counteracting global climate changes and in improving the management of the worlds forest resources. 89 refs., 3 figs., 7 tabs.
7. Imitatio Christi in the fourth gospel
Directory of Open Access Journals (Sweden)
D.G. van der Merwe
2001-08-01
Full Text Available Imitatio Christi is a concept which, although not referred to explicitly in the Fourth Gospel, is clearly spelled out in relation to the agency motif occurring in the Gospel. The disciples of Jesus have been appointed as his agents to continue his mission after his departure to his Father. In giving this message through to his readers, the Fourth Evangelist refers to Jesus' calling of his disciples, [Foreign font omitted]; pictures Jesus as [Foreign font omitted]; uses [Foreign font omitted] (the particle of comparison to compare the lives of the disciples with that of Jesus; points out the tasks the disciples had to perform after Jesus' ascension and, finally, indicates how Jesus dwells in his disciples through the Paraclete.
8. Global estimates of carbon stock changes in living forest biomass: EDGARv4.3 - time series from 1990 to 2010
Science.gov (United States)
Petrescu, A. M. R.; Abad-Viñas, R.; Janssens-Maenhout, G.; Blujdea, V. N. B.; Grassi, G.
2012-08-01
While the Emissions Database for Global Atmospheric Research (EDGAR) focuses on global estimates for the full set of anthropogenic activities, the Land Use, Land-Use Change and Forestry (LULUCF) sector might be the most diverse and most challenging to cover consistently for all countries of the world. Parties to United Nations Framework Convention on Climate Change (UNFCCC) are required to provide periodic estimates of greenhouse gas (GHG) emissions, following the latest approved methodological guidance by the International Panel on Climate Change (IPCC). The current study aims to consistently estimate the carbon (C) stock changes from living forest biomass for all countries of the world, in order to complete the LULUCF sector in EDGAR. In order to derive comparable estimates for developing and developed countries, it is crucial to use a single methodology with global applicability. Data for developing countries are generally poor, such that only the Tier 1 methods from either the IPCC Good Practice Guide for Land Use, Land-Use Change and Forestry (GPG-LULUCF) 2003 or the IPCC 2006 Guidelines can be applied to these countries. For this purpose, we applied the IPCC Tier 1 method at global level following both IPCC GPG-LULUCF 2003 and IPCC 2006, using spatially coarse activity data (i.e. area, obtained combining two different global forest maps: the Global Land Cover map and the eco-zones subdivision of the Global Ecological Zone (GEZ) map) in combination with the IPCC default C stocks and C stock change factors. Results for the C stock changes were calculated separately for gains, harvest, fires (Global Fire Emissions Database version 3, GFEDv.3) and net deforestation for the years 1990, 2000, 2005 and 2010. At the global level, results obtained with the two sets of IPCC guidance differed by about 40 %, due to different assumptions and default factors. The IPCC Tier 1 method unavoidably introduced high uncertainties due to the "globalization" of parameters. When the
9. Fourth international conference on Networks & Communications
CERN Document Server
Meghanathan, Natarajan; Nagamalai, Dhinaharan; Computer Networks & Communications (NetCom)
2013-01-01
Computer Networks & Communications (NetCom) is the proceedings from the Fourth International Conference on Networks & Communications. This book covers theory, methodology and applications of computer networks, network protocols and wireless networks, data communication technologies, and network security. The proceedings will feature peer-reviewed papers that illustrate research results, projects, surveys and industrial experiences that describe significant advances in the diverse areas of computer networks & communications.
10. Fourth World Theory: The Evolution of . . .
OpenAIRE
Olon F. Dotson
2014-01-01
Fourth World theory is a methodology for examining and developing greater understanding of the extent of the distress and abandonment commonly found in the cores of American cities resulting from de-industrialization, historic segregation and discrimination patterns, suburban sprawl, erosion of a viable tax base, racism, inability to embrace the concept of desegregation and civil rights legislation, fear, despair, crumbling infrastructure systems, disinvestment in urban school systems, and en...
11. Flavour democracy calls for the fourth generation
International Nuclear Information System (INIS)
It is argued with the help of an illustrative mode, that the inter species hierarchy among the fermion masses and the quark mixing angles can be accommodated naturally in the standard model with (approximate) flavour democracy provided there are four families of sequential quark-leptons with all members of the fourth family having roughly equal masses. The special problem of light neutrino masses (if any) and possible solutions are also discussed. (author). 15 refs
12. Flavour democracy calls for the fourth generation
International Nuclear Information System (INIS)
It is argued with the help of an illustrative model, that the inter species hierarchy among the fermion masses and the quark mixing angles can be accommodated naturally in the standard model with (approximate) flavor democracy provided there are three exactly massless neutrinos and four families of sequential quark-leptons with all members of the fourth family having roughly equal masses. The special problem of light neutrino masses (if any) and possible solutions are also discussed. (author). 22 refs
13. Flavour Democracy Calls for the Fourth Generation
CERN Document Server
Datta, A
1993-01-01
It is argued with the help of an illustrative model, that the inter--species hierarchy among the fermion masses and the quark mixing angles can be accommodated naturally in the standard model with (approximate) flavor democracy provided there are four families of sequential quark--leptons with all members of the fourth family having roughly equal masses. The special problem of light neutrino masses (if any) and possible solutions are also discussed.
14. A Comprehensive Modeling Study on Regional Climate Model (RCM Application — Regional Warming Projections in Monthly Resolutions under IPCC A1B Scenario
Directory of Open Access Journals (Sweden)
Md. Mujibur Rahman
2012-10-01
Full Text Available Some of the major dimensions of climate change include increase in surface temperature, longer spells of droughts in significant portions of the world, associated higher evapotranspiration rates, and so on. It is therefore essential to comprehend the future possible scenario of climate change in terms of global warming. A high resolution limited area Regional Climate Model (RCM can produce reasonably appropriate projections to be used for climate-scenario generation in country-scale. This paper features the development of future surface temperature projections for Bangladesh on monthly resolution for each year from 2011 to 2100 applying Providing Regional Climates for Impacts Studies (PRECIS, and it explains in detail the modeling processes including the model features, domain size selection, bias identification as well as construction of change field for the concerned climatic variable, in this case, surface temperature. PRECIS was run on a 50 km horizontal grid-spacing under the Intergovernmental Panel on Climate Change (IPCC A1B scenario and it was found to perform reasonably well in simulating future surface temperature of Bangladesh. The linear regression between observed and model simulated results of monthly average temperatures, within the 30-year period from 1971 to 2000, gives a high correlation of 0.93. The applied change field in average annual temperature shows only 0.5 °C–1 °C deviation from the observed values over the period from 2005 to 2008. Eventually, from the projected average temperature change during the years 1971–2000, it is apparent that warming in Bangladesh prevails invariably every month, which might eventually result in an average annual increase of 4 °C by the year 2100. Calculated anomalies in country-average annual temperature mostly remain on the positive side throughout the period of 2071–2100 indicating an overall up-shift. Apart from these quantitative analyses of temporal changes of temperature
15. Fourth order difference methods for hyperbolic IBVP's
Science.gov (United States)
Gustafsson, Bertil; Olsson, Pelle
1994-01-01
Fourth order difference approximations of initial-boundary value problems for hyperbolic partial differential equations are considered. We use the method of lines approach with both explicit and compact implicit difference operators in space. The explicit operator satisfies an energy estimate leading to strict stability. For the implicit operator we develop boundary conditions and give a complete proof of strong stability using the Laplace transform technique. We also present numerical experiments for the linear advection equation and Burgers' equation with discontinuities in the solution or in its derivative. The first equation is used for modeling contact discontinuities in fluid dynamics, the second one for modeling shocks and rarefaction waves. The time discretization is done with a third order Runge-Kutta TVD method. For solutions with discontinuities in the solution itself we add a filter based on second order viscosity. In case of the non-linear Burger's equation we use a flux splitting technique that results in an energy estimate for certain different approximations, in which case also an entropy condition is fulfilled. In particular we shall demonstrate that the unsplit conservative form produces a non-physical shock instead of the physically correct rarefaction wave. In the numerical experiments we compare our fourth order methods with a standard second order one and with a third order TVD-method. The results show that the fourth order methods are the only ones that give good results for all the considered test problems.
16. Art meets mathematics in the fourth dimension
CERN Document Server
Lipscomb, Stephen Leon
2014-01-01
To see objects that live in the fourth dimension we humans would need to add a fourth dimension to our three-dimensional vision. An example of such an object that lives in the fourth dimension is a hyper-sphere or “3-sphere”. The quest to imagine the elusive 3-sphere has deep historical roots: medieval poet Dante Alighieri, in his circa 1300 AD Divine Comedy, used a 3-sphere to convey his allegorical vision of the Christian afterlife. In 1917, Albert Einstein visualized the universe, at each instant in time, as a 3-sphere. He described his representation as “…the place where the reader’s imagination boggles. Nobody can imagine this thing.” Over time, however, our understanding of the concept of dimension evolved. By 2003, a researcher had successfully rendered into human vision the structure of a 4-web (think of an every increasingly-dense spider’s web). In this text Stephen Lipscomb takes his innovative dimension theory research a step further, using the 4-web to reveal a new partial image of a...
17. Al Gore' Inconvenient Truth, where? The IPCC report provides the answer
International Nuclear Information System (INIS)
The climate is changing. Extreme weather conditions occur with greater frequency. In the Netherlands, for example, there have been heat waves, the drought of 2003, exceptional rainstorms and, in 2004, the wettest August in 150 years. The populations of developing countries and vulnerable areas such as the Netherlands will bear the consequences of climate change. The impact on the natural environment will also be severe. Some flora and fauna will extinct while others will proliferate. The outcome will be a strong decline in biodiversity
18. Global estimates of C stock changes in living forest biomass: EDGARv4.3 – 5FL1 time series from 1990 to 2010
Directory of Open Access Journals (Sweden)
A. M. R. Petrescu
2012-03-01
Full Text Available While the Emissions Database for Global Atmospheric Research (EDGAR focuses on global estimates for the full set of anthropogenic activities, the Land-Use, Land-Use Change and Forestry (LULUCF sector might be the most diverse and most challenging to cover consistently for all world countries. Parties to UNFCCC are required to provide periodic estimates of GHG emissions, following the latest approved methodological guidance by the International Panel on Climate Change (IPCC. The aim of the current study is comparing the IPCC GPG 2003 and the IPCC AFOLU 2006 by calculating the C stock changes in living forest biomass, and then using computed results to extend the EDGAR database. For this purpose, we applied the IPCC Tier 1 method at global level, i.e. using spatially coarse activity data (i.e. area, obtained combining two different global forest maps: the Global Land Cover map and the eco-zones subdivision of the GEZ Ecological Zone map in combination with the IPCC default C stocks and C stock change factors. Results for the C stock changes were calculated separately for Gains, Harvest, Net Deforestation and Fires (GFED3, for the years 1990, 2000, 2005 and 2010. At the global level, results obtained with the two set of IPCC guidance differed by about 40%, due to different assumptions and default factors. The IPCC Tier 1 method unavoidably introduced high uncertainties due to the "globalization" of parameters. When the results using IPCC AFOLU 2006 for Annex I countries are compared to other international datasets (UNFCCC, FAO or scientific publications, it emerges a significant overestimation of the sink. For developing countries, we conclude that C stock change in forest remaining forest can hardly be estimated with Tier 1 method. Overall, confronting the IPCC 2003 and 2006 methodologies we conclude that IPCC 2006 suits best the needs of EDGAR and provide a consistent global picture of C stock changes in living forest biomass independent of
19. Global estimates of C stock changes in living forest biomass: EDGARv4.3 - 5FL1 time series from 1990 to 2010
Science.gov (United States)
Petrescu, A. M. R.; Abad-Viñas, R.; Janssens-Maenhout, G.; Blujdea, V.; Grassi, G.
2012-03-01
While the Emissions Database for Global Atmospheric Research (EDGAR) focuses on global estimates for the full set of anthropogenic activities, the Land-Use, Land-Use Change and Forestry (LULUCF) sector might be the most diverse and most challenging to cover consistently for all world countries. Parties to UNFCCC are required to provide periodic estimates of GHG emissions, following the latest approved methodological guidance by the International Panel on Climate Change (IPCC). The aim of the current study is comparing the IPCC GPG 2003 and the IPCC AFOLU 2006 by calculating the C stock changes in living forest biomass, and then using computed results to extend the EDGAR database. For this purpose, we applied the IPCC Tier 1 method at global level, i.e. using spatially coarse activity data (i.e. area, obtained combining two different global forest maps: the Global Land Cover map and the eco-zones subdivision of the GEZ Ecological Zone map) in combination with the IPCC default C stocks and C stock change factors. Results for the C stock changes were calculated separately for Gains, Harvest, Net Deforestation and Fires (GFED3), for the years 1990, 2000, 2005 and 2010. At the global level, results obtained with the two set of IPCC guidance differed by about 40%, due to different assumptions and default factors. The IPCC Tier 1 method unavoidably introduced high uncertainties due to the "globalization" of parameters. When the results using IPCC AFOLU 2006 for Annex I countries are compared to other international datasets (UNFCCC, FAO) or scientific publications, it emerges a significant overestimation of the sink. For developing countries, we conclude that C stock change in forest remaining forest can hardly be estimated with Tier 1 method. Overall, confronting the IPCC 2003 and 2006 methodologies we conclude that IPCC 2006 suits best the needs of EDGAR and provide a consistent global picture of C stock changes in living forest biomass independent of country estimates.
20. Greenhouse Gas Emissions in the Netherlands: summary Report 1990-1997 (IPCC Tables 7A)
NARCIS (Netherlands)
Olivier JGJ; Spakman J; Berg JC van den; LAE
1999-01-01
This report documents the Netherlands' annual greenhouse gas emission inventory submitted for 1998 according to the European Union's Greenhouse Gas Monitoring Mechanism and the United Nation's Framework Convention on Climate Change (FCCC). Emissions of CO2 and N2O increased from 1990 to 1997 (with a
1. Greenhouse Gas Emissions in the Netherlands: Summary Report 1990-1998 (IPCC Tables 1-7)
NARCIS (Netherlands)
Olivier JGJ; Berg JC van den; Peters JAHW; LAE
2000-01-01
This report documents the 1998 Netherlands' annual submission of its greenhouse gas emission inventory according to the European Union's Greenhouse Gas Monitoring Mechanism and to the United Nation's Framework Convention on Climate Change (FCCC). From the inventories it can be concluded that emissio
2. The impact of climate change on the drought variability over Australia
Science.gov (United States)
Kirono, D. G. C.; Hennessy, K.; Mpelasoka, F.; Bathols, J.; Kent, D.
2009-04-01
of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change (eds. Solomon, S. et al.). Cambridge University Press, Cambridge, United Kingdom and New York, NY, USA, www.ipcc.ch
3. Rice production simulation under IPCC SRES A1B scenario in Fuiian Province%基于IPCC SRES A1B情景下的福建省水稻生产模拟研究
Institute of Scientific and Technical Information of China (English)
江敏; 金之庆; 杨慧; 朱朝枝; 林文雄
2012-01-01
region in Northwest Fujian, single cropping rice regions in the mountain areas of Northwest Fujian. 17 sampling sites and 9 representative rice varieties from the three regions were studied. Daily weather data for 2006—2007 and regional test yield data for the period were used to initialize CERES-Rice model parameters. Based on Special Report on Emission Scenario (SRES) on A1B scenario of IPCC, data from the PRECIS were integrated with CERES-Rice model to predict the influence of climate change on rice production in Fujian Province in the 2020s and 2040s. The simulation considered enhanced CO2-fertilization effects and two cultivation modes under rain-fed and irrigated conditions. The results suggested that irrespective of rain-fed or irrigated rice, future growth durations shortened for three rice cropping regions.Growth duration of single cropping rice shortened the most (over 20 days) under the 2040s scenario. Future yields for single cropping rice and early rice dropped compared with yields under the baseline weather conditions. Yields for rain-fed and irrigated early rice in double cropping rice region in Southeast Fujian dropped by 12.4% and 11.3%, respectively, while that for early rice in double cropping rice region in Northwest Fujian was minimal under the 2020s scenario. Because of negative contributions of mid-season varieties, yields for rain-fed and irrigated single cropping rice region in mountain regions of Northwest Fujian dropped by 7.1% and 2.1%, respectively. Reductions in output of all the treatments increased under the 2040s scenario. On the contrary, yields of late rice over the same period were enhanced even though yield stability was the worst. Under the 2020s scenario, increasing yield rate of 21.0% of later rice in irrigated double cropping rice region in Northwest Fujian and 10.6% in Southeast Fujian were noted, but slight yield increases for rain-fed these rice regions. The amplitude of rice yield increase dropped for all rice regions under
4. Implications from the climatic change dynamics for research and development concerning renewable energies in Germany; Implikationen aus der Dynamik des Klimawandels fuer Forschung und Entwicklung erneuerbarer Energien in Deutschland
Energy Technology Data Exchange (ETDEWEB)
Lewerenz, Jana Celeste
2009-03-15
The climatic change is a fact. The contribution of the fourth IPCC (international panel on climate change) report is covered in the first chapter of the volume, describing the increase of anthropogenic greenhouse gas production during the last centuries, possible mitigation strategies, unavoidable consequences, macroeconomic costs for stabilizing the greenhouse gas emissions and possibilities of renewable energy resources. The second chapter is the Stern report 2006 on the existing scientific literature concerning the climatic change and its consequences, risk analyses and cost estimations for greenhouse gas reduction strategies on the one hand and costs of the possibly catastrophic consequences of the climate change on the other hand. The third chapter is the pilot study of the Federal Ministry of environment, nature conservation and nuclear safety 2007/2008. The fourth chapter is the contribution of the scientific commission of the Federal government on global environmental changes (WBGU) no 5: new stimuli for the climate policy: chances of the German double presidency. The fifth chapter covers the Meseberg topics - the integrated energy and climate program (IEKP) followed by a discussion of the described contributions.
5. IPCC气候情景下全球海平面长期趋势变化%Long term trends in global sea level under IPCC SRES A2 scenario
Institute of Scientific and Technical Information of China (English)
陈长霖; 左军成; 杜凌; 何倩倩
2012-01-01
利用CCSM3(Community Climate System Model version 3)气候系统模式模拟20世纪海平面变化,在IPCC SRES A2(IPCC,2001)情景假设下预测21世纪全球海平面长期趋势变化.模拟显示20世纪海平面上升约4.0 cm,且存在0.0048 mm/a2的加速度,这个结果仅为热盐比容的贡献.在A2情景假设下,21世纪海平面上升存在很大的区域特征,呈纬向带状分布;总体上北冰洋上升大,南大洋高纬度海区上升小,大西洋上升值比太平洋的大;整个21世纪全球平均比容海平面上升了约30 cm,且呈加速上升的趋势.同时发现,中深层水温度和盐度变化对区域比容海平面变化具有重要贡献.北太平洋增暖主要集中在上层700m以内,而北大西洋的增暖可达2 500 m的深度,南大洋南极绕极流海区热盐变化则是发生在整个深度.%Based on CCSM3 (Community Climate System Model version 3) model, the long-term sea level trends of global ocean in the 20th and 21st century under IPCC SRES A2 scenario (IPCC, 2001) are ana-lyzed in this paper. The result shows that global sea level rises about 4. 0 cm during 20th century through steric expansion, with an acceleration of 0. 004 8 mm/a2; eustatic changes are not included in this simulation. CCSM3 simulation also shows that in the 21st century sea level will rise in acceleration; the global sea level will rise 30 cm during the whole century through steric expansion. The vertical distribution of thermosteric and halosteric anomalies that contribute to sea level change is very different between ocean basins. Significant warming of the North Atlantic Ocean can extend to 2 500 m depth, while the salinity change partially counteracts sea level rise due to this warming. The steric anomalies in the Pacific Ocean occur mainly in the upper 700 m. In the Southern Ocean, steric change can extend to the entire water column.
6. Schneider lecture: From climate change impacts to climate change risks
Science.gov (United States)
Field, C. B.
2014-12-01
Steve Schneider was a strong proponent of considering the entire range of possible climate-change outcomes. He wrote and spoke frequently about the importance of low probability/high consequence outcomes as well as most likely outcomes. He worked tirelessly on communicating the risks from overlapping stressors. Technical and conceptual issues have made it difficult for Steve's vision to reach maturity in mainstream climate-change research, but the picture is changing rapidly. The concept of climate-change risk, considering both probability and consequence, is central to the recently completed IPCC Fifth Assessment Report, and the concept frames much of the discussion about future research agendas. Framing climate change as a challenge in managing risks is important for five core reasons. First, conceptualizing the issue as being about probabilities builds a bridge between current climate variability and future climate change. Second, a formulation based on risks highlights the fact that climate impacts occur primarily in extremes. For historical variability and future impacts, the real concern is the conditions under which things break and systems fail, namely, in the extremes. Third, framing the challenge as one of managing risks puts a strong emphasis on exploring the full range of possible outcomes, including low-probability, high/consequence outcomes. Fourth, explaining climate change as a problem in managing risks links climate change to a wide range of sophisticated risk management tools and strategies that underpin much of modern society. Fifth, the concept of climate change as a challenge in managing risks helps cement the understanding that climate change is a threat multiplier, adding new dimensions and complexity to existing and emerging problems. Framing climate change as a challenge in managing risks creates an important but difficult agenda for research. The emphasis needs to shift from most likely outcomes to most risky outcomes, considering the full
7. Regional Climate Change Hotspots over Africa
Science.gov (United States)
Anber, U.; Zakey, A.; Abd El Wahab, M.
2009-04-01
Regional Climate Change Index (RCCI), is developed based on regional mean precipitation change, mean surface air temperature change, and change in precipitation and temperature interannual variability. The RCCI is a comparative index designed to identify the most responsive regions to climate change, or Hot- Spots. The RCCI is calculated for Seven land regions over North Africa and Arabian region from the latest set of climate change projections by 14 global climates for the A1B, A2 and B1 IPCC emission scenarios. The concept of climate change can be approaches from the viewpoint of vulnerability or from that of climate response. In the former case a Hot-Spot can be defined as a region for which potential climate change impacts on the environment or different activity sectors can be particularly pronounced. In the other case, a Hot-Spot can be defined as a region whose climate is especially responsive to global change. In particular, the characterization of climate change response-based Hot-Spot can provide key information to identify and investigate climate change Hot-Spots based on results from multi-model ensemble of climate change simulations performed by modeling groups from around the world as contributions to the Fourth Assessment Report of Intergovernmental Panel on Climate Change (IPCC). A Regional Climate Change Index (RCCI) is defined based on four variables: change in regional mean surface air temperature relative to the global average temperature change ( or Regional Warming Amplification Factor, RWAF ), change in mean regional precipitation (P % , of present day value ), change in regional surface air temperature interannual variability (T % ,of present day value), change in regional precipitation interannual variability (P % ,of present day value ). In the definition of the RCCI it is important to include quantities other than mean change because often mean changes are not the only important factors for specific impacts. We thus also include inter
8. Geometry, relativity and the fourth dimension
CERN Document Server
Rucker, Rudolf
1977-01-01
This is a highly readable, popular exposition of the fourth dimension and the structure of the universe. A remarkable pictorial discussion of the curved space-time we call home, it achieves even greater impact through the use of 141 excellent illustrations. This is the first sustained visual account of many important topics in relativity theory that up till now have only been treated separately.Finding a perfect analogy in the situation of the geometrical characters in Flatland, Professor Rucker continues the adventures of the two-dimensional world visited by a three-dimensional being to expl
9. Fourth international seminar on horizontal steam generators
Energy Technology Data Exchange (ETDEWEB)
Tuomisto, H. [ed.] [IVO Group, Vantaa (Finland); Purhonen, H. [ed.] [VTT, Espoo (Finland); Kouhia, V. [ed.] [Lappeenranta Univ. of Technology (Finland)
1997-12-31
The general objective of the International Seminars of Horizontal Steam Generator Modelling has been the improvement in understanding of realistic thermal hydraulic behaviour of the generators when performing safety analyses for VVER reactors. The main topics presented in the fourth seminar were: thermal hydraulic experiments and analyses, primary collector integrity, feedwater distributor replacement, management of primary-to-secondary leakage accidents and new developments in the VVER safety technology. The number of participants, representing designers and manufacturers of the horizontal steam generators, plant operators, engineering companies, research organizations, universities and regulatory authorities, was 70 from 10 countries.
10. The fourth pattern of attachment: Disorganized / disoriented
Directory of Open Access Journals (Sweden)
Zlatka Cugmas
2003-09-01
Full Text Available On the basis of the study of recent scientific literature about the development of attachment behavior, the author answers the following questions about: the behavior children categorized as Disorganized/disorientated or Controlling in the procedure of the Strange situation; the life circumstances, in which these children live; the reasons for lack of balanced strategies of attachment and characteristics of their general manner of adaptation. The author finds the characteristics of the mothers' (insensitivity to be significantly influential for the emergence of the fourth pattern of attachment. These children are heterogeneous regarding adaptation in general. Professional help preceding negative consequences for their socioemotional development is neccesary.
11. Fourth international seminar on horizontal steam generators
International Nuclear Information System (INIS)
The general objective of the International Seminars of Horizontal Steam Generator Modelling has been the improvement in understanding of realistic thermal hydraulic behaviour of the generators when performing safety analyses for VVER reactors. The main topics presented in the fourth seminar were: thermal hydraulic experiments and analyses, primary collector integrity, feedwater distributor replacement, management of primary-to-secondary leakage accidents and new developments in the VVER safety technology. The number of participants, representing designers and manufacturers of the horizontal steam generators, plant operators, engineering companies, research organizations, universities and regulatory authorities, was 70 from 10 countries
12. Variability of the Atlantic meridional overturning circulation in the last millennium and two IPCC scenarios
Energy Technology Data Exchange (ETDEWEB)
Ortega, Pablo; Montoya, Marisa; Gonzalez-Rouco, Fidel [Universidad Complutense de Madrid, Ciudad Universitaria, Dpto. Astrofisica y Ciencias de la Atmosfera/Instituto de Geociencias, Facultad de Ciencias Fisicas, Madrid (Spain); Universidad Complutense de Madrid, Ciudad Universitaria, Instituto de Geociencias (UCM-CSIC), Facultad de Ciencias Fisicas, Madrid (Spain); Mignot, Juliette [IPSL/LOCEAN, UPMC/CNRS/IRD/MNHN, Universite Pierre et Marie Curie, Paris Cedex 05 (France); Legutke, Stephanie [Deutsches Klimarechenzentrum (DKRZ), Hamburg (Germany)
2012-05-15
The variability of the Atlantic meridional overturning circulation (AMOC) is investigated in several climate simulations with the ECHO-G atmosphere-ocean general circulation model, including two forced integrations of the last millennium, one millennial-long control run, and two future scenario simulations of the twenty-first century. This constitutes a new framework in which the AMOC response to future climate change conditions is addressed in the context of both its past evolution and its natural variability. The main mechanisms responsible for the AMOC variability at interannual and multidecadal time scales are described. At high frequencies, the AMOC is directly responding to local changes in the Ekman transport, associated with three modes of climate variability: El Nino-Southern Oscillation (ENSO), the North Atlantic Oscillation (NAO), and the East Atlantic (EA) pattern. At low frequencies, the AMOC is largely controlled by convection activity south of Greenland. Again, the atmosphere is found to play a leading role in these variations. Positive anomalies of convection are preceded in 1 year by intensified zonal winds, associated in the forced runs to a positive NAO-like pattern. Finally, the sensitivity of the AMOC to three different forcing factors is investigated. The major impact is associated with increasing greenhouse gases, given their strong and persistent radiative forcing. Starting in the Industrial Era and continuing in the future scenarios, the AMOC experiences a final decrease of up to 40% with respect to the preindustrial average. Also, a weak but significant AMOC strengthening is found in response to the major volcanic eruptions, which produce colder and saltier surface conditions over the main convection regions. In contrast, no meaningful impact of the solar forcing on the AMOC is observed. Indeed, solar irradiance only affects convection in the Nordic Seas, with a marginal contribution to the AMOC variability in the ECHO-G runs. (orig.)
13. Field Studies Show That In Situ Greenhouse Gas Emission Factors for East African Agriculture Are Less Than IPCC Values
Science.gov (United States)
Pelster, D.; Butterbach-Bahl, K.; Rufino, M.; Rosenstock, T. S.; Wanyama, G.
2015-12-01
Greenhouse gas (GHG) emissions from African agricultural systems are thought to comprise a large portion of total emissions from the continent, however these estimates have been calculated using emission factors (EF) from other regions due to the lack of field studies in Africa, which results in large uncertainties for these estimates. Field measurements from western Kenya calculating emissions over a year in 59 different sites found that GHG emissions from typical smallholder farms ranged from 2.8 to 15.0 Mg CO2-C ha-1, -6.0 to 2.4 kg CH4-C ha-1 and -0.1 to 1.8 kg N2O-N ha-1, and were not affected by management intensity. The lack of a response in N2O emissions to N fertilization suggests that the EF currently used in national inventories overestimates N2O emissions from typical smallholder agriculture. Another study measuring N2O and CH4 emissions from manure deposited by grazing cattle found that the N2O EF ranged from 0.1 to 0.2%, while the CH4 EF ranged from 0.04 to 0.14 Kg CH4-C per 173 kg animal. These suggest that the current IPCC EF overestimate agricultural soil and manure GHG emissions for Kenya, and likely for much of East Africa.
14. How good are the simulations of tropical SST–rainfall relationship by IPCC AR4 atmospheric and coupled models?
Indian Academy of Sciences (India)
K Rajendran; Ravi S Nanjundiah; Sulochana Gadgil; J Srinivasan
2012-06-01
The failure of atmospheric general circulation models (AGCMs) forced by prescribed SST to simulate and predict the interannual variability of Indian/Asian monsoon has been widely attributed to their inability to reproduce the actual sea surface temperature (SST)–rainfall relationship in the warm Indo-Pacific oceans. This assessment is based on a comparison of the observed and simulated correlation between the rainfall and local SST. However, the observed SSTconvection/rainfall relationship is nonlinear and for this a linear measure such as the correlation is not an appropriate measure. We show that the SST–rainfall relationship simulated by atmospheric and coupled general circulation models in IPCC AR4 is nonlinear, as observed, and realistic over the tropical West Pacific (WPO) and the Indian Ocean (IO). The SST–rainfall pattern simulated by the coupled versions of these models is rather similar to that from the corresponding atmospheric one, except for a shift of the entire pattern to colder/warmer SSTs when there is a cold/warm bias in the coupled version.
15. Testing an astronomically-based decadal-scale empirical harmonic climate model versus the IPCC (2007) general circulation climate models
CERN Document Server
Scafetta, Nicola
2012-01-01
We compare the performance of a recently proposed empirical climate model based on astronomical harmonics against all available general circulation climate models (GCM) used by the IPCC (2007) to interpret the 20th century global surface temperature. The proposed model assumes that the climate is resonating with, or synchronized to a set of natural harmonics that have been associated to the solar system planetary motion, mostly determined by Jupiter and Saturn. We show that the GCMs fail to reproduce the major decadal and multidecadal oscillations found in the global surface temperature record from 1850 to 2011. On the contrary, the proposed harmonic model is found to well reconstruct the observed climate oscillations from 1850 to 2011, and it is able to forecast the climate oscillations from 1950 to 2011 using the data covering the period 1850-1950, and vice versa. The 9.1-year cycle is shown to be likely related to a decadal Soli/Lunar tidal oscillation, while the 10-10.5, 20-21 and 60-62 year cycles are sy...
16. Predicting fire activity in the US over the next 50 years using new IPCC climate projections
Science.gov (United States)
Wang, D.; Morton, D. C.; Collatz, G. J.
2012-12-01
Fire is an integral part of the Earth system with both direct and indirect effects on terrestrial ecosystems, the atmosphere, and human societies (Bowman et al. 2009). Climate conditions regulate fire activities through a variety of ways, e.g., influencing the conditions for ignition and fire spread, changing vegetation growth and decay and thus the accumulation of fuels for combustion (Arora and Boer 2005). Our recent study disclosed the burned area (BA) in US is strongly correlated with potential evaporation (PE), a measurement of climatic dryness derived from National Centers for Environmental Prediction (NCEP) North American Regional Reanalysis (NARR) climate data (Morton et al. 2012). The correlation varies spatially and temporally. With regard to fire of peak fire seasons, Northwestern US, Great Plains and Alaska have the strongest BA/PE relationship. Using the recently released the Global Fire Emissions Database (GFED) Version 3 (van der Werf et al. 2010), we showed increasing BA in the last decade in most of NCA regions. Longer time series of Monitoring Trends in Burn Severity (MTBS) (Eidenshink et al. 2007) data showed the increasing trends occurred in all NCA regions from 1984 to 2010. This relationship between BA and PE provides us the basis to predict the future fire activities in the projected climate conditions. In this study, we build spatially explicit predictors using the historic PE/BA relationship. PE from 2011 to 2060 is calculated from the Coupled Model Intercomparison Project Phase 5 (CMIP5) data and the historic PE/BA relationship is then used to estimate BA. This study examines the spatial pattern and temporal dynamics of the future US fires driven by new climate predictions for the next 50 years. Reference: Arora, V.K., & Boer, G.J. (2005). Fire as an interactive component of dynamic vegetation models. Journal of Geophysical Research-Biogeosciences, 110 Bowman, D.M.J.S., Balch, J.K., Artaxo, P., Bond, W.J., Carlson, J.M., Cochrane, M.A., D
17. IPCC 排放因子法探究 CO2排放量%CO 2 emission of Zhejiang and Guizhou province by the method of emission factor(IPCC)from 2005 to 2012
Institute of Scientific and Technical Information of China (English)
杜赟
2015-01-01
this research explored the CO2 emissions from fossil fuels in Zhejiang and Guizhou province within the time period from 2005 to 2012 by IPCC emission factor method. In addition,CO2 / GDP was also investigated to discuss the relationship between CO2 emission and economic development. The outcomes showed,CO2 emissions from the two areas have been increasing linearly over the past 8 years. However,CO2 / GDP displayed descending trends in the two provinces. This indicates that, with technical innovation and efficiency improvement,fossil fuels consumption is on its way to sus-tainable development. Generally speaking,the economic status of Zhejiang is much better than that of Guizhou. This is not only reflected on its higher economic volume,but also due to its lower CO2 /GDP. In addition,Guizhou has also made great progress in its sustainable economic development, reflected by its rising GDP and decreasing CO2 / GDP during this period. However,huge space for its further improvement still remains in terms of sustainable development.%本课题收集了浙江省和贵州省2005—2012年的化石能源消耗量相关数据,运用 IPCC 排放因子法计算了两省的 CO2排放情况,并结合两省经济数据,考查了单位 GDP CO2排放量。结果显示,两省 CO2排放量均为线性增长趋势,而两省的单位 GDP CO2排放量却逐渐降低。说明随着经济的发展,生产技术的改革,效率的提高,化石能源的消耗逐步迈入了可持续发展的道路。综合看来,浙江省经济的发展要优于贵州省,这不仅体现在经济总量远远高于贵州省,而且其单位 GDP CO2排放量低于贵州省。另外,贵州省经济状况虽然落后,但是从其不断上升的 GDP 总量和不断下降的单位 GDP CO2排放量可知,其经济发展处于不断进步阶段。贵州省在可持续经济发展道路上取得的成绩是毋庸置疑的,继续提升的空间也是巨大的。
18. Aerosol Direct, Indirect, Semidirect, and Surface Albedo Effects from Sector Contributions Based on the IPCC AR5 Emissions for Preindustrial and Present-day Conditions
Science.gov (United States)
Bauer, Susanne E.; Menon, Surabi
2012-01-01
The anthropogenic increase in aerosol concentrations since preindustrial times and its net cooling effect on the atmosphere is thought to mask some of the greenhouse gas-induced warming. Although the overall effect of aerosols on solar radiation and clouds is most certainly negative, some individual forcing agents and feedbacks have positive forcing effects. Recent studies have tried to identify some of those positive forcing agents and their individual emission sectors, with the hope that mitigation policies could be developed to target those emitters. Understanding the net effect of multisource emitting sectors and the involved cloud feedbacks is very challenging, and this paper will clarify forcing and feedback effects by separating direct, indirect, semidirect and surface albedo effects due to aerosols. To this end, we apply the Goddard Institute for Space Studies climate model including detailed aerosol microphysics to examine aerosol impacts on climate by isolating single emission sector contributions as given by the Coupled Model Intercomparison Project Phase 5 (CMIP5) emission data sets developed for Intergovernmental Panel on Climate Change (IPCC) AR5. For the modeled past 150 years, using the climate model and emissions from preindustrial times to present-day, the total global annual mean aerosol radiative forcing is -0.6 W/m(exp 2), with the largest contribution from the direct effect (-0.5 W/m(exp 2)). Aerosol-induced changes on cloud cover often depends on cloud type and geographical region. The indirect (includes only the cloud albedo effect with -0.17 W/m(exp 2)) and semidirect effects (-0.10 W/m(exp 2)) can be isolated on a regional scale, and they often have opposing forcing effects, leading to overall small forcing effects on a global scale. Although the surface albedo effects from aerosols are small (0.016 W/m(exp 2)), triggered feedbacks on top of the atmosphere (TOA) radiative forcing can be 10 times larger. Our results point out that each
19. Coffee Beverage Quality Assessment Based on ETA/CPTEC-HadCM3 Model (A1B-IPCC/SRES Scenario), Southeastern Brazil
Science.gov (United States)
Giarolla, A.; Resende, N.; Chou, S. C.; Tavares, P. S.; Rodrigues, D. C.
2012-04-01
Environmental factors influence the coffee beverage quality and air temperature has a significant importance in this process. The grain maturation occurs very quickly in regions that present high temperatures and sometimes there is not enough time to complete all this phase adequately. In the other hand, with mild temperatures, the grain maturation occurs more slowly and it promotes a better quality beverage. The aim of this study was to assess the coffee beverage quality in the southeastern Brazil, based on climate projections using the Eta-CPTEC regional model driven by four members of an ensemble of the Met Office Hadley Centre Global Coupled climate model (HadCM3). The global model ensemble was run over the 21st century according to IPCC SRES, A1B emissions scenario. Each ensemble member presented different climate sensitivity in the analysis. The Eta-CPTEC-HadCM3 model was configured with a 40-km grid size and was run over the period of 1961-90 to represent a baseline climate, and over the period of 2011-2100 to simulate possible future changes and the effects on the coffee beverage quality. A coffee beverage quality classification, which depends on the annual air temperature proposed by Bressani (2007) and also, a quality coffee beverage sensory classification, based on Camargo and Cortez (1998) were considered in this study. An evaluation of the systematic errors (BIAS) for each member for the period from 1961 to 1990 was made. The results presented by Eta/CPTEC-HadCM3 model indicated that in the case of an occurrence of A1B emission scenario, the coffee beverage quality could be affected in this region due to the fact that the flavor may become stronger and unpleasant caused by rising air temperatures. The BIAS evaluation and subsequent errors removal demonstrated improvement in the scenarios simulations. A short review concerning agronomic techniques to mitigate extreme meteorological events or global warming on coffee crop based on Camargo (2010) also is
20. Climate change negotiations. COP-2 and beyond
Energy Technology Data Exchange (ETDEWEB)
NONE
1996-09-01
The Second Conference of the Parties to the UN Framework Convention on Climate Change (COP-2), which met in Geneva during July, 1996, was only a partial success when considered in relation to its avowed aims, gaining acceptance of the Second Assessment Report by IPCC (Intergovernmental Panel on Climate Change), producing an agreed Ministerial Declaration, making real advances towards a protocol, and agreeing Rules of Procedure. This paper describes the main aims of COP-2, consideration of and response to the IPCCs Second Assessment Report, the COP-2 Ministerial Declaration, some significant statements by individual country delegations at COP-2, lack of progress on Rules of Procedure for the Conference, realization of returning the greenhouse gas emissions in industrialized countries based on the Montreal Protocol, differing views among countries to the Convention on a protocol, prospects for achieving agreement on a legally binding protocol at COP-3 planned for Kyoto, Japan in December, 1997, and recent scientific and technical findings.
1. Fourth International Conference on Complex Systems
CERN Document Server
Minai, Ali A; Unifying Themes in Complex Systems IV
2008-01-01
In June of 2002, over 500 professors, students and researchers met in Boston, Massachusetts for the Fourth International Conference on Complex Systems. The attendees represented a remarkably diverse collection of fields: biology, ecology, physics, engineering, computer science, economics, psychology and sociology, The goal of the conference was to encourage cross-fertilization between the many disciplines represented and to deepen understanding of the properties common to all complex systems. This volume contains 43 papers selected from the more than 200 presented at the conference. Topics include: cellular automata, neurology, evolution, computer science, network dynamics, and urban planning. About NECSI: For over 10 years, The New England Complex Systems Institute (NECSI) has been instrumental in the development of complex systems science and its applications. NECSI conducts research, education, knowledge dissemination, and community development around the world for the promotion of the study of complex sys...
2. Fourth harmonic generation of DKDP crystal
International Nuclear Information System (INIS)
Deuterated dihydrogen phosphate (DKDP) crystals of different deuterium contents were cut for type-Ⅰ non-critical phase-matching with the direction at 90° to the crystal z axis and at 45° to the crystal x axis, and then measured in the fourth harmonic generation experiments with fundamental wavelengths of 1 064 nm and 1 053 nm. respectively. The optimum deuterium content of DKDP crystal was confirmed by measuring the relation between deuterium content and phase-matching angle, The results indicate that DKDP crystal can not achieve non-critical phase matching by adjusting its deuterium content at room temperature with the fundamental wavelength of 1 064 nm. However, at the fundamental wavelength of 1053 nm. the optimum deuterium content of DKDP crystal is 85%, to achieve non-critical phase matching at room temperature. (authors)
3. Fourth Thematic CERN School of Computing
CERN Multimedia
Alberto Pace, CSC Director
2016-01-01
The Fourth Thematic School of Computing (tCSC2016) takes place this year in Split, Croatia, from 22 to 28 May 2016. The theme is "Efficient and Parallel Processing of Scientific Data", looking at: The challenge of scientific data processing: commonalities, analogies and the main differences between different sciences. Size of scientific software projects. Parallelism and asynchronism: computation and I/O. The School is open to postgraduate students and research workers with a few years' experience in elementary particle physics, computing, engineering or related fields. All applicants are welcome, including former and future participants in the main CSC summer school. Registration will close on 15 February and participation is limited to 24 students. To register, please go here. About: The Thematic Schools are part of the annual series of CERN Schools of Computing, to promote advanced learning and knowledge exchange on the subject of scientific compu...
4. Fourth ventricle meningiomas: a rare entity.
Science.gov (United States)
Pichierri, Angelo; Ruggeri, Andrea; Morselli, Carlotta; Delfini, Roberto
2011-08-01
Fourth ventricle meningiomas (FVMs) are rare, often misdiagnosed, lesions. To the best of our knowledge, 47 cases have been reported in the literature: we describe our series of three cases treated at our Institution, focusing on some diagnostic tips and intraoperative features of these tumours. Our three patients have a history of headache. Gait disturbances, vomiting and/or diplopia complicated the clinical picture before the referral at our Department. The operations were uneventful, and the patients fully recovered from neurological symptoms. They are free of recurrence at a median follow-up of 19 years. FVMs are rare lesions, which are difficult to differentiate preoperatively from the much more common ependymomas. A preoperative distinction would be extremely advantageous: indeed, although both tumours share similar radiological and clinical patterns, they clearly differ as to surgical difficulty and outcome. In fact, meningiomas are comparatively easier to remove, granting better clinical results.
5. Fourth World Theory: The Evolution of . . .
Directory of Open Access Journals (Sweden)
Olon F. Dotson
2014-05-01
Full Text Available Fourth World theory is a methodology for examining and developing greater understanding of the extent of the distress and abandonment commonly found in the cores of American cities resulting from de-industrialization, historic segregation and discrimination patterns, suburban sprawl, erosion of a viable tax base, racism, inability to embrace the concept of desegregation and civil rights legislation, fear, despair, crumbling infrastructure systems, disinvestment in urban school systems, and environmental justice issues. This article uses the analytical lens of Fourth World theory to examine how such structural and cultural forces contributed to the severely distressed conditions now found in the city of Gary, Indiana. Tracking its one-hundred-year history, from its founding as an industrial town through its post-industrial decline occurring during the city’s first African-American mayor’s five terms in office, the methodology clearly demonstrates how the social construction of race has systematically undermined every aspect of Gary’s overall quality of life. To illustrate that this city is not an anomaly but rather reflects a typical pattern of disparity and uneven development arising from racist practices, Gary is compared to other cities of similar size and also to the much larger Detroit. The article triangulates academic literature, news media archives, and an oral history provided by the mayor to show how Gary evolved from being a model industrial city to a cauldron of racial disparity. The paper concludes by arguing that continued absence of reflection on the nation’s historical racialization of place threatens not just impoverished communities of color, but also the sustainability of the entire nation.
6. The fourth revolution how the infosphere is reshaping human reality
CERN Document Server
Floridi, Luciano
2014-01-01
Who are we, and how do we relate to each other? Luciano Floridi, one of the leading figures in contemporary philosophy, argues that the explosive developments in Information and Communication Technologies (ICTs) is changing the answer to these fundamental human questions. As the boundaries between life online and offline break down, and we become seamlessly connected to each other and surrounded by smart, responsive objects, we are all becoming integrated into an ". Personas we adopt in social media, for example, feed into our 'real' lives so that we begin to live, as Floridi puts in, ". Following those led by Copernicus, Darwin, and Freud, this metaphysical shift represents nothing less than a fourth revolution. " defines more and more of our daily activity - the way we shop, work, learn, care for our health, entertain ourselves, conduct our relationships; the way we interact with the worlds of law, finance, and politics; even the way we conduct war. In every department of life, ICTs have become environmenta...
7. Effects of an Intergenerational Friendly Visit Program on the Attitudes of Fourth Graders toward Elders.
Science.gov (United States)
Schwalbach, Eileen; Kiernan, Sharon
2002-01-01
Instruction on aging and stereotypes was followed by weekly visits by 22 fourth-graders to elderly nursing home residents. Students' attitudes toward elders were mixed, but their attitudes toward their own aging changed positively and empathy increased. (Contains 24 references.) (SK)
8. Developing Cross-Cultural Competence: A Guide for Working with Children and Their Families. Fourth Edition
Science.gov (United States)
Lynch, Eleanor W., Ed.; Hanson, Marci J., Ed.
2011-01-01
As the U.S. population grows more and more diverse, how can professionals who work with young children and families deliver the best services while honoring different customs, beliefs, and values? The answers are in the fourth edition of this bestselling textbook, fully revised to reflect nearly a decade of population changes and best practices in…
9. Climate change. Scientific background and process
Energy Technology Data Exchange (ETDEWEB)
Alfsen, Knut H.; Fuglestvedt, Jan; Seip, Hans Martin; Skodvin, Tora
1999-07-01
The paper describes briefly the natural and man-made forces behind climate change and outlines climate variations in the past. It also discusses the future impact of anthropogenic emission of greenhouse gases, and the background, organisation and functioning of the Intergovernmental Panel on Climate Change (IPCC)
10. Development of a risk database for the establishment of invasive mosquito species under impacts of climate change
Science.gov (United States)
Tagaris, Efthimios; Sotiropoulou, Rafaella-Eleni; Sotiropoulos, Andreas; Spanos, Ioannis; Milonas, Panayiotis; Michaelakis, Antonios
2016-04-01
Climate models suggest changes in future temperature and precipitation rates, the main climatic parameters that are related to the suitability of a region for the establishment and seasonal abundance of the Invasive Mosquito Species (IMS). In this work the future potentiality of IMS spread and establishment over Greece and Italy is assessed following four steps. In the first step current Spatial Risk Databases for the establishment of IMS over Greece and Italy are developed using the meteorological parameters from the ECA&D project. In the second step changes in the climatic parameters in 2050's are estimated using the NASA GISS GCM ModelE under the IPCC-A1B emissions scenarios. In the third step, the mesoscale meteorological model WRF is used, to simulate the changes in the meteorological fields caused by climate change in a finer grid size using dynamical regional downscaling. Finally, in the fourth step the estimated changes in the meteorological parameters from step three are combined with the observation data from step one in order to estimate the future level of the climatic parameters of interest. The final product is spatial distribution maps presenting the future suitability of a region for the establishment and seasonal abundance of the IMS over Greece and Italy. Acknowledgement: LIFE CONOPS project "Development & demonstration of management plans against - the climate change enhanced - invasive mosquitoes in S. Europe" (LIFE12 ENV/GR/000466).
11. Biogeochemical responses of shallow coastal lagoons to Climate Change
Science.gov (United States)
Brito, A.; Newton, A.; Tett, P.; Fernandes, T.
2009-04-01
carefully monitored so that appropriate responses can be timely to mitigate the impacts from global change. References: Eisenreich, S.J. (2005). Climate Change and the European Water Dimension - A report to the European Water Directors. Institute for Environment and Sustainability, European Comission-Joint Research Centre. Ispra, Italy. 253pp. Kerr, R. (2008). Global warming throws some curves in the Atlantic Ocean. Science, 322, 515. IPCC (2007). Climate Change 2007: The Physical Science Basis. Contribution of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change [Solomon, S., Qin, D., Manning, M., Chen, Z., Marquis, M., Averyt, K., Tignor, M., Miller, H. (eds.)]. Cambridge University Press. Cambridge, United Kingdom and New York, NY, USA, 996pp. Lloret, J., Marín, A., Marín-Guirao, L. (2008). Is coastal lagoon eutrophication likely to be aggravated by global climate change? Estuarine, Coastal and Shelf Science, 78, 403-412. Snoussi, M., Ouchani, T., Niazi, S. (2008). Vulnerability assessment of the impact of sea-level rise and flooding on the Moroccan coast: The case of the Mediterranean eastern zone. Estuarine, Coastal and Shelf Science, 77, 206-213.
12. Confronting Climate Change
Science.gov (United States)
Mintzer, Irving M.
1992-06-01
This book, which was published in time for the Earth Summit in Brazil in June 1992, is likely to make a huge impact on the political and economic agendas of international policy makers. It summarizes the scientific findings of Working Group I of the IPCC in the first part of the book. While acknowledging the uncertainties in subsequent chapters, it challenges and expands upon the existing views on how we should tackle the problems of climate change.
13. UK-China Dialogue on Climate Ethics:Finding Consensus Review of the Fourth International Conference on Climate Change and Public Policy%中英气候伦理对话:从分歧走向共识--第四届气候变化与公共政策国际学术会议综述
Institute of Scientific and Technical Information of China (English)
史军
2014-01-01
2014年9月21日至23日,第四届气候变化与公共政策国际学术会议暨中英气候伦理与政策研讨会在英国雷丁大学和牛津大学召开。会议围绕五组主题展开:代际气候正义与代内气候正义,气候贫困与消费者选择,平等与发展,国际法律与国际制度,新问题与新技术方案。此次会议对中英学者在应对气候变化问题上增进了解、求同存异、寻找共识起到了重要作用,进一步提升了我国学者参与国际气候问题交流与合作的水平和能力。%The Fourth International Conference on Climate Change and Public Policy, and UK⁃China Net⁃work on Climate Ethics was held on 21 and 22 September 2014 at University of Reading and on 23 September 2014 at the University of Oxford. The five groups of topics were as follow: intergenerational and intra⁃genera⁃tional Climate justice; climate poverty and consumer choice; equity and development; international institutions;new problems and new solutions. There was also extensive discussion on Chinese moral perspective on climate governance between UK and Chinese scholars.
14. The Fourth World Conference on Women.
Science.gov (United States)
1995-01-01
15. Fourth generation electron cyclotron resonance ion sources.
Science.gov (United States)
Lyneis, Claude M; Leitner, D; Todd, D S; Sabbi, G; Prestemon, S; Caspi, S; Ferracin, P
2008-02-01
The concepts and technical challenges related to developing a fourth generation electron cyclotron resonance (ECR) ion source with a rf frequency greater than 40 GHz and magnetic confinement fields greater than twice B(ECR) will be explored in this article. Based on the semiempirical frequency scaling of ECR plasma density with the square of operating frequency, there should be significant gains in performance over current third generation ECR ion sources, which operate at rf frequencies between 20 and 30 GHz. While the third generation ECR ion sources use NbTi superconducting solenoid and sextupole coils, the new sources will need to use different superconducting materials, such as Nb(3)Sn, to reach the required magnetic confinement, which scales linearly with rf frequency. Additional technical challenges include increased bremsstrahlung production, which may increase faster than the plasma density, bremsstrahlung heating of the cold mass, and the availability of high power continuous wave microwave sources at these frequencies. With each generation of ECR ion sources, there are new challenges to be mastered, but the potential for higher performance and reduced cost of the associated accelerator continues to make this a promising avenue for development. PMID:18315111
16. Examination of nuclear systems of fourth generation
International Nuclear Information System (INIS)
This report proposes a detailed discussion of the six nuclear systems selected by the Generation IV International Forum with the objective of coordinating research and development activities which should result in the deployment of nuclear systems (reactors and associated fuel cycle installations) of fourth generation by the second half of the 21. century. These systems are: sodium cooled fast reactors (SFR), very high temperature reactors (VHTR), gas cooled fast reactors (GFR), lead cooled fast reactors (LFR) or lead bismuth eutectic reactors (LBE), molten salt reactors (MSR), and supercritical water reactors (SCWR). Fast systems are interesting as they favour the transmutation of fertile materials into fissile materials. History and perspectives of development, main characteristics, management of safety functions, risk analysis, impact on the environment, radiation protection and decommissioning, concept maturity and R and D needs are discussed for each of these systems. A comparison is reported in terms of main characteristics of reactors, of neutron characteristics and reactivity control, of sensitivity to cooling losses, of confinement function, of exploitation safety, of in-service inspection, of behaviour in case of severe accident, of toxicity of chemical substances, of sensitivity to aggressions (seism), of concept maturity and technological difficulties. The report also proposes a review of the various fuels which can be used in these different systems and which have been considered as eligible by the International Forum: oxides, carbides, nitrides, metals, waste processing. The last part addresses the transmutation of long life radioactive elements: physics, context, assessment of scenarios soundness, influence of transmutation on installations and transports
17. IAEA research contracts. Fourth annual report
International Nuclear Information System (INIS)
This volume represents the fourth annual report on the results obtained under the Agency's research contract programme. During the short life of this programme, which is not quite six years old, a total investment of more than three million dollars has been made to support research in selected fields at institutes in 50 Member States. Extensive summaries are presented herein for all final reports relating to contracts which were completed during 1963. As it is the policy of the Agency to encourage publication in the open scientific literature of the results of work done under research contracts, a number of papers have also appeared in the appropriate journals - the Agency having been notified of 75 such publications in 1963. A complete list of references to these is given at the end of this report. The scientific data presented in the summaries of course remain the responsibility of the contractor. The Agency, however, is responsible for any additional observations. The reports presented are related to research in the field of radioactive waste management and environmental sciences; health physics and radiation protection; radiobiology; nuclear reactors physics and nuclear fuels; radioisotope applications in agriculture and medicine
18. Neural crest: The fourth germ layer
Directory of Open Access Journals (Sweden)
K Shyamala
2015-01-01
Full Text Available The neural crest cells (NCCs, a transient group of cells that emerges from the dorsal aspect of the neural tube during early vertebrate development has been a fascinating group of cells because of its multipotency, long range migration through embryo and its capacity to generate a prodigious number of differentiated cell types. For these reasons, although derived from the ectoderm, the neural crest (NC has been called the fourth germ layer. The non neural ectoderm, the neural plate and the underlying mesoderm are needed for the induction and formation of NC cells. Once formed, NC cells start migrating as a wave of cells, moving away from the neuroepithelium and quickly splitting into distinct streams. These migrating NCCs home in to different regions and give rise to plethora of tissues. Umpteen number of signaling molecules are essential for formation, epithelial mesenchymal transition, delamination, migration and localization of NCC. Authors believe that a clear understanding of steps and signals involved in NC formation, migration, etc., may help in understanding the pathogenesis behind cancer metastasis and many other diseases. Hence, we have taken this review to discuss the various aspects of the NC cells.
19. Analyses of the predicted changes of the global oceans under the increased greenhouse gases scenarios
Institute of Scientific and Technical Information of China (English)
MU Lin; WU Dexing; CHEN Xue'en; J Jungclaus
2006-01-01
A new climate model (ECHAM5/MPIOM1) developed for the fourth assessment report of the Intergovernmental Panel on Climate Change (IPCC) at Max-Planck Institute for Meteorology is used to study the climate changes under the different increased CO2 scenarios (B1, A1B and A2). Based on the corresponding model results, the sea surface temperature and salinity structure, the variations of the thermohaline circulation (THC) and the changes of sea ice in the northern hemisphere are analyzed. It is concluded that from the year of 2000 to 2100, under the B1, A1B and A2 scenarios, the global mean sea surface temperatures (SST) would increase by 2.5℃, 3.5℃ and 4.0℃ respectively, especially in the region of the Arctic, the increase of SST would be even above 10.0℃; the maximal negative value of the variation of the fresh water flux is located in the subtropical oceans, while the precipitation in the eastern tropical Pacific increases. The strength of THC decreases under the B1, A1B and A2 scenarios, and the reductions would be about 20%, 25% and 25.1% of the present THC strength respectively. In the northern hemisphere, the area of the sea ice cover would decrease by about 50% under the A1B scenario.
20. MANAGING PLANET EARTH TO MAKE FUTURE DEVELOPMENT MORE SUSTAINABLE: CLIMATE CHANGE AND HONG KONG
Institute of Scientific and Technical Information of China (English)
W.W.-S.Yim; C.D.Ollier
2009-01-01
School of Earth and Geographical Sciences, The University of Western Australia, Crawley, WA 6009, Australia)Selected recent findings related to climate change in Hong Kong include: (1)The Hong Kong seafloor has yielded a ca.0.5-million year record of climate and sea-level changes.(2)Greenhouse gases produced naturally from sub-aerially exposed continental shelves and oceanic islands were a probable forcing mechanism in triggering the abrupt termination of past ice ages. (3)An analysis of annual mean temperature records has revealed that the urban heat island effect has contributed ca.75% of the warming. (4)Past volcanic eruptions are found to lower Hong Kong's temperature and to cause extremely dry and wet years. (5)No evidence can be found for an increase in frequency and intensity of typhoons based on the instrumental record since the end of the Second World War. (6)The observed rate of sea-level rise in the South China Sea is much slower than the predictions of the IPCC Fourth Assessment. For the Earth's management, population growth and the depletion of non-renewable resources must be recognized as unsustainable. The human impact on the natural hydrological cycle is an important forcing mechanism in climate change. In order to delay the demise of the human race, management must include curbing population growth and much more waste recycling than at present.
1. Human Resources Administration: A School-Based Perspective. Fourth Edition
Science.gov (United States)
Smith, Richard
2009-01-01
Enhanced and updated, this Fourth Edition of Richard E. Smith's highly successful text examines the growing role of the principal in planning, hiring, staff development, supervision, and other human resource functions. The Fourth Edition includes new sections on ethics, induction, and the role of the mentor teacher. This edition also introduces…
2. An isolated fourth ventricle in neurosarcoidosis: MRI findings
Energy Technology Data Exchange (ETDEWEB)
Hesselmann, Volker; Terstegge, Klaus; Schulte, Oliver; Krug, Barbara; Lackner, Klaus [Department of Radiology, University of Cologne, Joseph Stelzmann Strasse 9, 50924 Cologne (Germany); Wedekind, Christoph [Department of Neurosurgery, University of Cologne, Joseph Stelzmann Strasse 9, 50924 Cologne (Germany); Voges, Juergen [Department of Stereotaxy und Functional Neurosurgery, University of Cologne, Joseph Stelzmann Strasse 9, 50924 Cologne (Germany)
2002-07-01
We report on an isolated enlargement of the fourth ventricle in a patient with neurosarcoidosis which developed 3 years after the insertion of a ventriculo-atrial shunt. Repeated MRI images were obtained in a patient with known neurosarcoidosis between 1995 and 2000. Imaging findings were correlated to the medical course of the patient, who developed a hydrocephalus and a trapped fourth ventricle consecutively. The isolation was presumably due to granulomatous inflammation of the ependyma surrounding the fourth ventricular outlets. The isolated fourth ventricle was responsible for a deterioration of neurological status. Neurosarcoidosis is a severe complication in sarcoidosis patients. An isolated enlargement of the fourth ventricle is a rare complication in clinically deteriorated patients with neurosarcoidosis and ventricular drainage, which may require neurosurgical treatment. (orig.)
3. Proceedings: Fourth Workshop on Mining Scientific Datasets
Energy Technology Data Exchange (ETDEWEB)
Kamath, C
2001-07-24
Commercial applications of data mining in areas such as e-commerce, market-basket analysis, text-mining, and web-mining have taken on a central focus in the JCDD community. However, there is a significant amount of innovative data mining work taking place in the context of scientific and engineering applications that is not well represented in the mainstream KDD conferences. For example, scientific data mining techniques are being developed and applied to diverse fields such as remote sensing, physics, chemistry, biology, astronomy, structural mechanics, computational fluid dynamics etc. In these areas, data mining frequently complements and enhances existing analysis methods based on statistics, exploratory data analysis, and domain-specific approaches. On the surface, it may appear that data from one scientific field, say genomics, is very different from another field, such as physics. However, despite their diversity, there is much that is common across the mining of scientific and engineering data. For example, techniques used to identify objects in images are very similar, regardless of whether the images came from a remote sensing application, a physics experiment, an astronomy observation, or a medical study. Further, with data mining being applied to new types of data, such as mesh data from scientific simulations, there is the opportunity to apply and extend data mining to new scientific domains. This one-day workshop brings together data miners analyzing science data and scientists from diverse fields to share their experiences, learn how techniques developed in one field can be applied in another, and better understand some of the newer techniques being developed in the KDD community. This is the fourth workshop on the topic of Mining Scientific Data sets; for information on earlier workshops, see http://www.ahpcrc.org/conferences/. This workshop continues the tradition of addressing challenging problems in a field where the diversity of applications is
4. IPCC-AR4模式对影响西北太平洋热带气旋的大气动力环境场的气候特征模拟性能%Assessments on climatologic simulation of atmospheric dynamic environment of tropical cyclone over western North Pacific in IPCC -AR4 models
Institute of Scientific and Technical Information of China (English)
余锦华; 赵晓彤; 陈成
2014-01-01
评估了23个IPCC-AR4模式在低纬地区1948-1999年7-9月大尺度环流场的模拟性能,重点关注西北太平洋区域的西太副高、季风槽以及台风活动海域的垂直风切变。结果显示,绝大多数模式的7-9月低纬地区500 hPa平均高度场、850 hPa风场空间分布与NCEP都具有很高的相似性,但大多模式500 hPa高度场存在系统性偏低,而850 hPa风场偏强。所有模式模拟的西北太平洋副高脊线与NCEP都有一致的西南-东北走向,但有些模式的脊线位置偏离NCEP的较远。有4个模式没有模拟出类似于NCEP的季风槽线。综合模式对夏季热带环流场、西北太平洋副热带高压、季风槽以及西北太平洋热带气旋活动关键区域垂直风切变气候特征的模拟性能,按性能优劣,排在前10的模式依次是mpi_echam5、cccma_t63、gfdl_cm2_1、cnrm_cm3、cccma_t47、ukmo_hadgem1、ingv_echam4、ncar_ccsm3_0、csiro_mk3_5、mri_cgcm2_3_2a;排在后6位的模式是inmcm3_0、iap_fgoals1_0_g、ipsl_cm4、miroc3_2_medres、giss_eh、giss_er。%The fidelity of atmospheric circulation over tropical regions ,especially subtropical high ,monsoon trough and vertical wind shear over the regions of tropical cyclone activity in the coupled general circulation models (CGCM) participating in the Forth Assessment Report (AR4) of Intergovernmental Panel on Climate Change (IPCC)is accessed by virtue of comparing simulated climate field with that of National Center Environmental Pre-diction(NCEP)/National Center Atmospheric Research (NCAR) during 1948-1999 from July to September. It is found that vast majority of 23 IPCC-AR4 models show a good performance in mean state in 500 hPa height field over the domain 0°-40°N ,0°E-180°-0°W and 850 hPa wind field over the region 20°S-40°N ,0°E-180°-0°W-180°-0°W ,but in comparison to that of NCEP/NCAR ,most of models show a low bias in 500 hPa height
5. NTFP and REDD at the Fourth World Conservation Congress: What is In and What is Not
Directory of Open Access Journals (Sweden)
Peña Pablo
2010-01-01
Full Text Available While the Fourth World Conservation Congress (WCC was effective in bringing together different participants to discuss climate change, the discussion of potential mitigation mechanisms was dominated by the Reducing Emissions for Deforestation and Degradation (REDD initiative, to the exclusion of other possibilities, including Non-timber Forest Products (NTFP-there was a notable lack of venues for discussing the relevance of NTFP projects for biodiversity conservation, climate change mitigation and poverty alleviation. This paper contrasts the treatment of NTFP and REDD at the WCC and discusses how the exclusion of NTFP from these discussions will probably affect its inclusion in the conservation agenda and the future design and funding of conservation projects. The paper also shares some ideas on unexplored complementarities between NTFP and REDD for climate change mitigation, showing that an opportunity was lost at the Fourth WCC for promoting NTFP as an additional market-based approach to conservation.
6. Le changement climatique d'origine humaine. Rappel de quelques résultats générauxAnthropogenic climate change: general results
Science.gov (United States)
Petit, Michel
1999-02-01
The IPCC (International Panel on Climate Change) reports have highlighted major results, which constitute a relevant framework for the specific papers in this issue. Both facts established with large confidence level and models results are presented.
7. Future Changes in Surface Runoff over Korea Projected by a Regional Climate Model under A1B Scenario
Directory of Open Access Journals (Sweden)
Ji-Woo Lee
2014-01-01
Full Text Available This study assesses future change of surface runoff due to climate change over Korea using a regional climate model (RCM, namely, the Global/Regional Integrated Model System (GRIMs, Regional Model Program (RMP. The RMP is forced by future climate scenario, namely, A1B of Intergovernmental Panel on Climate Change (IPCC Fourth Assessment Report (AR4. The RMP satisfactorily reproduces the observed seasonal mean and variation of surface runoff for the current climate simulation. The distribution of monsoonal precipitation-related runoff is adequately captured by the RMP. In the future (2040–2070 simulation, it is shown that the increasing trend of temperature has significant impacts on the intra-annual runoff variation. The variability of runoff is increased in summer; moreover, the strengthened possibility of extreme occurrence is detected in the future climate. This study indicates that future climate projection, including surface runoff and its variability over Korea, can be adequately addressed on the RMP testbed. Furthermore, this study reflects that global warming affects local hydrological cycle by changing major water budget components. This study adduces that the importance of runoff should not be overlooked in regional climate studies, and more elaborate presentation of fresh-water cycle is needed to close hydrological circulation in RCMs.
8. Five year ahead prediction of Sea Surface Temperature in the Tropical Atlantic: a comparison between IPCC climate models and simple statistical methods
CERN Document Server
Laepple, T; Laepple, Thomas; Jewson, Stephen
2007-01-01
There is a clear positive correlation between boreal summer tropical Atlantic sea-surface temperature and annual hurricane numbers. This motivates the idea of trying to predict the sea-surface temperature in order to be able to predict future hurricane activity. In previous work we have used simple statistical methods to make 5 year predictions of tropical Atlantic sea surface temperatures for this purpose. We now compare these statistical SST predictions with SST predictions made by an ensemble mean of IPCC climate models.
9. Study of cysticercosis in the fourth ventricle by CSF cinema MRI
International Nuclear Information System (INIS)
Purpose: To evaluate the diagnostic value of cysticercosis in the fourth ventricle by CSF cinema MRI. Materials and methods: Nine patients with intraventricular cysticercosis in the fourth ventricle were studied. The diagnosis was confirmed by surgery in all cases. All of these patients were examined systematically before the operation and studied with CSF cinema MRI in mid sagittal section and finger-gated scan technique. Results: (1) The path of CSF flow was directly displayed. All cysticercosis presented as a filling defect, and a cyst with a smooth wall. (2) The ventricular compliance was normal in cysticercosis. (3) The cysticercosis in active stage was free in the fourth ventricle and could be rolled over, its shape might change slightly within a cardiac cycle. In the degenerative stage, its wall could adhere to the ependyma and obstruct the CSF flow. Conclusion: CSF cinema MRI can demonstrate the degree of obstruction and pattern of CSF flow in cysticercosis of the fourth ventricle, thereby providing useful information for proper management
10. The endoscopic trans-fourth ventricle aqueductoplasty and stent placement for the treatment of trapped fourth ventricle: Long-term results in a series of 18 consecutive patients
Directory of Open Access Journals (Sweden)
Pasquale Gallo
2012-01-01
Full Text Available Background: Different surgical approaches have been described in the past to treat a trapped fourth ventricle (TFV but, unfortunately, these techniques showed a high rate of dysfunction and complications. During the last 10 years the development of neuroendoscopy has dramatically changed the outcome of these patients. Materials and Methods: We conducted a retrospective evaluation of the safety, effectiveness, and long-term outcome of endoscopic aqueductoplasty and stent placement, performed in 18 consecutive patients with symptomatic TFV through a trans-fourth ventricle approach between 1994 and 2010. Thirteen patients underwent endoscopic aqueductoplasty and stent placement and 5 patients underwent aqueductoplasty alone using a tailored suboccipital approach through the foramen of Magendie in prone or sitting position. Results: The mean age of the patients at the time of surgery was 15.2 years. All patients but 3 had a supratentorial ventriculoperitoneal shunt. Fifteen patients presented with slit supratentorial ventricles. At a mean followup of 90.8 months all patients experienced a stable clinical improvement. Only two complications were observed: A transient diplopia due to dysconjugate eye movements in one patient and a transient trochlear palsy in another one. Conclusions: Our experience and the literature review suggest that endoscopic trans-fourth ventricle aqueductoplasty and stent placement is a minimally invasive, safe, and effective technique for the treatment of TFV and should be strongly recommended, especially in patients with supratentorial slit ventricles.
11. The radial velocity experiment (RAVE) : Fourth data release
NARCIS (Netherlands)
Kordopatis, G.; Gilmore, G.; Steinmetz, M.; Boeche, C.; Seabroke, G. M.; Siebert, A.; Zwitter, T.; Binney, J.; de Laverny, P.; Recio-Blanco, A.; Williams, M. E. K.; Piffl, T.; Enke, H.; Roeser, S.; Bijaoui, A.; Wyse, R. F. G.; Freeman, K.; Munari, U.; Carrillo, I.; Anguiano, B.; Burton, D.; Campbell, R.; Cass, C. J. P.; Fiegert, K.; Hartley, M.; Parker, Q. A.; Reid, W.; Ritter, A.; Russell, K. S.; Stupar, M.; Watson, F. G.; Bienayme, O.; Bland-Hawthorn, J.; Gerhard, O.; Gibson, B. K.; Grebel, E. K.; Helmi, A.; Navarro, J. F.; Conrad, C.; Famaey, B.; Faure, C.; Just, A.; Kos, J.; Matijevic, G.; McMillan, P. J.; Minchev, I.; Scholz, R.; Sharma, S.; Siviero, A.; de Boer, E. Wylie; Zerjal, M.
2013-01-01
We present the stellar atmospheric parameters (effective temperature, surface gravity, overall metallicity), radial velocities, individual abundances, and distances determined for 425,561 stars, which constitute the fourth public data release of the RAdial Velocity Experiment (RAVE). The stellar atm
12. Solar Energy Technologies Program Newsletter - Fourth Quarter 2009
Energy Technology Data Exchange (ETDEWEB)
DOE Solar Energy Technologies Program
2009-12-31
The Fourth Quarter 2009 edition of the Solar Energy Technologies Program newsletter summarizes the activities for the past three months, funding opportunities, highlights from the national labs, and upcoming events.
13. The Fourth SM Family Neutrino at Future Linear Colliders
CERN Document Server
Çiftçi, A K; Sultansoy, S
2005-01-01
It is known that Flavor Democracy favors the existence of the fourth standard model (SM) family. In order to give nonzero masses for the first three family fermions Flavor Democracy has to be slightly broken. A parametrization for democracy breaking, which gives the correct values for fundamental fermion masses and, at the same time, predicts quark and lepton CKM matrices in a good agreement with the experimental data, is proposed. The pair productions of the fourth SM family Dirac $(\ 14. Comparative evaluation of the IPCC AR5 CMIP5 versus the AR4 CMIP3 model ensembles for regional precipitation and their extremes over South America Science.gov (United States) Tolen, J.; Kodra, E. A.; Ganguly, A. R. 2011-12-01 The assertion that higher-resolution experiments or more sophisticated process models within the IPCC AR5 CMIP5 suite of global climate model ensembles improves precipitation projections over the IPCC AR4 CMIP3 suite remains a hypothesis that needs to be rigorously tested. The questions are particularly important for local to regional assessments at scales relevant for the management of critical infrastructures and key resources, particularly for the attributes of sever precipitation events, for example, the intensity, frequency and duration of extreme precipitation. Our case study is South America, where precipitation and their extremes play a central role in sustaining natural, built and human systems. To test the hypothesis that CMIP5 improves over CMIP3 in this regard, spatial and temporal measures of prediction skill are constructed and computed by comparing climate model hindcasts with the NCEP-II reanalysis data, considered here as surrogate observations, for the entire globe and for South America. In addition, gridded precipitation observations over South America based on rain gage measurements are considered. The results suggest that the utility of the next-generation of global climate models over the current generation needs to be carefully evaluated on a case-by-case basis before communicating to resource managers and policy makers. 15. Statistical downscaling of IPCC sea surface wind and wind energy predictions for U.S. east coastal ocean, Gulf of Mexico and Caribbean Sea Science.gov (United States) Yao, Zhigang; Xue, Zuo; He, Ruoying; Bao, Xianwen; Song, Jun 2016-08-01 A multivariate statistical downscaling method is developed to produce regional, high-resolution, coastal surface wind fields based on the IPCC global model predictions for the U.S. east coastal ocean, the Gulf of Mexico (GOM), and the Caribbean Sea. The statistical relationship is built upon linear regressions between the empirical orthogonal function (EOF) spaces of a cross- calibrated, multi-platform, multi-instrument ocean surface wind velocity dataset (predictand) and the global NCEP wind reanalysis (predictor) over a 10 year period from 2000 to 2009. The statistical relationship is validated before applications and its effectiveness is confirmed by the good agreement between downscaled wind fields based on the NCEP reanalysis and in-situ surface wind measured at 16 National Data Buoy Center (NDBC) buoys in the U.S. east coastal ocean and the GOM during 1992-1999. The predictand-predictor relationship is applied to IPCC GFDL model output (2.0°×2.5°) of downscaled coastal wind at 0.25°×0.25° resolution. The temporal and spatial variability of future predicted wind speeds and wind energy potential over the study region are further quantified. It is shown that wind speed and power would significantly be reduced in the high CO2 climate scenario offshore of the mid-Atlantic and northeast U.S., with the speed falling to one quarter of its original value. 16. Combined search for the quarks of a sequential fourth generation CERN Document Server Chatrchyan, Serguei; Sirunyan, Albert M; Tumasyan, Armen; Adam, Wolfgang; Aguilo, Ernest; Bergauer, Thomas; Dragicevic, Marko; Erö, Janos; Fabjan, Christian; Friedl, Markus; Fruehwirth, Rudolf; Ghete, Vasile Mihai; Hammer, Josef; Hörmann, Natascha; Hrubec, Josef; Jeitler, Manfred; Kiesenhofer, Wolfgang; Knünz, Valentin; Krammer, Manfred; Krätschmer, Ilse; Liko, Dietrich; Mikulec, Ivan; Pernicka, Manfred; Rahbaran, Babak; Rohringer, Christine; Rohringer, Herbert; Schöfbeck, Robert; Strauss, Josef; Taurok, Anton; Waltenberger, Wolfgang; Walzel, Gerhard; Widl, Edmund; Wulz, Claudia-Elisabeth; Mossolov, Vladimir; Shumeiko, Nikolai; Suarez Gonzalez, Juan; Bansal, Monika; Bansal, Sunil; Cornelis, Tom; De Wolf, Eddi A; Janssen, Xavier; Luyckx, Sten; Mucibello, Luca; Ochesanu, Silvia; Roland, Benoit; Rougny, Romain; Selvaggi, Michele; Staykova, Zlatka; Van Haevermaet, Hans; Van Mechelen, Pierre; Van Remortel, Nick; Van Spilbeeck, Alex; Blekman, Freya; Blyweert, Stijn; D'Hondt, Jorgen; Gonzalez Suarez, Rebeca; Kalogeropoulos, Alexis; Maes, Michael; Olbrechts, Annik; Van Doninck, Walter; Van Mulders, Petra; Van Onsem, Gerrit Patrick; Villella, Ilaria; Clerbaux, Barbara; De Lentdecker, Gilles; Dero, Vincent; Gay, Arnaud; Hreus, Tomas; Léonard, Alexandre; Marage, Pierre Edouard; Mohammadi, Abdollah; Reis, Thomas; Thomas, Laurent; Vander Marcken, Gil; Vander Velde, Catherine; Vanlaer, Pascal; Wang, Jian; Adler, Volker; Beernaert, Kelly; Cimmino, Anna; Costantini, Silvia; Garcia, Guillaume; Grunewald, Martin; Klein, Benjamin; Lellouch, Jérémie; Marinov, Andrey; Mccartin, Joseph; Ocampo Rios, Alberto Andres; Ryckbosch, Dirk; Strobbe, Nadja; Thyssen, Filip; Tytgat, Michael; Verwilligen, Piet; Walsh, Sinead; Yazgan, Efe; Zaganidis, Nicolas; Basegmez, Suzan; Bruno, Giacomo; Castello, Roberto; Ceard, Ludivine; Delaere, Christophe; Du Pree, Tristan; Favart, Denis; Forthomme, Laurent; Giammanco, Andrea; Hollar, Jonathan; Lemaitre, Vincent; Liao, Junhui; Militaru, Otilia; Nuttens, Claude; Pagano, Davide; Pin, Arnaud; Piotrzkowski, Krzysztof; Schul, Nicolas; Vizan Garcia, Jesus Manuel; Beliy, Nikita; Caebergs, Thierry; Daubie, Evelyne; Hammad, Gregory Habib; Alves, Gilvan; Correa Martins Junior, Marcos; De Jesus Damiao, Dilson; Martins, Thiago; Pol, Maria Elena; Henrique Gomes E Souza, Moacyr; Aldá Júnior, Walter Luiz; Carvalho, Wagner; Custódio, Analu; Da Costa, Eliza Melo; De Oliveira Martins, Carley; Fonseca De Souza, Sandro; Matos Figueiredo, Diego; Mundim, Luiz; Nogima, Helio; Oguri, Vitor; Prado Da Silva, Wanda Lucia; Santoro, Alberto; Soares Jorge, Luana; Sznajder, Andre; Souza Dos Anjos, Tiago; Bernardes, Cesar Augusto; De Almeida Dias, Flavia; Tomei, Thiago; De Moraes Gregores, Eduardo; Lagana, Caio; Da Cunha Marinho, Franciole; Mercadante, Pedro G; Novaes, Sergio F; Padula, Sandra; Genchev, Vladimir; Iaydjiev, Plamen; Piperov, Stefan; Rodozov, Mircho; Stoykova, Stefka; Sultanov, Georgi; Tcholakov, Vanio; Trayanov, Rumen; Vutova, Mariana; Dimitrov, Anton; Hadjiiska, Roumyana; Kozhuharov, Venelin; Litov, Leander; Pavlov, Borislav; Petkov, Peicho; Bian, Jian-Guo; Chen, Guo-Ming; Chen, He-Sheng; Jiang, Chun-Hua; Liang, Dong; Liang, Song; Meng, Xiangwei; Tao, Junquan; Wang, Jian; Wang, Xianyou; Wang, Zheng; Xiao, Hong; Xu, Ming; Zang, Jingjing; Zhang, Zhen; Asawatangtrakuldee, Chayanit; Ban, Yong; Guo, Shuang; Guo, Yifei; Li, Wenbo; Liu, Shuai; Mao, Yajun; Qian, Si-Jin; Teng, Haiyun; Wang, Dayong; Zhang, Linlin; Zhu, Bo; Zou, Wei; Avila, Carlos; Gomez, Juan Pablo; Gomez Moreno, Bernardo; Osorio Oliveros, Andres Felipe; Sanabria, Juan Carlos; Godinovic, Nikola; Lelas, Damir; Plestina, Roko; Polic, Dunja; Puljak, Ivica; Antunovic, Zeljko; Kovac, Marko; Brigljevic, Vuko; Duric, Senka; Kadija, Kreso; Luetic, Jelena; Morovic, Srecko; Attikis, Alexandros; Galanti, Mario; Mavromanolakis, Georgios; Mousa, Jehad; Nicolaou, Charalambos; Ptochos, Fotios; Razis, Panos A; Finger, Miroslav; Finger Jr, Michael; Assran, Yasser; Elgammal, Sherif; Ellithi Kamel, Ali; Khalil, Shaaban; Mahmoud, Mohammed; Radi, Amr; Kadastik, Mario; Müntel, Mait; Raidal, Martti; Rebane, Liis; Tiko, Andres; Eerola, Paula; Fedi, Giacomo; Voutilainen, Mikko; Härkönen, Jaakko; Heikkinen, Mika Aatos; Karimäki, Veikko; Kinnunen, Ritva; Kortelainen, Matti J; Lampén, Tapio; Lassila-Perini, Kati; Lehti, Sami; Lindén, Tomas; Luukka, Panja-Riina; Mäenpää, Teppo; Peltola, Timo; Tuominen, Eija; Tuominiemi, Jorma; Tuovinen, Esa; Ungaro, Donatella; Wendland, Lauri; Banzuzi, Kukka; Karjalainen, Ahti; Korpela, Arja; Tuuva, Tuure; Besancon, Marc; Choudhury, Somnath 2012-01-01 Results are presented from a search for a fourth generation of quarks produced singly or in pairs in a data set corresponding to an integrated luminosity of 5 inverse femtobarns recorded by the CMS experiment at the LHC in 2011. A novel strategy has been developed for a combined search for quarks of the up- and down-type in decay channels with at least one isolated muon or electron. Limits on the mass of the fourth-generation quarks and the relevant CKM matrix elements are derived in the context of a simple extension of the standard model with a sequential fourth generation of fermions. The existence of mass-degenerate fourth-generation quarks with masses below 685 GeV is excluded at 95% confidence level for minimal off-diagonal mixing between the third- and the fourth-generation quarks. With a mass difference of 25 GeV between the quark masses, the obtained limit on the masses of the fourth-generation quarks shifts by about +/- 20 GeV. This result significantly reduces the allowed parameter space for a fourt... 17. Adaptation to Climate Variability and Change. Methodological Issues International Nuclear Information System (INIS) The Intergovernmental Panel on Climate Change (IPCC) convened a Workshop on Adaptation to Climate Variability and Change in Costa Rica in 1998 that involved more than 200 expects and incorporated views from many research communities. This paper summarizes the recommendations from the Workshop and profiles the contributions to the advancement of methodologies for adaptation science. 25 refs 18. New climate change scenarios reveal uncertain future for Central Asian glaciers Directory of Open Access Journals (Sweden) A. F. Lutz 2012-11-01 Full Text Available Central Asian water resources largely depend on (glacier melt water generated in the Pamir and Tien Shan mountain ranges, located in the basins of the Amu and Syr Darya rivers, important life lines in Central Asia and the prominent water source of the Aral Sea. To estimate future water availability in the region, it is thus necessary to project the future glacier extent and volume in the Amu and Syr Darya river basins. The aim of this study is to quantify the impact of uncertainty in climate change projections on the future glacier extent in the Amu and Syr Darya river basins. The latest climate change projections provided by the fifth Coupled Model Intercomparison Project (CMIP5 generated for the upcoming fifth assessment report of the Intergovernmental Panel on Climate Change (IPCC are used to model future glacier extent in the Central Asian region for the two large river basins. The outcomes are compared to model results obtained with the climate change projections used for the fourth IPCC assessment (CMIP3. We use a regionalized glacier mass balance model to estimate changes in glacier extent as a function of glacier size and projections of temperature and precipitation. The model is developed for implementation in (large scale hydrological models, when the spatial model resolution does not allow for modelling of individual glaciers and data scarcity is an issue. Both CMIP3 and CMIP5 model simulations point towards a strong decline in glacier extent in Central Asia. However, compared to the CMIP3 projections, the CMIP5 projections of future glacier extent in Central Asia provide a wider range of outcomes, mostly owing to greater variability in precipitation projections among the latest suite of climate models. These findings have great impact on projections of the timing and quantity of water availability in glacier melt dominated rivers in the region. Uncertainty about the size of the decline in glacier extent remains large, making 19. Calculation of probability density functions for temperature and precipitation change under global warming International Nuclear Information System (INIS) Full text: he IPCC Fourth Assessment Report (Meehl ef al. 2007) presents multi-model means of the CMIP3 simulations as projections of the global climate change over the 21st century under several SRES emission scenarios. To assess the possible range of change for Australia based on the CMIP3 ensemble, we can follow Whetton etal. (2005) and use the 'pattern scaling' approach, which separates the uncertainty in the global mean warming from that in the local change per degree of warming. This study presents several ways of representing these two factors as probability density functions (PDFs). The beta distribution, a smooth, bounded, function allowing skewness, is found to provide a useful representation of the range of CMIP3 results. A weighting of models based on their skill in simulating seasonal means in the present climate over Australia is included. Dessai ef al. (2005) and others have used Monte-Carlo sampling to recombine such global warming and scaled change factors into values of net change. Here, we use a direct integration of the product across the joint probability space defined by the two PDFs. The result is a cumulative distribution function (CDF) for change, for each variable, location, and season. The median of this distribution provides a best estimate of change, while the 10th and 90th percentiles represent a likely range. The probability of exceeding a specified threshold can also be extracted from the CDF. The presentation focuses on changes in Australian temperature and precipitation at 2070 under the A1B scenario. However, the assumption of linearity behind pattern scaling allows results for different scenarios and times to be simply obtained. In the case of precipitation, which must remain non-negative, a simple modification of the calculations (based on decreases being exponential with warming) is used to avoid unrealistic results. These approaches are currently being used for the new CSIRO/ Bureau of Meteorology climate projections 20. Climate Change in Colorado: Developing a Synthesis of the Science to Support Water Resources Management and Adaptation Science.gov (United States) Ray, A. J.; Barsugli, J. J.; Averyt, K. B.; Deheza, V.; Udall, B. 2008-12-01 peer-reviewed regional studies, and conducted new analyses derived from existing datasets and model projections, and took advantage of new regional analyses. In addition to the IPCC Fourth Assessment, we also took advantage of very new Climate Change Science Program Assessments. Many water managers, although often technically savvy engineers, hydrologists and other professionals, but are not trained as climate or atmospheric scientists, and seeks to complexity by using Fahrenheit units, minimizing use of or defining jargon terms, and re-plotting published figures/data for simplicity. The report is written at a less technical level than the IPCC reports, and some features are intended to raise the level of climate literacy of our audience about climate and how climate science is done. For example, the report includes a primer on climate models and theory that situates Colorado in the context of global climate change and describes how the unique features of the state -- such as the complex topography -- relate to interpreting and using climate change projections. This report responds to Colorado state agencies' and water management community needs to understanding of climate change and is an initial step in establishing Colorado's water-related adaptation needs. Another impact of this report is as an experiment in climate services for climate change information and exploring the challenges of communicating the information to diverse decisionmakers. 1. Climate Change 2013. The Physical Science Basis. Working Group I Contribution to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change - Abstract for decision-makers International Nuclear Information System (INIS) The Working Group I contribution to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change (IPCC) provides a comprehensive assessment of the physical science basis of climate change. It builds upon the Working Group I contribution to the IPCC's Fourth Assessment Report in 2007 and incorporates subsequent new findings from the Special Report on Managing the Risks of Extreme Events and Disasters to Advance Climate Change Adaptation, as well as from research published in the extensive scientific and technical literature. The assessment considers new evidence of past, present and projected future climate change based on many independent scientific analyses from observations of the climate system, paleo-climate archives, theoretical studies of climate processes and simulations using climate models. During the process of scoping and approving the outline of its Fifth Assessment Report, the IPCC focussed on those aspects of the current understanding of the science of climate change that were judged to be most relevant to policy-makers. In this report, Working Group I has extended coverage of future climate change compared to earlier reports by assessing near-term projections and predictability as well as long-term projections and irreversibility in two separate chapters. Following the decisions made by the Panel during the scoping and outline approval, a set of new scenarios, the Representative Concentration Pathways, are used across all three Working Groups for projections of climate change over the 21. century. The coverage of regional information in the Working Group I report is expanded by specifically assessing climate phenomena such as monsoon systems and their relevance to future climate change in the regions. The Working Group I Report is an assessment, not a review or a text book of climate science, and is based on the published scientific and technical literature available up to 15 March 2013. Underlying all aspects of the report is a 2. Costing issues for mitigation and adaptation to climate change: what policy makers need Energy Technology Data Exchange (ETDEWEB) Estrada-Oyuele, R.A. [Ministry of Foreign Affairs of the Argentine Republic, Buenos Aires (Argentina) 2000-07-01 As a supreme body on climate change, IPCC (Intergovernmental Panel on Climate Change) discusses related issues in conventions and originates scientific information which are needed by policy makers. This article highlights the problems usually faced by IPCC in originating such information. The author critically assesses the problems faces by climate change policy markers. He suggests that the scientific community should be actively involved in climate change policy formulation rather than providing warning against the risk of climate change. Plain language and clarity in preparing IPCC reports will improve the general understanding of the assessment. The author also points out that proliferation of scenarios is another source of many problems for policy makers. He argues out that the scenarios should be organized and presented according to some order of probability, as there is a tendency to confuse scenarios with forecasting. 3. Climate Change in China : Exploring Informants' Perceptions of Climate Change through a Qualitative Approach OpenAIRE Lipin, Tan 2016-01-01 Climate change is not only a natural phenomenon, but also a global social issue. Many studies try to explore the mechanisms behind climate change and the consequences of climate change, and provide information for developing the measures to mitigate or adapt to it. For example, the IPCC reviews and assesses climate-change-related scientific information produced worldwide, thus aiming to support decision-making from a scientific perspective. However, though various international and regional c... 4. Multinational enterprises and climate change strategies OpenAIRE Kolk, A.; Pinkse, J. 2012-01-01 Climate change is often perceived as the most pressing environmental problem of our time, as reflected in the large public, policy, and corporate attention it has received, and the concerns expressed about the (potential) consequences. Particularly due to temperature increases, climate change affects physical and biological systems by changing ecosystems and causing extinction of species, and is expected to have a negative social impact and adversely affect human health (IPCC, 2007). Moreover... 5. Uncertain outcomes and climate change policy OpenAIRE Robert S. Pindyck 2011-01-01 Focusing on tail effects, I incorporate distributions for temperature change and its economic impact in an analysis of climate change policy. I estimate the fraction of consumption w*(tau) that society would be willing to sacrifice to ensure that any increase in temperature at a future point is limited to tau. Using information on the distributions for temperature change and economic impact from studies assembled by the IPCC and from "integrated assessment models" (IAMs), I fit displaced gamm... 6. Rosette-forming glioneuronal tumor of the fourth ventricle Directory of Open Access Journals (Sweden) Da-chun ZHAO 2015-11-01 Full Text Available Objective To explore the clinicopathological features of rosette-forming glioneuronal tumor (RGNT of the fourth ventricle. Methods The clinical manifestations, neuroimaging, histopathological and immunohistochemical features were analysed in one case of RGNT of the fourth ventricle, and related literatures were reviewed. Results A 24-year-old female presented with progressive dizziness under no obvious predisposing causes and dyskinesia such as stumbling. MRI revealed expansion of the fourth ventricle, and a mass with long T1WI and T2WI signal and clear boundary could be seen within the fourth ventricle. The border of tumor showed slight enhancement. At surgery, it was observed that the solitary tumor arised from the fourth ventricle and appeared well demarcated with rhomboid fossa. The tumor was blocking the aqueduct of sylvius before it was removed. Microscopically, the tumor exhibited both neuronic and astrocytic components. In the neuronic components, neurocytes formed neurocytic rosettes and perivascular pseudorosettes. At the center of the neurocytic rosettes, there was an eosinophilic core and some region consisted of microcysts. While the astrocytic components of the tumor revealed typical pilocytic astrocytoma structure. The center of neuronic rosettes and perivascular pseudorosettes displayed strong positive staining with synaptophysin (Syn and oligodendrocytes transcription factor-2 (Olig-2. The astrocytic components showed positive immunostaining of glial fibrillary acidic protein (GFAP. There were focal and partial positive immunostaining of neuron-specific enolase (NSE in both components of the tumor. The Ki-67 labeling index was 1.50%-2.00% in two components. Conclusions Rosette-forming glioneuronal tumor of the fourth ventricle is an unusual neuronal and mixed neuronal-glial tumors. The imaging examination showed solid or mixed solid-cystic mass at the fourth ventricle with well demarcated border. The lesion has two 7. Tropical cyclones and climate change; Les cyclones tropicaux et le changement climatique Energy Technology Data Exchange (ETDEWEB) Andre, J.C. [Centre Europeen de Recherches Avancees en Calcul Scientifique, 31 - Toulouse (France); Royer, J.F.; Chauvin, F. [Meteo-France, Centre National de Recherches Meteorologique (CNRM), 31 - Toulouse (France) 2008-09-15 Results from observations and modelling studies, a number of which having been used to support the conclusions of the IPCC fourth assessment report, are presented. For the past and present-day (since 1970) periods, the increase of strong cyclonic activity over the North Atlantic Ocean appears to be in good correlation with increasing temperature of the ocean surface. For regions where observational data are of lesser quality, the increasing trend is less clear. In fact, assessing long-term changes is made difficult due to both the multi-decennial natural variability and the lesser coverage of observations before satellites were made available. Indirect observational data, such as those derived from quantitative estimations of damage caused by tropical cyclones, suffer from many artefacts and do not allow the resolving of the issue either. For the future, only numerical three-dimensional climate models can be used. They nevertheless run presently with too-large grid-sizes, so that their results are still not converging. Various simulations lead indeed to different results, and it is very often difficult to find the physical reasons for these differences. One concludes by indicating some ways through which numerical simulations could be improved, leading to a decrease of uncertainties affecting the prediction of cyclonic activity over the next decades. (authors) 8. Effects of Climate Change on Urban Rainwater Harvesting in Colombo City, Sri Lanka Directory of Open Access Journals (Sweden) Kwong Fai A. Lo 2015-03-01 Full Text Available Cities are becoming increasingly vulnerable to water-related issues due to rapid urbanization, installation of complex infrastructure and changes in rainfall patterns. This study aims at assessing the impacts of climate change on rainwater harvesting systems (RWH in the tropical urban city, Colombo, Sri Lanka. The future climate change projections are downscaled from global circulation models to the urban catchment scale using the Long Ashton Research Station Weather Generator (LARS-WG, described in the Intergovernmental Panel on Climate Change (IPCC Fourth Assessment Report (AR4, coupled with Inter Comparison Project (CMIP3 model results. Historical rainfall data from 1981–2010 is used to simulate long-term future rainfall data from 2011–2099. The percentage change of the rainfall is calculated. The rainfall patterns are analyzed based on the daily, monthly, seasonal and annual time scales. Water requirements are calculated based on the selected scenario types. Rainfall and water demand data are incorporated into a water balance model. Climate change impacts for the selected RWH scenarios are calculated based on the water security analysis for each scenario. Analysis of the future rainfall data of Colombo reveals that several extreme weather events with very heavy rainfall may occur in the future. However, the frequency of these big events may not occur too often. Most of the selected global circulation models (GCMs in this study predict that there will be more rainfall towards the end of this century (2080-2099. Residential RWH systems will be more affected than non-residential systems. RWH systems in Colombo should include potential future climate changes in their future design and planning and be prepared for excess runoff and additional measures against potential overflow and urban floods. 9. Research on the Impacts of Climate Change on State Maritime Strategy%全球气候变化对国家海洋战略的影响初探 Institute of Scientific and Technical Information of China (English) 李倩; 张韧; 金生龙; 徐海斌 2011-01-01 Based on IPCC Fourth Assessment Report, the paper generalizes the ocean responses to the global climate change, and then the influence mechanism of climate change on state maritime strategy is studied and expounded. The analysis shows that the climate change influences state maritime strategy in two steps. First, the climate change causes natural responses through physical mechanism, and then the environmental changes cause several social responses through social mechanism.%基于IPCC第四次评估报告中海洋对气候变化的响应,探讨了全球气候变化对国家海洋战略的影响机制。分析表明,气候变化对国家海洋战略的影响体现在两个阶段,即气候变化通过物理机制引起自然生态系统响应阶段和自然环境演变通过社会机制引起社会响应阶段。 10. Dark Coulomb binding of heavy neutrinos of fourth family Science.gov (United States) Belotsky, K. M.; Esipova, E. A.; Khlopov, M. Yu.; Laletin, M. N. 2015-11-01 Direct dark matter searches put severe constraints on the weakly interacting massive particles (WIMPs). These constraints cause serious troubles for the model of stable neutrino of fourth generation with mass around 50GeV. Though the calculations of primordial abundance of these particles make them in the charge symmetric case a sparse subdominant component of the modern dark matter, their presence in the universe would exceed the current upper limits by several orders of the magnitude. However, if quarks and leptons of fourth generation possess their own Coulomb-like y-interaction, recombination of pairs of heavy neutrinos and antineutrinos and their annihilation in the “neutrinium” atoms can play important role in their cosmological evolution, reducing their modern abundance far below the experimental upper limits. The model of stable fourth generation assumes that the dominant part of dark matter is explained by excessive Ū antiquarks, forming (ŪŪŪ)-- charged clusters, bound with primordial helium in nuclear-interacting O-helium (OHe) dark atoms. The y charge conservation implies generation of the same excess of fourth generation neutrinos, potentially dangerous WIMP component of this scenario. We show that due to y-interaction recombination of fourth neutrinos with OHe hides these WIMPs from direct WIMP searches, leaving the negligible fraction of free neutrinos, what makes their existence compatible with the experimental constraints. 11. Fourth standard model family neutrino at future linear colliders International Nuclear Information System (INIS) It is known that flavor democracy favors the existence of the fourth standard model (SM) family. In order to give nonzero masses for the first three-family fermions flavor democracy has to be slightly broken. A parametrization for democracy breaking, which gives the correct values for fundamental fermion masses and, at the same time, predicts quark and lepton Cabibbo-Kobayashi-Maskawa (CKM) matrices in a good agreement with the experimental data, is proposed. The pair productions of the fourth SM family Dirac (ν4) and Majorana (N1) neutrinos at future linear colliders with √(s)=500 GeV, 1 TeV, and 3 TeV are considered. The cross section for the process e+e-→ν4ν4(N1N1) and the branching ratios for possible decay modes of the both neutrinos are determined. The decays of the fourth family neutrinos into muon channels (ν4(N1)→μ±W±) provide cleanest signature at e+e- colliders. Meanwhile, in our parametrization this channel is dominant. W bosons produced in decays of the fourth family neutrinos will be seen in detector as either di-jets or isolated leptons. As an example, we consider the production of 200 GeV mass fourth family neutrinos at √(s)=500 GeV linear colliders by taking into account di-muon plus four jet events as signatures 12. Internet usage of fourth-grader primary school pupils Directory of Open Access Journals (Sweden) Peter Kristof Nagy 2014-08-01 Full Text Available The development of information and communication technologies has resulted in revolutionary changes, which seem to continue in economy, society and many areas of life. The most important resource of the new social structure emerging as a result of technologies is information, thus the development of human capacity related to the use of information is extremely important. It may not be easy for parents and teachers to face the fact that the majority of their children/pupils know more about cyberspace than themselves. But are primary school pupils really better versed? Does digital literacy indeed automatically evolve? Does the use of the internet by all means improve children’s academic performance? In our current study we are researching how primary school pupils in the easternmost part of Hungary use the internet, and what effects it has on their studies. We would also like to emphasize the role of teachers working at the lower grades of the primary school in the formation of pupils’ digital literacy. In 2013 190 fourth-grader primary school pupils of eight schools filled in our questionnaire. Based on the data negative correlation was found between the time spent on web browsing and the academic average. Analysis of variance also revealed that those pupils have a higher academic average whose parents regularly check the purpose of their children’s internet usage. Our research results showed that the pupils who use the internet and the library equally to search for new information have a higher academic average compared to those pupils who only make use of either the traditional or the online option only. The most important result of the research is that in the case of adequate parental or professional (teacher control internet may have fruitful effects on pupils’ studies, but in the lack of such control the opposite is more likely. 13. Global climate change: an unequivocal reality; Cambio climatico global: una realidad inequivoca Energy Technology Data Exchange (ETDEWEB) Raynal-Villasenor, J.A. [Universidad de las Americas, Puebla, Puebla (Mexico)]. E-mail: [email protected] 2011-10-15 During several years, a long discussion has taken place over the reality of global climate change phenomenon and, if there is one, what could be its cause. Once the 4th Assessment Report of the Intergovernmental Panel on Climatic Change (IPCC, 2007) - IPCC is part the United Nations Organization (UN) - was published, it was stated that there is a developing global climatic change and that the cause is unequivocally related with the human activity in the planet Earth. In this paper, relevant information is given about the development of global climatic change issues and some actions are mentioned that each human being of this planet can implement to mitigate it, since it has been accepted that it's impossible to stop it. [Spanish] Durante varios anos se ha discutido si existe un cambio climatico global y, si lo hay, cual es su causa. Una vez publicado el 4o. Reporte de Valoracion del Panel Intergubernamental sobre Cambio Climatico (IPCC, 2007) - el IPCC es parte de la Organizacion de las Naciones Unidas (ONU) - se preciso que hay un cambio climatico global en desarrollo y la causa inequivoca que lo esta produciendo es la actividad humana en el planeta Tierra, tambien se hablo en el IPCC de las causas naturales por las cuales el planeta se esta calentando. En el presente articulo, se da informacion relevante al cambio climatico global en desarrollo y se mencionan algunas acciones que cada ser humano de este planeta puede implementar para mitigarlo, ya que es imposible detenerlo. 14. Hubbert's Peak, The Coal Question, and Climate Change Science.gov (United States) Rutledge, D. 2008-12-01 The United Nations Intergovernmental Panel on Climate Change (IPCC) makes projections in terms of scenarios that include estimates of oil, gas, and coal production. These scenarios are defined in the Special Report on Emissions Scenarios or SRES (Nakicenovic et al., 2000). It is striking how different these scenarios are. For example, total oil production from 2005 to 2100 in the scenarios varies by 5:1 (Appendix SRES Version 1.1). Because production in some of the scenarios has not peaked by 2100, this ratio would be comparable to 10:1 if the years after 2100 were considered. The IPCC says "... the resultant 40 SRES scenarios together encompass the current range of uncertainties of future GHG [greenhouse gas] emissions arising from different characteristics of these models ..." (Nakicenovic et al., 2000, Summary for Policy Makers). This uncertainty is important for climate modeling, because it is larger than the likely range for the temperature sensitivity, which the IPCC gives as 2.3:1 (Gerard Meehl et al., 2007, the Fourth Assessment Report, Chapter 10, Global Climate Projections, p. 799). The uncertainty indicates that we could improve climate modeling if we could make a better estimate of future oil, gas, and coal production. We start by considering the two major fossil-fuel regions with substantial exhaustion, US oil and British coal. It turns out that simple normal and logistic curve fits to the cumulative production for these regions give quite stable projections for the ultimate production. By ultimate production, we mean total production, past and future. For US oil, the range for the fits for the ultimate is 1.15:1 (225- 258 billion barrels) for the period starting in 1956, when King Hubbert made his prediction of the peak year of US oil production. For UK coal, the range is 1.26:1 for the period starting in 1905, at the time of a Royal Commission on coal supplies. We extend this approach to find fits for world oil and gas production, and by a regional 15. Simulation of the Atlantic meridional overturning circulation in an atmosphere-ocean global coupled model. Part II: weakening in a climate change experiment: a feedback mechanism Energy Technology Data Exchange (ETDEWEB) Guemas, Virginie [Meteo-France, CNRS, Centre National de Recherches Meteorologiques/Groupe d' Etude de l' Atmosphere Meteorologique (CNRM/GMGEC), Toulouse Cedex (France); CEA-CNRS-UVSQ, Laboratoire des Sciences du Climat et de l' Environnement, UMR 1572, Gif-sur-Yvette (France); Salas-Melia, David [Meteo-France, CNRS, Centre National de Recherches Meteorologiques/Groupe d' Etude de l' Atmosphere Meteorologique (CNRM/GMGEC), Toulouse Cedex (France) 2008-06-15 Most state-of-the art global coupled models simulate a weakening of the Atlantic meridional overturning circulation (MOC) in climate change scenarios but the mechanisms leading to this weakening are still being debated. The third version of the CNRM (Centre National de Recherches Meteorologiques) global atmosphere-ocean-sea ice coupled model (CNRM-CM3) was used to conduct climate change experiments for the Intergovernmental Panel on Climate Change Fourth Assessment Report (IPCC AR4). The analysis of the A1B scenario experiment shows that global warming leads to a slowdown of North Atlantic deep ocean convection and thermohaline circulation south of Iceland. This slowdown is triggered by a freshening of the Arctic Ocean and an increase in freshwater outflow through Fram Strait. Sea ice melting in the Barents Sea induces a local amplification of the surface warming, which enhances the cyclonic atmospheric circulation around Spitzberg. This anti-clockwise circulation forces an increase in Fram Strait outflow and a simultaneous increase in ocean transport of warm waters toward the Barents Sea, favouring further sea ice melting and surface warming in the Barents Sea. Additionally, the retreat of sea ice allows more deep water formation north of Iceland and the thermohaline circulation strengthens there. The transport of warm and saline waters toward the Barents Sea is further enhanced, which constitutes a second positive feedback. (orig.) 16. Temporal stability of the Dutch version of the Wechsler Memory Scale - Fourth Edition (WMS-IV-NL) NARCIS (Netherlands) Bouman, Z.; Hendriks, M.P.H.; Aldenkamp, A.P.; Kessels, R.P.C. 2015-01-01 Objective: The Wechsler Memory Scale - Fourth Edition (WMS-IV) is one of the most widely used memory batteries. We examined the test–retest reliability, practice effects, and standardized regression-based (SRB) change norms for the Dutch version of the WMS-IV (WMS-IV-NL) after both short and long re 17. Bergman kernel function on Hua construction of the fourth type Institute of Scientific and Technical Information of China (English) 2007-01-01 This paper introduces the Hua construction and presents the holomorphic automorphism group of the Hua construction of the fourth type. Utilizing the Bergman kernel function, under the condition of holomorphic automorphism and the standard complete orthonormal system of the semi-Reinhardt domain, the infinite series form of the Bergman kernel function is derived. By applying the properties of polynomial and Γ functions, various identification relations of the aforementioned form are developed and the explicit formula of the Bergman kernel function for the Hua construction of the fourth type is obtained, which suggest that many of the previously-reported results are only the special cases of our findings. 18. Towards Reviving Electroweak Baryogenesis with a Fourth Generation Directory of Open Access Journals (Sweden) Wei-Shu Hou 2013-01-01 universe. However, it does not work within the standard model due to two reasons: (1 the strength of CP violation from the Kobayashi-Maskawa mechanism with three generations is too small; (2 the electroweak phase transition is not first order for the experimentally allowed Higgs boson mass. We discuss possibilities to solve these problems by introducing a fourth generation of fermions and how electroweak baryogenesis might be revived. We also discuss briefly the recent observation of a Higgs-like boson with mass around 125 GeV, which puts the fourth generation in a difficult situation, and the possible way out. 19. The effect of climate change and natural variability on wind loading values for buildings NARCIS (Netherlands) Steenbergen, R.D.J.M.; Koster, T.; Geurts, C.P.W. 2012-01-01 Since 2006, a number of countries developed reports on climate change following the IPCC 4th assessment reports. For the Netherlands, the Royal Netherlands Meteorological Institute (KNMI) presented four new climate scenarios. Typically, climate change is described in terms of average changes, but mu 20. A Search for the Fourth SM Family: Tevatron still has a Chance CERN Document Server Sahin, M; Turkoz, S 2010-01-01 Existence of the fourth family follows from the basics of the Standard Model. We discuss possible manifestations of the fourth SM family at existing and future colliders. The LHC and Tevatron potentials to discover the fourth SM family have been compared. The scenario with dominance of the anomalous decay modes of the fourth family quarks has been considered in details. 1. Urban focus in climate change adaptation and risk reduction OpenAIRE Wamsler, Christine 2014-01-01 Urban communities will face increased risks, such as floods, landslides, heat stress and fires and water scarcity, as a consequence of climate change. The latest IPCC report (AR5) has for the first time devoted a whole chapter to urban areas. The assessment stresses the need to tackle urban risk through more effective adaptation planning. 2. Fourth Generation Instructional Design Model: An Elaboration on Authoring Activities. Science.gov (United States) Christensen, Dean L. This paper presents the updated (fourth generation) version of the instructional design (ID) model, noting its emphasis on a scientific, iterative approach based upon research and theory in learning and instruction and upon applied development experience. Another important trend toward a scientific approach to instructional design is the increased… 3. Integrating Fourth-Generation Tools Into the Applications Development Environment. Science.gov (United States) Litaker, R. G.; And Others 1985-01-01 Much of the power of the "information center" comes from its ability to effectively use fourth-generation productivity tools to provide information processing services. A case study of the use of these tools at Western Michigan University is presented. (Author/MLW) 4. Installing an Integrated System and a Fourth-Generation Language. Science.gov (United States) Ridenour, David; Ferguson, Linda 1987-01-01 In the spring of 1986 Indiana State University converted to the Series Z software of Information Associates, an IBM mainframe, and Information Builders' FOCUS fourth-generation language. The beginning of the planning stage to product selection, training, and implementation is described. (Author/MLW) 5. Transactions of the fourth symposium on space nuclear power systems Energy Technology Data Exchange (ETDEWEB) El-Genk, M.S.; Hoover, M.D. (eds.) 1987-01-01 This paper contains the presented papers at the fourth symposium on space nuclear power systems. Topics of these papers include: space nuclear missions and applications, reactors and shielding, nuclear electric and nuclear propulsion, refractory alloys and high-temperature materials, instrumentation and control, energy conversion and storage, space nuclear fuels, thermal management, nuclear safety, simulation and modeling, and multimegawatt system concepts. (LSP) 6. Constraints on Majorana dark matter from a fourth lepton family DEFF Research Database (Denmark) Hapola, T.; Jarvinen, M.; Kouvaris, C.; 2014-01-01 We study the possibility of dark matter in the form of heavy neutrinos from a fourth lepton family with helicity suppressed couplings such that dark matter is produced thermally via annihilations in the early Universe. We present all possible constraints for this scenario coming from LHC...... account for the dark matter abundance.... 7. Dietary Behaviors among Fourth Graders: A Social Cognitive Theory Approach. Science.gov (United States) Corwin, Sara J.; Sargent, Roger G.; Rheaume, Carol E.; Saunders, Ruth P. 1999-01-01 Examined the impact of behavioral, personal, and environmental factors on fourth graders' dietary practices, using a social cognitive theory framework. Survey results highlighted social cognitive variables that significantly influenced dietary behaviors: gender, race, socioeconomic status, fruit and vegetable availability at home, nutrition… 8. Fourth annual conference on materials for coal conversion and utilization Energy Technology Data Exchange (ETDEWEB) 1979-01-01 The fourth annual conference on materials for coal conversion and utilization was held October 9 to 11, 1979, at the National Bureau of Standards, Gaithersburg, Maryland. It was sponsored by the National Bureau of Standards, the Electric Power Research Institute, the US Department of Energy, and the Gas Research Institute. The papers have been entered individually into EDB and ERA. (LTN) 9. Fourth annual report to Congress, Federal Alternative Motor Fuels Programs Energy Technology Data Exchange (ETDEWEB) NONE 1995-07-01 This annual report to Congress presents the current status of the alternative fuel vehicle programs being conducted across the country in accordance with the Alternative Motor Fuels Act of 1988. These programs, which represent the most comprehensive data collection effort ever undertaken on alternative fuels, are beginning their fifth year. This report summarizes tests and results from the fourth year. 10. Fourth-order discrete anisotropic boundary-value problems Directory of Open Access Journals (Sweden) Maciej Leszczynski 2015-09-01 Full Text Available In this article we consider the fourth-order discrete anisotropic boundary value problem with both advance and retardation. We apply the direct method of the calculus of variations and the mountain pass technique to prove the existence of at least one and at least two solutions. Non-existence of non-trivial solutions is also undertaken. 11. Image Inpainting Using a Fourth-Order Total Variation Flow OpenAIRE Schönlieb, Carola-Bibiane; Bertozzi, Andrea; Burger, Martin; He, Lin 2009-01-01 International audience We introduce a fourth-order total variation flow for image inpainting proposed in [5]. The well-posedness of this new inpainting model is discussed and its efficient numerical realization via an unconditionally stable solver developed in [15] is presented. 12. Short-term energy outlook: Quarterly projections, fourth quarter 1997 Energy Technology Data Exchange (ETDEWEB) NONE 1997-10-14 The Energy Information Administration (EIA) prepares quarterly short-term energy supply, demand, and price projections for printed publication in January, April, July, and October in the Short-Term Energy Outlook. The details of these projections, as well as monthly updates on or about the 6th of each interim month, are available on the internet at: www.eia.doe.gov/emeu/steo/pub/contents.html. The forecast period for this issue of the Outlook extends from the fourth quarter of 1997 through the fourth quarter of 1998. Values for the fourth quarter of 1997, however, are preliminary EIA estimates (for example, some monthly values for petroleum supply and disposition are derived in part from weekly data reported in EIAs Weekly Petroleum Status Report) or are calculated from model simulations that use the latest exogenous information available (for example, electricity sales and generation are simulated by using actual weather data). The historical energy data, compiled in the fourth quarter 1997 version of the Short-Term Integrated Forecasting System (STIFS) database, are mostly EIA data regularly published in the Monthly Energy Review, Petroleum Supply Monthly, and other EIA publications. Minor discrepancies between the data in these publications and the historical data in this Outlook are due to independent rounding. The STIFS model is driven principally by three sets of assumptions or inputs: estimates of key macroeconomic variables, world oil price assumptions, and assumptions about the severity of weather. 19 tabs. 13. A curve flow evolved by a fourth order parabolic equation Institute of Scientific and Technical Information of China (English) LIU YanNan; JIAN HuaiYu 2009-01-01 We study a fourth order curve flow,which is the gradient flow of a functional describing the shapes of human red blood cells.We prove that for any smooth closed initial curve in R2,the flow has a smooth solution for all time and the solution subconverges to a critical point of the functional. 14. Did that Dog Sniff Violate the Fourth Amendment? Science.gov (United States) Hawke, Catherine; Middleton, Tiffany 2012-01-01 Is sniffing at the front door of a private home by a trained narcotics detection dog a Fourth Amendment search requiring probable cause? Is a "drug dog" somehow like a manmade technology, such as a thermal imaging device? These were a couple of the questions recently presented to the U.S. Supreme Court during arguments in "Florida v. Jardines."… 15. North Carolina Tales Fly with Fourth Grade Tellers Science.gov (United States) Westman, Gretchen Daub 2008-01-01 In fourth grade, North Carolina students are required to write their own personal narratives. The teachers felt that telling a story would be a great stepping stone toward writing one. Rather than focusing on grammar and the mechanics of writing, students could focus on story development and creativity. In this article, the author describes how… 16. Engineering alumni present fourth$10,000 prize
OpenAIRE
Nystrom, Lynn A.
2005-01-01
A group of seven Virginia Tech alumni, in conjunction with the university's College of Engineering, has awarded its fourth 10,000 prize to university's Department of Computer Science in recognition of improvements to the department's Ph.D. program during the past year. 17. Discrimination Evidence for Examining Fourth Grade Students' Learning Disability Problems Science.gov (United States) Hassan, Abdulhameed S.; Al-Harthy, Ibrahim S. 2014-01-01 This study investigated the ability of discriminate variables (perceptual-motor, hyperactivity disorder, neurological and psychological skills) to distinguish between normal (n = 68) and students with learning disabilities (n = 72) in fourth grade. Three instruments were developed: perceptual-motor scale, hyperactivity disorder scale, skills test… 18. The Value of the Fourth Year of Mathematics. Math Works Science.gov (United States) Achieve, Inc., 2013 2013-01-01 Too many students and educators view the senior year and graduation from high school as an end point, rather than one vital step along the education pipeline. Students who engage in a fourth year of math tap into and build upon their advanced analytic skills and are more likely to have better success in postsecondary course work, as they have… 19. Gender Differences in Inference Generation by Fourth-Grade Students Science.gov (United States) Clinton, Virginia; Seipel, Ben; Broek, Paul; McMaster, Kristen L.; Kendeou, Panayiota; Carlson, Sarah E.; Rapp, David N. 2014-01-01 The purpose of this study was to determine if there are gender differences among elementary school-aged students in regard to the inferences they generate during reading. Fourth-grade students (130 females; 126 males) completed think-aloud tasks while reading one practice and one experimental narrative text. Females generated a larger number and a… 20. A curve flow evolved by a fourth order parabolic equation Institute of Scientific and Technical Information of China (English) 2009-01-01 We study a fourth order curve flow, which is the gradient flow of a functional describing the shapes of human red blood cells. We prove that for any smooth closed initial curve in R2, the flow has a smooth solution for all time and the solution subconverges to a critical point of the functional. 1. Climate change: The necessary, the possible and the desirable Earth League climate statement on the implications for climate policy from the 5th IPCC Assessment OpenAIRE Rockstrom, J.; Brasseur, G; Hoskins, B.; Lucht, W.; Schellnhuber, J.; P. Kabat; Nakicenovic, N.; P. Gong; P. Schlosser; Costa, M; Humble, A.; Eyre, N.; Gleick, P.; James, R.; Lucena, A. 2014-01-01 The development of human civilisations has occurred at a time of stable climate. This climate stability is now threatened by human activity. The rising global climate risk occurs at a decisive moment for world development. World nations are currently discussing a global development agenda consequent to the Millennium Development Goals (MDGs), which ends in 2015. It is increasingly possible to envisage a world where absolute poverty is largely eradicated within one generation and where ambitio... 2. Climate change: the necessary, the possible and the desirable Earth League climate statement on the implications for climate policy from the 5th IPCC Assessment OpenAIRE Rockström, Johan; Brasseur, Guy; Hoskins, Brian; Lucht, Wolfgang; Schellnhuber, John; Kabat, Pavel; Nakicenovic, Nebojsa; Gong, Peng; Schlosser, Peter; Máñez Costa, Maria; Humble, April; Eyre, Nick; Gleick, Peter; James, Rachel; Lucena, Andre 2014-01-01 The development of human civilisations has occurred at a time of stable climate. This climate stability is now threatened by human activity. The rising global climate risk occurs at a decisive moment for world development. World nations are currently discussing a global development agenda consequent to the Millennium Development Goals (MDGs), which ends in 2015. It is increasingly possible to envisage a world where absolute poverty is largely eradicated within one generation and where ambitio... 3. The changing world of climate change: Oregon leads the states International Nuclear Information System (INIS) Following on the heels of recent national and international developments in climate change policy, Oregon's open-quote best-of-batch close-quote proceeding has validated the use of CO2 offsets as a cost-effective means of advancing climate change mitigation goals. The proceeding was a first in several respects and represents a record commitment of funds to CO2 mitigation by a private entity. In December 1995, the Intergovernmental Panel on Climate Change (IPCC), issued its Second Assessment Report. The IPCC's conclusion that open-quotes[t]he balance of evidence suggests a discernible human influence on global climateclose quotes fundamentally changed the tenor of the policy debate regarding potential threats associated with global climate change. At the Climate Change Convention's Conference of the Parties (COP) in Geneva in July 1996, most countries, including the United States, advocated adopting the IPCC report as the basis for swift policy movement toward binding international emissions targets. The next COP, in December 1997, is scheduled to be the venue for the signing of a treaty protocol incorporating such targets. Binding targets would have major consequences for power plant operators in the US and around the world. Recent developments in the state of Oregon show the kinds of measures that may become commonplace at the state level in addressing climate change mitigation. First, Oregon recently completed the first administrative proceeding in the US aimed at offsetting the greenhouse gas emissions of a new power plant. Second, a legislatively mandated energy facility siting task force recently recommended that Oregon adopt a carbon dioxide (CO2) standard for new power plant construction and drop use of the open-quotes need for powerclose quotes standard. This article reviews these two policy milestones and their implications for climate change mitigation in the United States 4. Climate variations in the Northern Hemisphere based on the use of an atmosphere-ocean IPCC model International Nuclear Information System (INIS) Forced and natural variability of modelled and observed Atlantic Ocean temperature and Atlantic Meridional Overturning Circulation (AMOC) is studied. In the observations and in a forced climate model run, we find increasing temperature at 1000m in the Atlantic (20N). SVD analysis shows that, for both model data and observations, a high index of North Atlantic Oscillation (NAO) correspond to negative temperature anomaly at 1000m to the north of 55N, although geographical details of temperature anomaly distribution are different for the model and observations. Particular attention has been paid to the influence of the fresh water flux due to the present global warming on the slowing down of the AMOC. It is shown that fresh water flux change is only a secondary cause of reduced AMOC in global warming conditions, while heat flux change is probably the main reason. Finally, it is shown that internal model AMOC variability is positively correlated with the near-surface air temperature in Atlantic-European Arctic sector on a 10-year time scale. 5. Climate change and marine life DEFF Research Database (Denmark) Richardson, Anthony J.; Brown, Christopher J.; Brander, Keith; 2012-01-01 A Marine Climate Impacts Workshop was held from 29 April to 3 May 2012 at the US National Center of Ecological Analysis and Synthesis in Santa Barbara. This workshop was the culmination of a series of six meetings over the past three years, which had brought together 25 experts in climate change...... ecology, analysis of large datasets, palaeontology, marine ecology and physical oceanography. Aims of these workshops were to produce a global synthesis of climate impacts on marine biota, to identify sensitive habitats and taxa, to inform the current Intergovernmental Panel on Climate Change (IPCC......) process, and to strengthen research into ecological impacts of climate change... 6. Methotrexate administration directly into the fourth ventricle in children with malignant fourth ventricular brain tumors: a pilot clinical trial. Science.gov (United States) Sandberg, David I; Rytting, Michael; Zaky, Wafik; Kerr, Marcia; Ketonen, Leena; Kundu, Uma; Moore, Bartlett D; Yang, Grace; Hou, Ping; Sitton, Clark; Cooper, Laurence J; Gopalakrishnan, Vidya; Lee, Dean A; Thall, Peter F; Khatua, Soumen 2015-10-01 We hypothesize that chemotherapy can be safely administered directly into the fourth ventricle to treat recurrent malignant brain tumors in children. For the first time in humans, methotrexate was infused into the fourth ventricle in children with recurrent, malignant brain tumors. A catheter was surgically placed into the fourth ventricle and attached to a ventricular access device. Cerebrospinal fluid (CSF) flow was confirmed by CINE MRI postoperatively. Each cycle consisted of 4 consecutive daily methotrexate infusions (2 milligrams). Disease response was monitored with serial MRI scans and CSF cytologic analysis. Trough CSF methotrexate levels were sampled. Five patients (3 with medulloblastoma and 2 with ependymoma) received 18, 18, 12, 9, and 3 cycles, respectively. There were no serious adverse events or new neurological deficits attributed to methotrexate. Two additional enrolled patients were withdrawn prior to planned infusions due to rapid disease progression. Median serum methotrexate level 4 h after infusion was 0.04 µmol/L. Range was 0.02-0.13 µmol/L. Median trough CSF methotrexate level 24 h after infusion was 3.18 µmol/L (range 0.53-212.36 µmol/L). All three patients with medulloblastoma had partial response or stable disease until one patient had progressive disease after cycle 18. Both patients with ependymoma had progressive disease after 9 and 3 cycles, respectively. Low-dose methotrexate can be infused into the fourth ventricle without causing neurological toxicity. Some patients with recurrent medulloblastoma experience a beneficial anti-tumor effect both within the fourth ventricle and at distant sites. 7. The understanding of world climate change; Les connaissances sur le changement climatique mondial Energy Technology Data Exchange (ETDEWEB) Petit, M. 2008-07-01 After having recalled that the problem of global warming in relationship with human activities has been studied since the end of the nineteenth century and since then by different scientific programs, the author describes how the IPCC's or Intergovernmental Panel on Climate Change's report is produced. He briefly comments how Earth's temperature is determined and the various natural parameters which influence the climate on Earth. He recalls how the IPCC showed the actual influence of human activities, and which changes have actually been observed 8. Field assessment of semi-aerobic condition and the methane correction factor for the semi-aerobic landfills provided by IPCC guidelines Energy Technology Data Exchange (ETDEWEB) Jeong, Sangjae [Department of Civil and Environmental Engineering, College of Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 151-742 (Korea, Republic of); Nam, Anwoo [Korea Environment Corporation, 42 Hwangyeong-ro, Seo-gu, Incheon 404-170 (Korea, Republic of); Yi, Seung-Muk [Department of Environmental Health, School of Public Health, Seoul National University, Seoul 151-742 (Korea, Republic of); Kim, Jae Young, E-mail: [email protected] [Department of Civil and Environmental Engineering, College of Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 151-742 (Korea, Republic of) 2015-02-15 Highlights: • CH{sub 4}/CO{sub 2} and CH{sub 4} + CO{sub 2}% are proposed as indices to evaluate semi-aerobic landfills. • A landfill which CH{sub 4}/CO{sub 2} > 1.0 is difficult to be categorized as semi-aerobic landfill. • Field conditions should be carefully investigated to determine landfill types. • The MCF default value for semi-aerobic landfills underestimates the methane emissions. - Abstract: According to IPCC guidelines, a semi-aerobic landfill site produces one-half of the amount of CH{sub 4} produced by an equally-sized anaerobic landfill site. Therefore categorizing the landfill type is important on greenhouse gas inventories. In order to assess semi-aerobic condition in the sites and the MCF value for semi-aerobic landfill, landfill gas has been measured from vent pipes in five semi-aerobically designed landfills in South Korea. All of the five sites satisfied requirements of semi-aerobic landfills in 2006 IPCC guidelines. However, the ends of leachate collection pipes which are main entrance of air in the semi-aerobic landfill were closed in all five sites. The CH{sub 4}/CO{sub 2} ratio in landfill gas, indicator of aerobic and anaerobic decomposition, ranged from 1.08 to 1.46 which is higher than the values (0.3–1.0) reported for semi-aerobic landfill sites and is rather close to those (1.0–2.0) for anaerobic landfill sites. The low CH{sub 4} + CO{sub 2}% in landfill gas implied air intrusion into the landfill. However, there was no evidence that air intrusion has caused by semi-aerobic design and operation. Therefore, the landfills investigated in this study are difficult to be classified as semi-aerobic landfills. Also MCF of 0.5 may significantly underestimate methane emissions compared to other researches. According to the carbon mass balance analyses, the higher MCF needs to be proposed for semi-aerobic landfills. Consequently, methane emission estimate should be based on field evaluation for the semi-aerobically designed landfills. 9. facing the challenges of climate change and food security : the role of research, extension and communication for development NARCIS (Netherlands) Leeuwis, C.; Hall, A.; Weperen, van W.; Preissing, J. 2013-01-01 In line with the Intergovernmental Panel on Climate Change (IPCC) this study defines climate change as any change in climate over time, whether due to natural variability or as a result of human activity. This report is a shortened version of the final study report, produced on request of FAO. The p 10. The fourth edition of the ASP took place in Kigali CERN Multimedia CERN Bulletin 2016-01-01 The fourth biennial African School of Fundamental Physics and Applications (ASP) took place on 1-19 August in Africa’s cleanest city, Kigali in Rwanda. The students of the fourth African School of Fundamental Physics and Applications pose for the traditional conference group picture. Many lecturers flew in from CERN to give lectures and mentor students. (Photo: Gilbert Tekoute) Seventy-five students from around the African continent, chosen from 439 applicants, were hosted in the University of Rwanda’s College of Sciences and Technology for about 3 weeks. The school received financial support from CERN and 19 other institutions in total, including the International Centre for Theoretical Physics (ICTP), Brookhaven National Laboratory, the South African National Research Foundation and Department of Technology, the Rwandan Ministry of Education, INFN, and other major particle physics laboratories, as well as governmental institutions in Africa, Europe and the United States.... 11. Spinal accessory nerve schwannomas masquerading as a fourth ventricular lesion Directory of Open Access Journals (Sweden) Shyam Sundar Krishnan 2015-01-01 Full Text Available Schwannomas are benign lesions that arise from the nerve sheath of cranial nerves. The most common schwannomas arise from the 8 th cranial nerve (the vestibulo-cochlear nerve followed by trigeminal and facial nerves and then from glossopharyngeal, vagus, and spinal accessory nerves. Schwannomas involving the oculomotor, trochlear, abducens and hypoglossal nerves are very rare. We report a very unusual spinal accessory nerve schwannoma which occupied the fourth ventricle and extended inferiorly to the upper cervical canal. The radiological features have been detailed. The diagnostic dilemma was due to its midline posterior location mimicking a fourth ventricular lesion like medulloblastoma and ependymoma. Total excision is the ideal treatment for these tumors. A brief review of literature with tabulations of the variants has been listed. 12. Estimation of the fourth-order dispersion coefficient β4 Institute of Scientific and Technical Information of China (English) Jing Huang; Jianquan Yao 2012-01-01 The fourth-order dispersion coefficient of fibers are estimated by the iterations around the third-order dispersion and the high-order nonlinear items in the nonlinear Schordinger equation solved by Green's function approach.Our theoretical evaluation demonstrates that the fourth-order dispersion coefficient slightly varies with distance.The fibers also record β4 values of about 0.002,0.003,and 0.00032 ps4/km for SMF,NZDSF and DCF,respectively.In the zero-dispersion regime,the high-order nonlinear effect (higher than self-steepening) has a strong impact on the transmitted short pulse.This red-shifts accelerates the symmetrical split of the pulse,although this effect is degraded rapidly with the increase of β2.Thus,the contributions to β4 of SMF,NZDSF,and DCF can be neglected. 13. New Efficient Fourth Order Method for Solving Nonlinear Equations Directory of Open Access Journals (Sweden) Farooq Ahmad 2013-12-01 Full Text Available In a paper [Appl. Math. Comput., 188 (2 (2007 1587--1591], authors have suggested and analyzed a method for solving nonlinear equations. In the present work, we modified this method by using the finite difference scheme, which has a quintic convergence. We have compared this modified Halley method with some other iterative of fifth-orders convergence methods, which shows that this new method having convergence of fourth order, is efficient. 14. Fourth Generation Nuclear Weapons: Military effectiveness and collateral effects OpenAIRE Gsponer, Andre 2005-01-01 The paper begins with a general introduction and update to Fourth Generation Nuclear Weapons (FGNW), and then addresses some particularly important military aspects on which there has been only limited public discussion so far. These aspects concern the unique military characteristics of FGNWs which make them radically different from both nuclear weapons based on previous-generation nuclear-explosives and from conventional weapons based on chemical-explosives: yields in the 1 to 100 tons rang... 15. Proceedings of the fourth workshop on grand unification Energy Technology Data Exchange (ETDEWEB) Weldon, H.A.; Langacker, P.; Steinhardt, P.J. 1983-01-01 This book compiles the papers presented at the fourth conference of grand unified theories of nuclear physics held in University of Pennsylvania April 1983. The topics covered were proton decay theory; angular distribution and flux of atmospheric neutrinos; atmospheric neutrinos and astrophysical neutrinos in proton decay experiments; review of future nucleon decay experiments; monopole experiments; searches for magnetic monopole; monopoles, gauge, fields and anomalies; darkmatter, galaxies and voids; adiabatic fluctuations; supersymmetry, supergravity, and Kaluza-Klein theories; superstring theory and superunification. 16. Fourth Beijing Human Rights Forum Held in Beijing Institute of Scientific and Technical Information of China (English) OUR STAFF REPORTER 2012-01-01 The Fourth Beijing Forum on Human Rights was held in Beijing from September 21-23,2011.Jointly sponsored by the China Society for Human Rights Studies and the China Human Rights Development Foundation,the forum was centered on the theme of "Cultural Tradition,Concept of Values and Human Rights." Attending were nearly 100senior human rights officials,specialists and scholars from 26 countries and regions as well as the United Nations and other international organizations. 17. Applied Meteorology Unit (AMU) Quarterly Report Fourth Quarter FY-04 Science.gov (United States) Bauman, William; Wheeler, Mark; Lambert, Winifred; Case, Jonathan; Short, David 2004-01-01 This report summarizes the Applied Meteorology Unit (A MU) activities for the fourth quarter of Fiscal Year 2004 (July -Sept 2004). Tasks covered are: (1) Objective Lightning Probability Forecast: Phase I, (2) Severe Weather Forecast Decision Aid, (3) Hail Index, (4) Shuttle Ascent Camera Cloud Obstruction Forecast, (5) Advanced Regional Prediction System (ARPS) Optimization and Training Extension and (5) User Control Interface for ARPS Data Analysis System (ADAS) Data Ingest. 18. Pseudo-randomness of the fourth class of GSS sequences Institute of Scientific and Technical Information of China (English) HU Yupu; XIAO Guozhen 2004-01-01 This paper discusses pseudo-randomness of a periodic sequence, named the fourth class of GSS sequence. We get the ollowing results: ① Its least period always reaches the maximum (that is, 2n-1). ② Its least period and linear complexity keep robust under single-symbol-substitution. ③ It has good Iow-degree-auto-correlation feature. ④It has good short-length-run-distribution. 19. Fourth China-South Asia International Cultural Forum Institute of Scientific and Technical Information of China (English) 2013-01-01 <正>The Fourth China-South Asia International Cultural Forum on "revitalizing people-to-people cultural exchanges for peace and prosperity",co-sponsored by the CPAFFC and Shenzhen University,was held in Shenzhen from November 14 to 17,2012. More than 60 experts and scholars from over 40 Chinese research institutes and their colleagues from India,the United States and Singapore engaged in in-depth discussions on economic cooperation and cultural development,the present condi- 20. A rare case of fourth consecutive fallopian tube ectopic pregnancy Directory of Open Access Journals (Sweden) Divya Pandey 2015-04-01 Full Text Available We present a case of consecutive fourth tubal ectopic in a patient which was managed by emergency laparotomy and salpingectomy. The first two tubal ectopic pregnancies occurred on the right side while the subsequent two occurred on the contralateral side. The patient was planned for IVF-ET. [Int J Reprod Contracept Obstet Gynecol 2015; 4(2.000: 490-492 1. The Fourth Attempt to Construct a Politics of Welfare Obligations OpenAIRE Fitzpatrick, Tony 2005-01-01 Since the 1980s there have been three main attempts to ground citizenship upon the principles of duty, obligation and responsibility: conservative, communitarian and Third Way. Each of these are reviewed below. The principal task of this article, though, is to examine the emergence of a fourth attempt which, by relating duty to equality through the principle of reciprocity, represents a synthesis of traditional social democracy with the new politics of obligation. Our focus will be upon The... 2. Final Report for CORBA for Fourth Generation Language Energy Technology Data Exchange (ETDEWEB) Svetlana Shasharina 2005-06-28 The standard for object based networking is the Common Object Request Broker Architecture (CORBA). However, CORBA is not available for Fourth Generation Languages (4GL's) such as Visual Numerics? PV-WAVE or Research Systems? Interactive Data Language (RSI-IDL), which are widely used by scientists and engineers for data visualization and analysis. The proposed work would provide a set of tools to allow 4GL's to interoperate with CORBA. 3. Fourth China-Arab Friendship Conference Held in Yinchuan Institute of Scientific and Technical Information of China (English) 2012-01-01 <正>The Fourth China-Arab Friendship Conference (CAFC) with the theme of "promoting cooperation through friendship and pursuing development through exchanges", jointly sponsored by the CPAFFC, the China-Arab Friendship Association (CAFA) and the League of Arab States (LAS) and hosted by the People’s Government of the Ningxia Hui Autonomous Region,was held in Yinchuan from September 13 to 14, 2012.Before the opening ceremony,Vice Premier Li Keqiang met 4. Regional forecasting with global atmospheric models; Fourth year report Energy Technology Data Exchange (ETDEWEB) Crowley, T.J.; North, G.R.; Smith, N.R. [Applied Research Corp., College Station, TX (United States) 1994-05-01 The scope of the report is to present the results of the fourth years work on the atmospheric modeling part of the global climate studies task. The development testing of computer models and initial results are discussed. The appendices contain studies that provide supporting information and guidance to the modeling work and further details on computer model development. Complete documentation of the models, including user information, will be prepared under separate reports and manuals. 5. Fourth Meeting of China-Spain Forum Held in Madrid Institute of Scientific and Technical Information of China (English) 2008-01-01 <正>Chairman Hu Qili and Executive Chairman Chen Haosu of the Chinese Committee of the China-Spain Forum(CSF) led a Chinese delegation to attend the Fourth Meeting of the CSF in Madrid from November 27 to 28,2007.The meeting with the theme of "Harmony,Development and Gaining Win-Win Results" was jointly sponsored by the CPAFFC,the CSF Chinese Committee,the Spanish Ministry of Foreign Affairs and the Spain China Council Foundation. 6. Fourth Regional Meeting: Nuclear Energy in Central Europe, Proceedings International Nuclear Information System (INIS) Fourth Regional Meeting for Nuclear Energy in Central Europe is an annual meeting of the Nuclear Society of Slovenia. The proceedings contain 89 articles from Slovenia, surrounding countries and countries of the Central and Eastern European Region. Topics are: Research Reactors, Reactor Physics, Probabilistic Safety Assessment, Severe Accidents, Ageing and Integrity, Thermal Hydraulics, NPP Operation Experiance, Radioactive Waste Management, Environment and Other Aspects, Public and Nuclear Energy, SG Replacement and Plant Uprating. 7. Framing adaptation: three aspects for climate change risk management International Nuclear Information System (INIS) Full text: Substantial resources are being allocated to adaptation research and implementation. To use these resources wisely, framing the context within which adaptation decisions are made is critical. Three aspects are: Methods for assessing how much climate change to adapt to by when; Understanding the dynamic between different conceptual models for framing adaptation based on: a. Damages increasing proportionally with change, or b. Ricardian models that require adjustments to attain the 'new normal'; Adopting staged management strategies that depend on system status, which may range from business-as-usual to critical. General adaptation requirements and planning horizons need to have already been identified in scoping studies. Planning horizons include both operational and aspirational targets. Incremental adaptation can be informed by an aspirational goal far off into the future, but is undertaken through a shorter term operational approach. The need to anticipate long-term outcomes in advance is most relevant to measures that require large initial planning and investment, those with long lifetimes, or those where potential damages are irreversible and unacceptable. Five major sources of climate change uncertainty are relevant to assessing how much climate change to adapt to by when: ongoing climate variability and rate of change; past and future commitments to climate change; regional climate change projections; climate sensitivity; greenhouse gas emission scenarios and radiative forcing. These factors combine with different levels of importance depending on the relevant planning horizon. Short-term adaptation is most sensitive to the first and second factors, and long-term adaptation to the last three factors. These factors can be assessed within a probabilistic framework. Two conceptual models dominate assessments designed to inform adaptation. The IPCC Third and Fourth Assessment Reports clearly show that a great many risks increase proportionally with 8. Fourth-Order Difference Methods for Hyperbolic IBVPs Science.gov (United States) Gustafsson, Bertil; Olsson, Pelle 1995-03-01 In this paper we consider fourth-order difference approximations of initial-boundary value problems for hyperbolic partial differential equations. We use the method of lines approach with both explicit and compact implicit difference operators in space. The explicit operator satisfies an energy estimate leading to strict stability. For the implicit operator we develop boundary conditions and give a complete proof of strong stability using the Laplace transform technique. We also present numerical experiments for the linear advection equation and Burgers' equation with discontinuities in the solution or in its derivative. The first equation is used for modeling contact discontinuities in fluid dynamics; the second one is used for modeling shocks and rarefaction waves. The time discretization is done with a third-order Runge-Kutta TVD method. For solutions with discontinuities in the solution itself we add a filter based on second-order viscosity. In case of the non-linear Burgers' equation we use a flux splitting technique that results in an energy estimate for certain difference approximations, in which case also an entropy condition is fulfilled. In particular we shall demonstrate that the unsplit conservative form produces a non-physical shock instead of the physically correct rarefaction wave. In the numerical experiments we compare our fourth-order methods with a standard second-order one and with a third-order TVD method. The results show that the fourth-order methods are the only ones that give good results for all the considered test problems. 9. Is a Breakthrough on Climate Change Governance on the Horizon? DEFF Research Database (Denmark) Figueroa, Maria Josefina The recently released Fifth Assessment report of the IPCC has highlighted again with unprecedented scope and insight the urgency of addressing climate change. The international community has pledged to devise the next international agreement on climate change by 2015, while the EU and in particul...... great challenges for developing a pathway from knowledge to governance and climate action that can lead to an effective global response to climate change.......The recently released Fifth Assessment report of the IPCC has highlighted again with unprecedented scope and insight the urgency of addressing climate change. The international community has pledged to devise the next international agreement on climate change by 2015, while the EU and in particular...... the Scandinavian countries have forged ahead advancing a variety of policies to respond to climate change. Similarly, regions, municipalities, and private actors across the world are also contributing to climate governance. This paper asks whether the world is reaching a tipping point where a breakthrough... 10. Teasing out the impacts of climate change on agricultural development OpenAIRE Knox, Jerry W.; Kay, Melvyn G. 2010-01-01 plethora of articles, books, and academic papers. Not least are the detailed and extensive publications of the Inter-Governmental Panel on Climate Change (IPCC) which set out in their latest assessment (AR4), the scientific, technical, and socio-economic information relevant for understanding the risks posed by human- induced climate change, and the policy options for dealing with it. Although it is useful to study and identify the specific benefits and risks of a changing c... 11. Climate change impact on China food security in 2050 OpenAIRE Ye, Liming; Xiong, Wei; Li, Zhengguo; Yang, Peng; Wu, Wenbin; Yang, Guixia; Fu, Yijiang; zou, Jinqiu; Chen, Zhongxin; Van Ranst, Eric; Tang, Huajun 2013-01-01 International audience Climate change is now affecting global agriculture and food production worldwide. Nonetheless the direct link between climate change and food security at the national scale is poorly understood. Here we simulated the effect of climate change on food security in China using the CERES crop models and the IPCC SRES A2 and B2 scenarios including CO2 fertilization effect. Models took into account population size, urbanization rate, cropland area, cropping intensity and te... 12. OSCILLATION CRITERIA FOR A FOURTH ORDER SUBLINEAR DYNAMIC EQUATION ON TIME SCALE Institute of Scientific and Technical Information of China (English) 2011-01-01 Some new criteria for the oscillation of a fourth order sublinear and/or linear dynamic equation on time scale are established. Our results are new for the corresponding fourth order differential equations as well as difference equations. 13. 77 FR 37604 - Safety Zone; Fourth of July Fireworks, Berkeley Marina, Berkeley, CA Science.gov (United States) 2012-06-22 ... SECURITY Coast Guard 33 CFR Part 165 Safety Zone; Fourth of July Fireworks, Berkeley Marina, Berkeley, CA... enforce the safety zone for the Berkeley Marina Fourth of July Fireworks display in the Captain of the... Marina Fourth of July Fireworks display in 33 CFR 165.1191. This safety zone will be in effect from... 14. 78 FR 29022 - Safety Zone; Fourth of July Fireworks, Berkeley Marina, Berkeley, CA Science.gov (United States) 2013-05-17 ... SECURITY Coast Guard 33 CFR Part 165 Safety Zone; Fourth of July Fireworks, Berkeley Marina, Berkeley, CA... enforce the safety zone for the Berkeley Marina Fourth of July Fireworks display in the Captain of the...'19'' W (NAD 83) for the Berkeley Marina Fourth of July Fireworks display listed in 33 CFR... 15. A Search for the Fourth SM Family Fermions and E_6 Quarks at\\mu ^{+}\\mu ^{-}$Colliders CERN Document Server Çiftçi, A K; Sultansoy, S F 2002-01-01 The potential of$\\mu ^{+}\\mu ^{-}$colliders to investigate the fourth SM family fermions predicted by flavour democracy has been analyzed. It is shown that muon colliders are advantageous for both pair production of fourth family fermions and resonance production of fourth family quarkonia. Also isosinglet quarks production at$\\mu ^{+}\\mu ^{-}$colliders has been investigated. 16. 76 FR 37007 - Safety Zone; Stockton Ports Baseball Club Fourth of July Fireworks Display, Stockton, CA Science.gov (United States) 2011-06-24 ... SECURITY Coast Guard 33 CFR Part 165 RIN 1625-AA00 Safety Zone; Stockton Ports Baseball Club Fourth of July... Stockton Ports Baseball Club will sponsor the Stockton Ports Baseball Club Fourth of July Fireworks Display... read as follows: Sec. 165.T11-422 Safety Zone; Stockton Ports Baseball Club Fourth of July... 17. Can the Ethics of the Fourth Estate Persevere in a Global Age? DEFF Research Database (Denmark) Hansen, Ejvind 2014-01-01 for ethically good journalism. I argue that ethical evaluations should focus upon the meeting between normative ideals and factual realities. This meeting is always open because ideals can challenge reality, just as reality can challenge ideals. Ethical questions are thus always raising a fundamental “maybe......”. Traditionally the ideals of journalists have been articulated in close affiliation with ideas of the Fourth Estate. However, due to our globalised communicative structure, this articulation is in need of revision. I argue that the ethical requests change because the structure of Internet-based publics changes....... Departing from this situation I suggest that journalistic products are ethically urgent insofar as they both bring communities together and give voice to the inarticulate or voiceless. I argue that in order to substantiate this approach it is important to articulate rules, because rules further... 18. Fourth-order gravity as the inflationary model revisited CERN Document Server Kaneda, S; Watanabe, N 2010-01-01 We revisit the simplest (fourth-order or quadratically generated) modified gravity model in four space-time dimensions. It is equivalent to the certain quintessence model via a Legendre-Weyl transform. By using the quintessence scalar potential we compute the (CMB) observables of inflation associated with curvature perturbations (namely, the scalar and tensor spectral indices, and the tensor-to-scalar ratio) by using the most recent WMAP5 experimental bound. Our results include the next-to-leading terms with respect to the inverse number of e-foldings. 19. Symmetries, Large Leptonic Mixing and a Fourth Generation CERN Document Server Silva-Marcos, Joaquim I 2002-01-01 We show that large leptonic mixing occurs most naturally in the framework of the Sandard Model just by adding a fourth generation. One can then construct a small$Z_4$discrete symmetry, instead of the large$S_{4L}\\times S_{4R}$, which requires that the neutrino as well as the charged lepton mass matrices be proportional to a$4\\times 4$democratic mass matrix, where all entries are equal to unity. Without considering the see-saw mechanism, or other more elaborate extensions of the SM, and contrary to the case with only 3 generations, large leptonic mixing is obtained when the symmetry is broken. 20. Nanopore-based Fourth-generation DNA Sequencing Technology Institute of Scientific and Technical Information of China (English) Yanxiao Feng; Yuechuan Zhang; Cuifeng Ying; Deqiang Wang; Chunlei Du 2015-01-01 Nanopore-based sequencers, as the fourth-generation DNA sequencing technology, have the potential to quickly and reliably sequence the entire human genome for less than$1000, and possibly for even less than$100. The single-molecule techniques used by this technology allow us to further study the interaction between DNA and protein, as well as between protein and protein. Nanopore analysis opens a new door to molecular biology investigation at the single-molecule scale. In this article, we have reviewed academic achievements in nanopore technology from the past as well as the latest advances, including both biological and solid-state nanopores, and discussed their recent and potential applications. 1. Applied Meteorology Unit (AMU) Quarterly Report - Fourth Quarter FY-09 Science.gov (United States) Bauman, William; Crawford, Winifred; Barrett, Joe; Watson, Leela; Wheeler, Mark 2009-01-01 This report summarizes the Applied Meteorology Unit (AMU) activities for the fourth quarter of Fiscal Year 2009 (July - September 2009). Tasks reports include: (1) Peak Wind Tool for User Launch Commit Criteria (LCC), (2) Objective Lightning Probability Tool. Phase III, (3) Peak Wind Tool for General Forecasting. Phase II, (4) Update and Maintain Advanced Regional Prediction System (ARPS) Data Analysis System (ADAS), (5) Verify MesoNAM Performance (6) develop a Graphical User Interface to update selected parameters for the Hybrid Single-Particle Lagrangian Integrated Trajectory (HYSPLlT) 2. Rulison Site groundwater monitoring report. Fourth quarter, 1997 International Nuclear Information System (INIS) This report summarizes the results of the fourth quarter 1997 groundwater sampling event for the Rulison Site, which is located approximately 65 kilometers (km) (40 miles [mi]) northeast of Grand Junction, Colorado. This is the eighth and final sampling event of a quarterly groundwater monitoring program implemented by the U.S. Department of Energy (DOE). This program monitored the effectiveness of remediation of a drilling effluent pond that had been used to store drilling mud during drilling of the emplacement hole for a 1969 gas stimulation test conducted by the U.S. Atomic Energy Commission (AEC) (the predecessor agency to the DOE) and Austral Oil Company (Austral) 3. Fourth Generation Nuclear Weapons: Military effectiveness and collateral effects CERN Document Server Gsponer, A 2005-01-01 The paper begins with a general introduction and update to Fourth Generation Nuclear Weapons (FGNW), and then addresses some particularly important military aspects on which there has been only limited public discussion so far. These aspects concern the unique military characteristics of FGNWs which make them radically different from both nuclear weapons based on previous-generation nuclear-explosives and from conventional weapons based on chemical-explosives: yields in the 1 to 100 tons range, greatly enhanced coupling to targets, possibility to drive powerful shaped charged jets and forged fragments, enhanced prompt radiation effects, reduced collateral damage and residual radioactivity, etc. 4. Constraints on Majorana Dark Matter from a Fourth Lepton Family CERN Document Server Hapola, Tuomas; Kouvaris, Chris; Panci, Paolo; Virkajarvi, Jussi 2013-01-01 We study the possibility of dark matter in the form of heavy neutrinos from a fourth lepton family with helicity suppressed couplings such that dark matter is produced thermally via annihilations in the early Universe. We present all possible constraints for this scenario coming from LHC and collider physics, underground direct detectors, neutrino telescopes, and indirect astrophysical searches. Although we embed the WIMP candidate within a model of composite dynamics, the majority of our results are model independent and applicable to all models where heavy neutrinos with suppressed couplings account for the dark matter abundance. 5. Proceedings Fourth Workshop on Mathematically Structured Functional Programming CERN Document Server Chapman, James; 10.4204/EPTCS.76 2012-01-01 This volume contains the proceedings of the Fourth Workshop on Mathematically Structured Functional Programming (MSFP 2012), taking place on 25 March, 2012 in Tallinn, Estonia, as a satellite event of the European Joint Conferences on Theory and Practice of Software, ETAPS 2012. MSFP is devoted to the derivation of functionality from structure. It highlights concepts from algebra, semantics and type theory as they are increasingly reflected in programming practice, especially functional programming. The workshop consists of two invited presentations and eight contributed papers on a range of topics at that interface. 6. The virtual return of "fourth" and "fifth" generation migrants Directory of Open Access Journals (Sweden) Gredelj Stjepan 2006-01-01 Full Text Available The main aims of the project are twofold: the first one is to get insight into scope, structure, everyday life, opinions and plans of our people who live abroad as the fourth and fifth generations of emigrants, specially those who left the country during 90s. The second is checking and recording their preparedness for "return" to mother-country through complex set of activities and arrangements: return (repatriation, capital investments, know-how and skills investments, preservation and strengthening cultural identity of our people abroad and development their links with the ethnic/cultural background and inheritance. 7. Environmental health risk assessment and management for global climate change Science.gov (United States) Carter, P. 2014-12-01 This environmental health risk assessment and management approach for atmospheric greenhouse gas (GHG) pollution is based almost entirely on IPCC AR5 (2014) content, but the IPCC does not make recommendations. Large climate model uncertainties may be large environmental health risks. In accordance with environmental health risk management, we use the standard (IPCC-endorsed) formula of risk as the product of magnitude times probability, with an extremely high standard of precaution. Atmospheric GHG pollution, causing global warming, climate change and ocean acidification, is increasing as fast as ever. Time is of the essence to inform and make recommendations to governments and the public. While the 2ºC target is the only formally agreed-upon policy limit, for the most vulnerable nations, a 1.5ºC limit is being considered by the UNFCCC Secretariat. The Climate Action Network International (2014), representing civil society, recommends that the 1.5ºC limit be kept open and that emissions decline from 2015. James Hansen et al (2013) have argued that 1ºC is the danger limit. Taking into account committed global warming, its millennial duration, multiple large sources of amplifying climate feedbacks and multiple adverse impacts of global warming and climate change on crops, and population health impacts, all the IPCC AR5 scenarios carry extreme environmental health risks to large human populations and to the future of humanity as a whole. Our risk consideration finds that 2ºC carries high risks of many catastrophic impacts, that 1.5ºC carries high risks of many disastrous impacts, and that 1ºC is the danger limit. IPCC AR4 (2007) showed that emissions must be reversed by 2015 for a 2ºC warming limit. For the IPCC AR5 only the best-case scenario RCP2.6, is projected to stay under 2ºC by 2100 but the upper range is just above 2ºC. It calls for emissions to decline by 2020. We recommend that for catastrophic environmental health risk aversion, emissions decline 8. Scale-invariant scalar spectrum from the nonminimal derivative coupling with fourth-order term CERN Document Server Myung, Yun Soo 2015-01-01 An exactly scale-invariant spectrum of scalar perturbation generated during de Sitter spacetime is found from the gravity model of the nonminimal derivative coupling with fourth-order term. The nonminimal derivative coupling term generates a healthy (ghost-free) fourth-order derivative term, while the fourth-order term provides an unhealthy (ghost) fourth-order derivative term. The Harrison-Zel'dovich spectrum obtained from Fourier transforming the fourth-order propagator in de Sitter space is recovered by computing the power spectrum in its momentum space directly. It shows that this model provides a truly scale-invariant spectrum, in addition to the Lee-Wick scalar theory. 9. An automatic seismic signal detection method based on fourth-order statistics and applications Institute of Scientific and Technical Information of China (English) Liu Xi-Qiang; Cai Yin; Zhao Rui; Zhao Yin-Gang; Qu Bao-An; Feng Zhi-Jun; Li Hong 2014-01-01 Real-time, automatic, and accurate determination of seismic signals is critical for rapid earthquake reporting and early warning. In this study, we present a correction trigger function (CTF) for automatically detecting regional seismic events and a fourth-order statistics algorithm with the Akaike information criterion (AIC) for determining the direct wave phase, based on the differences, or changes, in energy, frequency, and amplitude of the direct P- or S-waves signal and noise. Simulations suggest for that the proposed fourth-order statistics result in high resolution even for weak signal and noise variations at different amplitude, frequency, and polarization characteristics. To improve the precision of establishing the S-waves onset,fi rst a specifi c segment of P-wave seismograms is selected and the polarization characteristics of the data are obtained. Second, the S-wave seismograms that contained the specifi c segment of P-wave seismograms are analyzed by S-wave polarizationfi ltering. Finally, the S-wave phase onset times are estimated. The proposed algorithm was used to analyze regional earthquake data from the Shandong Seismic Network. The results suggest that compared with conventional methods, the proposed algorithm greatly decreased false and missed earthquake triggers, and improved the detection precision of direct P- and S-wave phases. 10. Projected climate change and the changing biogeography of coastal Mediterranean fishes OpenAIRE Albouy, C; Guilhaumon, François; Leprieur, F; Lasram, F. B.; Somot, S.; Aznar, R; Velez, Laure; Le Loc'h, François; Mouillot, D. 2013-01-01 Aim To forecast the potential effects of climate change in the Mediterranean Sea on the species richness and mean body size of coastal fish assemblages. Location The Mediterranean Sea. Methods Using an ensemble forecasting approach, we used species distribution modelling to project the potential distribution of 288 coastal fish species by the middle and end of the 21st century based on the IPCC A2 scenario implemented with the Mediterranean climatic model NEMOMED8. Results A mean rise of 1.4 ... 11. Exploring the Linkages between Climate Change and Sustainable Development: A Challenge for Transdisciplinary Research Directory of Open Access Journals (Sweden) Mohan Munasinghe 2001-06-01 Full Text Available In recent years, both sustainable development and climate change have become well known worldwide, and the work of the Intergovernmental Panel on Climate Change (IPCC has also focused on the nexus of these two key topics. The IPCC third assessment report confirms that global mean temperatures will rise 1.5-6 degrees Celsius during the next century. Furthermore, climate change will significantly affect the economic, social, and environmental dimensions of sustainable development, as well as key issues like poverty and equity. Therefore, the IPCC is seeking answers to important questions: how future development patterns will affect climate change; how climate change impacts, adaptation, and mitigation will affect future sustainable development prospects; and how climate change responses might be better integrated into emerging sustainable development strategies. Some key lessons have emerged from these efforts. The IPCC intellectual community has already proved to be quite cohesive and resilient in the face of determined attacks by powerful and well-financed “anti-climate change” lobbies. While addressing sustainable development issues, adaptation and learning within the IPCC have further strengthened the network. First, fresh ideas have been brought in to catalyze change. Transdisciplinary approaches are essential to deal with large-scale, long-term, complex, and interlinked issues like sustainable development and climate change. Second, the disciplinary mix has continued to evolve to meet the challenge. However, crossing disciplinary and cultural boundaries requires sound knowledge of one’s own discipline (especially its limitations, open-mindedness, great patience, and sincere effort on all sides. Third, IPCC internal processes have adjusted to facilitate beneficial changes, while limiting harmful dissension. E-mail has proved to be a powerful, but potentially risky tool. How something is said could be as important as what is said, to 12. Towards a fourth spatial dimension of brain activity. Science.gov (United States) Tozzi, Arturo; Peters, James F 2016-06-01 Current advances in neurosciences deal with the functional architecture of the central nervous system, paving the way for general theories that improve our understanding of brain activity. From topology, a strong concept comes into play in understanding brain functions, namely, the 4D space of a "hypersphere's torus", undetectable by observers living in a 3D world. The torus may be compared with a video game with biplanes in aerial combat: when a biplane flies off one edge of gaming display, it does not crash but rather it comes back from the opposite edge of the screen. Our thoughts exhibit similar behaviour, i.e. the unique ability to connect past, present and future events in a single, coherent picture as if we were allowed to watch the three screens of past-present-future "glued" together in a mental kaleidoscope. Here we hypothesize that brain functions are embedded in a imperceptible fourth spatial dimension and propose a method to empirically assess its presence. Neuroimaging fMRI series can be evaluated, looking for the topological hallmark of the presence of a fourth dimension. Indeed, there is a typical feature which reveal the existence of a functional hypersphere: the simultaneous activation of areas opposite each other on the 3D cortical surface. Our suggestion-substantiated by recent findings-that brain activity takes place on a closed, donut-like trajectory helps to solve long-standing mysteries concerning our psychological activities, such as mind-wandering, memory retrieval, consciousness and dreaming state. PMID:27275375 13. Welcome to pandoraviruses at the 'Fourth TRUC’ club Directory of Open Access Journals (Sweden) Vikas eSharma 2015-05-01 Full Text Available Nucleocytoplasmic large DNA viruses (NCLDVs, or representatives of the proposed order Megavirales, belong to families of giant viruses that infect a broad range of eukaryotic hosts. Megaviruses have been previously described to comprise a fourth monophylogenetic TRUC (Things Resisting Uncompleted Classification together with cellular domains in the universal tree of life. Recently described pandoraviruses have large (1.9-2.5 MB and highly divergent genomes. In the present study, we updated the classification of pandoraviruses and other reported giant viruses. Phylogenetic trees were constructed based on six informational genes. Hierarchical clustering was performed based on a set of informational genes from Megavirales members and cellular organisms. Homologous sequences were selected from cellular organisms using TimeTree software, comprising comprehensive and representative sets of members from Bacteria, Archaea and Eukarya. Phylogenetic analyses based on three conserved core genes clustered pandoraviruses with phycodnaviruses, exhibiting their close relatedness. Additionally, hierarchical clustering analyses based on informational genes grouped pandoraviruses with Megavirales members as a super group distinct from cellular organisms. Thus, the analyses based on core conserved genes revealed that pandoraviruses are new genuine members of the ‘Fourth TRUC’ club, encompassing distinct life forms compared with cellular organisms. 14. Pseudospectral collocation methods for fourth order differential equations Science.gov (United States) Malek, Alaeddin; Phillips, Timothy N. 1994-01-01 Collocation schemes are presented for solving linear fourth order differential equations in one and two dimensions. The variational formulation of the model fourth order problem is discretized by approximating the integrals by a Gaussian quadrature rule generalized to include the values of the derivative of the integrand at the boundary points. Collocation schemes are derived which are equivalent to this discrete variational problem. An efficient preconditioner based on a low-order finite difference approximation to the same differential operator is presented. The corresponding multidomain problem is also considered and interface conditions are derived. Pseudospectral approximations which are C1 continuous at the interfaces are used in each subdomain to approximate the solution. The approximations are also shown to be C3 continuous at the interfaces asymptotically. A complete analysis of the collocation scheme for the multidomain problem is provided. The extension of the method to the biharmonic equation in two dimensions is discussed and results are presented for a problem defined in a nonrectangular domain. 15. Short-Term Energy Outlook: Quarterly projections. Fourth quarter 1993 Energy Technology Data Exchange (ETDEWEB) 1993-11-05 The Energy Information Administration (EIA) prepares quarterly, short-term energy supply, demand, and price projections for publication in February, May, August, and November in the Short-Term Energy Outlook (Outlook). An annual supplement analyzes the performance of previous forecasts, compares recent cases with those of other forecasting services, and discusses current topics related to the short-term energy markets. (See Short-Term Energy Outlook Annual Supplement, DOE/EIA-0202.) The forecast period for this issue of the Outlook extends from the fourth quarter of 1993 through the fourth quarter of 1994. Values for the third quarter of 1993, however, are preliminary EIA estimates (for example, some monthly values for petroleum supply and disposition are derived in part from weekly data reported in the Weekly Petroleum Status Report) or are calculated from model simulations using the latest exogenous information available (for example, electricity sales and generation are simulated using actual weather data). The historical energy data are EIA data published in the Monthly Energy Review, Petroleum Supply Monthly, and other EIA publications. 16. Saving the fourth generation Higgs with radion mixing CERN Document Server Frank, Mariana; Toharia, Manuel 2012-01-01 We study Higgs-radion mixing in a warped extra dimensional model with Standard Model fields in the bulk, and we include a fourth generation of chiral fermions. The main problem with the fourth generation is that, in the absence of Higgs-radion mixing, it produces a large enhancement in the Higgs production cross-section, now severely constrained by LHC data. We analyze the production and decay rates of the two physical states emerging from the mixing and confront them with present LHC data. We show that the current signals observed can be compatible with the presence of one, or both, of these Higgs-radion mixed states (the$\\phi$and the$h$), although with a severely restricted parameter space. In particular, the radion interaction scale must be quite low, Lambda_\\phi ~ 1-1.3 TeV. If m_\\phi ~ 125 GeV, the$h$state must be heavier (m_h>320 GeV). If m_h ~ 125 GeV, the$\\phi$state must be quite light or close in mass (m_\\phi ~ 120 GeV). We also present the modified decay branching ratios of the mixed Higgs-ra... 17. Long-Term Climate Change Commitment and Reversibility: An EMIC Intercomparison DEFF Research Database (Denmark) Zickfeld, K.; Eby, M.; Weaver, A. J.; 2013-01-01 This paper summarizes the results of an intercomparison project with Earth System Models of Intermediate Complexity (EMICs) undertaken in support of the Intergovernmental Panel on Climate Change (IPCC) Fifth Assessment Report (AR5). The focus is on long-term climate projections designed to 1... 18. Assessing Water Management Impacts of Climate Change for a semi-arid Watershed in the Southwestern US Science.gov (United States) Rajagopal, S.; Dominguez, F.; Gupta, H. V.; Castro, C. L.; Troch, P. A. 2011-12-01 Water managers for the City of Phoenix currently face the need to make informed policy decisions regarding long-term impacts of climate change on the Salt-Verde River basin in central Arizona. To provide a scientifically informed basis for this, we estimate the evolution of important components of the basin-scale water balance through the end of the 21st century. Bias-corrected and spatially downscaled climate projections from the Phase-3 Coupled Model Intercomparison Project of the World Climate Research Programme were used to drive a spatially distributed variable infiltration capacity model of the hydrologic processes in the basins. Of the Global Climate Model's participating in the IPCC fourth assessment we selected a five-model ensemble, including the three that best reproduce the historical climatology for our study region, plus two others to represent wetter and drier than model average conditions; the latter two were requested by City of Phoenix water managers to more fully represent the full range of GCM prediction uncertainty. For each GCM, data for three emission scenarios (A1B, A2, B1) was used to drive the hydrologic model into the future. The model projections indicate a 25% statistically significant decrease in streamflow by the end of the 21st century. This is primarily caused by decreased winter precipitation accompanied by significant (temperature driven) reductions in storage of snow. From the analysis of the future period, a synthetic climate dataset was created to reflect future changes in magnitude while preserving the correlation by perturbing the historical observed precipitation and temperature. This dataset was used to evaluate climate elasticity and improve water managers understanding of the impacts to the basin. The results shown in this presentation clearly indicate the manner in which water management in central Arizona is likely to be impacted by changes in regional climate. 19. Communicating Uncertainty about Climate Change for Application to Security Risk Management Science.gov (United States) Gulledge, J. M. 2011-12-01 -management framework for climate security. The IPCC's Fourth Assessment Report concluded that "Responding to climate change involves an iterative risk management process that includes both adaptation and mitigation and takes into account climate change damages, co-benefits, sustainability, equity and attitudes to risk." In risk management, key uncertainties guide action aimed at reducing risk and cannot be ignored or used to justify inaction. Security policies such as arms control and counter-terrorism demonstrate that high-impact outcomes matter to decision makers even if they are likely to be rare events. In spite of this fact, the long tail on the probability distribution of climate sensitivity was largely ignored by the climate science community until recently and its implications for decision making are still not receiving adequate attention. Informing risk management requires scientists to shift from a singular aversion to type I statistical error (i.e. false positive) to a balanced presentation of both type I error and type II error (i.e. false negative) when the latter may have serious consequences. Examples from national security, extreme weather, and economics illustrate these concepts. 20. Measuring Engagement with the Potential Consequences of Climate Change Science.gov (United States) Young, N.; Danielson, R. W.; Lombardi, D. 2015-12-01 Across three studies, we investigated engagement with the consequences of climate change. We drew from the conceptual change and risk analysis literatures to find the factors that determine how much people will care about future risks. Questions derived from these factors were then asked about many hypothesized consequences of climate change. These consequences were drawn from an Intergovernmental Panel on Climate Change special report (IPCC, 2012) and, in the third study, additionally from the IPCC AR5 (IPCC, 2014). The first two studies, using undergraduate students, demonstrated that some consequences were indeed considerably more engaging than others. The third study used a more representative sample of American adults, drawn from Amazon Mechanical Turk and used the Global Warming's Six Americas Screening Tool (Maibach, Leiserowitz, Roser-Renouf, Mertz, & Akerlof, 2011) in a large screening survey to find 20 participants in each of the six audiences defined by this tool. These participants were then asked about the potential consequences of climate change. Results again showed that some consequences are considered more engaging than others, and also showed the ways in which members of these six audiences perceive the consequences of climate change differently. 1. Development of a fourth generation predictive capability maturity model. Energy Technology Data Exchange (ETDEWEB) Hills, Richard Guy; Witkowski, Walter R.; Urbina, Angel; Rider, William J.; Trucano, Timothy Guy 2013-09-01 The Predictive Capability Maturity Model (PCMM) is an expert elicitation tool designed to characterize and communicate completeness of the approaches used for computational model definition, verification, validation, and uncertainty quantification associated for an intended application. The primary application of this tool at Sandia National Laboratories (SNL) has been for physics-based computational simulations in support of nuclear weapons applications. The two main goals of a PCMM evaluation are 1) the communication of computational simulation capability, accurately and transparently, and 2) the development of input for effective planning. As a result of the increasing importance of computational simulation to SNLs mission, the PCMM has evolved through multiple generations with the goal to provide more clarity, rigor, and completeness in its application. This report describes the approach used to develop the fourth generation of the PCMM. 2. Fourth ICT Innovations conference ”Secure and Intelligent Systems” CERN Document Server Gusev, Marjan; ICT Innovations 2012 : Secure and Intelligent Systems 2013-01-01 The present stage of the human civilization is the e-society, which is build over the achievements obtained by the development of the information and communication technologies. It affects everyone, from ordinary mobile phone users to designers of high quality industrial products, and every human activity, from taking medical care to improving the state governing. The science community working in computer sciences and informatics is therefore under constant challenge; it has to solve the new appeared theoretical problem as well as to find new practical solutions. \\\\ The fourth ICT Innovations Conference, held in September 2012 in Ohrid, Macedonia, was one of the several world-wide forums where academics, professionals and practitioners presented their last scientific results and development applications in the fields of high performance and parallel computing, bioinformatics, human computer interaction, security and cryptography, computer and mobile networks, neural networks, cloud computing, process verifica... 3. PRIST: a fourth-generation tool for medical information systems. Science.gov (United States) Cristiani, P; Larizza, C 1990-04-01 PRIST is a fourth-generation software package purposely oriented to development and management of medical applications, running under MS/DOS IBM compatible personal computers. The tool has been developed on the top of DBIII Plus language utilizing the Clipper Compiler networking features for the integration in a LAN environment. Several routines written in C and BASIC Microsoft languages integrated this DBMS-kernel system providing I/O, graphics, statistics, retrieval utilities. To increase the interactivity of the system both menu-driven and windowing interfaces have been implemented. PRIST has been utilized to develop a wide variety of small medical applications ranging from research laboratories to intensive care units. The great majority of reactions from the use of these applications were positive, confirming that PRIST is able to assist in practice management and patient care as well as research purposes. PMID:2345045 4. Wormhole geometries in fourth-order conformal Weyl gravity Science.gov (United States) Varieschi, Gabriele U.; Ault, Kellie L. 2016-04-01 We present an analysis of the classic wormhole geometries based on conformal Weyl gravity, rather than standard general relativity. The main characteristics of the resulting traversable wormholes remains the same as in the seminal study by Morris and Thorne, namely, that effective super-luminal motion is a viable consequence of the metric. Improving on previous work on the subject, we show that for particular choices of the shape and redshift functions the wormhole metric in the context of conformal gravity does not violate the main energy conditions at or near the wormhole throat. Some exotic matter might still be needed at the junction between our solutions and flat spacetime, but we demonstrate that the averaged null energy condition (as evaluated along radial null geodesics) is satisfied for a particular set of wormhole geometries. Therefore, if fourth-order conformal Weyl gravity is a correct extension of general relativity, traversable wormholes might become a realistic solution for interstellar travel. 5. Wormhole geometries in fourth-order conformal Weyl gravity CERN Document Server Varieschi, Gabriele U 2015-01-01 We present an analysis of the classic wormhole geometries based on conformal Weyl gravity, rather than standard general relativity. The main characteristics of the resulting traversable wormholes remain the same as in the seminal study by Morris and Thorne, namely, that effective super-luminal motion is a viable consequence of the metric. Improving on previous work on the subject, we show that for particular choices of the shape and redshift functions, the wormhole metric in the context of conformal gravity does not violate the main energy conditions, as was the case of the original solutions. In particular, the resulting geometry does not require the use of exotic matter at or near the wormhole throat. Therefore, if fourth-order conformal Weyl gravity is a correct extension of general relativity, traversable wormholes might become a realistic solution for interstellar travel. 6. Proceedings of the fourth annual conference on fossil energy materials Energy Technology Data Exchange (ETDEWEB) Judkins, R.R.; Braski, D.N. (comps.) 1990-08-01 The Fourth Annual Conference on Fossil Energy Materials was held in Oak Ridge, Tennessee, on may 15--17, 1990. The meeting was sponsored by the US Department of Energy's Office of Fossil Energy through the Advanced Research and Technology Development (AR TD) Materials Program, and ASM International. The objective of the AR TD Materials Program is to conduct research and development on materials for longer-term fossil energy applications as well as for generic needs of various fossil fuel technologies. The work is divided into the following categories: (1) Ceramics, (2) New Alloys, (3) Corrosion and Erosion, and (4) Technology Assessment and Technology Transfer. Individual projects are processed separately for the data bases. 7. CAS CERN Accelerator School: Fourth general accelerator physics course International Nuclear Information System (INIS) The fourth CERN Accelerator School (CAS) basic course on General Accelerator Physics was given at KFA, Juelich, from 17 to 28 September 1990. Its syllabus was based on the previous similar courses held at Gif-sur-Yvette in 1984, Aarhus 1986, and Salamanca 1988, and whose proceedings were published as CERN Reports 85-19, 87-10, and 89-05, respectively. However, certain topics were treated in a different way, improved or extended, while new subjects were introduced. All of these appear in the present proceedings, which include lectures or seminars on the history and applications of accelerators, phase space and emittance, chromaticity, beam-beam effects, synchrotron radiation, radiation damping, tune measurement, transition, electron cooling, the designs of superconducting magnets, ring lattices, conventional RF cavities and ring RF systems, and an introduction to cyclotrons. (orig.) 8. The Fourth Gravity Test and Quintessence Matter Field CERN Document Server Liu, Molin; Yu, Fei; Gui, Yuanxing 2010-01-01 After the previous work on gravitational frequency shift, light deflection (arXiv:1003.5296) and perihelion advance (arXiv:0812.2332), we calculate carefully the fourth gravity test, i.e. radar echo delay in a central gravity field surrounded by static free quintessence matter, in this paper. Through the Lagrangian method, we find the influence of the quintessence matter on the time delay of null particle is presence by means of an additional integral term. When the quintessence field vanishes, it reduces to the usual Schwarzschild case naturally. Meanwhile, we also use the data of the Viking lander from the Mars and Cassini spacecraft to Saturn to constrain the quintessence field. For the Viking case, the field parameter$\\alpha$is under the order of$10^{-9}$. However,$\\alpha$is under$10^{-18}$for the Cassini case. 9. Minutes of the fourth SALE program participants meeting Energy Technology Data Exchange (ETDEWEB) 1981-10-01 This report is a documentation of the presentations made to the Fourth Safeguards Analytical Laboratory Evaluation (S.A.L.E.) Program Participants Meeting at Argonne, Illinois, July 8-9, 1981. The meeting was sponsored by the US Department of Energy and was coordinated by the S.A.L.E. Program of the New Brunswick Laboratory. The objective of the meeting was to provide a forum through which administration of the Program and methods appropriate to the analysis of S.A.L.E. Program samples could be discussed. The Minutes of the Meeting is a collection of presentations by the speakers at the meeting and of the discussions following the presentations. The presentations are included as submitted by the speakers. The discussion sections were transcribed from tape recordings of the meeting and were edited to clarify and emphasize important comments. Seventeen papers have been abstracted and indexed. 10. The RAdial Velocity Experiment (RAVE): Fourth data release CERN Document Server Kordopatis, G; Steinmetz, M; Boeche, C; Seabroke, G M; Siebert, A; Zwitter, T; Binney, J; de Laverny, P; Recio-Blanco, A; Williams, M E K; Piffl, T; Enke, H; Roeser, S; Bijaoui, A; Wyse, R F G; Freeman, K; Munari, U; Carillo, I; Anguiano, B; Burton, D; Campbell, R; Cass, C J P; Fiegert, K; Hartley, M; Parker, Q A; Reid, W; Ritter, A; Russell, K S; Stupart, M; Watson, F G; Bienayme, O; Bland-Hawthorn, J; Gerhard, O; Gibson, B K; Grebel, E K; Helmi, A; Navarro, J F; Conrad, C; Famaey, B; Faure, C; Just, A; Kos, J; Matijevic, G; McMillan, P J; Minchev, I; Scholz, R; Sharma, S; Siviero, A; de Boer, E Wylie; Zerjal, M 2013-01-01 We present the stellar atmospheric parameters (effective temperature, surface gravity, overall metallicity), radial velocities, individual abundances and distances determined for 425 561 stars, which constitute the fourth public data release of the RAdial Velocity Experiment (RAVE). The stellar atmospheric parameters are computed using a new pipeline, based on the algorithms of MATISSE and DEGAS. The spectral degeneracies and the 2MASS photometric information are now better taken into consideration, improving the parameter determination compared to the previous RAVE data releases. The individual abundances for six elements (magnesium, aluminum, silicon, titanium, iron and nickel) are also given, based on a special-purpose pipeline which is also improved compared to that available for the RAVE DR3 and Chemical DR1 data releases. Together with photometric information and proper motions, these data can be retrieved from the RAVE collaboration website and the Vizier database. 11. Fourth international workshop on human chromosome 5. Final progress report Energy Technology Data Exchange (ETDEWEB) McPherson, J.D. 1996-12-31 The Fourth International Workshop on Human Chromosome 5 was held in Manchester, UK on November 9--10, 1996 and was hosted by the University of Manchester. The major goals of the workshop were: (1) to collate the various genetic, cytogenetic and physical maps of human chromosome 5; (2) to integrate these maps and identify/correct discrepancies between them wherever possible; (3) to catalogue the sequence-ready contigs of the chromosome; (4) to co-ordinate the various sequencing efforts to avoid future duplication; (5) to establish the first (to the authors knowledge) web site for the human chromosome 5 community which contains the above information in a readily accessible form. 12. Statement of incidents at nuclear installations fourth quarter 1985 International Nuclear Information System (INIS) The three incidents which are reported for the fourth quarter of 1985 occurred at Transfynydd and Hinkley Point-B reactors (operated by CEGB) and there is a follow-up report on an earlier incident at Hunterston A reactor (operated by the South of Scotland Electricity Board). The Transfynydd incident concerned a fire in conventional plant and although the reactor was shut down as a precaution, the damage was limited to conventional plant and there was no radiation hazard. The two incidents at Hinkley Point-B concerned a boiler tube leak which allowed high pressure steam into the coolant gas circuit and a leakage of reactor coolant gas which caused the site emergency arrangements to be invoked but which caused no contamination outside the reactor building. The Hunterston-A report concerns the replacement of an active effluent discharge pipeline which had been found to have a leak. (UK) 13. Development of a fourth generation predictive capability maturity model. Energy Technology Data Exchange (ETDEWEB) Hills, Richard Guy; Witkowski, Walter R.; Urbina, Angel; Rider, William J.; Trucano, Timothy Guy 2013-09-01 The Predictive Capability Maturity Model (PCMM) is an expert elicitation tool designed to characterize and communicate completeness of the approaches used for computational model definition, verification, validation, and uncertainty quantification associated for an intended application. The primary application of this tool at Sandia National Laboratories (SNL) has been for physics-based computational simulations in support of nuclear weapons applications. The two main goals of a PCMM evaluation are 1) the communication of computational simulation capability, accurately and transparently, and 2) the development of input for effective planning. As a result of the increasing importance of computational simulation to SNL's mission, the PCMM has evolved through multiple generations with the goal to provide more clarity, rigor, and completeness in its application. This report describes the approach used to develop the fourth generation of the PCMM. 14. A novel circuit architecture for fourth subharmonie mixers Institute of Scientific and Technical Information of China (English) Yao Changfei; Xu Conghai; Zhou Ming; Luo Yunsheng 2012-01-01 A circuit topology for high-order subharmonic (SH) mixers is described.By phase cancellation of idle frequency components,the SH mixer circuit can eliminate the complicated design procedure of idle frequency circuits.Similarly,the SH mixer circuit can achieve a high port isolation by phase cancellation of the leakage LO,RF and idle frequency signals.Based on the high-order SH mixer architecture,a new Ka-band fourth SH mixer is analyzed and designed,it shows the lowest measured conversion loss of 8.3 dB at 38.4 GHz and the loss is lower than 10.3 dB in 34-39 GHz.Measured LO-IF,RF-LO,RF-IF port isolation are better than 30.7 dB,22.9dB and 46.5 dB,respectively. 15. Dynamical Properties in a Fourth-Order Nonlinear Difference Equation Directory of Open Access Journals (Sweden) Xianyi Li 2010-01-01 Full Text Available The rule of trajectory structure for fourth-order nonlinear difference equation xn+1=(xan−2+xn−3/(xan−2xn−3+1, n=0,1,2,…, where a∈[0,1 and the initial values x−3,x−2,x−1,x0∈[0,∞, is described clearly out in this paper. Mainly, the lengths of positive and negative semicycles of its nontrivial solutions are found to occur periodically with prime period 15. The rule is 4+,3−,1+,2−,2+,1−,1+, 1− in a period. By utilizing this rule its positive equilibrium point is verified to be globally asymptotically stable. 16. Nanopore-based Fourth-generation DNA Sequencing Technology Directory of Open Access Journals (Sweden) Yanxiao Feng 2015-02-01 Full Text Available Nanopore-based sequencers, as the fourth-generation DNA sequencing technology, have the potential to quickly and reliably sequence the entire human genome for less than$1000, and possibly for even less than $100. The single-molecule techniques used by this technology allow us to further study the interaction between DNA and protein, as well as between protein and protein. Nanopore analysis opens a new door to molecular biology investigation at the single-molecule scale. In this article, we have reviewed academic achievements in nanopore technology from the past as well as the latest advances, including both biological and solid-state nanopores, and discussed their recent and potential applications. 17. The software factory: A fourth generation software engineering environment Energy Technology Data Exchange (ETDEWEB) Evans, M.W. 1989-01-01 The software-development process and its management are examined in a text intended for engineering managers and students of computer science. A unified concept based on the principle that software design is an engineering science rather than an art is applied, and a software engineering environment (SEE) analogous to an industrial plant is proposed. Chapters are devoted to the classical software environment, the history of software engineering, the evolution of the SEE, the fourth-generation SEE, the engineering process, software-data relationships, the SEE data base, data control in the SEE, software life cycles, information-system product assurance, business management and control, and automating and adapting the SEE. 143 refs. 18. The fourth-generation Water Vapor Millimeter-Wave Spectrometer Science.gov (United States) Gomez, R. Michael; Nedoluha, Gerald E.; Neal, Helen L.; McDermid, I. Stuart 2012-02-01 For 20 years the Naval Research Laboratory has been making continuous water vapor profile measurements at 22.235 GHz with the Water Vapor Millimeter-Wave Spectrometer (WVMS) instruments, with the program expanding from one to three instruments in the first 6 years. Since the initial deployments there have been gradual improvements in the instrument design which have improved data quality and reduced maintenance requirements. Recent technological developments have made it possible to entirely redesign the instrument and improve not only the quality of the measurements but also the capability of the instrument. We present the fourth-generation instrument now operating at Table Mountain, California, which incorporates the most recent advances in microwave radiometry. This instrument represents the most significant extension of our measurement capability to date, enabling us to measure middle atmospheric water vapor from ˜26-80 km. 19. Minutes of the fourth SALE program participants meeting International Nuclear Information System (INIS) This report is a documentation of the presentations made to the Fourth Safeguards Analytical Laboratory Evaluation (S.A.L.E.) Program Participants Meeting at Argonne, Illinois, July 8-9, 1981. The meeting was sponsored by the US Department of Energy and was coordinated by the S.A.L.E. Program of the New Brunswick Laboratory. The objective of the meeting was to provide a forum through which administration of the Program and methods appropriate to the analysis of S.A.L.E. Program samples could be discussed. The Minutes of the Meeting is a collection of presentations by the speakers at the meeting and of the discussions following the presentations. The presentations are included as submitted by the speakers. The discussion sections were transcribed from tape recordings of the meeting and were edited to clarify and emphasize important comments. Seventeen papers have been abstracted and indexed 20. Fourth International Conference on Complex Systems Design & Management CERN Document Server Boulanger, Frédéric; Krob, Daniel; Marchal, Clotilde 2014-01-01 This book contains all refereed papers that were accepted to the fourth edition of the « Complex Systems Design & Management » (CSD&M 2013) international conference which took place in Paris (France) from December 4-6, 2013. These proceedings cover the most recent trends in the emerging field of complex systems sciences & practices from an industrial and academic perspective, including the main industrial domains (transport, defense & security, electronics, energy & environment, e-services), scientific & technical topics (systems fundamentals, systems architecture & engineering, systems metrics & quality, systemic tools) and system types (transportation systems, embedded systems, software & information systems, systems of systems, artificial ecosystems). The CSD&M 2013 conference is organized under the guidance of the CESAMES non-profit organization 1. Black hole shadows in fourth-order conformal Weyl gravity CERN Document Server Mureika, Jonas R 2016-01-01 We calculate the characteristics of the "black hole shadow" for a rotating, neutral black hole in fourth-order conformal Weyl gravity. It is shown that the morphology is not significantly affected by the underlying framework, except for very large masses. Conformal gravity black hole shadows would also significantly differ from their general relativistic counterparts if the values of the main conformal gravity parameters,$\\gamma$and$\\kappa$, were increased by several orders of magnitude. Such increased values for$\\gamma$and$\\kappa$are currently ruled out by gravitational phenomenology. Therefore, it is unlikely that these differences in black hole shadows will be detected in future observations, carried out by the Event Horizon Telescope or others such experiments. 2. 第四核心资源%The Fourth Core Resource Institute of Scientific and Technical Information of China (English) 张基温 2009-01-01 Productivity is the ability of resource exploitation.The improvement of resource exploitation capabilities forms an important indicator of human social development.At present,human society has already witnessed the three phases of land resource,natural resource and the information resource exploitation,and is entering into the era of service resource exploitation.Service embraces some special functions such as optimizing the allocation of resources,relocating social distribution,reducing trade costs,improving product quality and storing the developing power.Every new growth area in great social change comes from the service field.As informationization increases,the service resource will become the fourth core resource in human society after material,energy and information.The service resource exploitation ability is sure to be a sign of advanced productivity and superiority in social competition. 3. Climate change impacts on working people : how to develop prevention policies OpenAIRE Nilsson, Maria; Kjellström, Tord 2010-01-01 The evidence on negative consequences from climate change on human health and well-being is growing. The Intergovernmental Panel on Climate Change (IPCC) described climate change as a threat to the climate system that sets the basis for life and human health conditions. The changing climate is expected to affect basic requirements needed to support and sustain human health such as good food, clean water, and unpolluted air, with negative effects that are expected to be unequally distributed. 4. Invited Editorial: Climate change impacts on working people: how to develop prevention policies OpenAIRE Nilsson, Maria; Kjellstrom, Tord 2010-01-01 The evidence on negative consequences from climate change on human health and well-being is growing. The Intergovernmental Panel on Climate Change (IPCC) described climate change as a threat to the climate system that sets the basis for life and human health conditions. The changing climate is expected to affect basic requirements needed to support and sustain human health such as good food, clean water, and unpolluted air, with negative effects that are expected to be unequally distributed. ... 5. Recent advances on thrombosis and haemostasis in Asian Pacific region: report of the Fourth Asian Pacific Congress on Thrombosis and Haemostasis Institute of Scientific and Technical Information of China (English) WANG Zhao-yue 2006-01-01 @@ The Fourth Asian Pacific Congress on Thrombosis and Haemostasis (APCTH) was held from September 21 to 23, 2006 in Suzhou, China and organized by the Asian Pacific Society on Thrombosis and Haemostasis and the Chinese Society of Haematology with professor RUAN Chang-geng as the chairperson. 6. The fourth shift: exploring the gendered nature of sleep disruption among couples with children. Science.gov (United States) Venn, Susan; Arber, Sara; Meadows, Robert; Hislop, Jenny 2008-03-01 The study of sleep has been neglected within sociology, yet may provide insights into fundamental aspects of the nature of gender inequalities. This article examines how, for couples with children, sleep is influenced by the gendered nature of caring. A key concern is not only who gets up to care for children's physical needs at night, but whether this changes with women's increased role in the labour market. Of concern also is how changes in the nature of caring for older children, as opposed to young children, may impact on parents' sleep. This article analyses qualitative data from an ESRC funded multi-disciplinary project on couples' sleep based on in-depth audio-tape recorded interviews with 26 couples (aged 20-59) with younger and older children. Additionally, one week's audio sleep diaries were completed and follow up in-depth interviews were undertaken with each partner on an individual basis. Physical and emotional care for young children at night was largely provided by women, with a lack of explicit negotiation between partners about who provides this care, even when women return to employment. Thus, considerably more women than men continued their daytime and evening shifts, as well as undertaking an ongoing third shift of sentient activity for their family, into the night. This resulted in a fourth night-time shift where physical caring, and sentient activities continued. As a consequence, women were more likely to subjugate their own sleep needs to those of their family. Fathers did not, in general, undertake this fourth night-time shift. Those that did were more likely to be the fathers of young adult children who were staying out late at night, with the focus of their concerns being the safety of their children. PMID:18321332 7. Weight Loss Composition is One-Fourth Fat-Free Mass: A Critical Review and Critique of This Widely Cited Rule OpenAIRE Heymsfield, Steven B; Cristina Gonzalez, M. C.; Shen, Wei; Redman, Leanne; Thomas, Diana 2014-01-01 Maximizing fat loss while preserving lean tissue mass and function is a central goal of modern obesity treatments. A widely cited rule guiding expected loss of lean tissue as fat-free mass (FFM) states that approximately one-fourth of weight loss will be FFM (i.e., ΔFFM/ΔWeight = ~0.25) with the remaining three-fourths fat mass. This review examines the dynamic relations between FFM, fat mass, and weight changes that follow induction of negative energy balance with hypocaloric dieting and/or ... 8. Opportunities and threats of the fourth industrial revolution and their reflection in the selection of innovative growth strategies Directory of Open Access Journals (Sweden) S.M. Illiashenko 2016-03-01 Full Text Available The aim of the article. Analysis and systematization of possible effects of the fourth industrial revolution both positive and negative, development of recommendations for innovative development strategies formation which would allow the use of opportunities for social and economic growth and will prevent treats. The results of the analysis. Based on a systematic analysis of the literature and practice of management allocated the positive and negative effects of the introduction of innovative technologies (both existing and forecasted, created in line with the fourth industrial revolution. The results of their systematization can be used as basis for the formation of an information base to determine the priorities of innovation. It is shown that distribution of the changes caused by the fourth industrial revolution and the completion of the fifth technological way and the transition to the sixth provide chances to individual institutions and national economies to move to outstripping innovative development. On the example of Ukraine and modern forms of work organization (freelance for the various sectors of activity it is shown that domestic experts have leading positions in global markets, they are successfully implementing the technology generated by the fourth industrial revolution. It demonstrates the significant potential of transition to the sixth technological way. The generalized scheme of formation of Ukrainian innovative development priority directions in line with the concept of technological advance is developed. It is shown that Ukraine has considerable potential for innovation growth which is relevant to the terms of the fourth industrial revolution. In particular, the 2015 world rankings in knowledge and innovation, it had high enough position: knowledge creation – 14; innovation effectiveness – 15; spending on education – 18; number of applications for patents – 19; number of graduates in science and technology – 20 9. A rare case of retained fourth molar teeth in maxilla and mandible. Case report OpenAIRE Rahnama Mansur; Szyszkowska Anna; Pulawska Marta; Szczerba-Gwozdz Joanna 2014-01-01 The study presents a case of the rarely occurring totally retained fourth molar teeth simultaneously in maxilla and mandible. The appearance of supernumerary teeth is a relatively uncommon dental anomaly and it is rare for patients to have impacted fourth molars in two quadrant. The aim of this work is to describe the presence of unilateral (right) fourth molars in the maxilla and the mandible in a young female patient aged 24 years. Orthopantomogram revealed impacted lower third molars but a... 10. Solvability of a fourth order boundary value problem with periodic boundary conditions Directory of Open Access Journals (Sweden) Chaitan P. Gupta 1988-01-01 Full Text Available Fourth order boundary value problems arise in the study of the equilibrium of an elastaic beam under an external load. The author earlier investigated the existence and uniqueness of the solutions of the nonlinear analogues of fourth order boundary value problems that arise in the equilibrium of an elastic beam depending on how the ends of the beam are supported. This paper concerns the existence and uniqueness of solutions of the fourth order boundary value problems with periodic boundary conditions. 11. Solvability of a fourth order boundary value problem with periodic boundary conditions OpenAIRE Chaitan P. Gupta 1988-01-01 Fourth order boundary value problems arise in the study of the equilibrium of an elastaic beam under an external load. The author earlier investigated the existence and uniqueness of the solutions of the nonlinear analogues of fourth order boundary value problems that arise in the equilibrium of an elastic beam depending on how the ends of the beam are supported. This paper concerns the existence and uniqueness of solutions of the fourth order boundary value problems with periodic boundary co... 12. Reply to "Comment on 'Cosmic-ray-driven reaction and greenhouse effect of halogenated molecules: Culprits for atmospheric ozone depletion and global climate change' by Dana Nuccitelli et al." Science.gov (United States) Lu, Q.-B. 2014-04-01 In the Comment by Nuccitelli et al., they make many false and invalid criticisms of the CFC-warming theory in my recent paper, and claim that their anthropogenic forcings including CO2 would provide a better explanation of the observed global mean surface temperature (GMST) data over the past 50 years. First, their arguments for no significant discrepancy between modeled and observed GMST changes and for no pause in recent global warming contradict the widely accepted fact and conclusion that were reported in the recent literature extensively. Second, their criticism that the key data used in my recent paper would be "outdated" and "flawed" is untrue as these data are still used in the recent or current literature including the newest (2013) IPCC Report and there is no considerable difference between the UK Met Office HadRCUT3 and HadRCUT4 GMST datasets. The use of even more recently computer-reconstructed total solar irradiance data (whatever have large uncertainties) for the period prior to 1976 would not change any of the conclusions in my paper, where quantitative analyses were emphasized on the influences of humans and the Sun on global surface temperature after 1970 when direct measurements became available. For the latter, the solar effect has been well shown to play only a negligible role in global surface temperature change since 1970, which is identical to the conclusion made in the 2013 IPCC Report. Third, their argument that the solar effect would not play a major role in the GMST rise of 0.2°C during 1850-1970 even contradicts the data and conclusion presented in a recent paper published in their Skeptical Science by Nuccitelli himself. Fourth, their comments also indicate their lack of understandings of the basic radiation physics of the Earth system as well as of the efficacies of different greenhouse gases in affecting global surface temperature. Their listed "methodological errors" are either trivial or non-existing. Fifth, their assertion that 13. Monitoring Network Confirms Land Use Change is a Substantial Component of the Forest Carbon Sink in the eastern United States OpenAIRE C. W. Woodall; Walters, B. F.; J. W. Coulston; A. W. D’Amato; G. M. Domke; Russell, M. B; P. A. Sowers 2015-01-01 Quantifying forest carbon (C) stocks and stock change within a matrix of land use (LU) and LU change is a central component of large-scale forest C monitoring and reporting practices prescribed by the Intergovernmental Panel on Climate Change (IPCC). Using a region–wide, repeated forest inventory, forest C stocks and stock change by pool were examined by LU categories. In eastern US forests, LU change is a substantial component of C sink strength (~37% of forest sink strength) only secondary ... 14. Changes in the Alpine environment Directory of Open Access Journals (Sweden) Philippe Schoeneich 2009-03-01 Full Text Available L’évolution de l’environnement alpin au XXIe siècle sera conditionnée par le changement climatique. Celui-ci pourrait conduire à des climats inconnus à ce jour dans les Alpes, avec comme conséquence une crise environnementale majeure et durable. Face à ces défis, les financements de recherche restent insuffisants pour la recherche appliquée aux milieux de montagne. Les financements nationaux privilégient souvent la recherche polaire au détriment des hautes altitudes, alors que les financements de type Interreg prennent insuffisamment en compte les besoins de recherche fondamentale, préalable nécessaire à l’élaboration de scénarios. Une évolution se dessine depuis deux ou trois ans vers des projets en réseau à l’échelle alpine. Le présent article fait le point sur les principaux enjeux qui attendent la recherche environnementale alpine et sur la capacité des programmes de recherche à répondre aux besoins. La première partie sur les changements climatiques est fondée sur les rapports récents : rapport de synthèse IPCC 2007 (IPCC 2007, rapport IPCC sur l’Europe (Alcamo et al. 2007, rapport de synthèse du programme ClimChAlp (Prudent-Richard et al., 2008. On y trouvera des bibliographies complètes et circonstanciées. La deuxième partie se base sur une analyse des appels d’offres récents ou en cours, et des projets soumis et financés.The way the Alpine environment will evolve in the 21st century depends upon climate change. This could lead to climates never before seen in the Alps, resulting in a major and lasting environmental crisis. In the face of these challenges, funding is still insufficient for specialised research on mountain environments. State funding often prioritises polar research at the expense of high altitude areas, whereas funding schemes from bodies such as Interreg do not sufficiently address the need for fundamental research, which is nevertheless a necessary first step prior to 15. [Study on the characters of phytoplankton chlorophyll fluorescence excitation spectra based on fourth-derivative]. Science.gov (United States) Lu, Lu; Su, Rong-Guo; Wang, Xiu-Lin; Zhu, Chen-Jian 2007-11-01 Chlorophyll fluorescence excitation spectra of six phytoplankton species, belonging to Bacillariophyta and Dinophyta, were dealt by fourth-derivative analysis with the Matlab program. The results show that between 350 nm and 550 nm six fluorescence peaks were found in the fourth-derivative spectra, which are representatives of non-pigments, chlorophylls and carotenoides respectively. The method makes Bacillariophyta and Dinophyta more distinguishable when the fourth-derivative spectra are compared with the chlorophyll fluorescence excitation spectra. It can be used not only to discriminate the two groups of algaes, but also to reduce the effect of noise. The fluorescence peaks in the fourth-derivative spectra are proved to be stable. 16. Fourth Computational Aeroacoustics (CAA) Workshop on Benchmark Problems Science.gov (United States) Dahl, Milo D. (Editor) 2004-01-01 This publication contains the proceedings of the Fourth Computational Aeroacoustics (CAA) Workshop on Benchmark Problems. In this workshop, as in previous workshops, the problems were devised to gauge the technological advancement of computational techniques to calculate all aspects of sound generation and propagation in air directly from the fundamental governing equations. A variety of benchmark problems have been previously solved ranging from simple geometries with idealized acoustic conditions to test the accuracy and effectiveness of computational algorithms and numerical boundary conditions; to sound radiation from a duct; to gust interaction with a cascade of airfoils; to the sound generated by a separating, turbulent viscous flow. By solving these and similar problems, workshop participants have shown the technical progress from the basic challenges to accurate CAA calculations to the solution of CAA problems of increasing complexity and difficulty. The fourth CAA workshop emphasized the application of CAA methods to the solution of realistic problems. The workshop was held at the Ohio Aerospace Institute in Cleveland, Ohio, on October 20 to 22, 2003. At that time, workshop participants presented their solutions to problems in one or more of five categories. Their solutions are presented in this proceedings along with the comparisons of their solutions to the benchmark solutions or experimental data. The five categories for the benchmark problems were as follows: Category 1:Basic Methods. The numerical computation of sound is affected by, among other issues, the choice of grid used and by the boundary conditions. Category 2:Complex Geometry. The ability to compute the sound in the presence of complex geometric surfaces is important in practical applications of CAA. Category 3:Sound Generation by Interacting With a Gust. The practical application of CAA for computing noise generated by turbomachinery involves the modeling of the noise source mechanism as a 17. Proceedings of the Fourth Glacier Bay Science Symposium Science.gov (United States) Piatt, John F.; Gende, Scott M. 2007-01-01 Foreword Glacier Bay was established as a National Monument in 1925, in part to protect its unique character and natural beauty, but also to create a natural laboratory to examine evolution of the glacial landscape. Today, Glacier Bay National Park and Preserve is still a place of profound natural beauty and dynamic landscapes. It also remains a focal point for scientific research and includes continuing observations begun decades ago of glacial processes and terrestrial ecosystems. In recent years, research has focused on glacial-marine interactions and ecosystem processes that occur below the surface of the bay. In October 2004, Glacier Bay National Park convened the fourth in a series of science symposiums to provide an opportunity for researchers, managers, interpreters, educators, students and the general public to share knowledge about Glacier Bay. The Fourth Glacier Bay Science Symposium was held in Juneau, Alaska, rather than at the Park, reflecting a desire to maximize attendance and communication among a growing and diverse number of stakeholders interested in science in the park. More than 400 people attended the symposium. Participants provided 46 oral presentations and 41 posters covering a wide array of disciplines including geology, glaciology, oceanography, wildlife and fisheries biology, terrestrial and marine ecology, socio-cultural research and management issues. A panel discussion focused on the importance of connectivity in Glacier Bay research, and keynote speakers (Gary Davis and Terry Chapin) spoke of long-term monitoring and ecological processes. These proceedings include 56 papers from the symposium. A summary of the Glacier Bay Science Plan-itself a subject of a meeting during the symposium and the result of ongoing discussions between scientists and resource managers-also is provided. We hope these proceedings illustrate the diversity of completed and ongoing scientific studies, conducted within the Park. To this end, we invited all 18. EDITORIAL: The Fourth International Workshop on Microfactories (IWMF'04) Science.gov (United States) Chu, Jiaru; Maeda, Ryutaro 2005-10-01 This special section of Journal of Micromechanics and Microengineering is devoted to the fourth International Workshop on Microfactories. After the first three successful Workshops, which took place in Tsukuba, Japan in 1998, Fribourg, Switzerland in 2000 and Minneapolis, USA in 2002, the fourth (IWMF'04) was held in Shanghai, China on 15-17 October 2004. The concept of the microfactory' and miniaturized production systems was first proposed by the Mechanical Engineering Laboratory in Japan who demonstrated the feasibility of downsizing energy-saving, distributed and eventually environmentally conscious manufacturing systems. There is incredible potential in reducing the physical scale of numerous processes related to the manufacture of many forms of future dense mechatronic' products and in the manipulation of microscopic and nanoscopic objects and materials for the benefit of mankind. Small systems capable of these operations can be referred to as microfactories'. A worldwide effort is currently underway to bring such microfactories to fruition. MEMS, MST and micromachines are regarded as synonyms, but they do not necessarily have the same meaning. In particular, differences can be found in their technological approaches. Roughly speaking, research and development in the USA is based primarily on surface or bulk silicon micromachining processes, and the ideal realization of MEMS seems to be the monolithic device. The European approach also focuses on integration between electronics and mechanics but, especially in connection with the development of µTAS', it also demands the integration of non-mechanical components into the system. In Japan the approaches for MST are said to be rather less focused. Besides the above-mentioned classical' microsystem technologies, down-scaling of conventional manufacturing methods, or non-silicon based device processing within this field, completely new technological methods are also considered as microsystem technologies 19. Climate change and its linkages with development, equity and sustainability International Nuclear Information System (INIS) The IPCC Working Groups 2 and 3 (on impacts and adaptation, and on mitigation, respectively) have taken the initiative to organise a number of expert meetings. The main goals of the meetings are: (a) to increase co-ordination among, and inform lead authors about development, equity and sustainability (DES) issues and climate change; and (b) to better place the IPCC Third Assessment Report in the context of DES, through enhanced access to the best available scientific, technical economic and social information from across the world. The rapidly growing economies of Asia will play a key role in determining future climate change policy. Thus, the organisation of the first expert meeting in Colombo, Sri Lanka, under the able guidance of Professor Mohan Munasinghe, is very appropriate. The conference has provided a crucial first step towards broadening the scope of the IPCC reports, by integrating DES issues into climate change response strategies. This Proceedings volume is also important as a key policy-relevant (but not policy-prescriptive) vehicle that will better inform worldwide networks of scientists from different regional, cultural and disciplinary backgrounds, in their pursuit of the best scientific information on the linkages between climate change and sustainable development. refs 20. Somali Jet Changes under the Global Warming Institute of Scientific and Technical Information of China (English) LIN Meijing; FAN Ke; WANG Huijun 2008-01-01 Somali Jet changes will influence the variability of Asian monsoon and climate. How would Somali Jet changes respond to the global warming in the future climate? To address this question, we first evaluate the ability of IPCC-AR4 climate models and perform the 20th century climate in coupled models (20C3M) experiments to reproduce the observational features of the low level Somali Jet in JJA (June-July-August) for the period 1976-1999. Then, we project and discuss the changes of Somali Jet under the climate change of Scenario A2 (SRESA2) for the period 2005-2099. The results show that 18 IPCC-AR4 models have performed better in describing the climatological features of Somali Jet in the present climate simulations. Analysis of Somali Jet intensity changes from the multi-model ensemble results for the period 2005-2099 shows a weakened Somali Jet in the early 21st century (2010-2040), the strongest Somali Jet in the middle 21st century (2050-2060), as well as the weakest Somali Jet at the end of the 21st century (2070-2090). Compared with the period 1976-1999, the intensity of Somali Jet is weakening in general, and it becomes the weakest at the end of the 21st century. The results also suggest that the relationship between the intensity of Somali Jet in JJA and the increment of global mean surface air temperature is nonlinear, which is reflected differently among the models, suggesting the uncertainty of the IPCC-AR4 models. Considering the important role of Somali Jet in the Indian monsoon and East Asian monsoon and climate of China, the variability of Somali Jet and its evolvement under the present climate or future climate changes need to be further clarified. 1. Hanford Seismic Annual Report and Fourth Quarter Report for Fiscal Year 1999 Energy Technology Data Exchange (ETDEWEB) AC Rohay; DC Hartshorn; SP Reidel 1999-12-07 Hanford Seismic Monitoring provides an uninterrupted collection of high-quality raw and processed seismic data from the Hanford Seismic Network (HSN) for the U.S. Department of Energy and its contractors. Hanford Seismic Monitoring also locates and identifies sources of seismic activity and monitors changes in the historical pattern of seismic activity at the Hanford Site. The data are compiled, archived, and published for use by the Hanford Site for waste management, Natural Phenomena Hazards assessments, and engineering design and construction. In addition, the seismic monitoring organization works with the Hanford Site Emergency Services Organization to provide assistance in the event of a significant earthquake on the Hanford Site. The HSN and the Eastern Washington Regional Network. (EWRN) consist of 40 individual sensor sites and 15 radio relay sites maintained by the Hanford Seismic Monitoring staff. A major reconfiguration of the HSN was initiated at the end of this quarter and the results will be reported in the first quarter report for next fiscal year (FY2000). For the HSN, there were 390 triggers during the fourth quarter of fiscal year(FY) 1999 on the primary recording system. With the implementation of dual backup systems during the second quarter of the fiscal year and an overall increase observed in sensitivity, a total of 1632 triggers were examined, identified, and processed during this fiscal year. During the fourth quarter, 24 seismic events were located by the HSN within the reporting region of 46 degrees to 47 degrees north latitude and 119 degrees to 120 degrees west longitude 9 were earthquakes in the Columbia River Basalt Group, 2 were earthquakes in the pre-basalt sediments, 10 were earthquakes in the crystalline basement; and 2 were quarry blasts. One earthquake appears to be related to a major geologic structure, 14 earthquakes occurred in known swarm areas, and 7 earthquakes were random occurrences. 2. CERN announces the fourth annual Beamline for Schools competition CERN Multimedia BL4S team 2016-01-01 CERN is pleased to announce the fourth annual Beamline for Schools (BL4S) competition. Once again, in 2017, a fully equipped beamline will be made available at CERN for students. As in previous years, two teams will be invited to the Laboratory to execute the experiments they proposed in their applications. The 2017 competition is being made possible thanks to support from the Alcoa Foundation for the second consecutive year. The competition is open to teams of high-school students aged 16 or older who, if they win, are invited (with two supervisors) to CERN to carry out their experiment. Teams must have at least five students but there is no upper limit to a team’s size (although just nine students per winning team will be invited to CERN). Teams may be composed of pupils from a single school, or from a number of schools working together. As science-loving mega-celebrity Will.I.Am told us: “If you’re interested in science, technology, engineering or ... 3. The Fourth US Naval Observatory CCD Astrograph Catalog (UCAC4) CERN Document Server Zacharias, Norbert; Girard, Terry; Henden, Arne; Bartlett, Jennifer; Monet, Dave; Zacharias, Marion 2012-01-01 The fourth United States Naval Observatory (USNO) CCD Astrograph Catalog, UCAC4 was released in August 2012 (double-sided DVD and CDS data center Vizier catalog I/322). It is the final release in this series and contains over 113 million objects; over 105 million of them with proper motions. UCAC4 is an updated version of UCAC3 with about the same number of stars also covering all-sky. Bugs were fixed, Schmidt plate survey data were avoided, and precise 5-band photometry were added. Astrograph observations have been supplemented for bright stars by FK6, Hipparcos and Tycho-2 data to compile a UCAC4 star catalog complete to about magnitude R = 16. Epoch 1998 to 2004 positions are obtained from observations with the 20 cm aperture USNO Astrograph's red lens, equipped with a 4k by 4k CCD. Mean positions and proper motions are derived by combining these observations with over 140 ground- and space-based catalogs, including Hipparcos/Tycho and the AC2000.2, as well as unpublished measures of over 5000 plates from ... 4. The fourth International Conference on Information Science and Cloud Computing Science.gov (United States) This book comprises the papers accepted by the fourth International Conference on Information Science and Cloud Computing (ISCC), which was held from 18-19 December, 2015 in Guangzhou, China. It has 70 papers divided into four parts. The first part focuses on Information Theory with 20 papers; the second part emphasizes Machine Learning also containing 21 papers; in the third part, there are 21 papers as well in the area of Control Science; and the last part with 8 papers is dedicated to Cloud Science. Each part can be used as an excellent reference by engineers, researchers and students who need to build a knowledge base of the most current advances and state-of-practice in the topics covered by the ISCC conference. Special thanks go to Professor Deyu Qi, General Chair of ISCC 2015, for his leadership in supervising the organization of the entire conference; Professor Tinghuai Ma, Program Chair, and members of program committee for evaluating all the submissions and ensuring the selection of only the highest quality papers; and the authors for sharing their ideas, results and insights. We sincerely hope that you enjoy reading papers included in this book. 5. Robust Optimization of Fourth Party Logistics Network Design under Disruptions Directory of Open Access Journals (Sweden) Jia Li 2015-01-01 Full Text Available The Fourth Party Logistics (4PL network faces disruptions of various sorts under the dynamic and complex environment. In order to explore the robustness of the network, the 4PL network design with consideration of random disruptions is studied. The purpose of the research is to construct a 4PL network that can provide satisfactory service to customers at a lower cost when disruptions strike. Based on the definition of β-robustness, a robust optimization model of 4PL network design under disruptions is established. Based on the NP-hard characteristic of the problem, the artificial fish swarm algorithm (AFSA and the genetic algorithm (GA are developed. The effectiveness of the algorithms is tested and compared by simulation examples. By comparing the optimal solutions of the 4PL network for different robustness level, it is indicated that the robust optimization model can evade the market risks effectively and save the cost in the maximum limit when it is applied to 4PL network design. 6. Fourth Data Challenge for the ALICE data acquisition system CERN Multimedia Maximilien Brice 2003-01-01 The ALICE experiment will study quark-gluon plasma using beams of heavy ions, such as those of lead. The particles in the beams will collide thousands of times per second in the detector and each collision will generate an event containing thousands of charged particles. Every second, the characteristics of tens of thousands of particles will have to be recorded. Thus, to be effective, the data acquisition system (DAQ) must meet extremely strict performance criteria. To this end, the ALICE Data Challenges entail step-by-step testing of the DAQ with existing equipment that is sufficiently close to the final equipment to provide a reliable indication of performance. During the fourth challenge, in 2002, a data acquisition rate of 1800 megabytes per second was achieved by using some thirty parallel-linked PCs running the specially developed DATE software. During the final week of tests in December 2002, the team also tested the Storage Tek linear magnetic tape drives. Their bandwidth is 30 megabytes per second a... 7. Direct detection of fourth generation Majorana neutrino dark matter CERN Document Server Zhou, Yu-Feng 2012-01-01 Heavy stable fourth generation Majorana neutrinos contribute to a small fraction of the relic density of dark matter (DM) in the Universe. Due to its relatively strong coupling to the standard model particles, it can be probed by the current direct and indirect DM detection experiments even it is a subdominant component of the halo DM. We show that the current Xenon100 data constrain the mass of the stable Majorana neutrino to be greater than the mass of the top quark. The effective spin-independent cross section for the neutrino elastic scattering off nucleon is predicted to be$\\sim 1.5\\times 10^{-44} cm^2$, which is insensitive to the neutrino mass and mixing and can be reached by the direct DM detection experiments in the near future. In the same mass region the predicted effective spin-dependent cross section for the heavy neutrino scattering off proton is in the range of$2\\times 10^{-40} cm^2\\sim 2\\times 10^{-39} cm^2$, which is within the reach of the ongoing DM indirect search experiments. We demonst... 8. Australian climate change impacts, adaptation and vulnerability International Nuclear Information System (INIS) Full text: Full text: The IPCC Fourth Assessment Report on impacts, adaptation and vulnerability made the following conclusions about Australia (Hennessy et al., 2007): Regional climate change has occurred. Since 1950, there has been 0.70C warming, with more heat waves, fewer frosts, more rain in north-west Australia, less rain in southern and eastern Australia, an increase in the intensity of Australian droughts and a rise in sea level of about 70 mm. Australia is already experiencing impacts from recent climate change. These are now evident in increasing stresses on water supply and agriculture, changed natural ecosystems, and reduced seasonal snow cover. Some adaptation has already occurred in response to observed climate change. Examples come from sectors such as water, natural ecosystems, agriculture, horticulture and coasts. However, ongoing vulnerability to extreme events is demonstrated by substantial economic losses caused by droughts, floods, fire, tropical cyclones and hail. The climate of the 21st century is virtually certain to be warmer, with changes in extreme events. Heat waves and fires are virtually certain to increase in intensity and frequency. Floods, landslides, droughts and storm surges are very likely to become more frequent and intense, and snow and frost are very likely to become less frequent. Large areas of mainland Australia are likely to have less soil moisture. Potential impacts of climate change are likely to be substantial without further adaptation; As a result of reduced precipitation and increased evaporation, water security problems are projected to intensify by 2030 in southern and eastern Australia; Ongoing coastal development and population growth, in areas such as Cairns and south-east Queensland, are projected to exacerbate risks from sea level rise and increases in the severity and frequency of storms and coastal flooding by 2050. Significant loss of biodiversity is projected to occur by 2020 in some ecologically rich 9. Implications of and possible responses to climate change OpenAIRE Kahiluoto, Helena; Rötter, Reimund 2009-01-01 Climate change is expected to worsen food insecurity and seriously undermines rural development prospects. It makes it harder to achieve the Millenium Development Goals and ensure a sustainable future beyond 2015. Findings from the recent 4th assessment report of IPCC, Working Group II indicate that already towards 2050 with respect to food crops yield losses between 10 and 30 % can be expected as compared to current conditions in large parts of Africa, including Western, Eastern and southern... 10. A coming anarchy? : Pathways from climate change to violent conflict in East Africa OpenAIRE van Baalen, Sebastian; Mobjörk, Malin 2016-01-01 The warming of the climate system is unequivocal according to the Intergovernmental Panel on Climate Change (IPCC), and will have a strong impact on the security of humans and states alike. In the past half-century the climate system has changed in unprecedented ways and future climate change and variability will include long-lasting alterations to all components of the climate system. With the warming of the climate system and the recognition of the implications that this has for the availab... 11. Updating soil CO2 emission experiments to assess climate change effects and extracellular soil respiration Science.gov (United States) Vidal Vazquez, Eva; Paz Ferreiro, Jorge 2014-05-01 emissions from sterilized soils and their unsterilized counterparts are compared. Moreover, different pH treatments are compared to analyze how soil pH affects extracellular CO2 release. Students benefit from experimental learning. Practical courses, being either in the field or indoors are of vital importance to bring soil processes to life and to evaluate implications for environment and climate change. IPCC, 2007: Climate Change 2007: The Physical Science Basis. Contribution of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change (Solomon, S., D. Qin, M. Manning, Z. Chen, M. Marquis, K.B. Averyt, M. Tignor and H.L. Miller (eds.). Cambridge University Press, Cambridge, United Kingdom, 996 pp. Maire, V., G. Alvarez, J. Colombet, A. Comby, R. Despinasse, E. Dubreucq, M. Joly, A.-C. Lehours, V. Perrier, T. Shahzad, and S. Fontaine. 2013. An unknown oxidative metabolism substantially contributes to soil CO2 emissions. Biogeochemistry, 10, 1155-1167, 2013 12. CLIMATE CHANGE IMPACTS ON WATER RESOURCES OpenAIRE T.M. CORNEA; Dima, M.; Roca, D. 2011-01-01 Climate change impacts on water resources – The most recent scientific assessment by the Intergovernmental Panel on Climate Change (IPCC) [6] concludes that, since the late 19th century, anthropogenic induced emissions of greenhouse gases have contributed to an increase in global surface temperatures of about 0.3 to 0.6o C. Based on the IPCC’s scenario of future greenhouse gas emissions and aerosols a further increase of 2o C is expected by the year 2100. Plants, animals, natural and managed ... 13. Climate change - The Macedonia's First National Communication International Nuclear Information System (INIS) Climate change impacts, consequences and concerns of the international community; United Nations Framework Convention on Climate Change (UNFCCC). Activities in the Republic of Macedonia, establishing the Climate Change Project Unit within the Ministry of Environment and Physical Planning and the National Climate Change Committee. Preparation of the Macedonia's First National Communications under the United Nations Framework Convention on Climate Change. Analyzing on the thematic areas of the Nationals Communications. The inventory of greenhouse gases(GHG) emissions was prepared according to IPCC Guidelines (IPCC), taking into consideration the three main GHGs:carbondioxide (CO2), methane (CH4), nitrous oxide (N2O). The main sources of CO2 emissions are the electricity production, the production and the transport. GHG abatement analysis and projections of emissions are prepared in accordance to the Macedonian economy and possibilities for development. The analysis of the energy sector is elaborated in a most advanced way, especially regarding the electricity production. According to the IS92a scenario (prepared by IPCC) the average annual temperature in Macedonia could arise for 4,6o C by 2100, and the average summer temperature could arise for 5.1o C. The average sum of precipitation will decrease for 6.3% in 2100, but the most alarming is the sum of precipitation in summer, which could decrease for 2.5%. Venerability assessment and adaptation measures are elaborated in the following sectors: agriculture, forestry, biodiversity, water resources and human health. The National Action Plan sets out the objectives and initial points for undertaking measures, contributing to the reduction of GHG emissions at national level. (Author) 14. The contribution of executive functions to narrative writing in fourth grade children NARCIS (Netherlands) Drijbooms, E.; Groen, M.A.; Verhoeven, L.T.W. 2015-01-01 The present study investigated the contribution of executive functions to narrative writing in fourth grade children, and evaluated to what extent executive functions contribute differentially to different levels of narrative composition. The written skills of 102 Dutch children in fourth grade were 15. The impact of the fourth SM family on the Higgs observability at the LHC CERN Document Server Arik, E; Sultansoy, S 2007-01-01 It is shown that if the fourth SM fermion family exists then the Higgs boson could be observed at the LHC with an integrated luminosity of few fb-1. The Higgs discovery potential for different channels is discussed in the presence of the fourth SM family. 16. ON THE INSTABILITY OF SOLUTIONS TO A NONLINEAR VECTOR DIFFERENTIAL EQUATION OF FOURTH ORDER Institute of Scientific and Technical Information of China (English) 2011-01-01 This paper presents a new result related to the instability of the zero solution to a nonlinear vector differential equation of fourth order.Our result includes and improves an instability result in the previous literature,which is related to the instability of the zero solution to a nonlinear scalar differential equation of fourth order. 17. An endoscopic, cadaveric analysis of the roof of the fourth ventricle. Science.gov (United States) Salma, Asem; Yeremeyeva, Esmiralda; Baidya, Nishanta B; Sayers, Martin Peter; Ammirati, Mario 2013-05-01 We performed endoscopic dissections of the roof of the fourth ventricle in eight fresh human cadaveric heads to characterize the endoscopic anatomy of the roof of the fourth ventricle and the anatomical configuration of the structures forming its roof. We also made three-dimensional (3D) silicone casts of the fourth ventricle in seven formalin-fixed specimens to evaluate the 3D configuration of the structures that create the roof of the fourth ventricle. The roof of the fourth ventricle can be divided into three zones. The upper zone is formed by the superior cerebellar peduncle and superior medullary velum and is associated with the lingula. The middle zone is formed by the inferior cerebellar peduncles and inferior medullary velum and is associated with the nodule in the midline and with the peduncle of the flocculus. The lower zone is formed by the tela choroidea and is associated with the tonsils. The 3D shape of the roof the fourth ventricle resembles that of a rhomboid-based pyramid; the edges of the base represent the borders of the ventricle, and the apex is the cerebellar fastigium. The lateral recess is shaped like a triangular-based pyramid, with its base connected to the cavity of the fourth ventricle and its tip opening into the lateral cerebellomedullary cistern through the foramen of Luschka. Our results may help in the endoscopic exploration of and microsurgical approaches to the fourth ventricle through its roof. PMID:23507044 18. Differentiated Jurisprudence? Examining Students' Fourth Amendment Court Decisions by Region of Country Science.gov (United States) Torres, Mario S., Jr. 2012-01-01 This study examined federal and state court decisions related to student Fourth Amendment rights following the "New Jersey v. T.L.O." ruling in 1985. There has been minimal research in judicial treatment of students' Fourth Amendment rights across regions of the country and less to what extent regional rulings implicitly or explicitly transmit… 19. PPN-limit of Fourth Order Gravity inspired by Scalar-Tensor Gravity OpenAIRE Capozziello, S.; Troisi, A. 2005-01-01 Based on the {\\it dynamical} equivalence between higher order gravity and scalar-tensor gravity the PPN-limit of fourth order gravity is discussed. We exploit this analogy developing a fourth order gravity version of the Eddington PPN-parameters. As a result, Solar System experiments can be reconciled with higher order gravity, if physical constraints descending from experiments are fulfilled. 20. Ten years of HIV testing with fourth generation assays: The Amsterdam experience NARCIS (Netherlands) S. Jurriaans; N.K.T. Back; K.C. Wolthers 2011-01-01 Serological HIV assays combining detection of HIV antigen and antibodies are referred to as fourth generation assays. Fourth generation assays were implemented in Europe for routine patient testing about 10 years ago. The Academic Medical Center is one of the main HIV treatment centers in the Nether 1. The Effects of a Teacher-Guided Library Selection Program on Fourth-Grade Reading Achievement Science.gov (United States) Weber, Sherri M. 2009-01-01 The purpose of this mixed methods study was to examine the impact that a teacher-guided library selection program had on overall reading scores and how this program may or may not promote reading improvement of fourth-grade students at a small suburban school in western New York. Forty fourth graders were used as the sample population, which was… 2. A rare case of retained fourth molar teeth in maxilla and mandible. Case report Directory of Open Access Journals (Sweden) Rahnama Mansur 2014-06-01 Full Text Available The study presents a case of the rarely occurring totally retained fourth molar teeth simultaneously in maxilla and mandible. The appearance of supernumerary teeth is a relatively uncommon dental anomaly and it is rare for patients to have impacted fourth molars in two quadrant. The aim of this work is to describe the presence of unilateral (right fourth molars in the maxilla and the mandible in a young female patient aged 24 years. Orthopantomogram revealed impacted lower third molars but also unerupted unilateral (right upper and lower fourth molars. Before orthodontic treatment, the patient was subsequently admitted for removal of third and fourth impacted upper and lower molars under local anesthesia. 3. On the fourth Diadema species (Diadema-sp from Japan. Directory of Open Access Journals (Sweden) Seinen Chow Full Text Available Four long-spined sea urchin species in the genus Diadema are known to occur around the Japanese Archipelago. Three species (D. savignyi, D. setosum, and D. paucispinum are widely distributed in the Indo-Pacific Ocean. The fourth species was detected by DNA analysis among samples originally collected as D. savignyi or D. setosum in Japan and the Marshall Islands and tentatively designated as Diadema-sp, remaining an undescribed species. We analyzed nucleotide sequences of the cytochrome oxidase I (COI gene in the "D. savignyi-like" samples, and found all 17 individuals collected in the mainland of Japan (Sagami Bay and Kyushu to be Diadema-sp, but all nine in the Ryukyu Archipelago (Okinawa and Ishigaki Islands to be D. savignyi, with large nucleotide sequence difference between them (11.0%±1.7 SE. Diadema-sp and D. savignyi shared Y-shaped blue lines of iridophores along the interambulacrals, but individuals of Diadema-sp typically exhibited a conspicuous white streak at the fork of the Y-shaped blue iridophore lines, while this feature was absent in D. savignyi. Also, the central axis of the Y-shaped blue lines of iridophores was approximately twice as long as the V-component in D. savignyi whereas it was of similar length in Diadema-sp. Two parallel lines were observed to constitute the central axis of the Y-shaped blue lines in both species, but these were considerably narrower in Diadema-sp. Despite marked morphological and genetic differences, it appears that Diadema-sp has been mis-identified as D. savignyi for more than half a century. 4. A guide to the evaluation of fourth cranial nerve palsies. Science.gov (United States) Lee; Hayman; Beaver; Prager; Kelder; Scasta; Avilla; von Noorden GK; Tang 1998-12-01 PURPOSE To devise a cost-effective guide for the evaluation of fourth nerve palsies (FNP). METHODS A review of the pertinent English language literature was performed to devise a guide for the evaluation (including neuroimaging) of FNP. The authors report a retrospective review of imaging studies performed on 206 patients with FNP. RESULTS The literature was used to develop the imaging guide. In the retrospective chart review of 206 patients from two tertiary care centers, 28 patients (13.6%) underwent a computed tomography scan and/or a magnetic resonance scan. Of these patients, five had associated neurological symptoms (non-isolated), one was traumatic, five were congenital, four were vasculopathic, eleven were non-vasculopathic, and two were progressive. Following the recommendations of the imaging guide, the five isolated congenital FNP and the four isolated vasculopathic FNP would not have undergone neuroimaging studies. The total costs of these neuroimaging studies in these nine patients were 19,000 dollars. Four patients in the retrospective review with associated neurological deficits (non-isolated) should have undergone neuroimaging according to the guide, but did not. CONCLUSIONS Although the evaluation of FNP can be difficult, the decision to order neuroimaging can be improved by using an imaging guide. An imaging guide for the evaluation of FNP may allow more appropriate and cost-effective imaging of these patients. Isolated congenital, old traumatic, or vasculopathic FNP do not require neuroimaging studies. Patients with non-isolated FNP should have directed neuroimaging studies based upon the results of clinical examination. 5. Preface to the first monograph: The Fourth Psychiatric Revolution Directory of Open Access Journals (Sweden) Ajai R. Singh 2004-05-01 Full Text Available At present, psychiatry is in the midst of a fourth revolution. The first revolution was the so-called Moral Treatment which involved the activism of Phileppe Pinel (1745-1826 and William Tuke (1732-1819, as also the efforts of Dorothea Dix (1802-1887. This resulted in destigmatization of the lunatic label which had earlier meant treating the insane in a dehumanizing manner e.g.. chaining them to walls, displaying them for money etc. It resulted in the transition to custodial care and the opening of mental hospitals. The second psychiatric revolution was the Mental Hygiene Movement heralded by the eyeopening works of Elizabeth Packard (Modern Persecution or Insane Hospital Unveiled and Clifford Beers (1876-1943; A Mind That Found Itself which was furthered by, amongst others, pioneers like Adolf Meyer (1866-1950 and William James (1842-1910. This was followed by the third Psychiatric revolution, that of the Community Psychiatry Movement. This involved community participation, removal of restrictions, comprehensive set of services multi-disciplinary in nature, active consumer participation, mental health consultancy and preventive measures. This well intentioned grand movement had its problems, as all such grand movement must indeed have. It became the fountainsource of a fresh crop of difficulties related to transinstitutionalization in boarding and halfway houses, with increased rates of hospital admission, and the 'revolving door syndrome'. Moreover, it lead to an ominous rise in contact between the criminal justice system and the mentally ill as they moved more freely in the community.Today, we are in the midst of a silent by strong fourth revolution. Firstly, this revolution reiterates its strong linkage with the mainstream of medicine. Secondly, it bases itself on strong, empirical findings based on rigorous methodological studies, mainly biological. The major paradigm shift of contemporary psychiatry is towards methodological rigour on 6. National report card on energy efficiency: Fourth annual report card on government activities Energy Technology Data Exchange (ETDEWEB) NONE 2003-04-01 The objective of this fourth annual report card produced by the Canadian Energy Efficiency Alliance is, as before, to identify government commitments to energy efficiency and to raise the awareness and importance of providing the necessary mechanisms that support energy efficiency in the market place. For the first time this year, the report card includes an assessment of the role of crown utilities and local distribution utilities in the promotion of energy efficiency. Their results were combined with those of the provincial government to develop a aggregate grade for each province. The ratings are based on ten categories, encompassing building codes, energy efficiency legislation, regulatory infrastructure, programs and public outreach, in-house programs, innovative new programs, performance and evaluation, climate change and Voluntary Change Registry reporting, and access to government information. Performance is rated by way of a narrative summary and a letter grade which represents the aggregate of performance in all ten categories. Best performance was recorded by Quebec, followed by the Yukon, the Federal Government, the Northwest Territories, British Columbia, Nova Scotia, New Brunswick, Manitoba, Alberta, Ontario, Prince Edward Island, Newfoundland and Labrador, and Saskatchewan, in that order. 7. Potential hydrologic changes in the Amazon by the end of the 21st century and the groundwater buffer International Nuclear Information System (INIS) This study contributes to the discussions on the future of the Amazon rainforest under a projected warmer-drier climate from the perspectives of land hydrology. Using IPCC HadGEM2-ES simulations of the present and future Amazon climate to drive a land hydrology model that accounts for groundwater constraint on land drainage, we assess potential hydrologic changes in soil water, evapotranspiration (ET), water table depth, and river discharge, assuming unchanged vegetation. We ask: how will ET regimes shift at the end of the 21st century, and will the groundwater help buffer the anticipated water stress in some places-times? We conducted four 10 yr model simulations, at the end of 20th and 21st century, with and without the groundwater. Our model results suggest that, first, over the western and central Amazon, ET will increase due to increased potential evapotranspiration (PET) with warmer temperatures, despite a decrease in soil water; that is, ET will remain PET or atmospheric demand-limited. Second, in the eastern Amazon dry season, ET will decrease in response to decreasing soil water, despite increasing PET demand; that is, ET in these regions-seasons will remain or become more soil water or supply-limited. Third, the area of water-limited regions will likely expand in the eastern Amazonia, with the dry season, as indicated by soil water store, even drier and longer. Fourth, river discharge will be significantly reduced over the entire Amazon but particularly so in the southeastern Amazon. By contrasting model results with and without the groundwater, we found that the slow soil drainage constrained by shallow groundwater can buffer soil water stress, particularly in southeastern Amazon dry season. Our model suggests that, if groundwater buffering effect is accounted for, the future Amazon water stress may be less than that projected by most climate models. (letter) 8. Potential Hydrologic Changes in the Amazon By the End of the 21st Century and the Groundwater Buffer Science.gov (United States) Pokhrel, Y. N.; Fan, Y.; Miguez-Macho, G. 2014-12-01 This study contributes to the discussions on the future of the Amazon rainforest under a projected warmer-drier climate from the perspectives of land hydrology. Using IPCC HadGEM2-ES simulations of the present and future Amazon climate to drive a land hydrology model that accounts for groundwater constraint on land drainage, we assess potential hydrologic changes in soil water, evapotranspiration (ET), water table depth, and river discharge, assuming unchanged vegetation. We ask: how will ET regimes shift at the end of the 21st century, and will the groundwater help buffer the anticipated water stress in some places-times? We conducted four 10yr model simulations, at the end of 20th and 21st century, with and without the groundwater. Our model results suggest that, first, over the western and central Amazon, ET will increase due to increased potential evapotranspiration (PET) with warmer temperatures, despite a decrease in soil water; that is, ET will remain atmosphere or demand-limited. Second, in the eastern Amazon dry season, ET will decrease in response to decreasing soil water, despite increasing PET demand; that is, ET in these regions-seasons will remain or become more soil water or supply-limited. Third, the area of water-limited regions will likely expand in the eastern Amazonia, with the dry season, as indicated by soil water store, even drier and longer. Fourth, river discharge will be significantly reduced over the entire Amazon but particularly so in the southeastern Amazon. By contrasting model results with and without the groundwater, we found that the slow soil drainage constrained by a shallow groundwater can buffer soil water stress, particularly in southeastern Amazon dry season. Our model suggests that, if the groundwater buffering effect is accounted for, the future Amazon water stress may be less than projected by most climate models. 9. The role of uncertainty in climate change adaptation strategies — A Danish water management example DEFF Research Database (Denmark) Refsgaard, J.C.; Arnbjerg-Nielsen, Karsten; Drews, Martin; 2013-01-01 be used to address concerns that the IPCC approach is oversimplified. We have studied the role of uncertainty in climate change adaptation planning using examples from four Danish water related sectors. The dominating sources of uncertainty differ greatly among issues; most uncertainties on impacts......We propose a generic framework to characterize climate change adaptation uncertainty according to three dimensions: level, source and nature. Our framework is different, and in this respect more comprehensive, than the present UN Intergovernmental Panel on Climate Change (IPCC) approach and could...... are epistemic (reducible) by nature but uncertainties on adaptation measures are complex, with ambiguity often being added to impact uncertainties. Strategies to deal with uncertainty in climate change adaptation should reflect the nature of the uncertainty sources and how they interact with risk level... 10. The GLOBE climate legislation study: a review of climate change legislation in 66 countries: fourth edition OpenAIRE Nachmany, Michal; Fankhauser, Samuel; Townshend, Terry; Collins, Murray; Landesman, Tucker; Matthews, Adam; Pavese, Carolina; Rietig, Katharina; SCHLEIFER, Philip; Setzer, Joana 2014-01-01 The GLOBE Climate Legislation Study is the most comprehensive audit of climate legislation across 66 countries, together responsible for around 88% of global manmade greenhouse gas emissions. It is produced by the Grantham Research Institute at the London School of Economics in collaboration with GLOBE International. The 4th edition of the Study was formally launched at the 2nd GLOBE Climate Legislation Summit held at the Senate of the United States of America and at the World Bank in Was... 11. Seismological analysis of the fourth North Korean nuclear test Science.gov (United States) Hartmann, Gernot; Gestermann, Nicolai; Ceranna, Lars 2016-04-01 The Democratic People's Republic of Korea has conducted its fourth underground nuclear explosions on 06.01.2016 at 01:30 (UTC). The explosion was clearly detected and located by the seismic network of the International Monitoring System (IMS) of the Comprehensive Nuclear-Test-Ban Treaty (CTBT). Additional seismic stations of international earthquake monitoring networks at regional distances, which are not part of the IMS, are used to precisely estimate the epicenter of the event in the North Hamgyong province (41.38°N / 129.05°E). It is located in the area of the North Korean Punggye-ri nuclear test site, where the verified nuclear tests from 2006, 2009, and 2013 were conducted as well. The analysis of the recorded seismic signals provides the evidence, that the event was originated by an explosive source. The amplitudes as well as the spectral characteristics of the signals were examined. Furthermore, the similarity of the signals with those from the three former nuclear tests suggests very similar source type. The seismograms at the 8,200 km distant IMS station GERES in Germany, for example, show the same P phase signal for all four explosions, differing in the amplitude only. The comparison of the measured amplitudes results in the increasing magnitude with the chronology of the explosions from 2006 (mb 4.2), 2009 (mb 4.8) until 2013 (mb 5.1), whereas the explosion in 2016 had approximately the same magnitude as that one three years before. Derived from the magnitude, a yield of 14 kt TNT equivalents was estimated for both explosions in 2013 and 2016; in 2006 and 2009 yields were 0.7 kt and 5.4 kt, respectively. However, a large inherent uncertainty for these values has to be taken into account. The estimation of the absolute yield of the explosions depends very much on the local geological situation and the degree of decoupling of the explosive from the surrounding rock. Due to the missing corresponding information, reliable magnitude-yield estimation for the 12. CLIMATE CHANGES: CAUSES AND IMPACT Directory of Open Access Journals (Sweden) Camelia Slave 2013-07-01 Full Text Available Present brings several environmental problems for people. Many of these are closely related, but by far the most important problem is the climate change. In the course of Earth evolution, climate has changed many times, sometimes dramatically. Warmer eras always replaced and were in turn replaced by glacial ones. However, the climate of the past almost ten thousand years has been very stable. During this period human civilization has also developed. In the past nearly 100 years - since the beginning of industrialization - the global average temperature has increased by approx. 0.6 ° C (after IPCC (Intergovernmental Panel on Climate Change, faster than at any time in the last 1000 years. 13. Effects of Climate Change on Hydrology and Water Management in the Skagit River Basin Science.gov (United States) LEE, S.; Hamlet, A. F. 2012-12-01 Based on global climate model (GCM) scenarios from the Intergovernmental Panel on Climate Change Fourth Assessment Report (IPCC AR4) and subsequent hydrologic modeling studies for the Pacific Northwest, the impacts of climate change on hydrology in the Skagit River Basin are likely to be substantial under natural (unregulated) conditions. To assess the combined effects of increasing extreme flows (floods and low flows) and dam operations that determine impacts to regulated flow, a new integrated daily time step reservoir operations model was built for the Skagit River Basin. The model simulates current reservoir operating policies for historical flow conditions and for projected flows for the 2040s and 2080s using five different GCMs with the A1B emissions scenarios, and estimates sediment loading. By simulating alternative reservoir operating policies that provide increased flood storage and start flood evacuation one month earlier, prospects for adaptation in response to increased flood risks are considered. Results from the daily time step reservoir operations modeling show that regulated hydrologic extremes in the basin will likely become more intense. The regulated 100-year flood is projected to increase substantially in the future in comparison with historical regulated 100-year flood (23% by the 2040s and 40% by the 2080s). The regulated extreme 7-day low flows (7Q10) are also projected to decrease by about 30 % in the future. Alternative flood control operations (earlier and larger drafting of reservoirs) are shown to be largely ineffective in mitigating the increased flood risks in the lower Skagit, supporting the argument that climate change adaptation efforts will need to focus primarily on improving management of the floodplain, rather than focusing on modifying flood control operations in existing headwater projects. Projected changes in the Skagit River's flow regime are shown to have dramatic effects on estimated total suspended sediment load in the 14. Climatic and ecological future of the Amazon: likelihood and causes of change Directory of Open Access Journals (Sweden) B. Cook 2010-05-01 Full Text Available Some recent climate modeling results suggested a possible dieback of the Amazon rainforest under future climate change, a prediction that raised considerable interest as well as controversy. To determine the likelihood and causes of such changes, we analyzed the output of 15 models from the Intergovernmental Panel on Climate Change Fourth Assessment Report (IPCC/AR4 and a dynamic vegetation model VEGAS driven by these climate output. Our results suggest that the core of the Amazon rainforest should remain largely stable as rainfall is projected to increase in nearly all models. However, the periphery, notably the southern edge of the Amazon and further south in central Brazil, are in danger of drying out, driven by two main processes. Firstly, a decline in precipitation of 22% in the southern Amazon's dry season (May–September reduces soil moisture, despite an increase in precipitation during the wet season, due to nonlinear responses in hydrology and ecosystem dynamics. Two dynamical mechanisms may explain the lower dry season rainfall: (1 a general subtropical drying under global warming when the dry season southern Amazon is under the control of the subtropical high pressure; (2 a stronger north-south tropical Atlantic sea surface temperature gradient, and to lesser degree a warmer eastern equatorial Pacific. Secondly, evaporation demand will increase due to the general warming, further reducing soil moisture. In terms of ecosystem response, higher maintenance cost and reduced productivity under warming may also have additional adverse impact. The drying corresponds to a lengthening of the dry season by 11 days. As a consequence, the median of the models projects a reduction of 20% in vegetation carbon stock in the southern Amazon, central Brazil, and parts of the Andean Mountains. Further, VEGAS predicts enhancement of fire risk by 10–15%. The increase in fire is primarily due to the reduction in soil moisture, and the decrease in dry 15. Introduction to the symposium theme : climate change in fragmented landscapes: can we develop spatial adaptation strategies? OpenAIRE Verboom, J.; Vos, C.C. 2007-01-01 The Intergovernmental Panel for Climate Change (IPCC) concluded that by increasing the concentration of greenhouse gasses, man has a discernible influence on climate, and this is expected to be a long-term phenomenon affecting the environment in the forthcoming decades or even centuries. Since climate is a key driving force for ecological processes, climate change is likely to exert considerable impact on ecosystems. Since nature policy worldwide is often based upon policy plans which do not ... 16. Xylem Variability as a Proxy for Environmental and Climate Change in Corsica During the Past Millennium OpenAIRE Hetzer, Timo 2013-01-01 The Mediterranean region is considered as one of the hotspots of future climate change (IPCC 2007; Christensen et al. 2007). To assess the impacts of climate change in different regions, it is helpful to have a look at the past. Proxies, i.e. data derived from different environmental archives, contain climatic or environmental information from the past. On the island of Corsica in the center of the western Mediterranean basin, high-resolution climate reconstructions for the past millennium ar... 17. Climate change and environmental impacts on maternal and newborn health with focus on Arctic populations OpenAIRE Rylander, Charlotta; Odland, Jon Ø; Sandanger, Torkjel M. 2011-01-01 Background: In 2007, the Intergovernmental Panel on Climate Change (IPCC) presented a report on global warming and the impact of human activities on global warming. Later the Lancet commission identified six ways human health could be affected. Among these were not environmental factors which are also believed to be important for human health. In this paper we therefore focus on environmental factors, climate change and the predicted effects on maternal and newborn health. Arctic issues are d... 18. Climate change and environmental impacts on maternal and newborn health with focus on Arctic populations. OpenAIRE Rylander, Charlotta; Odland, Jon Øyvind; Sandanger, Torkjel Manning 2011-01-01 In 2007, the Intergovernmental Panel on Climate Change (IPCC) presented a report on global warming and the impact of human activities on global warming. Later the Lancet commission identified six ways human health could be affected. Among these were not environmental factors which are also believed to be important for human health. In this paper we therefore focus on environmental factors, climate change and the predicted effects on maternal and newborn health. Arctic issues are discussed ... 19. CECILIA Regional Climate Simulations for Future Climate: Analysis of Climate Change Signal OpenAIRE Michal Belda; Petr Skalák; Aleš Farda; Tomáš Halenka; Michel Déqué; Gabriella Csima; Judit Bartholy; Csaba Torma; Constanta Boroneant; Mihaela Caian; Valery Spiridonov 2015-01-01 Regional climate models (RCMs) are important tools used for downscaling climate simulations from global scale models. In project CECILIA, two RCMs were used to provide climate change information for regions of Central and Eastern Europe. Models RegCM and ALADIN-Climate were employed in downscaling global simulations from ECHAM5 and ARPEGE-CLIMAT under IPCC A1B emission scenario in periods 2021–2050 and 2071–2100. Climate change signal present in these simulations is consistent with respective... 20. Full effects of land use change in the representative concentration pathways OpenAIRE Davies-Barnard, T.; P. J. Valdes; Singarayer, J. S.; Pacifico, F. M.; Jones, C. D. 2014-01-01 Future land use change (LUC) is an important component of the IPCC representative concentration pathways (RCPs), but in these scenarios' radiative forcing targets the climate impact of LUC only includes greenhouse gases. However, climate effects due to physical changes of the land surface can be as large. Here we show the critical importance of including non-carbon impacts of LUC when considering the RCPs. Using an ensemble of climate model simulations with and without LUC, we show that the n... 1. Under What Circumstances Do Wood Products from Native Forests Benefit Climate Change Mitigation? OpenAIRE Heather Keith; David Lindenmayer; Andrew Macintosh; Brendan Mackey 2015-01-01 Climate change mitigation benefits from the land sector are not being fully realised because of uncertainty and controversy about the role of native forest management. The dominant policy view, as stated in the IPCC's Fifth Assessment Report, is that sustainable forest harvesting yielding wood products, generates the largest mitigation benefit. We demonstrate that changing native forest management from commercial harvesting to conservation can make an important contribution to mitigation. Con... 2. The Fourth R: A School-Based Adolescent Dating Violence Prevention Program Directory of Open Access Journals (Sweden) David A. Wolfe 2011-07-01 Full Text Available This paper presents a school-based primary prevention program (The Fourth R to prevent adolescent dating violence, and related risk behaviors. The cornerstone of The Fourth R is a 21-lesson skillbased curriculum delivered by teachers who receive specialized training, that promotes healthy relationships, and targets violence, high-risk sexual behavior, and substance use among adolescents. The Fourth R was evaluated in a cluster randomized trial in 20 schools. Results indicated that teaching youth healthy relationships and skills as part of their curriculum reduced physical dating violence, and increased condom use 2.5 years later. 3. Bound Energy Masses of Mesons Containing the Fourth Generation and Iso-singlet Quarks OpenAIRE Ikhdair, Sameer M.; Sever, Ramazan 2005-01-01 The fourth Standard Model (SM) family quarks and weak iso-singlet quarks The fourth Standard Model (SM) family quarks and weak iso-singlet quarks predicted by${\\rm E}_{6\\text{}}$GUT are considered. The spin-average of the pseudoscalar$\\eta_{4}(n^{1}S_{0})$and vector$\\psi_{4}(n^{3}S_{1})$quarkonium binding masses of the new mesons formed by the fourth Standard Model (SM) family and iso-singlet${\\rm E}_{6\\text{}}$with their mixings to ordinary quarks are investigated. Further, the fine ... 4. Weldon Spring Site Remedial Action Project quarterly environmental data summary (QEDS) for fourth quarter 1998 Energy Technology Data Exchange (ETDEWEB) NONE 1999-02-01 This report contains the Quarterly Environmental Data Summary (QEDS) for the fourth quarter of 1998 in support of the Weldon Spring Site Remedial Action Project Federal Facilities Agreement. The data, except for air monitoring data and site KPA generated data (uranium analyses) were received from the contract laboratories, verified by the Weldon Spring Site verification group, and merged into the database during the fourth quarter of 1998. KPA results for on-site total uranium analyses performed during fourth quarter 1998 are included. Air monitoring data presented are the most recent complete sets of quarterly data. 5. Generation and synchronization of N-scroll chaotic and hyperchaotic attractors in fourth-order systems Institute of Scientific and Technical Information of China (English) Yu Si-Min; Ma Zai-Guang; Qiu Shui-Sheng; Peng Shi-Guo; Lin Qing-Hua 2004-01-01 Based on our previous works and Lyapunov stability theory, this paper studies the generation and synchronization of N-scroll chaotic and hyperchaotic attractors in fourth-order systems. A fourth-order circuit, by introducing additional breakpoints in the modified Chua oscillator, is implemented for the study of generation and synchronization of N-scroll chaotic attractors. This confirms the consistency of theoretical calculation, numerical simulation and circuit experiment.Furthermore, we give a refined and extended study of generating and synchronizing N-scroll hyperchaotic attractors in the fourth-order MCK system and report the new theoretical result, which is verified by computer simulations. 6. An integrated framework for assessing vulnerability to climate change and developing adaptation strategies for coffee growing families in Mesoamerica OpenAIRE Baca, María; Läderach, Peter; Haggar, Jeremy; Schroth, Götz; Ovalle, Oreana 2014-01-01 The Mesoamerican region is considered to be one of the areas in the world most vulnerable to climate change. We developed a framework for quantifying the vulnerability of the livelihoods of coffee growers in Mesoamerica at regional and local levels and identify adaptation strategies. Following the Intergovernmental Panel on Climate Change (IPCC) concepts, vulnerability was defined as the combination of exposure, sensitivity and adaptive capacity. To quantify exposure, changes in the climatic ... 7. PREFACE: Fourth Meeting on Constrained Dynamics and Quantum Gravity Science.gov (United States) Cadoni, Mariano; Cavaglia, Marco; Nelson, Jeanette E. 2006-04-01 The formulation of a quantum theory of gravity seems to be the unavoidable endpoint of modern theoretical physics. Yet the quantum description of the gravitational field remains elusive. The year 2005 marks the tenth anniversary of the First Meeting on Constrained Dynamics and Quantum Gravity, held in Dubna (Russia) due to the efforts of Alexandre T. Filippov (JINR, Dubna) and Vittorio de Alfaro (University of Torino, Italy). At the heart of this initiative was the desire for an international forum where the status and perspectives of research in quantum gravity could be discussed from the broader viewpoint of modern gauge field theories. Since the Dubna meeting, an increasing number of scientists has joined this quest. Progress was reported in two other conferences in this series: in Santa Margherita Ligure (Italy) in 1996 and in Villasimius (Sardinia, Italy) in 1999. After a few years of working silence'' the time was now mature for a new gathering. The Fourth Meeting on Constrained Dynamics and Quantum Gravity (QG05) was held in Cala Gonone (Sardinia, Italy) from Monday 12th to Friday 16th September 2005. Surrounded by beautiful scenery, 100 scientists from 23 countries working in field theory, general relativity and related areas discussed the latest developments in the quantum treatment of gravitational systems. The QG05 edition covered many of the issues that had been addressed in the previous meetings and new interesting developments in the field, such as brane world models, large extra dimensions, analogue models of gravity, non-commutative techniques etc. The format of the meeting was similar to the previous ones. The programme consisted of invited plenary talks and parallel sessions on cosmology, quantum gravity, strings and phenomenology, gauge theories and quantisation and black holes. A major goal was to bring together senior scientists and younger people at the beginning of their scientific career. We were able to give financial support to both 8. Fourth Annual Nursing Leadership Congress: "Driving Patient Safety Through Transformation" Conference proceedings. Science.gov (United States) Pinakiewicz, Diane; Smetzer, Judy; Thompson, Pamela; Navarra, Mary Beth; Lambert, Monique 2009-06-01 In September 2008, nurse executives from around the country met in Scottsdale, Ariz, to develop practical tools and recommendations for "Driving Patient Safety Through Transformation," the theme of the fourth annual Nursing Leadership Congress. The Congress was made possible through an educational grant from McKesson and Intel in collaboration with sponsorship from the American Organization of Nurse Executives, Institute for Safe Medication Practices and National Patient Safety Foundation. This paper summarizes the Congress plenary sessions and roundtable discussions. Plenaries included the following: *Transformational Leadership: The Role of Leaders in Managing Complex Problems *Using the Baldrige Business Model as the Infrastructure for Creating a Culture of Patient Safety *Prospects for Structural Reform in Health Care Roundtables included the following: *Joy and Meaning in Work *Managing Chronic Care Across the Continuum *The Future of Acute Care Delivery in Light of Changing Reimbursement* Leveraging Transparency to Drive Patient Safety *Collaborative Partnerships for Driving a Patient Safety Agenda *Innovative Solutions for Patient Safety *Implementing the Fundamentals of the Toyota Production Model forHealthcare 9. Comparison of e-mail communication skills among first- and fourth-year dental students. Science.gov (United States) Oakley, Marnie; Horvath, Zsuzsa; Weinberg, Seth M; Bhatt, Jaya; Spallek, Heiko 2013-11-01 As e-mail and other forms of electronic communication increase in popularity, it is important for dental schools to consider a curriculum that prepares their graduates to understand and apply effective electronic communication strategies to their patients. Reflecting this shift in communication behavior, the American Medical Association has developed specific e-mail communication guidelines. Some behavioral examples in these guidelines include protecting patients' protected health information (PHI), ensuring proper record keeping, and using professional, courteous, and understandable language. In this study, a sample of first- and fourth-year dental students (n=160) at the University of Pittsburgh School of Dental Medicine participated in an assignment assessing their patient-provider e-mail communication skills. A rubric was used to evaluate and compare the data between dental student classes. The results reveal a generalized lack of compliance with several of these guidelines by both classes (e.g., failure to protect PHI), despite efforts to expose students to these concepts in the curriculum. In an effort to train emerging dentists to function in a rapidly changing technological environment, these findings suggest a need for growth and development of curricula and perhaps guidelines/recommendations for behavioral competencies regarding dental students' use of electronic communication in the patient care environment. 10. A review of football injuries on third and fourth generation artificial turfs compared with natural turf. Science.gov (United States) Williams, Sean; Hume, Patria A; Kara, Stephen 2011-11-01 rotational stiffness properties of shoe-surface interfaces, decreased impact attenuation properties of surfaces, differing foot loading patterns and detrimental physiological responses. Changing between surfaces may be a precursor for injury in soccer. In conclusion, studies have provided strong evidence for comparable rates of injury between new generation artificial turfs and natural turfs. An exception is the likely increased risk of ankle injury on third and fourth generation artificial turfs. Therefore, ankle injury prevention strategies must be a priority for athletes who play on artificial turf regularly. Clarification of effects of artificial surfaces on muscle and knee injuries are required given inconsistencies in incidence rate ratios depending on the football code, athlete, gender or match versus training. PMID:21985213 11. Time-Periodic Solution of a 2D Fourth-Order Nonlinear Parabolic Equation Indian Academy of Sciences (India) Xiaopeng Zhao; Changchun Liu 2014-08-01 By using the Galerkin method, we study the existence and uniqueness of time-periodic generalized solutions and time-periodic classical solutions to a fourth-order nonlinear parabolic equation in 2D case. 12. Thermodynamic and classical instability of AdS black holes in fourth-order gravity International Nuclear Information System (INIS) We study thermodynamic and classical instability of AdS black holes in fourth-order gravity. These include the BTZ black hole in new massive gravity, Schwarzschild-AdS black hole, and higher-dimensional AdS black holes in fourth-order gravity. All thermodynamic quantities which are computed using the Abbot-Deser-Tekin method are used to study thermodynamic instability of AdS black holes. On the other hand, we investigate the s-mode Gregory-Laflamme instability of the massive graviton propagating around the AdS black holes. We establish the connection between the thermodynamic instability and the GL instability of AdS black holes in fourth-order gravity. This shows that the Gubser-Mitra conjecture holds for AdS black holes found from fourth-order gravity 13. Anomalous single production of the fourth SM family quarks at Tevatron CERN Document Server Arik, E; Sultansoy, S F 2003-01-01 Possible single productions of fourth family u_{4} and d_{4} quarks via anomalous q_{4}qV interactions at Tevatron are studied. Signature of such processes are discussed and compared with the recent results from Tevatron. 14. Search for anomalous single production of the fourth SM family quark decaying into a light scalar CERN Document Server Arik, E; Sultansoy, S F 2003-01-01 Superjet events observed by the CDF Collaboration are interpreted as anomalous single production of the fourth SM family$u_4$quark, decaying into a new light scalar particle. The specific predictions of the proposed mechanism are discussed. 15. Mixed Waste Management Facility Groundwater Monitoring Report, Fourth Quarter 1998 and 1998 Summary Energy Technology Data Exchange (ETDEWEB) Chase, J. 1999-04-29 During fourth quarter 1998, ten constituents exceeded final Primary Drinking Water Standards (PDWS) in groundwater samples from downgradient monitoring wells at the Mixed Waste Management Facility. No constituents exceeded final PDWS in samples from the upgradient monitoring wells. 16. The Savannah River Site Groundwater Monitoring Program Fourth Quarter 2000 (October thru December 2000) Energy Technology Data Exchange (ETDEWEB) Dukes, M.D. 2001-08-02 This report summarizes the Groundwater Monitoring Program conducted by SRS during fourth quarter 2000. It includes the analytical data, field data, data review, quality control, and other documentation for this program. 17. MULTIPLE POSITIVE SOLUTIONS TO FOURTH-ORDER SINGULAR BOUNDARY VALUE PROBLEMS Institute of Scientific and Technical Information of China (English) 2011-01-01 In this paper,using the Krasnaselskii's fixed point theory in cones and localization method,under more general conditions,the existence of n positive solutions to a class of fourth-order singular boundary value problems is considered. 18. Fourth-quarter Economic Growth and Time-varying Expected Returns DEFF Research Database (Denmark) Møller, Stig V.; Rangvid, Jesper not predict returns. Fourth-quarter economic growth rates contain considerably more information about expected returns than standard variables used in the literature, are robust to the choice of macro variable, and work in-sample, out-of-sample, and in subsamples. To help explain these results, we show...... that economic growth and growth in consumer confidence are correlated during the fourth quarter, but not during the other quarters: When economic growth is low during the fourth quarter, confidence in the economy is also low such that investors require higher future returns. We discuss rational and behavioral...... reasons why fourth-quarter economic growth, growth in consumer confidence, and expected returns are related.... 19. Which Fourth-Grade Children Participate in School Breakfast and Do Their Parents Know It? OpenAIRE Guinn, Caroline H; Baxter, Suzanne Domel; THOMPSON, WILLIAM O.; FRYE, FRANCESCA H. A.; KOPEC, CANDACE T. 2002-01-01 Objective: To explore fourth-graders' school breakfast participation by gender and race (black, white) and examine the extent to which parents' responses to “Does this child usually eat school breakfast?” reflected their children's participation. 20. Clinical, histological, and genetic features of fourth ventricle ependymoma in the elderly. Case report. Science.gov (United States) Hayashi, Takuro; Inamasu, Joji; Kanai, Ryuichi; Sasaki, Hikaru; Shinoda, Jun; Hirose, Yuichi 2012-01-01 A 71-year-old woman presented with a rare case of geriatric ependymoma originating from the fourth ventricle manifesting as progressive gait and memory disturbance. Imaging studies revealed an extraaxial mass in the fourth ventricle protruding into the right cerebellomedullary cistern, with concomitant obstructive hydrocephalus. Surgery achieved subtotal removal since the tumor tightly adhered to the right vestibular area of the fourth ventricular floor. The histological diagnosis was ependymoma, which was also confirmed by comparative genetic hybridization. Although she developed severe laryngeal edema and worsening of the hydrocephalus postoperatively which required additional treatment, she recovered with residual mild gait disturbance, and was transferred to a rehabilitation facility. Fourth ventricle ependymoma in the elderly is rare. Comparative genetic hybridization may be important in the diagnosis of geriatric ependymoma and in the choice for adjuvant therapy as well as in estimating the prognosis for patients with rare types of ependymoma. PMID:22976148 1. Fourth NASA Workshop on Computational Control of Flexible Aerospace Systems, part 2 Science.gov (United States) Taylor, Lawrence W., Jr. (Compiler) 1991-01-01 A collection of papers presented at the Fourth NASA Workshop on Computational Control of Flexible Aerospace Systems is given. The papers address modeling, systems identification, and control of flexible aircraft, spacecraft and robotic systems. 2. On the sign-regularity of positive fourth-order differential operators OpenAIRE Vladimirov, A. A. 2016-01-01 It is shown that positivity in$(0,1)\\times (0,1)of Green function of positively defined fourth-order ordinary differential operator (with separated boundary conditions) is a criterium of sign-regularity of this operator. 3. Adenocarcinoma of the third and fourth portions of the duodenum: The capsule endoscopy value OpenAIRE Paquissi, Feliciano Chanana; Lima, Ana Henriqueta Filipe Bunga Pimentel; Lopes, Maria de Fátima do Nascimento Vieira; Diaz, Francisco Viamontes 2015-01-01 Primary adenocarcinoma of the small intestine occurs in over 50% of cases in the duodenum. However, its location in the third and fourth duodenal portions occurs rarely and is a diagnostic challenge. The aim of this work is to report an adenocarcinoma of the third and fourth duodenal portions, emphasizing its diagnostic difficulty and the value of video capsule endoscopy. A man, 40 years old, with no medical history, with abdominal discomfort and progressive fatigue, presented four months ago... 4. The community's research and development programme on decommissioning of nuclear installations. Fourth annual progress report 1988 International Nuclear Information System (INIS) This is the fourth annual progress report on the European Community's programme (1984-88) of research on the decommissioning of nuclear installations. It shows the status of the programme at 31 December 1988. The fourth progress report describes the objectives, scope and work programme of the 72 research contracts concluded, as well as the progress of work achieved and the results obtained in 1988 5. Block Hybrid Collocation Method with Application to Fourth Order Differential Equations Directory of Open Access Journals (Sweden) Lee Ken Yap 2015-01-01 Full Text Available The block hybrid collocation method with three off-step points is proposed for the direct solution of fourth order ordinary differential equations. The interpolation and collocation techniques are applied on basic polynomial to generate the main and additional methods. These methods are implemented in block form to obtain the approximation at seven points simultaneously. Numerical experiments are conducted to illustrate the efficiency of the method. The method is also applied to solve the fourth order problem from ship dynamics. 6. Alveolar rhabdomyosarcoma originating between the fourth and fifth metatarsal--case report and literature review. LENUS (Irish Health Repository) Bolger, J C 2010-09-01 We report a case of alveolar rhabdomyosarcoma arising between the fourth and fifth metatarsal. A 13-year-old boy presented to outpatients with a history of pain and swelling in the lateral aspect of his left forefoot. Plain radiographs and MRI showed a soft tissue mass displacing the fourth metatarsal. Percutaneous biopsy revealed an alveolar rhabdomyosarcoma. Staging scans showed advanced metastatic disease. The patient was treated with chemotherapy. This highly malignant lesion remains challenging to diagnose, and difficult to treat successfully. 7. Joint Use of Third and Fourth Cumulants in Independent Component Analysis OpenAIRE Virta, Joni; Nordhausen, Klaus; Oja, Hannu 2015-01-01 The independent component model is a latent variable model where the components of the observed random vector are linear combinations of latent independent variables. The aim is to find an estimate for a transformation matrix back to independent components. In moment-based approaches third cumulants are often neglected in favor of fourth cumulants, even though both approaches have similar appealing properties. This paper considers the joint use of third and fourth cumulants in finding indepen... 8. An Inquiry Into How Fourth-Grade Students Investigate Their Theories For Learning Scientific Vocabulary OpenAIRE Miller, Tatiana Frederiksen 2015-01-01 ABSTRACTAN INQUIRY INTO HOW FOURTH-GRADE STUDENTS INVESTIGATE THEIR THEORIES FOR LEARNING SCIENTIFIC VOCABULARYByTatiana F. MillerWhile instructional practices that engage students in developing metacognitive skills and capabilities are likely to be beneficial for all students, they may be particularly beneficial for students who are traditionally less well-served by schools. In this research, two culturally and linguistically diverse groups of fourth-grade students engaged in developing and ... 9. Fourth-order 2N-storage Runge-Kutta schemes Science.gov (United States) Carpenter, Mark H.; Kennedy, Christopher A. 1994-01-01 A family of five-stage fourth-order Runge-Kutta schemes is derived; these schemes required only two storage locations. A particular scheme is identified that has desirable efficiency characteristics for hyperbolic and parabolic initial (boundary) value problems. This scheme is competitive with the classical fourth-order method (high-storage) and is considerably more efficient and accurate than existing third-order low-storage schemes. 10. Energy and Climate Change Energy Technology Data Exchange (ETDEWEB) NONE 2007-06-15 Climate change, and more specifically the carbon emissions from energy production and use, is one of the more vexing problems facing society today. The Intergovernmental Panel on Climate Change (IPCC) has just completed its latest assessment on the state of the science of climate change, on the potential consequences related to this change, and on the mitigation steps that could be implemented beginning now, particularly in the energy sector. Few people now doubt that anthropogenic climate change is real or that steps must be taken to deal with it. The World Energy Council has long recognized this serious concern and that in its role as the world's leading international energy organization, it can address the concerns of how to provide adequate energy for human well-being while sustaining our overall quality of life. It has now performed and published 15 reports and working papers on this subject. This report examines what has worked and what is likely to work in the future in this regard and provides policymakers with a practical roadmap to a low-carbon future and the steps needed to achieve it. 11. Radiative Forcing of Climate Change Energy Technology Data Exchange (ETDEWEB) Ramaswamy, V.; Boucher, Olivier; Haigh, J.; Hauglustaine, D.; Haywood, J.; Myhre, G.; Nakajima, Takahito; Shi, Guangyu; Solomon, S.; Betts, Robert E.; Charlson, R.; Chuang, C. C.; Daniel, J. S.; Del Genio, Anthony D.; Feichter, J.; Fuglestvedt, J.; Forster, P. M.; Ghan, Steven J.; Jones, A.; Kiehl, J. T.; Koch, D.; Land, C.; Lean, J.; Lohmann, Ulrike; Minschwaner, K.; Penner, Joyce E.; Roberts, D. L.; Rodhe, H.; Roelofs, G.-J.; Rotstayn, Leon D.; Schneider, T. L.; Schumann, U.; Schwartz, Stephen E.; Schwartzkopf, M. D.; Shine, K. P.; Smith, Steven J.; Stevenson, D. S.; Stordal, F.; Tegen, I.; van Dorland, R.; Zhang, Y.; Srinivasan, J.; Joos, Fortunat 2001-10-01 Chapter 6 of the IPCC Third Assessment Report Climate Change 2001: The Scientific Basis. Sections include: Executive Summary 6.1 Radiative Forcing 6.2 Forcing-Response Relationship 6.3 Well-Mixed Greenhouse Gases 6.4 Stratospheric Ozone 6.5 Radiative Forcing By Tropospheric Ozone 6.6 Indirect Forcings due to Chemistry 6.7 The Direct Radiative Forcing of Tropospheric Aerosols 6.8 The Indirect Radiative Forcing of Tropospheric Aerosols 6.9 Stratospheric Aerosols 6.10 Land-use Change (Surface Albedo Effect) 6.11 Solar Forcing of Climate 6.12 Global Warming Potentials hydrocarbons 6.13 Global Mean Radiative Forcings 6.14 The Geographical Distribution of the Radiative Forcings 6.15 Time Evolution of Radiative Forcings Appendix 6.1 Elements of Radiative Forcing Concept References. 12. Urban Vulnerability and Climate Change in Africa DEFF Research Database (Denmark) Jørgensen, Gertrud IPCC climate change scenarios, which also consider possible changes in urban population, have been developed. Innovative strategies to land use and spatial planning are proposed that seek synergies between the adaptation to climate change and the need to solve social problems. Furthermore, the book......Urbanisation and climate change are among the major challenges for sustainable development in Africa. The overall aim of this book is to present innovative approaches to vulnerability analysis and for enhancing the resilience of African cities against climate change-induced risks. Locally adapted...... explores the role of governance in successfully coping with climate-induced risks in urban areas. The book is unique in that it combines: a top-down perspective of climate change modeling with a bottom-up perspective of vulnerability assessment; quantitative approaches from engineering sciences... 13. Climate - Understanding climate change in order to act International Nuclear Information System (INIS) In a first part, the author proposes an overview of considerations about climate change and global warming. He discusses greenhouse gas emissions and their perspectives of evolution (IPCC scenarios, recent assessments, unreachable objectives). He comments and discusses the consequences and effects of climate change and global warming (impact on the biosphere and predictable consequences, the largely unknown issue of oceans). He comments the relationship between warming and meteorological evolutions (what is sure and what is not, what is due to climate change and what is not), and the associated risks and hazards 14. Modeling of climate change impacts on agriculture, forestry and fishery International Nuclear Information System (INIS) Changes in climate affect agriculture, forest and fisheries. This paper examines the climate change impact on crop production, fishery and forestry using state - of - the - art modeling technique. Crop growth model InfoCrop was used to predict the climate change impacts on the yields of rice, wheat and maize in Bangladesh. Historical climate change scenario has little or no negative impacts on rice and wheat yields in Mymensingh and Dinajpur but IPCC climate change scenario has higher negative impacts. There is almost no change in the yields of maize for the historical climate change scenario in the Chittagong, Hill Tracts of but there is a small decrease in the yields of rice and maize for IPCC climate change scenario. A new statistical model to forecast climate change impacts on fishery in the world oceans has been developed. Total climate change impact on fishery in the Indian Ocean is negative and the predictor power is 94.14% for eastern part and 98.59% for the western part. Two models are presented for the mangrove forests of the Sundarbans. To bole volumes of the pioneer, intermediate and climax are simulated for three different logging strategies and the results have been discussed in this paper. (author) 15. An 'agenda for change': Quantifying climate change impacts on natural resource-based economies Energy Technology Data Exchange (ETDEWEB) MacGregor, James; Reid, Hannah; Sahlen, Linda 2006-10-15 For climate change adaptation to be beneficial to developing countries, it must begin quickly and this will require domestic political will. The third assessment report of the Intergovernmental Panel on Climate Change (IPCC) made clear that even if the Kyoto Protocol is fully implemented, inertia in climatic systems means that some level of climate change is unavoidable. The countries most vulnerable to CC include many developing nations; while those better-able to adapt and less willing to mitigate are those most guilty of past pollution, including many developed nations. 16. Surgical Fixation of Fourth and Fifth Metacarpal Shaft Fractures with Flexible Intramedullary Absorbable Rods: Early Clinical Outcomes and Implications Institute of Scientific and Technical Information of China (English) Ge Xiong; Zi-Run Xiao; Shi-Gong Guo; Wei Zheng; Lu-Fei Dai 2015-01-01 Background:To avoid the irritation of tendons and soft tissues as well as hardware-related problems,we designed an intramedullary fixation with bioabsorbable rods for the treatment of the metacarpal shaft fractures.Methods:Five patients with nine shaft fractures of the fourth and fifth metacarpi were treated with intramedullary absorbable implants and followed up with an average of 4.2 months postoperatively.Results:At final follow-up,all patients achieved fracture union with no signs of inflammatory or subcutaneous effusion.There was no shortening,angulatory,or rotatory deformity.There was almost full active extension range of motion (ROM) of the metacarpophalangeal joints while the active flexion ROM of these joints was 80.7 ± 9.6°.Compared with the contralateral hand,the grip strength of the injured hand was 94.0 ± 9.6%.X-rays showed that the arch of the second to fifth metacarpal heads was smooth.There were no intramedullary lytic changes and soft tissue swellings.Conclusion:The intramedullary absorbable implants are a safe,simple,and practical treatment for fourth and fifth metacarpal fractures with good early clinical outcomes and no significant complications. 17. Prospective investigation of penile length with newborn male circumcision and second to fourth digit ratio Science.gov (United States) Park, Jong Kwan; Doo, A. Ram; Kim, Joo Heung; Park, Hyung Sub; Do, Jung Mo; Choi, Hwang; Park, Seung Chol; Kim, Myung Ki; Jeong, Young Beom; Kim, Hyung Jim; Kim, Young Gon; Shin, Yu Seob 2016-01-01 Introduction: We prospectively investigated the relationship between newborn male circumcision (NMC) and second to fourth digit ratio with penile length. Methods: As participants for our study, we identified already circumcised young patients who visited our hospital for urological treatment. The age at which the circumcision had been done was assessed. The patients’ height and weight were measured. Second to fourth digit ratio was calculated by measuring the second and fourth digit lengths. The flaccid and erectile penile lengths were measured from the base of the penis to the tip of the glans in standing position. Results: A total of 248 patients were included in our study. In univariate analysis, height, second to fourth digit ratio, flaccid penile length, and age of circumcision were associated with erectile penile length. Among these variables, second to fourth digit ratio, flaccid penile length, and age of circumcision were significant predictive factors for erectile penile length in multivariate analysis. The subjects were divided into two groups, including 72 patients in the NMC group and 176 patients in the non-NMC group. No significant difference was found in height, weight, and second to fourth digit ratio between both groups. However, flaccid (p<0.001) and erectile (p=0.001) penile lengths were shorter in the NMC group than in the non-NMC group. Conclusions: Despite the small number of subjects, this study shows that NMC was associated with shorter penile length. Second to fourth digit ratio, flaccid penile length, and age of circumcision were also significant predictive factors for erectile penile length. Further multicentre studies with larger number of subjects and biochemical analyses are needed for potential clinical applicability. PMID:27695583 18. Scientific consensus and climate change: the codification of a global research agenda International Nuclear Information System (INIS) The 'scientific consensus' which influenced the Framework Convention on Climate Change was carefully drafted by the Intergovernmental Panel on Climate Change (IPCC) between 1988 and 1992. In spite of it, there have been divergent national responses and policy controversy continues. The willingness of States to reduce the emission of greenhouse gases appears to be declining. An explanation for this is proposed which stresses the question of whether the nature of the scientific advice as sought and given bears some responsibility for the weak policy response. Institutional and personality factors in the formulation of IPCC advice are explored, as is the policy model upon which advice was given. It is concluded that this model is intrinsically unable to generate decisive environmental policy, but rather invites the institutions of the natural sciences and macro-economics to endow their research agendas with claims to policy relevance through the production of future 'findings'. (Author) 19. New climate change report turns up the heat on energy policymakers International Nuclear Information System (INIS) Any doubts about the role of nuclear power in fighting the damaging effects of climate change should be dispelled once and for all by the latest report from the Intergovernmental Panel on Climate Change (IPCC). The key findings of the IPCC's so-called 'Synthesis Report', published in November 2014, said that if left unchecked, climate change ''will increase the likelihood of severe, pervasive and irreversible impacts for people and ecosystems''. The Synthesis Report, which brought together the findings of the IPCC's Fifth Assessment Report produced by more than 800 scientists and released over more than one year, noted that multiple mitigation pathways are available that could limit warming to below 2 C relative to pre-industrial levels, all of which would need substantial cuts in emissions reductions over the coming few decades and near-zero emissions of CO2 and other long-lived greenhouse gases (GHGs) by the end of the century. As 2014 draws to a close, perhaps those with the power to effect energy policy change will take time to reflect on the mounting evidence of what is good, and what is not good, in terms of balancing the world's energy needs with the planet's overall health. 20. High resolution scenarios of land-use and land-cover change for the conterminous United States Science.gov (United States) Sleeter, B. M.; Sohl, T. L.; Bouchard, M. A.; Reker, R. R.; Sayler, K.; Sleeter, R.; Soulard, C. E.; Wilson, T. S. 2012-12-01 We describe a series of high resolution maps of past and projected changes in land use and land cover (LULC) for the conterminous United States for the period 1992 to 2100. Four scenarios from the Intergovernmental Panel on Climate Change's (IPCC) Special Report on Emission Scenarios (SRES) were used to create annual maps showing spatially explicit change in 15 LULC classes at a spatial resolution of 250 meters. A modular land-use modeling approach was utilized with distinct demand and spatial allocation components. To quantify demand for future LULC change (i.e. the quantity of changes in land use and land cover classes), a scenario downscaling model was developed to extend global scenarios from the IPCC to hierarchically nested ecoregions of the U.S. The Forecasting Scenarios (FORE-SCE) land use model was then employed to allocate scenario demand on the landscape. Both models were parameterized at the ecoregion scale and relied extensively on land use histories and expert knowledge. Results reveal large differences across IPCC-SRES scenarios. Scenarios prioritizing economic development over environmental protection result in the highest rates of LULC change, particularly in regions with extensive forest management, large urban areas, and/or large investments in agricultural land. Scenarios where environmental protection is emphasized result in slower rates of change and less intensity in regional land use patterns. 1. ATZ (3-amino-1,2,4-triazole injected into the fourth cerebral ventricle influences the Bezold-Jarisch reflex in conscious rats Directory of Open Access Journals (Sweden) Vitor E. Valenti 2010-01-01 Full Text Available OBJECTIVES: Many studies have investigated the importance of oxidative stress on the cardiovascular system. In this study we evaluated the effects of central catalase inhibition on cardiopulmonary reflex in conscious Wistar rats. METHODS: Male Wistar rats were implanted with a stainless steel guide cannula in the fourth cerebral ventricle. The femoral artery and vein were cannulated for mean arterial pressure and heart rate measurement and for drug infusion, respectively. After basal mean arterial pressure and heart rate recordings, the cardiopulmonary reflex was tested with a dose of phenylbiguanide (PBG, 8 μg/kg, bolus. Cardiopulmonary reflex was evaluated before and μl15 minutes after 1.0 μl 3-amino-1,2,4-triazole (ATZ, 0.01g/100μl0.01 g/100 μl injection into the fourth cerebral ventricle. Vehicle treatment did not change cardiopulmonary reflex responses. RESULTS: Central ATZ significantly increased hypotensive responses without influencing the bradycardic reflex. CONCLUSION: ATZ injected into the fourth cerebral ventricle increases sympathetic inhibition but does not change the parasympathetic component of the cardiopulmonary reflex in conscious Wistar rats. 2. PREFACE: Fourth Meeting on Constrained Dynamics and Quantum Gravity Science.gov (United States) Cadoni, Mariano; Cavaglia, Marco; Nelson, Jeanette E. 2006-04-01 The formulation of a quantum theory of gravity seems to be the unavoidable endpoint of modern theoretical physics. Yet the quantum description of the gravitational field remains elusive. The year 2005 marks the tenth anniversary of the First Meeting on Constrained Dynamics and Quantum Gravity, held in Dubna (Russia) due to the efforts of Alexandre T. Filippov (JINR, Dubna) and Vittorio de Alfaro (University of Torino, Italy). At the heart of this initiative was the desire for an international forum where the status and perspectives of research in quantum gravity could be discussed from the broader viewpoint of modern gauge field theories. Since the Dubna meeting, an increasing number of scientists has joined this quest. Progress was reported in two other conferences in this series: in Santa Margherita Ligure (Italy) in 1996 and in Villasimius (Sardinia, Italy) in 1999. After a few years of working silence'' the time was now mature for a new gathering. The Fourth Meeting on Constrained Dynamics and Quantum Gravity (QG05) was held in Cala Gonone (Sardinia, Italy) from Monday 12th to Friday 16th September 2005. Surrounded by beautiful scenery, 100 scientists from 23 countries working in field theory, general relativity and related areas discussed the latest developments in the quantum treatment of gravitational systems. The QG05 edition covered many of the issues that had been addressed in the previous meetings and new interesting developments in the field, such as brane world models, large extra dimensions, analogue models of gravity, non-commutative techniques etc. The format of the meeting was similar to the previous ones. The programme consisted of invited plenary talks and parallel sessions on cosmology, quantum gravity, strings and phenomenology, gauge theories and quantisation and black holes. A major goal was to bring together senior scientists and younger people at the beginning of their scientific career. We were able to give financial support to both 3. Discussion about"The Fourth Industry Revolution"%关于“第四次工业革命”的探讨 Institute of Scientific and Technical Information of China (English) 张海平 2014-01-01 介绍笔者在德国2013汉诺威工业博览会上所了解到的一些关于“第四次工业革命”的最新信息,包括:“第四次工业革命”的概念、特点,以及可能导致在工业的变革和对未来工业的影响。%This paper introduced the discussion about"The Fourth Industry Revolution", which the author read dur-ing the Hannover Messe 2013. And it includes the ideal, features of"The Fourth Industry Revolution". Moreover, some changes and influences over future industry were also proposed in this paper. 4. Significant reductions in oil quality and lipid content of oilseed rape (Brassica napus L.) under climate change DEFF Research Database (Denmark) Namazkar, Shahla; Egsgaard, Helge; Frenck, Georg; Despite of the potential importance to food and bioenergy purposes effects from climate change on plant oil quality have hardly been characterized. Worldwide Brassica napus, rapeseed or oilseed rape, is the second largest source of vegetable oil and the predominant oil crop in Europe. We found...... significant changes in oil quality and quantity of cultivars of oilseed rape grown in five future climate scenarios with elevated [CO2], [O3], temperature and combinations hereof (~RCP8.5, IPCC 2013).... 5. Choroid plexus of the fourth ventricle: Review and anatomic study highlighting anatomical variations. Science.gov (United States) Tubbs, R Shane; Shoja, Mohammadali M; Aggarwal, Anjali; Gupta, Tulika; Loukas, Marios; Sahni, Daisy; Ansari, Shaheryar F; Cohen-Gadol, Aaron A 2016-04-01 Relatively few studies have been performed that analyze the morphology of the choroid plexus of the fourth ventricle. Due to the importance of this tissue as a landmark on imaging and during surgical intervention of the fourth ventricle, the authors performed a cadaveric study to better characterize this important structure. The choroid plexus of the fourth ventricle of 60 formalin fixed adult human brains was examined and measured. The horizontal distance from the midline to the lateral most point of the protruding tip of the horizontal limbs was measured. In the majority of the 60 brain specimens, right and left horizontal limbs of the choroid plexus were seen extending from the midline and protruding out of their respective lateral apertures of the fourth ventricle and into the subarachnoid space. However, on 3.3% of sides, there was absence of an extension into the foramen of Luschka and in one specimen, this lack of extension into the foramen of Luschka was bilateral. On two sides, there was discontinuity between the midline choroid plexus and the tuft of choroid just outside the foramen of Luschka. For specimens in which the choroid plexus did protrude through the foramen of Luschka (96.7%), these tufts were located anterior to the flocculus and inferolateral to the facial/vestibulocochlear nerve complex and posterosuperior to the glossopharyngeal/vagal/accessory complex. A thorough understanding of the normal and variant anatomy of the fourth ventricular choroid plexus is necessary for those who operate in, or interpret imaging of, this region. 6. Climate Change and Water Resources Management: A Federal Perspective Science.gov (United States) Brekke, Levi D.; Kiang, Julie E.; Olsen, J. Rolf; Pulwarty, Roger S.; Raff, David A.; Turnipseed, D. Phil; Webb, Robert S.; White, Kathleen D. 2009-01-01 Many challenges, including climate change, face the Nation's water managers. The Intergovernmental Panel on Climate Change (IPCC) has provided estimates of how climate may change, but more understanding of the processes driving the changes, the sequences of the changes, and the manifestation of these global changes at different scales could be beneficial. Since the changes will likely affect fundamental drivers of the hydrological cycle, climate change may have a large impact on water resources and water resources managers. The purpose of this interagency report prepared by the U.S. Geological Survey (USGS), U.S. Army Corps of Engineers (USACE), Bureau of Reclamation (Reclamation), and National Oceanic and Atmospheric Administration (NOAA) is to explore strategies to improve water management by tracking, anticipating, and responding to climate change. This report describes the existing and still needed underpinning science crucial to addressing the many impacts of climate change on water resources management. 7. Ocean bottom pressure changes lead to a decreasing length-of-day in a warming climate OpenAIRE F. Landerer; Jungclaus, J.; Marotzke, J. 2007-01-01 We use a coupled climate model to evaluate ocean bottom pressure changes in the IPCC-A1B climate scenario. Ocean warming in the 21st and 22nd centuries causes secular oceanic bottom pressure anomalies. The essential feature is a net mass transfer onto shallow shelf areas from the deeper ocean areas, which exhibit negative bottom pressure anomalies. We develop a simple mass redistribution model that explains this mechanism. Regionally, however, distinct patterns of bottom pressure anomalies em... 8. Quarterly environmental radiological survey summary -- fourth quarter 1997: 100, 200, 300, and 600 areas Energy Technology Data Exchange (ETDEWEB) McKinney, S.M. 1998-01-27 Routine radiological surveys are part of near-facility environmental monitoring which monitors and helps direct the reduction of the radiological areas at the Hanford Site. The routine radiological surveys are performed by the Southern Area Remediation Support Group and the Site Support Services Radiological Control Group as directed by Environmental Monitoring and Investigations. The surveys performed have included inactive waste sites; outdoor radiological control areas; tank farm perimeters and associated diversion boxes, lift stations, and vent stations; perimeters of active or uncovered waste sites such as burial grounds, retention basins, ponds, process trenches, and ditches; underground pipelines; and road and rail surfaces. This report provides a synopsis of the radiological surveys performed in support of near-facility environmental monitoring at the Hanford Site during calendar year 1997. The Fourth Quarter 1997 survey results and the status of actions required are also discussed. A waste site survey schedule, Routine Environmental Monitoring Schedule Calendar Year 1997, WHC-SP-0098-8, was developed by Environmental Monitoring and Investigations and reviewed by the Southern Area Remediation Support Group and the Site Support Services Radiological Control Group. Environmental Monitoring and Investigations reviews the radiological survey reports and files a copy for historical purposes and reference. Radiological conditions are tracked and trends noted. All sites are surveyed at least once each year. The survey frequencies for particular sites are based on site history, radiological conditions, and general maintenance. Special surveys may be conducted at irregular frequencies if conditions warrant (e.g., growth of deep-rooted vegetation is noted at a waste site). Radiological surveys are conducted to detect surface contamination and document changes in vegetation growth, biological intrusion, erosion, and general site maintenance conditions. Survey data are 9. Fourth-Order Contour Mode ZnO-on-SOI Disk Resonators for Mass Sensing Applications Directory of Open Access Journals (Sweden) Ivan Rivera 2015-04-01 Full Text Available In this work, we have investigated the design, fabrication and testing of ZnO-on-SOI fourth-order contour mode disk resonators for mass sensing applications. This study aims to unveil the possibility for real-time practical mass sensing applications by using high-Q ZnO-on-SOI contour-mode resonators while taking into account their unique modal characteristics. Through focused ion beam (FIB direct-write metal deposition techniques, the effects of localized mass loading on the surface of three extensional mode devices have been investigated. Ten microfabricated 40 mm-radius disk resonators, which all have a 20 mm-thick silicon device layer and 1 mm-thick ZnO transducer layer but varied anchor widths and numbers, have exhibited resonant frequencies ranging from 84.9 MHz to 86.7 MHz with Q factors exceeding 6000 (in air and 10,000 (in vacuum, respectively. It has been found that the added mass at the nodal locations leads to noticeable Q-factor degradation along with lower induced frequency drift, thereby resulting in reduced mass sensitivity. All three measured devices have shown a mass sensitivity of ~1.17 Hz·fg−1 at the maximum displacement points with less than 33.3 ppm of deviation in term of fractional frequency change. This mass sensitivity is significantly higher than 0.334 Hz·fg−1 at the nodal points. Moreover, the limit of detection (LOD for this resonant mass sensor was determined to be 367 ag and 1290 ag (1 ag = 10−18 g for loaded mass at the maximum and minimum displacement points, accordingly. 10. Radiologic-pathologic Correlation-An Advanced Fourth-year Elective: How We Do It. Science.gov (United States) Hartman, Matthew; Silverman, Jan; Spruill, Laura; Hill, Jeanne 2016-07-01 Traditionally, the radiology elective has been designed to teach medical students the fundamentals of radiologic interpretation. When questioned, many students state that they want to take a radiology elective so they can "interpret images." For the students on radiology, rotation/elective education was often passive, consisting of didactic conferences and observational shadowing of radiologists as they interpreted images. Students had only a superficial appreciation of how radiologists interacted with clinical services, multidisciplinary teams, and pathology. There was very little emphasis on imaging appropriateness or the most efficient and effective imaging for various clinical problems. With the expansion of numerous imaging modalities and the emphasis on patient-centered care, including imaging safety and dose reduction, it is important to change the focus of radiology education from interpretation to the optimal integration of imaging into clinical medicine. Radiology-pathology (rad path) electives were created at Allegheny General Hospital and the Medical University of South Carolina as a new option to provide a high-quality advanced elective for fourth-year medical students. These electives enable students to correlate radiologic images with gross and microscopic pathology specimens, thus increasing their knowledge and understanding of both. The rad path elective combines aspects of surgery, radiology, and pathology and requires students to be active learners. The implementation of this elective is an exciting work in progress that has been evolving over the past 2 and 4 years at Medical University of South Carolina and Allegheny General Hospital, respectively. We will discuss the historical basis for the elective, the advantages and challenges of having such an integrated course, and some different strategies for creating a rad path elective. PMID:27311804 11. Radiologic-pathologic Correlation-An Advanced Fourth-year Elective: How We Do It. Science.gov (United States) Hartman, Matthew; Silverman, Jan; Spruill, Laura; Hill, Jeanne 2016-07-01 Traditionally, the radiology elective has been designed to teach medical students the fundamentals of radiologic interpretation. When questioned, many students state that they want to take a radiology elective so they can "interpret images." For the students on radiology, rotation/elective education was often passive, consisting of didactic conferences and observational shadowing of radiologists as they interpreted images. Students had only a superficial appreciation of how radiologists interacted with clinical services, multidisciplinary teams, and pathology. There was very little emphasis on imaging appropriateness or the most efficient and effective imaging for various clinical problems. With the expansion of numerous imaging modalities and the emphasis on patient-centered care, including imaging safety and dose reduction, it is important to change the focus of radiology education from interpretation to the optimal integration of imaging into clinical medicine. Radiology-pathology (rad path) electives were created at Allegheny General Hospital and the Medical University of South Carolina as a new option to provide a high-quality advanced elective for fourth-year medical students. These electives enable students to correlate radiologic images with gross and microscopic pathology specimens, thus increasing their knowledge and understanding of both. The rad path elective combines aspects of surgery, radiology, and pathology and requires students to be active learners. The implementation of this elective is an exciting work in progress that has been evolving over the past 2 and 4 years at Medical University of South Carolina and Allegheny General Hospital, respectively. We will discuss the historical basis for the elective, the advantages and challenges of having such an integrated course, and some different strategies for creating a rad path elective. 12. Comparison of temperature indices for three IPCC SRES scenarios based on RegCM simulations for Poland in 2011–2030 period Directory of Open Access Journals (Sweden) Adam Jaczewski 2015-01-01 Full Text Available The regional climate model RegCM3 is used to investigate potential future changes of temperature indices in Poland for the period 2011–2030. The model is forced by ECHAM5/MPI-OM GCM data from World Data Centre (WDCC database for the 1971–1990 reference period and 2011–2030 projection period under SRES B1, A1B and A2 emission scenarios. Model output statistics methods are used to transform simulated minimum and maximum temperature data into realistic data. Selected indices of temperature extremes and their differences between the scenario simulations and the reference were calculated, for all scenarios, for the entire period and for each season. Results show a mean yearly increase in the number of summer and hot days and a decrease in the number of frost and ice days. Highest decline in the number of frost and ice days in autumn and an increase in spring is noticed. An highest increase in the number of summer and hot days is seen in summer. Future projections of these indices are relevant for studies on climate change impact in agriculture, tourism, health, transportation, road and building infrastructure. 13. Global climate change and cryospheric evolution in China Directory of Open Access Journals (Sweden) Qin D. 2009-02-01 Full Text Available Major outcomes of Working Group I, IPCC AR4 (2007, as well as the recent understandings from our regional climatic assessments in China were summarized. Changes of cryosphere in China, one of the major components in regional climate system, is specifically reviewed. Under the global/regional warming, all components of cryosphere in China (Tibetan Plateau and surroundings including glaciers, frozen ground (including permafrost and snow cover show rapid decay in the last decades. These changes have big socioeconomic impacts in west China, thus encourages both government and scientists pay more and more attention to this field. 14. Anticipated changes in the Nordic seas marine climate OpenAIRE Furevik, Tore; Drange, Helge; Sorteberg, Asgeir 2002-01-01 Possible future changes in the Nordic Seas marine climate are here discussed. The wide range of climate models used in the Intergovernmental Panel on Climate Change predict a global mean temperature increase between 1- 6ºC by the end of this century, with the estimates using the intermediate IPCC B2 scenario being in the range 1.9-3.4ºC. For climate models forced by a 1percent per year CO2 increase only, the so-called CMIP2 integrations, the increase in temperatures is close to 2º... 15. Living with poverty and climate change – a study on vulnerability to climate-related shocks on household level DEFF Research Database (Denmark) Jakobsen, Kristian Thor , especially at high altitudes. As an effect of these increased risks, humanitarian disasters caused by weather-related shocks are likely to increase in both number and severity. In that sense, it is vital to understand how people living in disaster-prone areas are handling such changes, and how the risk......Projections from the Intergovernmental Panel on Climate Change (IPCC) show that likely increases in the frequencies and intensities of extreme weather events are expected to have mostly adverse effects on natural and human systems (IPCC, 2007). Thus, the risk of suffering from an extreme weather...... addressing the environmental risks alone, but instead explores options that could assist households in achieving persistent welfare gains, no matter whether the expected outcomes of climate change in terms of extreme weather events are realized or not. In doing so, this thesis represents an effort... 16. Fourth-order partial differential equation noise removal on welding images International Nuclear Information System (INIS) Partial differential equation (PDE) has become one of the important topics in mathematics and is widely used in various fields. It can be used for image denoising in the image analysis field. In this paper, a fourth-order PDE is discussed and implemented as a denoising method on digital images. The fourth-order PDE is solved computationally using finite difference approach and then implemented on a set of digital radiographic images with welding defects. The performance of the discretized model is evaluated using Peak Signal to Noise Ratio (PSNR). Simulation is carried out on the discretized model on different level of Gaussian noise in order to get the maximum PSNR value. The convergence criteria chosen to determine the number of iterations required is measured based on the highest PSNR value. Results obtained show that the fourth-order PDE model produced promising results as an image denoising tool compared with median filter 17. Constraining a fourth generation of quarks: non-perturbative Higgs boson mass bounds CERN Document Server Bulava, John; Nagy, Attila 2013-01-01 We present a non-perturbative determination of the upper and lower Higgs boson mass bounds with a heavy fourth generation of quarks from numerical lattice computations in a chirally symmetric Higgs-Yukawa model. We find that the upper bound only moderately rises with the quark mass while the lower bound increases significantly, providing additional constraints on the existence of a straight-forward fourth quark generation. We examine the stability of the lower bound under the addition of a higher dimensional operator to the scalar field potential using perturbation theory, demonstrating that it is not significantly altered for small values of the coupling of this operator. For a Higgs boson mass of\\sim125\\mathrm{GeV}$we find that the maximum value of the fourth generation quark mass is$\\sim300\\mathrm{GeV}\$, which is already in conflict with bounds from direct searches.
18. Blasius flow and heat transfer of fourth-grade fluid with slip
Institute of Scientific and Technical Information of China (English)
B SAHOO; S PONCET
2013-01-01
This investigation deals with the effects of slip, magnetic field, and non-Newtonian flow parameters on the flow and heat transfer of an incompressible, electrically conducting fourth-grade fluid past an infinite porous plate. The heat transfer analysis is carried out for two heating processes. The system of highly non-linear differential equations is solved by the shooting method with the fourth-order Runge-Kutta method for moderate values of the parameters. The effective Broyden technique is adopted in order to improve the initial guesses and to satisfy the boundary conditions at infinity. An exceptional cross-over is obtained in the velocity profile in the presence of slip. The fourth-grade fluid parameter is found to increase the momentum boundary layer thickness, whereas the slip parameter substantially decreases it. Similarly, the non-Newtonian fluid parameters and the slip have opposite effects on the thermal boundary layer thickness.
19. The Savannah River Sites Groundwater Monitoring Program. Fourth quarter, 1989
Energy Technology Data Exchange (ETDEWEB)
1989-12-31
The Environmental Monitoring Section of the Environmental and Health Protection (EHP) Department administers the Savannah River Site`s Groundwater Monitoring Program. During fourth quarter 1989 (October--December), EHP conducted routine sampling of monitoring wells and drinking water locations. EHP collected the drinking water samples from Savannah River Site (SRS) drinking water systems supplied by wells. EHP established two sets of flagging criteria in 1986 to assist in the management of sample results. The flagging criteria aid personnel in sample scheduling, interpretation of data, and trend identification. An explanation of flagging criteria for the fourth quarter is presented in the Flagging Criteria section of this document. All analytical results from fourth quarter 1989 are listed in this report, which is distributed to all waste-site custodians.
20. Fourth Ventricular Schwannoma: Identical Clinicopathologic Features as Schwann Cell-Derived Schwannoma with Unique Etiopathologic Origins
Directory of Open Access Journals (Sweden)
Tiffany R. Hodges
2011-01-01
Full Text Available Background. To our knowledge, this is the sixth reported case in the literature of fourth ventricular schwannoma. The etiology and natural history of intraventricular schwannomas is not well understood. A thorough review of potential etiopathogenic mechanisms is provided in this case report. Case Description. A 69-year-old man presented with an incidentally found fourth ventricular tumor during an evaluation for generalized weakness, gait instability, and memory disturbance. Magnetic resonance imaging (MRI revealed a heterogeneously enhancing lesion in the fourth ventricle. A suboccipital craniotomy was performed to resect the lesion. Histopathological examination confirmed the diagnosis of schwannoma (WHO grade I. Conclusions. Schwannomas should be considered in the differential diagnosis of intraventricular tumors. Although the embryologic origins may be different from nerve sheath-derived schwannomas, the histologic, clinical, and natural history appear identical and thus should be managed similarly.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5655562281608582, "perplexity": 5161.221154645316}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463612502.45/warc/CC-MAIN-20170529165246-20170529185246-00154.warc.gz"}
|
http://openstudy.com/updates/55cbe9cce4b0016bb016611a
|
## A community for students. Sign up today
Here's the question you clicked on:
55 members online
• 0 replying
• 0 viewing
## zmudz one year ago Find a closed form for $$S_n = 1 \cdot 1! + 2 \cdot 2! + \ldots + n \cdot n!.$$ for integer $$n \geq 1.$$ Your response should have a factorial. Delete Cancel Submit
• This Question is Closed
1. ganeshie8
• one year ago
Best Response
You've already chosen the best response.
3
\begin{align}S_n &= \color{red}{1} \cdot 1! + \color{red}{2} \cdot 2! + \ldots + \color{red}{n} \cdot n!\\~\\ &= \color{red}{(2-1)} \cdot 1! + \color{red}{(3-1)} \cdot 2! + \ldots + \color{red}{(n+1-1)} \cdot n!\\~\\ &=2!-1! + 3!-2! +\cdots + (n+1)!-n!\\~\\ &=-1! + (n+1)! \end{align}
2. anonymous
• one year ago
Best Response
You've already chosen the best response.
0
You might recall asking this about a month ago... http://openstudy.com/users/zmudz#/updates/55ad142ee4b0d48ca8ed35d0
3. Not the answer you are looking for?
Search for more explanations.
• Attachments:
#### Ask your own question
Sign Up
Find more explanations on OpenStudy
Privacy Policy
## Your question is ready. Sign up for free to start getting answers.
##### spraguer (Moderator) 5→ View Detailed Profile
23
• Teamwork 19 Teammate
• Problem Solving 19 Hero
• You have blocked this person.
• ✔ You're a fan Checking fan status...
Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 1, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9997920393943787, "perplexity": 13358.928114115957}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988720238.63/warc/CC-MAIN-20161020183840-00549-ip-10-171-6-4.ec2.internal.warc.gz"}
|
https://2021.help.altair.com/2021.2/hwsolvers/acusolve/topics/acusolve/udfgetsddatadim_acusolve_udf.htm
|
Return the dimension of a data variable.
## Type
AcuSolve User-Defined Subdomain Routine
## Parameters
udfHd
The opaque handle (pointer) which was passed to the user function.
dataName (integer)
Symbolic name corresponding to the data index.
UDF_SD_VELOCITY
Velocity.
UDF_SD_ACCELERATION
Acceleration.
UDF_SD_PRESSURE
Pressure.
UDF_SD_TEMPERATURE
Temperature.
UDF_SD_SPECIES
Species.
UDF_SD_EDDY_VISCOSITY
Turbulence eddy viscosity.
UDF_SD_KINETIC_ENERGY
Turbulence kinetic energy.
UDF_SD_EDDY_FREQUENCY
Turbulence eddy frequency.
UDF_SD_MESH_DISPLACEMENT
Mesh displacement.
UDF_SD_MESH_VELOCITY
Mesh velocity.
UDF_SD_TURBULENCE_Y
Distance to nearest turbulence wall.
UDF_SD_TURBULENCE_YPLUS
Turbulence y+ based on distance to nearest turbulence wall and shear at that wall.
## Return Value
Dimension of the data variable.
## Description
This routine returns the dimension of a variable associated with the given symbolic data name. For example,
Integer dataDim ;
...
dataDim = udfGetSdDataDim( udfHd, UDF_SD_VELOCITY ) ;
## Errors
• This routine expects a valid udfHd.
• This routine may only be called within an external code.
• dataName must be one of the values given above.
• The subdomain must contain the equation associated with the data variable.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3634110391139984, "perplexity": 13410.050085482}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296943750.71/warc/CC-MAIN-20230322051607-20230322081607-00294.warc.gz"}
|
http://swmath.org/software/6486
|
# svylogitgof
Goodness-of-fit tests for logistic regression models when data are collected using a complex sampling design Logistic regression models are frequently used in epidemiological studies for estimating associations that demographic, behavioral, and risk factor variables have on a dichotomous outcome, such as disease being present versus absent. After the coefficients in a logistic regression model have been estimated, goodness-of-fit of the resulting model should be examined, particularly if the purpose of the model is to estimate probabilities of event occurrences. While various goodness-of-fit tests have been proposed, the properties of these tests have been studied under the assumption that observations selected were independent and identically distributed. Increasingly, epidemiologists are using large-scale sample survey data when fitting logistic regression models, such as the National Health Interview Survey or the National Health and Nutrition Examination Survey. Unfortunately, for such situations no goodness-of-fit testing procedures have been developed or implemented in available software. To address this problem, goodness-of-fit tests for logistic regression models when data are collected using complex sampling designs are proposed. Properties of the proposed tests were examined using extensive simulation studies and results were compared to traditional goodness-of-fit tests. A Stata ado function svylogitgof for estimating the $F$-adjusted mean residual test after svylogit fit is available at the author’s website http://www.people.vcu.edu/ kjarcher/Research/Data.htm.
## Keywords for this software
Anything in here will be replaced on browsers that support the canvas element
## References in zbMATH (referenced in 3 articles , 1 standard article )
Showing results 1 to 3 of 3.
Sorted by year (citations)
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6256386041641235, "perplexity": 1294.591817144812}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400206133.46/warc/CC-MAIN-20200922125920-20200922155920-00040.warc.gz"}
|
https://www.scala-algorithms.com/CountInversions/
|
# Scala algorithm: Counting inversions of a sequence (array) using a Merge Sort
Published , last updated
## Algorithm goal
The number of inversions in a sequence is the number of pairs of elements that are out of order, ie$$|{(i, j) : i < j, A_i > A_j}|$$ (count of distinct $$(i,j)$$ such that $$i < j$$ and value of the $$i$$th element is greater than that of the $$j$$th one).
### Why is this algorithm useful?
One situation where this is very useful is to see how different preferences between two people are, when ranked. This sort of algorithm, or versions of it, would be useful in eCommerce, or even movie recommendations. While it is not as advanced as machine learning, it has the big advantage of high performance at $$O(n\log{n})$$, meaning it can be used as a tactical solution in a highly paced environment, where the cost of implementation of a machine learning solution could only become viable only once the commercial viability of the product has established.
The brute-force solution is $$O(n^2)$$ complexity, which eats up computation time very quickly
### Examples
$$[2,1]$$ has 1 inversion, because swapping $$1$$ with $$2$$ leads to array $$[1,2]$$ which is sorted.
2 1 1 2
Likewise, $$[3, 1, 8, 5, 6, 4, 7]$$ has 7 inversions: $$(1, 3)$$, $$(8, 5)$$, $$(8, 6)$$, $$(8, 4)$$, $$(8, 7)$$, $$(5, 4)$$.$$(6, 4)$$.
3 1 8 5 6 4 7 1 3 4 5 6 7 8
$$[1,2,4,3]$$ has 1 inversion $$(4, 3)$$:
1 2 4 3 1 2 3 4
What is the most curious about the above diagrams is that the number of crossings between arrows corresponds exactly to the number of inversions! What do you think this would mean?
It will be very helpful to first understand the problem of MergeSort first and compare the two. Although the merging function is different - because this time, we really have to count how many exchanges there would be, as we are sorting it. This is what leads us to a solution that is much more efficient than a typical brute-force solution (which, in fact, we include in the algorithm solution).
The algorithm code here, while a MergeSort, is implemented using the bottom-up approach of the algorithm MergeSortStackSafe, which is more stack-safe.
## Algorithm in Scala
79 lines of Scala (compatible versions 2.13 & 3.0).
## Explanation
The problem is actually closely related to the marge sort. In the merge sort, we approach the problem by the divide-and-conquer method, where we process one half of the array, the other half, and then we merge them.
If we divide our input sequence into 2 parts, through example, we will notice that the total number of inversions is equal to the number of inversions on the left-hand side (LHS), plus inversions on RHS, plus the number of inversions across the two sides. Example: (this is © from www.scala-algorithms.com)
$$[3,2,1,4]$$ has inversion $$(2,3)$$ on the LHS to make it sorted, no inversion on RHS, and 2 inversions across the half: $$(3,1)$$ and $$(2,1)$$, thus 3 inversions in total.
3 2 2 3 Counted 1 inversion
1 4 1 4 Counted 0 inversions
As we are sorting the data on a side, we record how many times it was out of order. Even in a sub-problem of 2, we count an inversion of a change between the left side and the right side.
## Scala concepts & Hints
1. ### Def Inside Def
A great aspect of Scala is being able to declare functions inside functions, making it possible to reduce repetition.
def exampleDef(input: String): String = {
def surroundInputWith(char: Char): String = s"$char$input\$char"
surroundInputWith('-')
}
assert(exampleDef("test") == "-test-")
It is also frequently used in combination with Tail Recursion.
2. ### Drop, Take, dropRight, takeRight
Scala's drop and take methods typically remove or select n items from a collection.
assert(List(1, 2, 3).drop(2) == List(3))
assert(List(1, 2, 3).take(2) == List(1, 2))
assert(List(1, 2, 3).dropRight(2) == List(1))
assert(List(1, 2, 3).takeRight(2) == List(2, 3))
assert((1 to 5).take(2) == (1 to 2))
3. ### For-comprehension
The for-comprehension is highly important syntatic enhancement in functional programming languages.
val Multiplier = 10
val result: List[Int] = for {
num <- List(1, 2, 3)
anotherNum <-
List(num * Multiplier - 1, num * Multiplier, num * Multiplier + 1)
} yield anotherNum + 1
assert(result == List(10, 11, 12, 20, 21, 22, 30, 31, 32))
4. ### Lazy List
The 'LazyList' type (previously known as 'Stream' in Scala) is used to describe a potentially infinite list that evaluates only when necessary ('lazily').
5. ### Pattern Matching
Pattern matching in Scala lets you quickly identify what you are looking for in a data, and also extract it.
assert("Hello World".collect {
case character if Character.isUpperCase(character) => character.toLower
} == "hw")
6. ### Range
The (1 to n) syntax produces a "Range" which is a representation of a sequence of numbers.
assert((1 to 5).toString == "Range 1 to 5")
assert((1 to 5).reverse.toString() == "Range 5 to 1 by -1")
assert((1 to 5).toList == List(1, 2, 3, 4, 5))
7. ### Stack Safety
Stack safety is present where a function cannot crash due to overflowing the limit of number of recursive calls.
This function will work for n = 5, but will not work for n = 2000 (crash with java.lang.StackOverflowError) - however there is a way to fix it :-)
In Scala Algorithms, we try to write the algorithms in a stack-safe way, where possible, so that when you use the algorithms, they will not crash on large inputs. However, stack-safe implementations are often more complex, and in some cases, overly complex, for the task at hand.
def sum(from: Int, until: Int): Int =
if (from == until) until else from + sum(from + 1, until)
def thisWillSucceed: Int = sum(1, 5)
def thisWillFail: Int = sum(1, 300)
8. ### Tail Recursion
In Scala, tail recursion enables you to rewrite a mutable structure such as a while-loop, into an immutable algorithm.
def fibonacci(n: Int): Int = {
@scala.annotation.tailrec
def go(i: Int, previous: Int, beforePrevious: Int): Int =
if (i >= n) previous else go(i + 1, previous + beforePrevious, previous)
go(i = 1, previous = 1, beforePrevious = 0)
}
assert(fibonacci(8) == 21)
# Scala Algorithms: The most comprehensive library of algorithms in standard pure-functional Scala
## Think in Scala & master the highest paid programming language in the US
Scala is used at many places, such as AirBnB, Apple, Bank of America, BBC, Barclays, Capital One, Citibank, Coursera, eBay, JP Morgan, LinkedIn, Morgan Stanley, Netflix, Singapore Exchange, Twitter.
### Study our 92 Scala Algorithms: 6 fully free, 74 published & 18 upcoming
Fully unit-tested, with explanations and relevant concepts; new algorithms published about once a week.
### Explore the 21 most useful Scala concepts
To save you going through various tutorials, we cherry-picked the most useful Scala concepts in a consistent form.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.35966578125953674, "perplexity": 2593.1996168150094}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057508.83/warc/CC-MAIN-20210924080328-20210924110328-00346.warc.gz"}
|
https://en.m.wikipedia.org/wiki/Hungarian_algorithm
|
# Hungarian algorithm
The Hungarian method is a combinatorial optimization algorithm that solves the assignment problem in polynomial time and which anticipated later primal-dual methods. It was developed and published in 1955 by Harold Kuhn, who gave the name "Hungarian method" because the algorithm was largely based on the earlier works of two Hungarian mathematicians: Dénes Kőnig and Jenő Egerváry.[1][2]
James Munkres reviewed the algorithm in 1957 and observed that it is (strongly) polynomial.[3] Since then the algorithm has been known also as the Kuhn–Munkres algorithm or Munkres assignment algorithm. The time complexity of the original algorithm was ${\displaystyle O(n^{4})}$, however Edmonds and Karp, and independently Tomizawa noticed that it can be modified to achieve an ${\displaystyle O(n^{3})}$ running time. Ford and Fulkerson extended the method to general transportation problems. In 2006, it was discovered that Carl Gustav Jacobi had solved the assignment problem in the 19th century, and the solution had been published posthumously in 1890 in Latin.[4]
## Simple explanation of the assignment problemEdit
In this simple example there are three workers: Armond, Francine, and Herbert. One of them has to clean the bathroom, another sweep the floors, and the third wash the windows, but they each demand different pay for the various tasks. The problem is to find the lowest-cost way to assign the jobs. The problem can be represented in a matrix of the costs of the workers doing the jobs. For example:
Clean bathroom Sweep floors Wash windows
Armond $2$3 $3 Francine$3 $2$3
Herbert $3$3 $2 The Hungarian method, when applied to the above table, would give the minimum cost: this is$6, achieved by having Armond clean the bathroom, Francine sweep the floors, and Herbert wash the windows.
## SettingEdit
We are given a nonnegative n×n matrix, where the element in the i-th row and j-th column represents the cost of assigning the j-th job to the i-th worker. We have to find an assignment of the jobs to the workers that has minimum cost. If the goal is to find the assignment that yields the maximum cost, the problem can be altered to fit the setting by replacing each cost with the maximum cost subtracted by the cost.
The algorithm is easier to describe if we formulate the problem using a bipartite graph. We have a complete bipartite graph ${\displaystyle G=(S,T;E)}$ with ${\displaystyle n}$ worker vertices (${\displaystyle S}$ ) and ${\displaystyle n}$ job vertices (${\displaystyle T}$ ), and each edge has a nonnegative cost ${\displaystyle c(i,j)}$ . We want to find a perfect matching with minimum cost.
Let us call a function ${\displaystyle y:(S\cup T)\to \mathbb {R} }$ a potential if ${\displaystyle y(i)+y(j)\leq c(i,j)}$ for each ${\displaystyle i\in S,j\in T}$ . The value of potential ${\displaystyle y}$ is ${\displaystyle \sum _{v\in S\cup T}y(v)}$ . It can be seen that the cost of each perfect matching is at least the value of each potential. The Hungarian method finds a perfect matching and a potential with equal cost/value which proves the optimality of both. In fact it finds a perfect matching of tight edges: an edge ${\displaystyle ij}$ is called tight for a potential ${\displaystyle y}$ if ${\displaystyle y(i)+y(j)=c(i,j)}$ . Let us denote the subgraph of tight edges by ${\displaystyle G_{y}}$ . The cost of a perfect matching in ${\displaystyle G_{y}}$ (if there is one) equals the value of ${\displaystyle y}$ .
## The algorithm in terms of bipartite graphsEdit
During the algorithm we maintain a potential y and an orientation of ${\displaystyle G_{y}}$ (denoted by ${\displaystyle {\overrightarrow {G_{y}}}}$ ) which has the property that the edges oriented from T to S form a matching M. Initially, y is 0 everywhere, and all edges are oriented from S to T (so M is empty). In each step, either we modify y so that its value increases, or modify the orientation to obtain a matching with more edges. We maintain the invariant that all the edges of M are tight. We are done if M is a perfect matching.
In a general step, let ${\displaystyle R_{S}\subseteq S}$ and ${\displaystyle R_{T}\subseteq T}$ be the vertices not covered by M (so ${\displaystyle R_{S}}$ consists of the vertices in S with no incoming edge and ${\displaystyle R_{T}}$ consists of the vertices in T with no outgoing edge). Let ${\displaystyle Z}$ be the set of vertices reachable in ${\displaystyle {\overrightarrow {G_{y}}}}$ from ${\displaystyle R_{S}}$ by a directed path only following edges that are tight. This can be computed by breadth-first search.
If ${\displaystyle R_{T}\cap Z}$ is nonempty, then reverse the orientation of a directed path in ${\displaystyle {\overrightarrow {G_{y}}}}$ from ${\displaystyle R_{S}}$ to ${\displaystyle R_{T}}$ . Thus the size of the corresponding matching increases by 1.
If ${\displaystyle R_{T}\cap Z}$ is empty, then let ${\displaystyle \Delta :=\min\{c(i,j)-y(i)-y(j):i\in Z\cap S,j\in T\setminus Z\}}$ . ${\displaystyle \Delta }$ is positive because there are no tight edges between ${\displaystyle Z\cap S}$ and ${\displaystyle T\setminus Z}$ . Increase y by ${\displaystyle \Delta }$ on the vertices of ${\displaystyle Z\cap S}$ and decrease y by ${\displaystyle \Delta }$ on the vertices of ${\displaystyle Z\cap T}$ . The resulting y is still a potential. The graph ${\displaystyle G_{y}}$ changes, but it still contains M. We orient the new edges from S to T. By the definition of ${\displaystyle \Delta }$ the set Z of vertices reachable from ${\displaystyle R_{S}}$ increases (note that the number of tight edges does not necessarily increase).
We repeat these steps until M is a perfect matching, in which case it gives a minimum cost assignment. The running time of this version of the method is ${\displaystyle O(n^{4})}$ : M is augmented n times, and in a phase where M is unchanged, there are at most n potential changes (since Z increases every time). The time needed for a potential change is ${\displaystyle O(n^{2})}$ .
## Matrix interpretationEdit
Given ${\displaystyle n}$ workers and tasks, and an n×n matrix containing the cost of assigning each worker to a task, find the cost minimizing assignment.
First the problem is written in the form of a matrix as given below
${\displaystyle {\begin{bmatrix}a1&a2&a3&a4\\b1&b2&b3&b4\\c1&c2&c3&c4\\d1&d2&d3&d4\end{bmatrix}}}$
where a, b, c and d are the workers who have to perform tasks 1, 2, 3 and 4. a1, a2, a3, a4 denote the penalties incurred when worker "a" does task 1, 2, 3, 4 respectively. The same holds true for the other symbols as well. The matrix is square, so each worker can perform only one task.
Step 1
Then we perform row operations on the matrix. To do this, the lowest of all ai (i belonging to 1-4) is taken and is subtracted from each element in that row. This will lead to at least one zero in that row (We get multiple zeros when there are two equal elements which also happen to be the lowest in that row). This procedure is repeated for all rows. We now have a matrix with at least one zero per row. Now we try to assign tasks to agents such that each agent is doing only one task and the penalty incurred in each case is zero. This is illustrated below.
0 a2' a3' a4' b1' b2' b3' 0 c1' 0 c3' c4' d1' d2' 0 d4'
The zeros that are indicated as 0' are the assigned tasks.
Step 2
Sometimes it may turn out that the matrix at this stage cannot be used for assigning, as is the case for the matrix below.
0 a2' a3' a4' b1' b2' b3' 0 0 c2' c3' c4' d1' 0 d3' d4'
In the above case, no assignment can be made. Note that task 1 is done efficiently by both agent a and c. Both can't be assigned the same task. Also note that no one does task 3 efficiently. To overcome this, we repeat the above procedure for all columns (i.e. the minimum element in each column is subtracted from all the elements in that column) and then check if an assignment is possible.
In most situations this will give the result, but if it is still not possible then we need to keep going.
Step 3
All zeros in the matrix must be covered by marking as few rows and/or columns as possible. The following procedure is one way to accomplish this:
First, assign as many tasks as possible.
• Row 1 has one zero, so it is assigned. The 0 in row 3 is crossed out because it is in the same column.
• Row 2 has one zero, so it is assigned.
• Row 3's only zero has been crossed out, so nothing is assigned.
• Row 4 has two uncrossed zeros. Either one can be assigned (both are optimum), and the other zero would be crossed out.
Alternatively, the 0 in row 3 may be assigned, causing the 0 in row 1 to be crossed instead.
0' a2' a3' a4' b1' b2' b3' 0' 0 c2' c3' c4' d1' 0' 0 d4'
Now to the drawing part.
• Mark all rows having no assignments (row 3).
• Mark all (unmarked) columns having zeros in newly marked row(s) (column 1).
• Mark all rows having assignments in newly marked columns (row 1).
• Repeat for all non-assigned rows.
× 0' a2' a3' a4' × b1' b2' b3' 0' 0 c2' c3' c4' × d1' 0' 0 d4'
Now draw lines through all marked columns and unmarked rows.
× 0' a2' a3' a4' × b1' b2' b3' 0' 0 c2' c3' c4' × d1' 0' 0 d4'
The aforementioned detailed description is just one way to draw the minimum number of lines to cover all the 0s. Other methods work as well.
Step 4
From the elements that are left, find the lowest value. Subtract this from every unmarked element and add it to every element covered by two lines.
Repeat steps 3–4 until an assignment is possible; this is when the minimum number of lines used to cover all the 0s is equal to max(number of people, number of assignments), assuming dummy variables (usually the max cost) are used to fill in when the number of people is greater than the number of assignments.
Basically you find the second minimum cost among the remaining choices. The procedure is repeated until you are able to distinguish among the workers in terms of least cost.
## BibliographyEdit
• R.E. Burkard, M. Dell'Amico, S. Martello: Assignment Problems (Revised reprint). SIAM, Philadelphia (PA.) 2012. ISBN 978-1-61197-222-1
• M. Fischetti, "Lezioni di Ricerca Operativa", Edizioni Libreria Progetto Padova, Italia, 1995.
• R. Ahuja, T. Magnanti, J. Orlin, "Network Flows", Prentice Hall, 1993.
• S. Martello, "Jeno Egerváry: from the origins of the Hungarian algorithm to satellite communication". Central European Journal of Operations Research 18, 47–58, 2010
## ReferencesEdit
1. ^ Harold W. Kuhn, "The Hungarian Method for the assignment problem", Naval Research Logistics Quarterly, 2: 83–97, 1955. Kuhn's original publication.
2. ^ Harold W. Kuhn, "Variants of the Hungarian method for assignment problems", Naval Research Logistics Quarterly, 3: 253–258, 1956.
3. ^ J. Munkres, "Algorithms for the Assignment and Transportation Problems", Journal of the Society for Industrial and Applied Mathematics, 5(1):32–38, 1957 March.
4. ^ http://www.lix.polytechnique.fr/~ollivier/JACOBI/jacobiEngl.htm
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 56, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7412024140357971, "perplexity": 1004.90565506182}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463608617.80/warc/CC-MAIN-20170525233846-20170526013846-00073.warc.gz"}
|
http://www.cs.ncsu.edu/research/faculty_projs/fundedprojects.php
|
Total: $45,271,018 ## Other Years # Current Computer Science Research Projects (by faculty) The funded projects listed below are active projects and the funded running total for the active projects is on the left navigational menu. Collaborative Research: Modeling Social Interaction and Performance in STEM Learning Tiffany Barnes$200,003 by National Science Foundation
09/ 1/2014 - 08/31/2017
Despite long-standing awareness that social interaction is an integral part of knowledge construction, efforts to study complex collaborative learning have traditionally been relegated to qualitative and small-scale methodologies. Relatively new data traces left by online learning environments, including massive open online courses (MOOCs), offer the first real hope for scaling up such analyses. The purpose of the proposed research is to develop comprehensive models for collaborative learning which in turn will enable instructional design and the authentic assessment of the individual within the group context. This task is undertaken by an interdisciplinary team of researchers with specializations in natural language processing, discourse analysis, social network analysis, educational data mining and psychometrics.
REU Site: Interactive and Intelligent Media
Tiffany Barnes
$359,999 by National Science Foundation 04/ 1/2013 - 03/31/2016 The REU Site at NC State University will immerse a diverse group of undergraduates in a vibrant research community of faculty and graduate students working on cutting-edge games, intelligent tutors, and mobile applications. We will recruit students from underrepresented groups and colleges and universities with limited research opportunities through the STARS Alliance, an NSF-funded national consortium of institutions dedicated to broadening participation in computing. Using the Affinity Research Groups and STARS Training for REUs models, we will engage faculty and graduate student mentors with undergraduates to create a supportive culture of collaboration while promoting individual contributions to research through just-in-time training for both mentors and students throughout the summer. Type I: Collaborative Research: FRABJOUS CS - Framing a Rigorous Approach to Beauty and Joy for Outreach to Underrepresented Students in Computing at Scale Tiffany Barnes$352,831 by National Science Foundation
02/ 1/2013 - 02/28/2016
In this FRABJOUS CS project, we will prepare 60 high school teachers to teach the Beauty and Joy of Computing (BJC) Computer Science Principles curriculum. The BJC course is a rigorous introductory computing course that highlights the meaning and applications of computing, while introducing low-threshold programming languages Snap-Scratch, GameMaker and AppInventor. BJC is informed and inspired by the Exploring Computer Science curriculum, that was explicitly designed to channel the interests of urban HS students with ?culturally relevant and meaningful curriculum? [Goode 2011][Margolis 2008]. The BJC course uses collaborative classroom methods including pair learning, and student-selected projects are geared toward leveraging students? knowledge of social media, games, devices, and the internet. At UNC Charlotte in 2010 and 2011, PI Barnes engaged college students in supporting the BJC course, and in after-school outreach and summer camps that excite middle and high school students about this curriculum at different levels. The project engages three university faculty members and 6 college students to help the high school teachers build a Computer Science Teachers Association chapter and provide ongoing professional development and support for the BJC course. The project also engages high school teachers and an education researcher to help refine and enriches the BJC curriculum to be easier to adopt and teach in high schools.
Type I: Collaborative Research: FRABJOUS CS - Framing a Rigorous Approach to Beauty and Joy for Outreach to Underrepresented Students in Computing at Scale (Supplement)
Tiffany Barnes
$86,000 by NSF 02/ 1/2013 - 02/28/2016 In this FRABJOUS CS project, we will prepare 60 high school teachers to teach the Beauty and Joy of Computing (BJC) Computer Science Principles curriculum. The BJC course is a rigorous introductory computing course that highlights the meaning and applications of computing, while introducing low-threshold programming languages Snap-Scratch, GameMaker and AppInventor. BJC is informed and inspired by the Exploring Computer Science curriculum, that was explicitly designed to channel the interests of urban HS students with ?culturally relevant and meaningful curriculum? [Goode 2011][Margolis 2008]. The BJC course uses collaborative classroom methods including pair learning, and student-selected projects are geared toward leveraging students? knowledge of social media, games, devices, and the internet. At UNC Charlotte in 2010 and 2011, PI Barnes engaged college students in supporting the BJC course, and in after-school outreach and summer camps that excite middle and high school students about this curriculum at different levels. The project engages three university faculty members and 6 college students to help the high school teachers build a Computer Science Teachers Association chapter and provide ongoing professional development and support for the BJC course. The project also engages high school teachers and an education researcher to help refine and enriches the BJC curriculum to be easier to adopt and teach in high schools. CAREER: Educational Data Mining for Student Support in Interactive Learning Environment Tiffany Barnes$237,770 by UNC Charlotte/NSF
11/ 1/2013 - 06/30/2015
Creating intelligent learning technologies from data has unique potential to transform the American educational system, by building a low cost way to adapt learning environments to individual students, while informing research on human learning. This project will create the technology for a new generation of data-driven intelligent tutors, enabling the rapid creation of individualized instruction to support learning in science, technology, engineering, and mathematics (STEM) fields. This has the potential to make individualized learning support accessible for a broad audience, from children to adults, including students that are traditionally underrepresented in STEM fields. This project will (1) develop computational methods to derive cognitive models from data that can be used to support individual learners through guidance, feedback, and help; (2) develop approaches to providing student support that leverage data to provide hints and guidance based on information such as frequency of student responses, probability of future errors, and solution efficiency; (3) develop interactive visualization tools for teachers to learn from student data in real time, to allow teachers and instructional designers to tailor instruction to address actual, rather than perceived, student problem areas; and (4) conduct formal empirical evaluations of the pedagogical effectiveness of our student support. Our software will construct adaptive support for teaching and learning in logic, discrete mathematics, and other STEM domains using a data-driven approach. From the extensive but tractable student performance data in computer-aided learning environments, we will automatically construct student cognitive models. Our cognitive models will build on our prior work using Markov Decision Processes and dimensionality reduction methods that leverage past data to assess student performance, direct a student’s learning path, and provide contextualized hints. We will use machine learning techniques to expand our problem-specific models into more general cognitive models to bootstrap the construction of new tutors and learn about student learning. For teachers and learning researchers, we will build a web-based visualization and analysis tool to graphically and interactively model student solutions annotated with performance data that reflects frequency, tendency to commit future errors, and closeness to a final solution. Through our new tutors and tools we will conduct experiments to understand student learning in a variety of contexts and domains, including logic, algebra, and chemistry. We will engage a team of diverse students and colleagues to bring interdisciplinary expertise to our research and share our findings broadly. This award is funded under the American Recovery and Reinvestment Act of 2009 (Public Law 111-5).
BPC-AE: Scaling the STARS Alliance: A National Community for Broadening Participation Through Regional Partnerships
Tiffany Barnes
$150,000 by UNC-UNC Charlotte ( NSF) 01/ 1/2013 - 12/31/2014 The Beauty and Joy of Computing project presents a unique opportunity to scale the STARS Alliance further while also enhancing national efforts to engage more high school teachers and students in teaching and learning computing and build stronger university/college/K12 partnerships. Through this supplement, we will extend the Alliance with at least three new STARS Computing Corps, providing leadership training to a group of 8-10 students in each Corps, all focused on supporting the BJC effort. New Corps will provide teaching assistance to high school teachers implementing the BJC course through classroom visits and monthly Computer Science Teacher Association chapter meetings. These new STARS Computing Corps will also teach BJC material either through in middle school Citizen Schools after-school programs, and K-12 summer camps. This will provide a vibrant community of support for high school teachers and students engaging the new BJC course. LAS DO3 Task Order 2.8 Analytic Workflow Kristy Boyer$33,853 by LAS
03/31/2014 - 12/31/2014
Internal award supplement
LAS DO3 Task Order 2.8 Analytic Workflow
Kristy Boyer
$16,529 by LAS 03/31/2014 - 12/31/2014 DO3 Task Order 2.8 Analytic Workflow Educational Data Mining for Individualized Instruction in STEM Learning Environments Min Chi ; Tiffany Barnes$639,401 by National Science Foundation
09/ 1/2014 - 08/31/2017
Human one-on-one tutoring is one of the most effective educational interventions. Tutored students often perform significantly better than students in classroom settings (Bloom, 1984; Cohen, Kulik, & Kulik, 1982). Computer learning environments that mimic aspects of human tutors have also been highly successful. Intelligent Tutoring Systems (ITSs) have been shown to be highly effective in improving students' learning at real classrooms (Anderson, Corbett, Koedinger, & Pelletier, 1995; Koedinger, Anderson, Hadley, & Mark, 1997; VanLehn et al., 2005). The development of ITSs has enabled schools and universities to reach out and educate students who otherwise would be unable to take advantage of one-on-one tutoring due to cost and time constraints (Koedinger, Anderson, Hadley, & Mark, 1997). Despite the high payoffs provided by ITSs, significant barriers remain. High development costs and the challenges of knowledge engineering have prevented their widespread deployment. A diverse team of software developers, domain experts, and educational theorists are required for development, testing and even maintenance. Generally speaking, it requires an average of 80 man-hours per hour of tutoring content. In this proposed work, our goal is to automatically design effective personalized ITSs directly from log data. We will combine co-pI Dr. Barnes data-driven approach on learning what to teach with PI Dr. Chi’s data-driven work on learning how to teach. More specifically, we will explore two important undergraduate stem domains: discrete math and probability; and employ two types of ITSs: an example-based ITS, the discrete math tutor, and a rule-based ITS, Pyrenees. The former can automatically generate hints directly from students’ prior solutions while the latter has hard-coded domain rules and teaches students a domain-general problem-solving strategy within the context of probability. For learning how to teach, we will apply reinforcement learning to induce adaptive pedagogical strategies directly from students’ log files and will focus on three levels of pedagogical decisions: 1) whether or not to give students hints and what level of hint (step-level); 2) whether to show students worked example or ask students to engage problem solving (problem-level); and 3) whether or not to teach students a meta-cognitive problem-solving strategy (metacognitive level).
I/UCRC Planning Grant: Site Addition to CHMPR I/UCRC
$13,779 by National Science Foundation 09/ 1/2014 - 08/31/2015 The objective of this letter of intent is to indicate that North Carolina State University (NCSU) will join, as a site, the Center of Hybrid Multicore Productivity Research (CHMPR) in Year 2 of its planned phase II I/UCRC renewal. CHMPR consists of sites at the University of Maryland, Baltimore County, and the University of California, San Diego. CHMPR is proposing a phase II extension of its successful conduct of research based on the continued challenges arising from the evolution of multicore technologies. The focus of NCSU within the center will be the science of technologies for end-to-end enablement of data. We are forming an NCSU-based site to join an existing I/UCRC Center for Hybrid Multicore Productivity (CHMPR). This planning grant will support meetings with potential industrial partners and CHMPR co-investigators from the University of Maryland Baltimore County. Planning activities will be facilitated by the NCSU CSC Director of Development & External Relations, Corporate & Alumni Relations Mr. Ken Tate, as well as by the Director of Institute for NEXT generation IT Systems Mr. John Streck. Meetings organized using the planning grant funds will operationalize, transform and integrate themes developed during the preparation of this planning grant, experiences from ongoing industry partnerships at the three centers, and concepts outlined in NSF’s ‘Purple Book’ entitled Managing the Industry/University Cooperative Research Center. The NCSU I/UCRC center site will represent an exciting and novel integration of approaches to Big Data in multiple disciplines, as we target technological innovations on extracting value from Big Data for a diverse pool of stakeholders. LAS DO3 Task Order 2.7 Data Readiness - Chirkova Rada Chirkova$74,660 by Laboratory for Analytic Sciences
03/31/2014 - 12/31/2014
DO 2 Task 3.6 - Doyle
Jon Doyle
$46,904 by Laboratory for Analytic Sciences 09/13/2013 - 12/31/2014 DO 2 task 3.6 activities LAS DO3 Task Order 2.9 KRM Jon Doyle$76,241 by LAS
03/31/2014 - 12/31/2014
NeTS: JUNO: Service Offering Model and Versatile Network Resource Grooming for Optical Packet and Circuit Integrated Networks
Rudra Dutta
$291,956 by National Science Foundation (NSF) 04/ 1/2014 - 03/31/2017 The explosive growth in bandwidth represented by advances in optical communication and networking technologies has underpinned the increasing reach and reliability of the Internet in the last two decades. However, the potential impact of increasingly sophisticated recent advances in optical technology, such as rapid switching and elastic wavelengths have not yet been realized. The main cause of this is that such technology, while possible to integrate into the data plane of planetary networking, is difficult to accommodate in the current planning, management, and control strategies. We propose in this project to work hand-in-hand with collaborating researchers from NICT, Japan, who are working to realize a novel technology of hybrid optical packet/circuit switching. Such a technology could be immensely useful to large transport network operators, but there are no existing algorithms that can easily determine how a provider can provision their resources between the circuit and packet possibilities on an ongoing dynamic basis. We envision a novel approach to this problem, where we utilize the concept of a "choice marketplace" that allows sophisticated rendezvous semantics between customer and provider, and allows them to cooperatively guide network resource provisioning to dynamically fulfill network objectives such as maximizing performance received by network traffic. Our approach also allows balancing of various objectives, such as network utilization, latency, and the increasingly important metric of energy expenditure in the network. NeTS: Small: Collaborative Research: Enabling Robust Communication in Cognitive Radio Networks with Multiple Lines of Defense Rudra Dutta ; Peng Ning$249,901 by National Science Foundation
10/ 1/2013 - 09/30/2016
Cognitive radio is an emerging advanced radio technology in wireless access, with many promising benefits including dynamic spectrum sharing, robust cross-layer adaptation, and collaborative networking. Opportunistic spectrum access (OSA) is at the core of cognitive radio technologies, which has received great attention recently, focusing on improving spectrum utilization efficiency and reliability. However, the state-of-the-art still suffers from one severe security vulnerability, which has been largely overlooked by the research community so far. That is, a malicious jammer can always disrupt the legitimate network communication by leveraging the public-available channel statistic information to effectively jam the channels and thus lead to serious spectrum underutilization. In this proposal, we propose to address the challenge of effective anti-jamming communication in cognitive radio networks (CRNs). We propose a multiple lines of defense approach, which considers and integrates defense technologies from different dimensions, including frequency hopping, power control, cooperative communication, and signal processing. The proposed defense approach enables both reactive and proactive protection, from evading jammers to competing against jammers, and to expelling jamming signals, and thus guarantees effective anti-jamming communication under a variety of network environments.
NeTS: Large: Collaborative Research: Network Innovation Through Choice
Rudra Dutta ; George Rouskas
$643,917 by National Science Foundation 09/15/2011 - 09/30/2015 This project builds on the SILO project that started in 2006 to design a new architecture for the Internet. In this new project, we will collaborate with teams of researchers from the University of Kentucky, the University of Massachusetts, and RENCI, to design critical parts of a new architecture for the Internet that will support the flexible use of widely applicable information transport and transformation modules to create good solutions for specific communication applications. The key idea is to allow a network to offer information transformation services at the edge or in the core transparently to the application, and creating a framework in which application can issue a request not only for communication but for specific reusable services. We also propose research tasks that will enable network virtualization and isolation seamlessly at many levels, currently a difficult but highly relevant problem in practical networking. CAREER: Secure OS Views for Modern Computing Platforms William Enck$400,000 by National Science Foundation
02/ 1/2013 - 01/31/2018
Controlling the access and use of information is a fundamental challenge of computer security. Emerging computing platforms such as Android and Windows 8 further complicate access control by relying on sharing and collaboration between applications. When more than two applications participate in a workflow, existing permission systems break down due to their boolean nature. In this proposal, we seek to provide applications with residual control of their data and its copies. To do this, we propose secure OS views, which combines a new abstraction for accessing data with whole-system information tracking. We apply secure OS views to modern operating systems (e.g., Android and Windows 8), which use database-like abstractions for sharing and accessing information. Similar to a database view, secure OS views uses runtime context to dynamically define the protection domain, allowing the return of the value, a fake value, or nonexistence of the record.
TWC: Small: Collaborative: Characterizing the Security Limitations of Accessing the Mobile Web
William Enck
$167,000 by NSF 10/ 1/2012 - 09/30/2015 Mobile browsers are beginning to serve as critical enablers of modern computing. With a combination of rich features that rival their desktop counterparts and strong security mechanisms such as TLS/SSL, these applications are becoming the basis of many other mobile apps. Unfortunately, the security guarantees provided by mobile browsers and the risks associated with today's mobile web have not been evaluated in great detail. In the proposed work, we will investigate the security of mobile browsers and the mobile specific websites they access. Characterizing and responding to the threats in this space is critical, especially given that cellular devices are used by more than five billion people around the world Refining Security Policy for Smartphone Applications William Enck$49,726 by US Army-Army Research Office
08/15/2014 - 05/14/2015
Title: Refining Security Policy for Smartphone Applications Abstract: Smartphones running Android and iOS have penetrated all areas of the public and private sectors. These platforms mark an important milestone in operating system security: they enforce security policies on applications, not users. Both platforms use least privilege security goals for user-applications downloaded from application markets. However, whereas Android largely enforces security goals with OS policy, iOS depends highly on a review process. Unfortunately, review alone can be circumvented, as demonstrated by multiple recent works. This proposal seeks to deepen our understanding of least privilege policy for user-applications by designing formal frameworks and algorithms for modeling, extracting, and enforcing policy. Successful execution of the proposed tasks enhance end-user security and privacy in Android and iOS devices.
Collaborative Research: Research in Student Peer Review: A Cooperative Web-Services Approach
Edward Gehringer
$1,034,166 by NSF 09/ 1/2014 - 08/31/2017 Peer review between students has a 40-year history in academia. During the last half of that period, web-based peer-review systems have been used in tens of thousands of classes. Many online systems have been developed, in diverse settings and with diverse purposes. The systems, however, have common concerns: assigning capable reviewers to each student submission, insuring review quality, and delivering reliable scores, in cases where the systems are used for summative review of student work. Many strategies have been proposed to meet those concerns, and tested in relatively small numbers of courses. The next step is to scale up the studies to learn how well they perform in diverse settings, and with large numbers of students. This project brings together researchers from several peer-review systems, including some of the largest, to build web services that can be incorporated into existing systems to test these strategies and visualize the results. CAREER: Enable Robust Virtualized Hosting Infrastructures via Coordinated Learning, Recovery, and Diagnosis Xiaohui (Helen) Gu$450,000 by National Science Foundation
01/ 1/2012 - 12/31/2016
Large-scale virtualized hosting infrastructures have become the fundamental platform for many real world systems such as cloud computing, enterprise data centers, and educational computing lab. However, due to their inherent complexity and sharing nature, hosting infrastructures are prone to various runtime problems such as performance anomalies and software/hardware failures. The overarching objective of this proposal is to systematically explore innovative runtime reliability management techniques for large-scale virtualized hosting infrastructures. Our research focuses on handling performance anomalies in distributed systems that are often very difficult to reproduce offline. Our approach combines the power of online learning, knowledge-driven first-response recovery, and in-situ diagnosis to handle unexpected system anomalies more efficiently and effectively. We aim at transforming the runtime system anomaly management from a trial-and-error guessing game into an efficient knowledge-driven self-healing process.
Modeling Context and Sentiment to Visualize Narrative Threads in Large Document Collections
Christopher Healey
$91,749 by SAS Institute, Inc. 08/16/2014 - 08/15/2015 This project will investigate the use of text analytics and information visualization for analyzing, visualizing, and exploring large document collections. Text will be analyzed using a combination of SAS Text Miner and new algorithms designed to identify, structure, and associate context, sentiment, and other properties in a document collection. Results will be segmented into narrative threads that identify conceptual storylines within the document collection. Threads will be visualized using two approaches: node-link graphs and adjacency matrices. For the graphs, graph analysis algorithms will be studied to determine methods to derive useful metrics and summaries on the document collection. For the adjacency matrices, algorithms will be developed to order of rows and columns in ways that expose patters in the document collection. For both visualization techniques, issues of multivariate data representation and level-of-detail hierarchies will also be considered. Transcriptional Nodes Coordinate Patterning and Cellular Proliferation During Carpel Margin Meristem Development Steffen Heber$71,066 by National Science Foundation
03/ 1/2014 - 02/28/2017
The coordination of spatial patterning cues and cellular proliferation underlies diverse processes from cancerous growth to reproductive development. A long-term objective of my research program is to understand how proliferative cues are coordinated with spatial information during organogenesis. In Arabidopsis thaliana this coordination of patterning and proliferation is necessary within the carpel margin meristem (CMM) to generate ovules that when fertilized will become seeds. In the previous funding period we demonstrated that the SEUSS (SEU) and AINTEGUMENTA (ANT) transcription factors regulate critical patterning events that support carpel margin meristem and ovule development. Our genetic analysis demonstrates that SEU and ANT share a partially redundant and overlapping function essential for proper seed formation. As SEU and ANT do not share sequence similarity, the molecular basis for this redundancy is not understood. We propose that the SEU and ANT activities synergistically converge at key transcriptional nodes. A node in this sense is a gene or a set of related genes that requires the combined activities of SEU and ANT for its expression. Our recently published transcriptomic analysis indicates that many of these nodes encode known transcriptional regulators. By studying these nodes we hope to better understand the transcriptional hierarchies that control CMM development and uncover the mechanistic basis of the synergistic action of SEU and ANT. Our transcriptomics study cannot determine if the nodes that we have identified are directly or indirectly regulated by SEU or ANT activity, However, even if these genes are indirectly controlled by SEU and ANT activity, their expression within the developing CMM suggests they may still play a critical functional role during CMM development. Furthermore, having now identified a set of genes that are enriched for CMM expression we are in a position to study the cis-regulatory elements that support gene expression within the CMM and medial gynoecial domain. Thus here we propose to: 1) Identify direct targets of SEU regulation within the CMM to further refine the transcriptional hierarchy required for CMM development; 2) assay the functional role of two of these nodes during CMM development; one encoded by the transcription factor PERIANTHIA and the second encoded by members of the REM family of B3 domain-containing proteins; 3) Identify cis-acting DNA regulatory elements required for CMM expression. Scientific significance: Understanding the coordination of cellular proliferation and spatial patterning during organogenesis is broadly of interest to scientists working in a diversity of fields. Completion of these specific aims will move us toward this future goal by illuminating the mechanistic basis for the overlapping functions of SEU and ANT during carpel margin and ovule development. Additionally, we expect that by elucidating the molecular mechanisms of the synergistic action of SEU and ANT upon key transcriptional nodes, we will engender a greater understanding of the molecular underpinnings of non-additivity within transcriptional networks and the complexity of developmental programs. Past NSF funding for this project (IOS-0821896) has resulted in the publication of five articles in well-respected journals (two in Plant Physiology, and one each in Developmental Biology, PLoS One, and BMC Plant Biology). Broader impacts: I ensure a broad societal impact from my program by integrating my research efforts with my teaching and training responsibilities and by widely disseminating materials and results. Furthermore, I initiated and continue to lead an outreach group that prepares and presents hands-on science demonstrations at local North Carolina schools. Our group has reached over 1500 Kindergarten through Grade 12 students over the past six years and continues to develop new demonstration modules inspired by our current work in developmental biology and genetics.
Comprehension-Driven Program Analysis (CPA) for Malware Detection in Android Phones
Xuxian Jiang
$556,488 by Iowa State University/US Air Force-Research Laboratory 02/ 3/2012 - 08/ 2/2016 Our goal is to develop new automated program analyses capable of proving that the application programs have security properties of interest to the DoD and demonstrate those analyses in the form of tools designed specifically to keep malicious code out of DoD Android-based mobile application marketplaces. CAREER: Towards Exterminating Stealthy Rootkits -- A Systematic Immunization Approach Xuxian Jiang$424,166 by the National Science Foundation
02/15/2010 - 01/31/2015
Stealthy rootkit attacks are one of the most foundational threats to cyberspace. With the capability of subverting the software root of trust of a computer system, i.e., the operating system (OS) or the hypervisor, a rootkit can instantly take over the control of the system and stealthily inhabit the victim. To effectively mitigate and defeat them, researchers have explored various solutions. Unfortunately, the state-of-the-art defense is mainly reactive and in a fundamentally disadvantageous position in the arms-race against these stealthy attacks. The proposed research aims to fundamentally change the arms-race by proposing a systematic immunization approach to proactively prevent and exterminate rootkit attacks. Inspired by our human immune system and fundamental biological design principles, the proposed approach transforms system software (i.e., the OS and the hypervisor) so that the new one will tip the balance of favor toward the rootkit defense. To accomplish that, we will investigate a suite of innovative techniques to preserve kernel/hypervisor control flow integrity and evaluate their effectiveness with real-world malware and infrastructures. The proposed education components include the creation and dissemination of unique hands-on course materials with live demos, lab sessions, and tutorials.
Secure Open Systems Initiative
Dennis Kekas ; Peng Ning ; Mladen Vouk ; Rudra Dutta
$5,644,306 by Army Research Office 04/ 3/2008 - 11/30/2014 This program will establish a national Secure Open Systems Institute (SOSI), located on North Carolina State’s premier Centennial Campus that will be a global center for Open Source security research and development. The goals are twofold. First, SOSI will significantly contribute to strengthening mission critical information technology infrastructures vital to the Department of Defense, state and nation. Second, SOSI will accelerate the creation and growth of high tech industries in North Carolina and beyond by providing a centralized repository of research results, testing tools and qualification services. Learning Environments Across Disciplines LEADS: Supporting Technology Rich Learning Across Disciplines: Affect Generation and Regulation During Co-Regulated Learning in Game-Based Learning Environments (Supplement James Lester$30,894 by McGill University/Social Sciences and Humanities Research Council of Canada
04/ 1/2012 - 02/28/2020
Contemporary research on multi-agent learning environments has focused on self-regulated learning (SRL) while relatively little effort has been made to use co-regulated learning as a guiding theoretical framework (Hadwin et al., 2011). This oversight needs to be addressed given the complex nature that self-and other-regulatory processes play when human learners and artificial pedagogical agents (APAs) interact to support learners? internalization of cognitive, affective, and metacognitive (CAM) SRL processes. We will use the Crystal Island learning environment to investigate these issues.
SCH: INT: Collaborative Research: A Self-Adaptive Personalized Behavior Change System for Adolescent Preventive Healthcare
James Lester
$952,818 by National Science Foundation 10/ 1/2013 - 09/30/2017 Although the majority of adolescent health problems are amenable to behavioral intervention, and most adolescents are comfortable using interactive computing technology, few health information technology interventions have been integrated into adolescent care. The objective of the proposed research is to design, implement, and investigate INSPIRE, a self-adaptive personalized behavior change system for adolescent preventive healthcare. With a focus on adolescents, INSPIRE will enable adolescents to be active participants in dynamically generated, personalized narrative experiences that operationalize theoretically grounded interventions for behavior change through interactive narratives? plot structures and virtual character interactions. Learning Environments Across Disciplines LEADS: Supporting Technology Rich Learning Across Disciplines: Affect Generation and Regulation During Co-Regulated Learning in Game-Based Learning Environments James Lester$77,864 by McGill University/Social Sciences and Humanities Research Council of Canada
04/ 1/2012 - 03/31/2015
Contemporary research on multi-agent learning environments has focused on self-regulated learning (SRL) while relatively little effort has been made to use co-regulated learning as a guiding theoretical framework (Hadwin et al., 2011). This oversight needs to be addressed given the complex nature that self-and other-regulatory processes play when human learners and artificial pedagogical agents (APAs) interact to support learners? internalization of cognitive, affective, and metacognitive (CAM) SRL processes. We will use the Crystal Island learning environment to investigate these issues.
CHS: Medium: Adapting to Affect in Multimodal Dialogue-Rich Interaction with Middle School Students
James Lester ; Kristy Boyer ; Bradford Mott ; Eric Wiebe
$1,184,073 by National Science Foundation 08/ 1/2014 - 07/31/2017 Despite the great promise offered by learning environments for K-12 science education, realizing its potential poses significant challenges. In particular, current learning environments do not adaptively support children's affect. This project investigates computational models of affect that support multimodal dialogue-rich interaction. With an emphasis on devising scalable solutions, the project focus on machine-learning techniques for automatically acquiring affect and dialogue models that can be widely deployed in natural classroom settings. Type I: ENGAGE: Immersive Game-Based Learning for Middle Grade Computational Fluency James Lester ; Kristy Boyer ; Bradford Mott ; Eric Wiebe$1,047,996 by National Science Foundation
01/ 1/2012 - 12/31/2014
The goal of the ENGAGE project is to develop a game-based learning environment that will support middle grade computer fluency education. It will be conducted by an interdisciplinary research team drawn from computer science, computer science education, and education. In collaboration with North Carolina middle schools, the research team will design, develop, deploy, and evaluate a game-based learning environment that enables middle school students to acquire computer fluency knowledge and skills. The ENGAGE project will be evaluated in middle grade classrooms with respect to both learning effectiveness and engagement.
Using Deep Learning to Build Context-Sensitive Language Models
$272,839 by SAS Institute, Inc. 09/ 1/2014 - 08/31/2015 An important problem in Natural Language Processing is creating a robust language model that represents how a set of documents is composed from a set of words. Having a robust language model is critical for many text analytic tasks, including topic modeling, information retrieval, text categorization, and sentiment analysis. Most language models utilize a bag-of-words representation that ignores word order. While these representations suffice for some text analytic tasks, a context-sensitive language model is essential for many tasks such as discovering semantic relationships, question answering, machine translation, and dialog modeling. This project will investigate the effectiveness of leveraging Deep Learning architectures to create context-sensitive language models. Using Deep Learning to Build Context-Sensitive Language Models James Lester ; Bradford Mott$272,839 by SAS Institute
09/ 1/2014 - 08/31/2015
An important problem in Natural Language Processing is creating a robust language model that represents how a set of documents is composed from a set of words. Having a robust language model is critical for many text analytic tasks, including topic modeling, information retrieval, text categorization, and sentiment analysis. Most language models utilize a bag-of-words representation that ignores word order. While these representations suffice for some text analytic tasks, a context-sensitive language model is essential for many tasks such as discovering semantic relationships, question answering, machine translation, and dialog modeling. This project will investigate the effectiveness of leveraging Deep Learning architectures to create context-sensitive language models.
Detection and Transition Analysis of Engagement and Affect in a Simulation-Based Combat Medic Training Environment
$478,592 by Columbia University/US Army Research Laboratory 12/19/2012 - 06/17/2015 The project will develop automated detectors that can infer the engagement and affect of trainees learning through the vMedic training system. This project will combine interaction-based methods for detecting these constructs (e.g., models making inference solely from the trainee?s performance history) with scalable sensor-based methods for detecting these constructs, towards developing models that can leverage sensor information when it is available, but which can still assess trainee engagement and affect effectively even when sensors are not available. The automated detectors will be developed, integrated together, and validated for accuracy when applied to new trainees. The Leonardo Project: An Intelligent Cyberlearning System for Interactive Scientific Modeling in Elementary Science Education James Lester ; Bradford Mott ; Michael Carter ; Eric Weibe$3,499,409 by National Science Foundation
08/15/2010 - 07/31/2015
The goal of the Leonardo project is to develop an intelligent cyberlearning system for interactive scientific modeling. Students will use Leonardo's intelligent virtual science notebooks to create and experiment with interactive models of physical phenomena. As students design and test their models, Leonardo's intelligent virtual tutors will engage them in problem-solving exchanges in which they will interactively annotate their models as they devise explanations and make predictions. During the project, the Leonardo virtual science notebook system will be rolled out to 60 classrooms in North Carolina, Texas, and California.
Co-Design of Hardware / Software for Predicting MAV Aerodynamics
Frank Mueller
$799,999 by Virginia Polytechnic Institute and State University (US Air Force) 09/ 1/2012 - 10/31/2017 This proposal provides subcontractor support to Virginia Tech for a proposal submitted under the Air Force's Basic Research Initiative. The proposal will focus on development of reconfigurable mapping strategies for porting multi-block structured and unstructured-mesh CFD codes to computing clusters containing CPU/GPU processing units. Hobbes: OS and Runtime Support for Application Composition Frank Mueller$300,000 by Sandia National Laboratories via US Dept of Energy
10/24/2013 - 10/23/2016
This project intends to deliver an operating system and runtime system (OS/R) environment for extreme-scale scientific computing. We will develop the necessary OS/R interfaces and lowlevel system services to support isolation and sharing functionality for designing and implementing applications as well as performance and correctness tools. We propose a lightweight OS/R system with the flexibility to custom build runtimes for any particular purpose. Each component executes in its own "enclave" with a specialized runtime and isolation properties. A global runtime system provides the software required to compose applications out of a collection of enclaves, join them through secure and low-latency communication, and schedule them to avoid contention and maximize resource utilization. The primary deliverable of this project is a full OS/R stack based on the Kitten operating system and Palacios virtual machine monitor that can be delivered to vendors for further enhancement and optimization.
CPS: Breakthrough: Collaborative Research: Bringing the Multicore Revolution to Safety-Critical Cyber-Physical Systems
Frank Mueller
$225,000 by National Science Foundation 02/ 1/2013 - 01/31/2016 Multicore platforms have the potential of revolutionizing the capabilities of embedded cyber-physical systems but lack predictability in execution time due to shared resources. Safety-critical systems require such predictability for certification. This research aims at resolving this multicore predictability problem.'' It will develop methods that enable to share hardware resources to be allocated and provide predictability, including support for real-time operating systems, middleware, and associated analysis tools. The devised methods will be evaluated through experimental research involving synthetic micro-benchmarks and code for unmanned air vehicles re-thinking'' their adapting to changing environmental conditions within Cyber-Physical Systems. SHF: Small: RESYST: Resilience via Synergistic Redundancy and Fault Tolerance for High-End Computing Frank Mueller$376,219 by National Science Foundation
10/ 1/2010 - 09/30/2015
In High-End Computing (HEC), faults have become the norm rather than the exception for parallel computation on clusters with 10s/100s of thousands of cores. As the core count increases, so does the overhead for fault-tolerant techniques that rely on checkpoint/restart (C/R) mechanisms. At 50% overheads, redundancy is a viable alternative to fault recovery and actually scales, which makes the approach attractive for HEC. The objective of this work to the develop a synergistic approach by combining C/R-based fault tolerance with redundancy in computer to achieve high levels of resilience. This work alleviates scalability limitations of current fault tolerant practices. It contributes to fault modeling as well as fault detection and recovery in significantly advancing existing techniques by controlling levels of redundancy and checkpointing intervals in the presence of faults. It is transformative in providing a model where users select a target failure probability at the price of using additional resources.
Frank Mueller
$153,934 by Lawrence Berkeley National Laboratory via US Dept of Energy 09/24/2013 - 08/15/2015 he objective of this work is to provide functionality for the BLCR Linux module under a PGAS runtime system (within the DEGAS software stack) to support advanced fault-tolerant capabilities, which are of specific value in the context of large-scale computational science codes running on high-end clusters and, ultimately, exascale facilities. Our proposal is to develop and integrate into DEGAS a set of advanced techniques to reduce the checkpoint/restart (C/R)overhead. SHF: Small: Scalable Trace-Based Tools for In-Situ Data Analysis of HPC Applications (ScalaJack) Frank Mueller$457,395 by National Science Foundation
06/ 1/2012 - 05/31/2015
This decade is projected to usher in the period of exascale computing with the advent of systems with more than 500 million concurrent tasks. Harnessing such hardware with coordinated computing in software poses significant challenges. Production codes tend to face scalability problems, but current performance analysis tools seldom operate effectively beyond 10,000 cores. We propose to combine trace analysis and in-situ data analysis techniques at runtime. Application developers thus create ultra low-overhead measurement and analysis facilities on-the-fly, customized for the performance problems of particular application. We propose an analysis generator called ScalaJack for this purpose. Results of this work will be contributed as open-source code to the research community and beyond as done in past projects. Pluggable, customization analysis not only allows other groups to build tools on top of our approach but to also contribute components to our framework that will be shared in a repository hosted by us.
Operating System Mechanisms for Many-Core Systems-Phase II (PICASO II) Pico-kernel Adaptive and Scalable Operating Systems Phase II
Frank Mueller
$225,000 by Securboration via US Air Force Research Laboratory 06/ 1/2013 - 05/31/2015 The objective of this work is to design and evaluate novel system and program abstractions for combined performance and scalability paving the path into a future of operating system supporting a massive number of cores on a single chip. CAREER:Expanding Developers' Usage of Software Tools by Enabling Social Learning Emerson Murphy-Hill$495,721 by National Science Foundation
08/ 1/2013 - 07/31/2018
Tools can help software developers alleviate the challenge of creating and maintaining software. Unfortunately, developers only use a small subset of the available tools. The proposed research investigates how social learning, an effective mechanism for discovering new tools, can help software developers to discover relevant tools. In doing so, developers will be able to increase software quality while decreasing development time.
TWC: Small: Collaborative: Discovering Software Vulnerabilities Through Interactive Static Analysis
Emerson Murphy-Hill
$249,854 by National Science Foundation 10/ 1/2013 - 09/30/2016 Software vulnerabilities originating from insecure code are one of the leading causes of security problems people face today. Current tool support for secure programming focuses on catching security errors after the program is written. We propose a new approach, interactive static analysis, to improve upon static analysis techniques by introducing a new mixed-initiative paradigm for interacting with developers to aid in the detection and prevention of security vulnerabilities. DO 2 Task 3.7 - Murphy Hill Emerson Murphy-Hill$49,486 by Laboratory for Analytic Sciences
09/13/2013 - 12/31/2014
LAS DO3 Task Order 2.11 Mission Enabling - Murphy-Hill
Emerson Murphy-Hill
$72,651 by Laboratory for Analytic Sciences 03/31/2014 - 12/31/2014 DO3 Task Order 2.11 Mission Enabling LAS DO3 Task Order 2.8 Analytic Workflow Emerson Murphy-Hill$16,686 by LAS
03/31/2014 - 12/31/2014
DO3 Task Order 2.8 Analytic Workflow
SHF: Small: Expressive and Scalable Notifications from Program Analysis Tools
Emerson Murphy-Hill ; Sarah Heckman
$250,000 by National Science Foundation 10/ 1/2012 - 09/30/2015 A wide variety of program analysis tools have been created to help software developers do their jobs, yet the output of these tools are often difficult to understand and vary significantly from tool to tool. As a result, software developers may waste time trying to interpret the output of these tools, instead of making their software more capable and reliable. This proposal suggests a broad investigation of several types of program analysis tools, with the end goal being an improved understanding of how program analysis tools can inform developers in the most expressive and uniform way possible. Once this goal is reached, we can create program analysis tools that enable developers to make tremendous strides towards more correct, more reliable, and more on-time software systems. CISCO-NCSU Internship Program Peng Ning$160,000 by Cisco Systems, Inc.
07/12/2011 - 07/14/2016
This is a pilot internship program between NCSU and Cisco for 4 undergraduate students to learn through working part-time on real life problems for Cisco with the hope that this pilot program can grow and develop into a long term working relationship. Specifically, NCSU students will participate in Cisco Software Application Support plus Upgrades (SASU) projects and/or conduct research for SASU. This will be done with an understanding that the interns are students, and as such are learning and being trained with the training coming from both the Cisco (for SASU-specific skills), and NCSU (through the undergraduate program they are enrolled in) in general relevant skills.
Is Wireless Channel Dependable for Security Provisioning?
Peng Ning (co-PI)
$350,000 by the National Science Foundation 08/12/2013 - 07/31/2016 Wireless security is receiving increasing attention as wireless systems become a key component in our daily life as well as critical cyber-physical systems. Recent progress in this area exploits physical layer characteristics to offer enhanced and sometimes the only available security mechanisms. The success of such security mechanisms depends crucially on the correct modeling of underlying wireless propagation. It is widely accepted that wireless channels decorrelate fast over space, and half a wavelength is the key distance metric used in existing wireless physical layer security mechanisms for security assurance. We believe that this channel correlation model is incorrect in general: it leads to wrong hypothesis about the inference capability of a passive adversary and results in false sense of security, which will expose the legitimate systems to severe threats with little awareness. In this project, we seek to understand the fundamental limits in passive inference of wireless channel characteristics, and further advance our knowledge and practice in wireless security. III: Small: Optimization Techniques for Scalable Semantic Web Data Processing in the Cloud Kemafor Ogan$446,942 by National Science Foundation
09/ 1/2012 - 08/31/2015
Achieving scalable processing of the increasing amount of publicly-available Semantic Web data will hinge on parallelization. The Map-Reduce programming paradigm recently emerged as a de-facto parallel data processing standard and has demonstrated effectiveness with respect to structured and unstructured data. However, Semantic Web data presents challenges not adequately addressed by existing techniques due to its flexible, fine-grained data model and the need to reason beyond explicitly represented data. This project will investigate optimization techniques that address these unique challenges based on rethinking Semantic Web data processing on Map-Reduce platforms from the ground, up - from query algebra to query execution.
LAS DO3 Task Order 2.9 KRM
Kemafor Ogan
$74,523 by LAS 03/31/2014 - 12/31/2014 DO3 Task Order 2.9 KRM TWC SBE: Medium: Collaborative: User-Centric Risk Communication and Control on Mobile Devices Douglas Reeves$267,096 by the National Science Foundation
09/ 1/2013 - 08/31/2016
Human-system interactions is an integral part of any system. Because the vast majority of ordinary users have limited technical knowledge and can easily be confused and/or worn out by repeated security notifications/questions, the quality of users? decisions tends to be very low. On the other hand, any system targeting end-users must have the flexibility to accommodate a wide spectrum of different users, and therefore needs to get the full range of users involved in the decision making loop. This dilemma between fallible human nature and inevitable human decision making is one main challenge to the goal of improving security. In this project, we aim at developing principles and mechanisms for usable risk communication and control. The major technical innovations include (1) multi-granularity risk communications; (2) relative risk information in the context of comparison with alternatives; (3) Discover and integrate risk information from multiple sources; (4) Expand opportunities for risk communication and control.
Runtime Enforcement of Security Policies
Douglas Reeves
$29,780 by US Army - Army Research Office 03/ 5/2014 - 12/ 4/2014 Android smartphones have grown in market share and have penetrated all corners of the market, including US Government and, in particular, DoD. The ecosystem of the Android App marketplace while encouraging creativity also has lax standards. Recent work by Aiken's group shows that it is possible to use static analysis techniques to identify vulnerabilities due to abuse of {\em permissions} afforded to the software app, by the user, but with potential for false positives and attendant necessity for manual analysis. In this preliminary investigation, we propose to investigate a run time monitor that could be used in combination with static analysis to enforce strict permission policies. The particular research questions we will consider are: x Design of a language for expressing positive (shall) and negative (should not) permission x Algorithms for instrumenting application code that would be used to maintain invariants implied by the permission policies set by the user x Algorithms for instrumenting application code to collect trace data that could be mined later for surreptitious violations of security policies, and algorithms for deleting applications automatically when policies are violated. The three parts of the research proposal, when taken together, correspond to traditional law enforcement strategies -- setting of the law, monitoring for compliance, and imposition of penalty when laws are broken. While the ultimate goal is to validate the proposed work in the context of the Android market place, the proposed preliminary investigation will be theoretical in nature. NeTS: Small: Investigation of Human Mobility: Measurement, Modeling,Analysis, Applications and Protocols Injong Rhee$298,356 by National Science Foundation
08/ 1/2010 - 07/31/2015
Simulating realistic mobility patterns of mobile devices is important for the performance study of mobile networks because deploying a real testbed of mobile networks is extremely difficult, and furthermore, even with such a testbed, constructing repeatable performance experiments using mobile devices is not trivial. Humans are a big factor in simulating mobile networks as most mobile nodes or devices (cell phones, PDAs and cars) are attached to or driven by humans. Emulating the realistic mobility patterns of humans can enhance the realism of simulation-based performance evaluation of human-driven mobile networks. Our NSF-funded research that ends this year has studied the patterns of human mobility using GPS traces of over 100 volunteers from five different sites including university campuses, New York City, Disney World, and State Fair. This research has revealed many important fundamental statistical properties of human mobility, namely heavy-tail flight distributions, self-similar dispersion of visit points, and least-action principle for trip planning. Most of all, it finds that people tend to optimize their trips in a way to minimize their discomfort or cost of trips (e.g., distance). No existing mobility models explicitly represent all of these properties. Our results are very encouraging and the proposed research will extend the work well beyond what has been accomplished so far. . We will perform a measurement study tracking the mobility of 100 or 200 students in a campus simultaneously, and analyze the mobility patterns associated with geo-physical and social contexts of participants including social networks, interactions, spatio-temporal correlations, and meetings. . We will cast the problem of mobility modeling as an optimization problem borrowing techniques from AI and Robotics which will make it easy to incorporate the statistical properties of mobility patterns commonly arising from group mobility traces. The realism of our models in expressing human mobility will surpass any existing human mobility models. . We will develop new routing protocols leveraging the researched statistical properties found in real traces to optimize delivery performance. The end products of the proposed research is (a) a new human mobility model that is capable of realistically expressing mobility patterns arising from reaction to social and geo-physical contexts, (b) their implementation in network simulators such as NS-2/3 and GloMoSim, (c) mobility traces that contain both trajectories of people in a university campus and contact times, (d) new efficient routing protocols for mobile networks
NetSE: Large: Collaborative Research: Platys: From Position to Place in Next Generation Networks
Injong Rhee ; Munindar Singh
$706,167 by National Science Foundation 09/ 1/2009 - 08/31/2015 This project develops a high-level notion of context that exploits the capabilities of next genera-tion networks to enable applications that deliver better user experiences. In particular, it exploits mobile devices always with a user to capture key elements of context: the user's location and, through localization, characteristics of the user's environment. RI: Small: Collaborative Research: Speeding Up Learning through Modeling the Pragmatics of Training David Roberts$156,203 by National Science Foundation
10/ 1/2013 - 09/30/2015
We propose to develop techniques that will enable humans to train computers efficiently and intuitively. In this proposed work, we draw inspiration from the ways that human trainers teach dogs complex behaviors to develop novel machine learning paradigms that will enable intelligent agents to learn from human trainers quickly, and in a way that humans can intuitively take advantage of. This research aims to return to the basics of programming---it seeks to develop novel methods that allow humans to tell computers what to do. More specifically, this research will develop learning techniques that explicitly model and leverage the implicit communication channel that humans use while training, a process akin to interpreting the pragmatic implicature of a natural language communication. We will develop algorithms that view the training process as an intentional communicative act, and can vastly outperform standard reward-seeking algorithms in terms of the speed and accuracy with which human trainers can generate desired behavior.
CPS: Synergy: Integrated Sensing and Control Algorithms for Computer-Assisted Training (Computer Assisted Training Systems (CATS) for Dogs)
David Roberts ; Alper Bozkurt ECE ; Barbara Sherman CVM
$999,103 by National Science Foundation 10/ 1/2013 - 09/30/2016 We propose to develop tools and techniques that will enable more effective two-way communication between dogs and handlers. We will work to create non-invasive physiological and inertial measuring devices that will transmit real-time information wirelessly to a computer. We will also develop technologies that will enable the computer to train desired behaviors using positive reinforcement without the direct input from humans. We will work to validate our approach using laboratory animals in the CVM as well as with a local assistance dog training organization working as a consultant. NeTS:Small: Computationally Scalable Optical Network Design George Rouskas$429,995 by NSF
08/ 1/2011 - 07/31/2015
Optical networking forms the foundation of the global network infrastructure, hence the planning and design of optical networks is crucial to the operation and economics of the Internet and its ability to support critical and reliable communication services. With this research project we aim to make contributions that will lead to a quantum leap in the ability to solve optimally a range of optical design problems. In particular, we will develop compact formulations and solution approaches that can be applied efficiently to instances encountered in Internet-scale environments. Our goal is to lower the barrier to entry in fully exploring the solution space and in implementing and deploying innovative designs. The solutions we will develop are "future-proof" with respect to advances in DWDM transmission technology, as the size of the corresponding problem formulations is independent of the number of wavelengths.
Graduate Industrial Traineeship for Vedika Seth
George Rouskas
$52,547 by SAS Institute, INC 05/11/2014 - 05/31/2015 NCSU through the SAS GA will provide research and analysis to SAS as set forth in this Agreement. Such research and analysis shall include, but is not limited to, research, generation, testing, and documentation of operations research software. SAS GA will provide such services for SAS' offices in Cary, North Carolina, at such times as have been mutually agreed upon by the parties. Graduate Industrial Traineeship for Ameeta Muralidharan George Rouskas$65,051 by SAS Institute, Inc.
01/27/2014 - 05/31/2015
NCSU through the SAS GA will provide research and analysis to SAS as set forth in this Agreement. Such research and analysis shall include, but is not limited to, research, generation, testing, and documentation of operations research software. SAS GA will provide such services for SAS' offices in Cary, North Carolina, at such times as have been mutually agreed upon by the parties.
Graduate Industrial Traineeship for Namita Shubhy
George Rouskas
$65,050 by SAS Institute, Inc. 01/27/2014 - 05/31/2015 NCSU through the SAS GA will provide research and analysis to SAS as set forth in this Agreement. Such research and analysis shall include, but is not limited to, research, generation, testing, and documentation of operations research software. SAS GA will provide such services for SAS' offices in Cary, North Carolina, at such times as have been mutually agreed upon by the parties. In Situ Indexing and Query Processing of AMR Data Nagiza Samatova$383,000 by US Department of Energy
09/ 1/2014 - 08/31/2017
One of the most significant advances for large-scale scientific simulations has been the advent of Adaptive Mesh Refinement, or AMR. By using dynamic gridding, AMR can achieve substantial savings in memory, computation, and disk resources while maintaining or even increasing simulation accuracy, relative to static, uniform gridding. However, the resultant non-uniform structure of the simulation mesh produced by AMR methods cause inefficient access patterns during data analysis and visualization. Given the exponential increase in simulation output, the massive I/O operations are becoming a substantial bottleneck in simulations and analysis. To efficiently analyze AMR data, we propose an integrated, three-prong approach that aims: (a) To devise an AMR query model; (b) To develop effective in situ indexing and query processing methods for AMR data analytics; and (c) To investigate data storage layout strategies for AMR data retrieval optimized for analytics-induced heterogeneous data access patterns. The results, algorithms, and software will be in the public domain.
Runtime System for I/O Staging in Support of In-Situ Processing of Extreme Scale Data
Nagiza Samatova
$286,140 by Oak Ridge National Loboratory/Dept. of Energy 01/31/2011 - 08/31/2015 Accelerating the rate of insight and scientific productivity demands new solutions to managing the avalanche of data expected in extreme-scale. Our approach is to use tools that can reduce, analyze, and index the data while it is still in memory (referred to as "in-situ" processing of data). ). In order to deal with the large amount of data generated by the simulations, our team has partnered with many application teams to deliver proven technology that can accelerate their knowledge discovery process. These technologies include ADIOS, FastBit, and Parallel R. In this proposal we wish to integrate these technologies together, and create a runtime system that will allow scientist to create an easy-to-use scientific workflow system, that will run in situ, in extra nodes on the system, which is used to not only accelerate their I/O speeds, but also to pre-analyze, index, visualize, and reduce the overall amount of information from these solutions. Joint Faculty Agreement For Nagiza Samatova Nagiza Samatova$507,294 by Oak Ridge National Laboratories - UT Battelle, LLC
08/ 9/2007 - 08/ 8/2015
Dr. Nagiza Samatova's joint work with NC State University and Oak Ridge National Laboratory (ORNL) will provide the interface between the two organizations aiming to collaboratively address computational challenges in the Scientific Data Management, Data-Intensive Computing for Understanding Complex Biologicial Systems, Knowledge Integration for the Shewanella Federation, and the Large-Scale Analysis of Biologicial Networks with Applications to Bioenergy Production.
Scalable and Power Efficient Data Analytics for Hybrid Exascale Systems
Nagiza Samatova
$364,944 by Oak Ridge National Laboratories/ US Dept. of Energy 01/31/2011 - 12/31/2014 The specific objectives of the proposal are as follows: 1. Design and develop data mining kernels and algorithms for acceleration on hybrid architectures which include many-core systems, GPUs, and other accelerators. 2. Design and develop approximate scalable algorithms for data mining and analysis kernels enabling faster exploration, more efficient resource usage, reduced memory footprint, and more power efficient computations. 3. Design and develop index-based data analysis and mining kernels and algorithms for performance and power optimizations including index-based perturbation analysis kernels for noisy and uncertain data. 4. Demonstrate the results of our project by enabling analytics at scale for selected applications on large-scale HPC systems. Analytics-driven Efficient Indexing and Query Processing of Extreme Scale AMR Data Nagiza Samatova$149,999 by National Science Foundation
05/ 1/2012 - 12/31/2014
One of the most significant advances for large-scale scientific simulations has been the advent of Adaptive Mesh Refinement, or AMR. By using dynamic gridding, AMR can achieve substantial savings in memory, computation, and disk resources while maintaining or even increasing simulation accuracy, relative to static, uniform gridding. However, the resultant non-uniform structure of the simulation mesh produced by AMR methods cause inefficient post-simulation access patterns during AMR data analytics that is becoming a substantial bottleneck given the exponential increase in simulation output. Toward bridging this gap in efficient analytics support for AMR data, we propose an integrated, three-prong approach that aims: (a) To devise an AMR query model; (b) To explore effective indexing methods for AMR data analytics; and (c) To investigate data storage layout strategies for AMR data retrieval optimized for analytics-induced heterogeneous data access patterns.
Nagiza Samatova
$49,585 by Laboratory for Analytic Sciences 09/13/2013 - 12/31/2014 DO 2 Task 3.7 activities Scalable Data Management, Analysis, and Visualization (SDAV) Institute Nagiza Samatova ; Anatoli Melechko$750,000 by US Department of Energy
02/15/2012 - 02/14/2017
The SDAV is a unique and comprehensive combination of scientific data management, analysis, and visualization expertise and technologies aimed at enabling scientific knowledge discovery for applications running on state-of-the-art computational platforms located at DOE's primary computing facilities. This integrated institute focuses on tackling key challenges facing applications in our three focus areas through a well-coordinated team and management organization that can respond to changing technical and programmatic objectives. The proposed work portfolio is a blend of applied research and development, aimed at having key software services operate effectively on large distributed memory multi-core, and many-core platforms and especially DOE's open high performance computing facilities. Our goal is to create an integrated, open source, sustainable framework and software tools for the science community.
Scalable Statistical Computing For Physical Science Applications
Nagiza Samatova ; Anatoli Melechko
$354,646 by US Department of Energy (DOE) 12/ 2/2011 - 06/30/2015 Physical science applications such as nanoscience, fusion science, climate and biology generate large-scale data sets from their simulations and high throughput technologies. This necessitates scalable technologies for processing and analyzing this data. We plan to research and develop advanced data mining algorithms for knowledge discovery from this complex, high-dimensional, and noisy data. We will apply these technologies to DOE-mission scientific applications related to fusion energy, bioenergy, understanding the impacts of climate extremes, and insider threat detection and mitigation. Ultrascale Computational Modeling of Phenotype-Specific Metabolic Processes in Microbial Communities Nagiza Samatova ; Anatoli Melechko$454,311 by Oak National Laboratories - UT Battelle (DOE)
01/15/2010 - 12/31/2014
Ultrascale computational modeling methods will be developed for revealing phenotype-specific metabolic processes and their cross-talks and applied to the critical DOE problem of acid mine drainage (AMD). The apex of the project will be a systematic and iterative computational procedure for: (1) identification and expression-level characterization of phenotype-related genes; (2) reconstruction of phenotype-specific metabolic pathways enriched by these genes; (3) elucidation of the symbiotic and/or competing interplay between these pathways across species; and (4) characterization of evolutionary and environmental adaptation of the community.
Collaborative Research: Understanding Climate Change: A Data Driven Approach
Nagiza Samatova ; Frederick Semazzi
$1,815,739 by National Science Foundation 09/ 1/2010 - 08/31/2015 The goal is to provide a computational capability for effective and efficient exploration of high-resolution climate networks derived from multivariate, uncertain, noisy and spatio-temporal climate data. We plan to increase the efficiency and climatologically relevancy of the network patterns identification through integrated research activities focused on: (a) supporting comparative analysis of multiple climate networks; (b) constraining the search space via exploiting the inherent structure (e.g., multi-partite) of climate networks; (c) establishing the foundation to efficiently update solutions for perturbed (changing) graphs; and (d) designing and implementing parallel algorithms scalable to thousands of processors on multi-node multi-core supercomputer architectures. LAS DO3 Task Order 2.6 Future States Nagiza Samatove$50,320 by LAS
03/31/2014 - 12/31/2014
DO3 Task Order 2.6 Future States
Lecture Hall Polytopes, Inversion Sequences, and Eulerian Polynomials
Carla Savage
$30,000 by Simons Foundation 09/ 1/2012 - 08/31/2017 Over the past ten years, lecture hall partitions have emerged as fundamental structures in combinatorics and number theory, leading to new generalizations and new interpretations of several classical theorems. This project takes a geometric view of lecture hall partitions and uses polyhedral geometry to investigate their remarkable properties. CCF:SHF:Small: Non-Uniformity-Centric Program Optimizations for Dynamic Computations on Chip Multiprocessors Xipeng Shen$404,956 by National Science Foundation
06/16/2014 - 08/31/2016
In this project, Dr. Xipeng Shen and his team are building a new paradigm of program optimizations. It is motivated by a growing gap between trends in processor development and needs of modern data-intensive dynamic applications. This class of applications, ranging from differential equation solvers to data mining tools to particle dynamics simulations, play an essential role in science and humanity. But they feature tremendous data accesses and complex patterns in data accesses or control flows. The properties make them a great challenge for modern processors which are evolving exactly opposite to these applications' needs: A chip's aggregate computing power is rapidly outgrowing memory bandwidth; the rise of throughput-oriented manycores makes system throughput even more sensitive to irregular computations. The paradigm being built by Dr. Shen's team, namely "non-uniformity--centric optimizations", distinctively takes the non-uniform inter-core relations in modern systems as the first-order constraint for program optimizations. Specifically, they are developing a framework named PipeReg as a new way to reorganize data accesses and threads during run time to reduce the influence of irregular computations on the throughput of massively parallel processors. Meanwhile, they are investigating a novel kind of program transformations called neighborhood-aware transformations, which exploits the non-uniform interactions among threads in on-chip storage (e.g., shared cache) in a multi-socket multicore system. Together, the two techniques will synergistically remove some important barriers for data-intensive dynamic applications to tap into the full power of future computing systems. The outcome from this research will provide essential support for enhancing the computing efficiency of data-intensive dynamic applications in the era of heterogeneous parallel systems. Because of the critical roles of these applications, this research will help foster sustained advancement in science, commerce, health, and so on. Beyond its technical content, this project stresses technology transfer, develops new teaching materials and tools, emphasizes demographic diversity, and improves the training of both graduate and undergraduate students.
CAREER: Input-Centric Program Behavior Analysis and Adaptation
Xipeng Shen
$266,165 by National Science Foundation 07/28/2014 - 08/31/2015 By analyzing and predicting program dynamic behaviors, program behavior analysis offers the fundamental support for program transformations and resource management. Its effectiveness is crucial for the maximization of computing efficiency. This research proposes to include program inputs---a so far virtually ignored dimension---into the focus of program behavior analysis, cultivating a new paradigm, namely input-centric program behavior analysis and adaptation. This input-centric paradigm will create many new opportunities for enhancing the matching between software and hardware, hence significantly improving the performance and power efficiency in modern computing. The proposed technique, input-centric program behavior analysis and adaptation, consists of three components. The first two components, program input characterization and input-behavior modeling, resolve the complexities of program inputs, extract important features, and recognize the correlations between characterized input features and program behaviors. The third component, input-centric adaptation, capitalizes on the novel opportunities that the first two components create, making dynamic optimizations proactive and holistic, but without losing the adaptivity to inputs and environmental changes. Together, the three components make evolvable programming systems more feasible than before. In such a system, the input-behavior models embody the central knowledge base, which grows incrementally across program production runs. As the knowledge base becomes larger, behavior prediction becomes more accurate, stimulating better software-hardware matching and making the program and runtime systems perform increasingly better. Because of the fundamental role of program behavior analysis in software-hardware matching, this research helps pave the way for advancing the optimizations in various layers in the software execution stack (compilers, virtual machines, OS, etc.). Policy-Based Governance for the OOI Cyberinfrastructure Munindar Singh$124,688 by Univ of Calif-San Diego/NSF
09/ 1/2009 - 02/25/2015
This project will develop policy-based service governance modules for the Oceanographic Observatories Initiative (OOI) Cyberinfrastructure. The main objectives of the proposed project include (1) formulating the key conceptual model underlying the patterns of governance; (2) formalizing "best practices" patterns familiar to the scientific community and seeding the cyberinfrastructure with them; (3) understanding user requirements for tools that support creating and editing patterns of governance
DO 2 Task 3.7 - Singh
Munindar Singh
$43,889 by Laboratory for Analytic Sciences 09/13/2013 - 12/31/2014 DO 2 Task 3.7 activities Student Support for Participation in the Symposium and Bootcamp on the Science of Security (HotSoS) Munindar Singh$5,000 by National Science Foundation (NSF)
01/ 1/2014 - 12/31/2014
This project will support travel by US student researchers to the Symposium and Bootcamp on the Science of Security (HotSoS), which will be held in April 2014 in Raleigh, North Carolina. Travel support will be critical in encouraging participation, which is especially important since HotSoS 2014 will be one of the first peer-reviewed events on the Science of Security.
LAS DO3 Task Order 2.11 Mission Enabling - Singh
Munindar Singh
$240,158 by Laboratory for Analytic Sciences 03/31/2014 - 12/31/2014 DO3 Task Order 2.11 Mission Enabling LAS DO3 Task Order 2.8 Analytic Workflow Munindar Singh$61,570 by LAS
03/31/2014 - 12/31/2014
DO3 Task Order 2.8 Analytic Workflow
LAS DO3 Task Order 2.9 KRM
Munindar Singh
$80,116 by LAS 03/31/2014 - 12/31/2014 DO3 Task Order 2.9 KRM DO 2 Task 3.8 - St. Amant Robert St. Amant$49,493 by Laboratory for Analytic Sciences
09/13/2013 - 12/31/2014
LAS DO3 Task Order 2.5 Cognitive Processing
Robert St. Amant
$90,860 by LAS 03/31/2014 - 12/31/2014 DO3 Task Order 2.5 Cognitive Processing EAGER: Cognitive Modeling of Strategies for Dealing with Errors in Mobile Touch Interfaces Robert St. Amant ; Emerson Murphy-Hill$281,076 by National Science Foundation
09/ 1/2014 - 08/31/2016
Touch interfaces on mobile phones and tablets are notoriously error prone in use. One plausible reason for slow progress in improving usability is that research and design efforts in HCI take a relatively narrow focus on isolating and eliminating human error. We take a different perspective: failure represents breakdowns in adaptations directed at coping with complexity. The key to improved usability is understanding the factors that contribute to both expertise and its breakdown. We propose to develop cognitive models of strategies for touch interaction. Our research will examine the detailed interactions between users perceptual, cognitive, and motor processes in recognizing, recovering from, and avoiding errors in touch interfaces. Our proposal is for three stages of research: exploratory experiments, analysis and modeling, and finally validation experiments.
CHS: SMALL: Direct Physical Grasping, Manipulation, and Tooling of Simulated Objects
Robert St.Amant ; Christopher Healey
$496,858 by National Science Foundation 08/ 1/2014 - 07/31/2017 This proposal is for the development and evaluation of CAPTIVE, a Cube with Augmented Physical Tools, to support exploration of three-dimensional information. The design of CAPTIVE is founded on the concept of tool use, in which physical objects (tools) are used to modify the properties or presentation of target objects. CAPTIVE integrates findings across a wide range of areas in human-computer interaction and visualization, from bimanual and tangible user interfaces to augmented reality. CAPTIVE is configured as a desktop augmented reality/fishtank virtual reality system [120], with a stereo- scopic display, a haptic pointing device, and a user-facing camera. In one hand the user holds a wireframe cube that contains virtual objects, in the other the pointing device, augmented to reflect its function as a tool: a probe probes for pointing at, choosing, and moving objects; a magnifying or semantic lens for filter- ing, recoding, and elaborating information; a cutting plane that shows slices or projection views. CAPTIVE supports visualization with more fluid and natural interaction techniques, improving the ability of users to explore and understand 3D information. LAS DO3 Task Order 2.2 Steck-Infrastructure John Streck ; John Bass$277,915 by Laboratory for Analytic Sciences
03/31/2014 - 12/31/2014
Joint Faculty Appointment For Vida Blair Sullivan
Vida Sullivan
$136,144 by Oak Ridge National Laboratories - UT-Battelle LLC 09/13/2013 - 08/15/2015 The PI's unique combination of expertise in structural graph theory and scalable graph algorithms/big data is necessary to ensure the success of ORNL-based projects using applied discrete mathematics to enable advances in graph generation and HPC benchmarking, social network analysis, and multi-scale graph decompositions for the Department of Energy and Department of Defense. The PI will direct and conduct fundamental research, collaborate with ORNL staff, write up research results for peer-reviewed publication, give presentations, and mentor students and postdoctoral staff as appropriate. NCDS Data Science Faculty Fellow-Tracking Community Evolution in Dynamic Graph Data Using Tree-Like Structure Vida Sullivan$30,000 by National Consortium for Data Science (UNC-UNC Chapel Hill)
01/ 1/2014 - 12/31/2014
"Big Data" sources for many real-world applications pose numerous challenges to understanding the complex and possibly hidden relationships between components of a complex network. Furthermore, these networks often consist of heterogeneous entities and types of relationships, and many existing algorithms for computing network features and similarity are not directly applicable. In order to draw actionable insights, analysis need to identify events of interest, place them in context, and understand their impact. Existing approaches which emphasize visualization (at the expense of analytics), struggle to coherently present networks with hundreds of entities, whereas practical applications require hundreds of thousands (or more). We propose to develop methods which integrate ideas from graph theory with multi-scale modeling (since events of interest may occur at different levels of granularity/contexts within the data) to improve comprehension of such relational data and form a foundation for novel methods of visualization and interaction.
Triangle Computer Science Distinquished Lecture Series
$6,700 by Duke University (Us Army-Army Research Office) 01/ 1/2014 - 12/31/2014 Since 1995, the Triangle Computer Science Distinguished Lecturer Series (TCSDLS) has been hosting influential university researchers and industry leaders from computer-related fields as speakers at the three universities within the Research Triangle Area. The lecturer series, sponsored by the Army Research Office (ARO), is organized and administered by the Computer Science departments at Duke University, NC State University, and the University of North Carolina at Chapel Hill. This proposal argues for continuation, for an additional 3 years, of this highly successful lecturer series which is being led by Duke University. Collaborative Research: CPATH II: Incorporating Communication Outcomes into the Computer Science Curriculum Mladen Vouk ; Michael Carter (co-PI). Grad$369,881 by National Science Foundation
10/ 1/2009 - 03/31/2015
In partnership with industry and faculty from across the country, this project will develop a transformative approach to developing the communication abilities (writing, speaking, teaming, and reading) of Computer Science and Software Engineering students. We will integrate communication instruction and activities throughout the curriculum in ways that enhance rather than replace their learning technical content and that supports development of computational thinking abilities of the students. We will implement the approach at two institutions. By creating concepts and resources that can be adapted by all CS and SE programs, this project also has the potential to increase higher education's ability nationwide to meet industry need for CS and SE graduates with much better communication abilities than, on average, is the case today. In addition, by using the concepts and resources developed in this project, CS and SE programs will be able to increase their graduates' mastery of technical content and computational thinking.
Investigation of a Novel Indoor Localization (Navigation) Technique For Smartphones
$75,000 by Samsung Telecommunications America, LLC - TX 01/ 2/2012 - 12/31/2014 In this project, we aim at developing a new indoor localization technique relying on low-frequency radio that can penetrate indoor obstacles (or detour obstacles by diffraction in the shortest path) by its long wave characteristics. The smartphone running this system would be able to identify its position by measuring straight-line distances from a few radio transmission towers deployed in a city scale (or in a district scale). Straight-line distances that have not been affected by indoor obstacles would be able to provide a three-dimensional position including floor information and position information in the floor (e.g., store information in a shopping complex). CSR: Small: Collaborative Research: Enabling Cost-effective Cloud HPC Mladen Vouk ; Xiaosong Ma$311,998 by National Science Foundation
10/ 1/2013 - 09/30/2016
The proposed work examines novel services built on top of public cloud infrastructure to enable cost-effective high-performance computing. We will explore the on-demand, elastic, and configurable features of cloud computing to complement the traditional supercomputer/cluster platforms. If successful, this research will result in tools that adaptively aggregate, configure, and re-configure cloud resources for different HPC needs, with the purpose of offering low-cost R&D environments for scalable parallel applications.
TWC: Frontier: Collaborative: Rethinking Security in the Era of Cloud Computing
$749,996 by National Science Foundation 09/ 1/2013 - 08/31/2018 Increased use of cloud computing services is becoming a reality in today's IT management. The security risks of this move are active research topics, yielding cautionary examples of attacks enabled by the co-location of competing tenants. In this project, we propose to mitigate such risks through a new approach to cloud architecture defined by leveraging cloud providers as trusted (but auditable) security enablers. We will exploit cooperation between cloud providers and tenants in preventing attacks as a means to tackle long-standing open security problems, including protection of tenants against outsider attacks, improved intrusion detection and security diagnosis, and security-monitoring inlays. DO 2 Task 3.5 Laurie Williams$62,901 by LAS
09/13/2013 - 12/31/2014
EDU: Motivating and Reaching Students and Professionals with Software Security Education
Laurie Williams ; Emerson Murphy-Hill ; Kevin Oliver (Education)
$300,000 by National Science Foundation 09/ 1/2013 - 08/31/2015 According to a 2010 report that was based on the interviews from 2,800 Information Technology professionals worldwide, the gap between hacker threats and suitable security defenses is widening, and the types and numbers of threats are changing faster than ever before . In 2010, Jim Gosler, a fellow at the Sandia National Laboratory who works on countering attacks on U.S. networks, claimed that there are approximately 1,000 people in the country with the skills needed for cyber defense. Gosler went on to say that 20 to 30 times that many are needed. Additionally, the Chief Executive Officer (CEO) of the Mykonos Software security firm indicated that today's graduates in software engineering are unprepared to enter the workforce because they lack a solid understanding of how to make their applications secure. Particularly due to this shortage of security expertise, education of students and professionals already in the workforce is paramount. In this grant we provide a plan for motivating and providing software security education to students and professionals. NSA / North Carolina State University Science of Security Lablet: Analytics Supporting Security Science Laurie Williams ; Michael Rappa$2,475,248 by National Security Agency via US Army Research Office
10/ 1/2011 - 11/30/2014
North Carolina State University (NCSU), led by the Department of Computer Science and the Institute of Advanced Analytics in conjunction with the Institute for Next Generation IT Systems (ITng), will create and manage a Science of Security Lablet (SOSL) research organization on behalf of the National Security Agency (NSA). The SOSL research organization will conduct and coordinate basic science research projects focused on the Science of Security. SOSL will coordinate with related Lablet activities sponsored by NSA at Carnegie Mellon University and at University of Illinois at Urbana Champaign (UIUC). SOSL will also coordinate with the Security Science Virtual Organization at Vanderbilt University. The coordination will be in the form of workshops and technical exchanges.
Science of Security
Laurie Williams ; Michael Rappa (joint coll)
$2,117,740 by US Army - Army Research Office (National Security Agency) 06/25/2013 - 06/24/2015 Critical cyber systems must inspire trust and confidence, protect the privacy and integrity of data resources, and perform reliably. Therefore, a more scientific basis for the design and analysis of trusted systems is needed. In this proposal, we aim to progress the Science of Security. The Science of Security entails the development of a body of knowledge containing laws, axioms and provable theories relating to some aspect of system security. Security science should give us an understanding of the limits of what is possible in some security domain, by providing objective and quantifiable descriptions of security properties and behaviors. The notions embodied in security science should have broad applicability - transcending specific systems, attacks, and defensive mechanisms. A major goal is the creation of a unified body of knowledge that can serve as the basis of a trust engineering discipline, curriculum, and rigorous design methodologies. As such, we provide eight hard problems in the science of security. We also present representative projects which we feel will make progress in the discipline of the science of security. Growing The Science Of Security Through Analytics Laurie Williams ; Munindar Singh$2,255,237 by NSA (US Dept of Defense)
03/28/2014 - 03/27/2015
Since August 2011, North Carolina State University (NCSU) analytics-focused Science of Security Lablet (SOSL) has embraced and helped build a foundation for the NSA vision of the Science of Security (SoS) and a SoS community. Jointly with other SOSLs, we formulated five SoS hard problems, which lie at the core of the BAA. At NCSU, data-driven discovery and analytics have been used to formulate, validate, evolve, and solidify security models and theories as well as the practice of cyber-security. We propose to (1) investigate solutions to five cross-dependent hard problems, building on our extensive experience and research, including in the current SOSL; (2) advance our SoS community development activities; and (3) enhance our evaluation efforts regarding progress on the hard problems by bringing in experts on science evaluation.
HCC:Small:Collaborative Research:Integrating Cognitive and Computational Models of Narrative for Cinematic Generation
R. Michael Young
$352,696 by the National Science Foundation 08/ 1/2013 - 07/31/2016 Virtual cinematography, the use of a virtual camera within a three dimensional virtual world, is being used increasingly to communicate both in entertainment contexts as well as serious ones (e.g., training, education, news reporting). While many aspects of the use of virtual cameras are currently automated, the control of the camera is currently determined using either a pre-programmed script or a human operator controlling the camera at the time of filming. This project seeks to develop a model of virtual cinematography that is both computational -- providing a software system capable of generating camera control directives automatically -- and cognitive -- capable of modeling a viewer's understanding of an unfolding cinematic. ALICE: A Model for Sustaining Technology-Rich Adaptive Learning Spaces and Interactive Content Environments R. Michael Young$285,321 by Institute of Museum & Library Services
11/ 1/2013 - 10/31/2014
This proposal would fund the research and design phase of building the ALICE engine. The project will focus on creating a conceptual model for how to build an adaptive learning space; an architecture for the Artificial Intelligence (AI) engine and technology core; a series of "proof of concept" functional prototypes for collecting data and creating content; and a continuous assessment model for measuring the success of the AI as far as the quality of its engagement with the community and the success of the engine at enhancing a given technology-rich research and learning space.
LAS DO3 Task Order 2.4 Narrative Processing
R. Michael Young ; Christopher Healey
\$170,941 by LAS
03/31/2014 - 12/31/2014
DO3 Task Order 2.4 Narrative Processing
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1912120282649994, "perplexity": 4469.701108221429}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1412037663467.44/warc/CC-MAIN-20140930004103-00491-ip-10-234-18-248.ec2.internal.warc.gz"}
|
http://chemistry.stackexchange.com/questions/2493/is-combustion-considered-a-redox-reaction
|
# Is combustion considered a redox reaction?
When carbon combusts with oxygen, is this considered a redox reaction since the oxygen atoms gain electrons and the carbon atoms lose them?
-
In general, yes. If the reaction involves oxygen going from the oxidation state of zero in $\ce{O_2}$ to an oxidation state of -2, then there is oxidation and the reaction is a redox reaction.
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9619551301002502, "perplexity": 1344.587330796074}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464051342447.93/warc/CC-MAIN-20160524005542-00016-ip-10-185-217-139.ec2.internal.warc.gz"}
|
https://rdrr.io/cran/lmomco/man/cdftexp.html
|
cdftexp: Cumulative Distribution Function of the Truncated Exponential... In lmomco: L-Moments, Censored L-Moments, Trimmed L-Moments, L-Comoments, and Many Distributions
Description
This function computes the cumulative probability or nonexceedance probability of the Truncated Exponential distribution given parameters (ψ and α) computed by partexp. The parameter ψ is the right truncation of the distribution and α is a scale parameter. The cumulative distribution function, letting β = 1/α to match nomenclature of Vogel and others (2008), is
F(x) = \frac{1-\mathrm{exp}(-β{t})}{1-\mathrm{exp}(-βψ)}\mbox{,}
where F(x) is the nonexceedance probability for the quantile 0 ≤ x ≤ ψ and ψ > 0 and α > 0. This distribution represents a nonstationary Poisson process.
The distribution is restricted to a narrow range of L-CV (τ_2 = λ_2/λ_1). If τ_2 = 1/3, the process represented is a stationary Poisson for which the cumulative distribution function is simply the uniform distribution and F(x) = x/ψ. If τ_2 = 1/2, then the distribution is represented as the usual exponential distribution with a location parameter of zero and a rate parameter β (scale parameter α = 1/β). These two limiting conditions are supported.
Usage
1 cdftexp(x, para)
Arguments
x A real value vector. para The parameters from partexp or vec2par.
Value
Nonexceedance probability (F) for x.
W.H. Asquith
References
Vogel, R.M., Hosking, J.R.M., Elphick, C.S., Roberts, D.L., and Reed, J.M., 2008, Goodness of fit of probability distributions for sightings as species approach extinction: Bulletin of Mathematical Biology, DOI 10.1007/s11538-008-9377-3, 19 p.
pdftexp, quatexp, lmomtexp, partexp
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 cdftexp(50,partexp(vec2lmom(c(40,0.38), lscale=FALSE))) ## Not run: F <- seq(0,1,by=0.001) A <- partexp(vec2lmom(c(100, 1/2), lscale=FALSE)) x <- quatexp(F, A) plot(x, cdftexp(x, A), pch=16, type='l') by <- 0.01; lcvs <- c(1/3, seq(1/3+by, 1/2-by, by=by), 1/2) reds <- (lcvs - 1/3)/max(lcvs - 1/3) for(lcv in lcvs) { A <- partexp(vec2lmom(c(100, lcv), lscale=FALSE)) x <- quatexp(F, A) lines(x, cdftexp(x, A), pch=16, col=rgb(reds[lcvs == lcv],0,0)) } # Vogel and others (2008) example sighting times for the bird # Eskimo Curlew, inspection shows that these are fairly uniform. # There is a sighting about every year to two. T <- c(1946, 1947, 1948, 1950, 1955, 1956, 1959, 1960, 1961, 1962, 1963, 1964, 1968, 1970, 1972, 1973, 1974, 1976, 1977, 1980, 1981, 1982, 1982, 1983, 1985) R <- 1945 # beginning of record S <- T - R lmr <- lmoms(S) PARcurlew <- partexp(lmr) # read the warning message and then force the texp to the # stationary process model (min(tau_2) = 1/3). lmr$ratios[2] <- 1/3 lmr$lambdas[2] <- lmr$lambdas[1]*lmr$ratios[2] PARcurlew <- partexp(lmr) Xmax <- quatexp(1, PARcurlew) X <- seq(0,Xmax, by=.1) plot(X, cdftexp(X,PARcurlew), type="l") # or use the MVUE estimator TE <- max(S)*((length(S)+1)/length(S)) # Time of Extinction lines(X, punif(X, min=0, max=TE), col=2) ## End(Not run)
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8776788711547852, "perplexity": 4930.767446599699}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676590711.36/warc/CC-MAIN-20180719070814-20180719090814-00270.warc.gz"}
|
https://www.aimsciences.org/article/doi/10.3934/ipi.2015.9.511
|
# American Institute of Mathematical Sciences
May 2015, 9(2): 511-550. doi: 10.3934/ipi.2015.9.511
## A study of the one dimensional total generalised variation regularisation problem
1 Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Wilberforce Road, CB3 0WA, Cambridge, United Kingdom 2 Institute for Mathematics and Scientific Computing, University of Graz, Heinrichstrasse 36, A-8010, Graz, Austria
Received October 2013 Revised October 2014 Published March 2015
In this paper we study the one dimensional second order total generalised variation regularisation (TGV) problem with $L^{2}$ data fitting term. We examine the properties of this model and we calculate exact solutions using simple piecewise affine functions as data terms. We investigate how these solutions behave with respect to the TGV parameters and we verify our results using numerical experiments.
Citation: Konstantinos Papafitsoros, Kristian Bredies. A study of the one dimensional total generalised variation regularisation problem. Inverse Problems & Imaging, 2015, 9 (2) : 511-550. doi: 10.3934/ipi.2015.9.511
##### References:
[1] L. Ambrosio, N. Fusco and D. Pallara, Functions of Bounded Variation and Free Discontinuity Problems, Oxford University Press, USA, 2000. Google Scholar [2] H. Attouch and H. Brezis, Duality for the sum of convex functions in general Banach spaces, North-Holland Mathematical Library, 34 (1986), 125-133. doi: 10.1016/S0924-6509(09)70252-1. Google Scholar [3] P. Belhumeur, A binocular stereo algorithm for reconstructing sloping, creased, and broken surfaces in the presence of half-occlusion, in Proceedings of the Fourth International Conference on Computer Vision, 1993, 431-438. doi: 10.1109/ICCV.1993.378184. Google Scholar [4] M. Benning, Singular Regularization of Inverse Problems, Bregman Distances and Applications to Variational Frameworks with Singular Regularization Energies, Ph.D thesis, University of Münster, 2011. Google Scholar [5] M. Benning, C. Brune, M. Burger and J. Müller, Higher-order TV methods - Enhancement via Bregman iteration, Journal of Scientific Computing, 54 (2013), 269-310. doi: 10.1007/s10915-012-9650-3. Google Scholar [6] M. Benning and M. Burger, Ground states and singular vectors of convex variational regularization methods, Methods and Applications of Analysis, 20 (2013), 295-334. doi: 10.4310/MAA.2013.v20.n4.a1. Google Scholar [7] K. Bredies, Recovering piecewise smooth multichannel images by minimization of convex functionals with total generalized variation penalty, in Efficient Algorithms for Global Optimization Methods in Computer Vision, Lecture Notes in Computer Science, Springer Berlin Heidelberg, 2014, 44-77. doi: 10.1007/978-3-642-54774-4_3. Google Scholar [8] K. Bredies and M. Holler, Artifact-free decompression and zooming of JPEG compressed images with total generalized variation, in Computer Vision, Imaging and Computer Graphics - Theory and Application, Communications in Computer and Information Science, Springer Berlin Heidelberg, 2013, 242-258. doi: 10.1007/978-3-642-38241-3_16. Google Scholar [9] K. Bredies, K. Kunisch and T. Pock, Total generalized variation, SIAM Journal on Imaging Sciences, 3 (2010), 492-526. doi: 10.1137/090769521. Google Scholar [10] K. Bredies, K. Kunisch and T. Valkonen, Properties of $L^1$-$TGV^{2}$: The one-dimensional case, Journal of Mathematical Analysis and Applications, 398 (2013), 438-454. doi: 10.1016/j.jmaa.2012.08.053. Google Scholar [11] K. Bredies and T. Valkonen, Inverse problems with second-order total generalized variation constraints, in Proceedings of SampTA 2011 - 9th International Conference on Sampling Theory and Applications, Singapore, 2011. Google Scholar [12] V. Caselles, A. Chambolle and M. Novaga, The discontinuity set of solutions of the {TV} denoising problem and some extensions, Multiscale Modeling & Simulation, 6 (2007), 879-894. doi: 10.1137/070683003. Google Scholar [13] A. Chambolle and P. Lions, Image recovery via total variation minimization and related problems, Numerische Mathematik, 76 (1997), 167-188. doi: 10.1007/s002110050258. Google Scholar [14] A. Chambolle and T. Pock, A first-order primal-dual algorithm for convex problems with applications to imaging, Journal of Mathematical Imaging and Vision, 40 (2011), 120-145. doi: 10.1007/s10851-010-0251-1. Google Scholar [15] T. Chan and S. Esedoglu, Aspects of total variation regularized $L^1$ function approximation, SIAM Journal on Applied Mathematics, 65 (2005), 1817-1837. doi: 10.1137/040604297. Google Scholar [16] G. Dal Maso, I. Fonseca, G. Leoni and M. Morini, A higher order model for image restoration: The one dimensional case, SIAM Journal on Mathematical Analysis, 40 (2009), 2351-2391. doi: 10.1137/070697823. Google Scholar [17] F. Demengel, Fonctions à Hessien borné, Annales de l'Institut Fourier, 34 (1984), 155-190. doi: 10.5802/aif.969. Google Scholar [18] V. Duval, J. Aujol and Y. Gousseau, The TVL1 model: A geometric point of view, SIAM Journal on Multiscale Modeling and Simulation, 8 (2009), 154-189. doi: 10.1137/090757083. Google Scholar [19] I. Ekeland and R. Temam, Convex Analysis and Variational Problems, Vol. 1, North Holland, 1976. Google Scholar [20] M. Grasmair, The equivalence of the taut string algorithm and BV-regularization, Journal of Mathematical Imaging and Vision, 27 (2007), 59-66. doi: 10.1007/s10851-006-9796-4. Google Scholar [21] W. Hinterberger, M. Hintermüller, K. Kunisch, M. Von Oehsen and O. Scherzer, Tube methods for BV regularization, Journal of Mathematical Imaging and Vision, 19 (2003), 219-235. doi: 10.1023/A:1026276804745. Google Scholar [22] F. Knoll, K. Bredies, T. Pock and R. Stollberger, Second order total generalized variation (TGV) for MRI, Magnetic Resonance in Medicine, 65 (2011), 480-491. doi: 10.1002/mrm.22595. Google Scholar [23] Y. Meyer, Oscillating Patterns in Image Processing and Nonlinear Evolution Equations: The Fifteenth Dean Jacqueline B. Lewis Memorial Lectures, Vol. 22, American Mathematical Society, 2001. Google Scholar [24] J. Müller, Advanced Image Reconstruction and Denoising - Bregmanized (Higher Order) Total Variation and Application in PET, Ph.D thesis, University of Münster, 2013. Google Scholar [25] M. Nikolova, Minimizers of cost-functions involving nonsmooth data-fidelity terms. Application to the processing of outliers, SIAM Journal on Numerical Analysis, 40 (2002), 965-994. doi: 10.1137/S0036142901389165. Google Scholar [26] K. Papafitsoros, Novel Higher Order Regularisation Methods for Image Reconstruction, Ph.D thesis, University of Cambridge, 2014. Google Scholar [27] C. Pöschl and O. Scherzer, Exact solutions of one-dimensional total generalized variation, Communications in Mathematical Sciences, 13 (2015), 171-202. doi: 10.4310/CMS.2015.v13.n1.a9. Google Scholar [28] C. Pöschl and O. Scherzer, Characterization of minimizers of convex regularization functionals, Contemporary mathematics, 451 (2008), 219-248. doi: 10.1090/conm/451/08784. Google Scholar [29] W. Ring, Structural properties of solutions to total variation regularization problems, ESAIM: Mathematical Modelling and Numerical Analysis, 34 (2000), 799-810. doi: 10.1051/m2an:2000104. Google Scholar [30] L. Rudin, S. Osher and E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica D: Nonlinear Phenomena, 60 (1992), 259-268. doi: 10.1016/0167-2789(92)90242-F. Google Scholar [31] D. Strong and T. Chan, Edge-preserving and scale-dependent properties of total variation regularization, Inverse Problems, 19 (2003), S165-S187. doi: 10.1088/0266-5611/19/6/059. Google Scholar [32] R. Temam, Mathematical Problems in Plasticity, Vol. 12, Gauthier-Villars Paris, 1983. Google Scholar [33] T. Valkonen, K. Bredies and F. Knoll, Total generalized variation in diffusion tensor imaging, SIAM Journal on Imaging Sciences, 6 (2013), 487-525. doi: 10.1137/120867172. Google Scholar
show all references
##### References:
[1] L. Ambrosio, N. Fusco and D. Pallara, Functions of Bounded Variation and Free Discontinuity Problems, Oxford University Press, USA, 2000. Google Scholar [2] H. Attouch and H. Brezis, Duality for the sum of convex functions in general Banach spaces, North-Holland Mathematical Library, 34 (1986), 125-133. doi: 10.1016/S0924-6509(09)70252-1. Google Scholar [3] P. Belhumeur, A binocular stereo algorithm for reconstructing sloping, creased, and broken surfaces in the presence of half-occlusion, in Proceedings of the Fourth International Conference on Computer Vision, 1993, 431-438. doi: 10.1109/ICCV.1993.378184. Google Scholar [4] M. Benning, Singular Regularization of Inverse Problems, Bregman Distances and Applications to Variational Frameworks with Singular Regularization Energies, Ph.D thesis, University of Münster, 2011. Google Scholar [5] M. Benning, C. Brune, M. Burger and J. Müller, Higher-order TV methods - Enhancement via Bregman iteration, Journal of Scientific Computing, 54 (2013), 269-310. doi: 10.1007/s10915-012-9650-3. Google Scholar [6] M. Benning and M. Burger, Ground states and singular vectors of convex variational regularization methods, Methods and Applications of Analysis, 20 (2013), 295-334. doi: 10.4310/MAA.2013.v20.n4.a1. Google Scholar [7] K. Bredies, Recovering piecewise smooth multichannel images by minimization of convex functionals with total generalized variation penalty, in Efficient Algorithms for Global Optimization Methods in Computer Vision, Lecture Notes in Computer Science, Springer Berlin Heidelberg, 2014, 44-77. doi: 10.1007/978-3-642-54774-4_3. Google Scholar [8] K. Bredies and M. Holler, Artifact-free decompression and zooming of JPEG compressed images with total generalized variation, in Computer Vision, Imaging and Computer Graphics - Theory and Application, Communications in Computer and Information Science, Springer Berlin Heidelberg, 2013, 242-258. doi: 10.1007/978-3-642-38241-3_16. Google Scholar [9] K. Bredies, K. Kunisch and T. Pock, Total generalized variation, SIAM Journal on Imaging Sciences, 3 (2010), 492-526. doi: 10.1137/090769521. Google Scholar [10] K. Bredies, K. Kunisch and T. Valkonen, Properties of $L^1$-$TGV^{2}$: The one-dimensional case, Journal of Mathematical Analysis and Applications, 398 (2013), 438-454. doi: 10.1016/j.jmaa.2012.08.053. Google Scholar [11] K. Bredies and T. Valkonen, Inverse problems with second-order total generalized variation constraints, in Proceedings of SampTA 2011 - 9th International Conference on Sampling Theory and Applications, Singapore, 2011. Google Scholar [12] V. Caselles, A. Chambolle and M. Novaga, The discontinuity set of solutions of the {TV} denoising problem and some extensions, Multiscale Modeling & Simulation, 6 (2007), 879-894. doi: 10.1137/070683003. Google Scholar [13] A. Chambolle and P. Lions, Image recovery via total variation minimization and related problems, Numerische Mathematik, 76 (1997), 167-188. doi: 10.1007/s002110050258. Google Scholar [14] A. Chambolle and T. Pock, A first-order primal-dual algorithm for convex problems with applications to imaging, Journal of Mathematical Imaging and Vision, 40 (2011), 120-145. doi: 10.1007/s10851-010-0251-1. Google Scholar [15] T. Chan and S. Esedoglu, Aspects of total variation regularized $L^1$ function approximation, SIAM Journal on Applied Mathematics, 65 (2005), 1817-1837. doi: 10.1137/040604297. Google Scholar [16] G. Dal Maso, I. Fonseca, G. Leoni and M. Morini, A higher order model for image restoration: The one dimensional case, SIAM Journal on Mathematical Analysis, 40 (2009), 2351-2391. doi: 10.1137/070697823. Google Scholar [17] F. Demengel, Fonctions à Hessien borné, Annales de l'Institut Fourier, 34 (1984), 155-190. doi: 10.5802/aif.969. Google Scholar [18] V. Duval, J. Aujol and Y. Gousseau, The TVL1 model: A geometric point of view, SIAM Journal on Multiscale Modeling and Simulation, 8 (2009), 154-189. doi: 10.1137/090757083. Google Scholar [19] I. Ekeland and R. Temam, Convex Analysis and Variational Problems, Vol. 1, North Holland, 1976. Google Scholar [20] M. Grasmair, The equivalence of the taut string algorithm and BV-regularization, Journal of Mathematical Imaging and Vision, 27 (2007), 59-66. doi: 10.1007/s10851-006-9796-4. Google Scholar [21] W. Hinterberger, M. Hintermüller, K. Kunisch, M. Von Oehsen and O. Scherzer, Tube methods for BV regularization, Journal of Mathematical Imaging and Vision, 19 (2003), 219-235. doi: 10.1023/A:1026276804745. Google Scholar [22] F. Knoll, K. Bredies, T. Pock and R. Stollberger, Second order total generalized variation (TGV) for MRI, Magnetic Resonance in Medicine, 65 (2011), 480-491. doi: 10.1002/mrm.22595. Google Scholar [23] Y. Meyer, Oscillating Patterns in Image Processing and Nonlinear Evolution Equations: The Fifteenth Dean Jacqueline B. Lewis Memorial Lectures, Vol. 22, American Mathematical Society, 2001. Google Scholar [24] J. Müller, Advanced Image Reconstruction and Denoising - Bregmanized (Higher Order) Total Variation and Application in PET, Ph.D thesis, University of Münster, 2013. Google Scholar [25] M. Nikolova, Minimizers of cost-functions involving nonsmooth data-fidelity terms. Application to the processing of outliers, SIAM Journal on Numerical Analysis, 40 (2002), 965-994. doi: 10.1137/S0036142901389165. Google Scholar [26] K. Papafitsoros, Novel Higher Order Regularisation Methods for Image Reconstruction, Ph.D thesis, University of Cambridge, 2014. Google Scholar [27] C. Pöschl and O. Scherzer, Exact solutions of one-dimensional total generalized variation, Communications in Mathematical Sciences, 13 (2015), 171-202. doi: 10.4310/CMS.2015.v13.n1.a9. Google Scholar [28] C. Pöschl and O. Scherzer, Characterization of minimizers of convex regularization functionals, Contemporary mathematics, 451 (2008), 219-248. doi: 10.1090/conm/451/08784. Google Scholar [29] W. Ring, Structural properties of solutions to total variation regularization problems, ESAIM: Mathematical Modelling and Numerical Analysis, 34 (2000), 799-810. doi: 10.1051/m2an:2000104. Google Scholar [30] L. Rudin, S. Osher and E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica D: Nonlinear Phenomena, 60 (1992), 259-268. doi: 10.1016/0167-2789(92)90242-F. Google Scholar [31] D. Strong and T. Chan, Edge-preserving and scale-dependent properties of total variation regularization, Inverse Problems, 19 (2003), S165-S187. doi: 10.1088/0266-5611/19/6/059. Google Scholar [32] R. Temam, Mathematical Problems in Plasticity, Vol. 12, Gauthier-Villars Paris, 1983. Google Scholar [33] T. Valkonen, K. Bredies and F. Knoll, Total generalized variation in diffusion tensor imaging, SIAM Journal on Imaging Sciences, 6 (2013), 487-525. doi: 10.1137/120867172. Google Scholar
[1] Mujibur Rahman Chowdhury, Jun Zhang, Jing Qin, Yifei Lou. Poisson image denoising based on fractional-order total variation. Inverse Problems & Imaging, 2020, 14 (1) : 77-96. doi: 10.3934/ipi.2019064 [2] Denis R. Akhmetov, Renato Spigler. $L^1$-estimates for the higher-order derivatives of solutions to parabolic equations subject to initial values of bounded total variation. Communications on Pure & Applied Analysis, 2007, 6 (4) : 1051-1074. doi: 10.3934/cpaa.2007.6.1051 [3] Yunho Kim, Paul M. Thompson, Luminita A. Vese. HARDI data denoising using vectorial total variation and logarithmic barrier. Inverse Problems & Imaging, 2010, 4 (2) : 273-310. doi: 10.3934/ipi.2010.4.273 [4] Haijuan Hu, Jacques Froment, Baoyan Wang, Xiequan Fan. Spatial-Frequency domain nonlocal total variation for image denoising. Inverse Problems & Imaging, 2020, 14 (6) : 1157-1184. doi: 10.3934/ipi.2020059 [5] Rinaldo M. Colombo, Francesca Monti. Solutions with large total variation to nonconservative hyperbolic systems. Communications on Pure & Applied Analysis, 2010, 9 (1) : 47-60. doi: 10.3934/cpaa.2010.9.47 [6] Nils Dabrock, Yves van Gennip. A note on "Anisotropic total variation regularized $L^1$-approximation and denoising/deblurring of 2D bar codes". Inverse Problems & Imaging, 2018, 12 (2) : 525-526. doi: 10.3934/ipi.2018022 [7] Rustum Choksi, Yves van Gennip, Adam Oberman. Anisotropic total variation regularized $L^1$ approximation and denoising/deblurring of 2D bar codes. Inverse Problems & Imaging, 2011, 5 (3) : 591-617. doi: 10.3934/ipi.2011.5.591 [8] Yuan Shen, Lei Ji. Partial convolution for total variation deblurring and denoising by new linearized alternating direction method of multipliers with extension step. Journal of Industrial & Management Optimization, 2019, 15 (1) : 159-175. doi: 10.3934/jimo.2018037 [9] Raymond H. Chan, Haixia Liang, Suhua Wei, Mila Nikolova, Xue-Cheng Tai. High-order total variation regularization approach for axially symmetric object tomography from a single radiograph. Inverse Problems & Imaging, 2015, 9 (1) : 55-77. doi: 10.3934/ipi.2015.9.55 [10] Xiaoqun Zhang, Tony F. Chan. Wavelet inpainting by nonlocal total variation. Inverse Problems & Imaging, 2010, 4 (1) : 191-210. doi: 10.3934/ipi.2010.4.191 [11] Sandra Lucente. Large data solutions for semilinear higher order equations. Discrete & Continuous Dynamical Systems - S, 2020, 13 (12) : 3525-3533. doi: 10.3934/dcdss.2020247 [12] Xianchao Xiu, Ying Yang, Wanquan Liu, Lingchen Kong, Meijuan Shang. An improved total variation regularized RPCA for moving object detection with dynamic background. Journal of Industrial & Management Optimization, 2020, 16 (4) : 1685-1698. doi: 10.3934/jimo.2019024 [13] Sören Bartels, Marijo Milicevic. Iterative finite element solution of a constrained total variation regularized model problem. Discrete & Continuous Dynamical Systems - S, 2017, 10 (6) : 1207-1232. doi: 10.3934/dcdss.2017066 [14] J. Mead. $\chi^2$ test for total variation regularization parameter selection. Inverse Problems & Imaging, 2020, 14 (3) : 401-421. doi: 10.3934/ipi.2020019 [15] Wei Wang, Ling Pi, Michael K. Ng. Saturation-Value Total Variation model for chromatic aberration correction. Inverse Problems & Imaging, 2020, 14 (4) : 733-755. doi: 10.3934/ipi.2020034 [16] Yunhai Xiao, Junfeng Yang, Xiaoming Yuan. Alternating algorithms for total variation image reconstruction from random projections. Inverse Problems & Imaging, 2012, 6 (3) : 547-563. doi: 10.3934/ipi.2012.6.547 [17] Juan C. Moreno, V. B. Surya Prasath, João C. Neves. Color image processing by vectorial total variation with gradient channels coupling. Inverse Problems & Imaging, 2016, 10 (2) : 461-497. doi: 10.3934/ipi.2016008 [18] Liyan Ma, Lionel Moisan, Jian Yu, Tieyong Zeng. A stable method solving the total variation dictionary model with $L^\infty$ constraints. Inverse Problems & Imaging, 2014, 8 (2) : 507-535. doi: 10.3934/ipi.2014.8.507 [19] Lu Liu, Zhi-Feng Pang, Yuping Duan. Retinex based on exponent-type total variation scheme. Inverse Problems & Imaging, 2018, 12 (5) : 1199-1217. doi: 10.3934/ipi.2018050 [20] Leyu Hu, Wenxing Zhang, Xingju Cai, Deren Han. A parallel operator splitting algorithm for solving constrained total-variation retinex. Inverse Problems & Imaging, 2020, 14 (6) : 1135-1156. doi: 10.3934/ipi.2020058
2020 Impact Factor: 1.639
|
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8350321054458618, "perplexity": 6160.996556602454}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363135.71/warc/CC-MAIN-20211205035505-20211205065505-00431.warc.gz"}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.