content
stringlengths
228
999k
pred_label
stringclasses
1 value
pred_score
float64
0.5
1
Product Documentation Secure peering Which Citrix SD-WAN WANOP features required secure peering? You need to establish secure peering between Citrix SD-WAN WANOP appliances at two ends of the link when you intend to use any of the following features: • SSL compression • Signed CIFS support • Encrypted MAPI support Do I need to consider anything before configuring a secure tunnel? Yes. You must order and receive a crypto license before you can configure a secure tunnel between the Citrix SD-WAN WANOP appliances at to ends of the link. What happens when you enable secure peering on an appliance at one end of the link? When you enable secure peering on a Citrix SD-WAN WANOP appliance at one end of the link, the other appliance detects it and attempts to open an SSL signaling tunnel. If the two appliances successfully authenticate each other over this tunnel, the appliances have a secure peering relationship. All accelerated connections between the two appliances are encrypted, and compression is enabled. What happens when I do not enable secure peering on the partner appliance? When an appliance has secure peering enabled, connections with a partner for which it does not have a secure peer relationship are not encrypted or compressed, though TCP flow-control acceleration is still available. Compression is disabled to ensure that data stored in compression history from secured partners cannot be shared with unsecured partners. Why do I need a keystore password? You need a keystore password to access the security parameters. This password is different from the administrator’s password and allows security administration to be separated from other tasks. If the keystore password is reset, all existing encrypted data and private keys are lost. To protect data even if the appliance is stolen, the keystore password must be reentered every time the appliance is restarted. Until this is done, secure peering and compression are disabled. Does the Citrix SD-WAN WANOP appliance I received from Citrix contain keys and certificate to set up secure tunnel? No. Citrix SD-WAN WANOP products are shipped without the required keys and certificates for the SSL signaling tunnel. You must generate them yourself. Secure peering In this article
__label__pos
0.996116
Contents assignDetectionsToTracks Assign detections to tracks for multiobject tracking Syntax • [assignments,unassignedTracks,unassignedDetections] = assignDetectionsToTracks( costMatrix,costOfNonAssignment) example • [assignments,unassignedTracks,unassignedDetections] = assignDetectionsToTracks(costMatrix, unassignedTrackCost,unassignedDetectionCost) Description example [assignments,unassignedTracks,unassignedDetections] = assignDetectionsToTracks( costMatrix,costOfNonAssignment) assigns detections to tracks in the context of multiple object tracking using the James Munkres's variant of the Hungarian assignment algorithm. It also determines which tracks are missing and which detections should begin new tracks. It returns the indices of assigned and unassigned tracks, and unassigned detections. The costMatrix must be an M-by-N matrix. In this matrix, M represents the number of tracks, and N is the number of detections. Each value represents the cost of assigning the Nth detection to the Mth track. The lower the cost, the more likely that a detection gets assigned to a track. The costOfNonAssignment scalar input represents the cost of a track or a detection remaining unassigned. [assignments,unassignedTracks,unassignedDetections] = assignDetectionsToTracks(costMatrix, unassignedTrackCost,unassignedDetectionCost) specifies the cost of unassigned tracks and detections separately. The unassignedTrackCost must be a scalar value, or an M-element vector, where M represents the number of tracks. For the M-element vector, each element represents the cost of not assigning any detection to that track. The unassignedDetectionCost must be a scalar value or an N-element vector, where N represents the number of detections. Code Generation Support: Compile-time constant input: No restriction Supports MATLAB Function block: Yes Code Generation Support, Usage Notes, and Limitations Examples expand all Assign Detections to Tracks in a Single Video Frame This example shows you how the assignDetectionsToTracks function works for a single video frame. Set the predicted locations of objects in the current frame. Obtain predictions using the Kalman filter System object. predictions = [1,1; 2,2]; Set the locations of the objects detected in the current frame. For this example, there are 2 tracks and 3 new detections. Thus, at least one of the detections is unmatched, which can indicate a new track. detections = [1.1, 1.1; 2.1, 2.1; 1.5, 3]; Preallocate a cost matrix. cost = zeros(size(predictions,1),size(detections,1)); Compute the cost of each prediction matching a detection. The cost here, is defined as the Euclidean distance between the prediction and the detection. for i = 1:size(predictions, 1) diff = detections - repmat(predictions(i, :), [size(detections, 1), 1]); cost(i, :) = sqrt(sum(diff .^ 2, 2)); end Associate detections with predictions. Detection 1 should match to track 1, and detection 2 should match to track 2. Detection 3 should be unmatched. [assignment, unassignedTracks, unassignedDetections] = assignDetectionsToTracks(cost, 0.2); figure; plot(predictions(:, 1), predictions(:, 2), '*',detections(:, 1),detections(:, 2), 'ro'); hold on; legend('predictions', 'detections'); for i = 1:size(assignment, 1) text(predictions(assignment(i, 1), 1)+0.1,predictions(assignment(i, 1), 2)-0.1,num2str(i)); text(detections(assignment(i, 2), 1)+0.1,detections(assignment(i, 2), 2)-0.1, num2str(i)); end for i = 1:length(unassignedDetections) text(detections(unassignedDetections(i), 1)+0.1,detections(unassignedDetections(i), 2)+0.1, 'unassigned'); end xlim([0, 4]); ylim([0, 4]); Input Arguments expand all costMatrix — Cost of assigning detection to trackM-by-N matrix Cost of assigning a detection to a track, specified as an M-by-N matrix, where M represents the number of tracks, and N is the number of detections. The cost matrix value must be real and nonsparse. The lower the cost, the more likely that a detection gets assigned to a track. Each value represents the cost of assigning the Nth detection to the Mth track. If there is no likelihood of an assignment between a detection and a track, the costMatrix input is set to Inf. Internally, this function pads the cost matrix with dummy rows and columns to account for the possibility of unassigned tracks and detections. The padded rows represent detections not assigned to any tracks. The padded columns represent tracks not associated with any detections. The function applies the Hungarian assignment algorithm to the padded matrix. Data Types: int8 | uint8 | int16 | uint16 | int32 | uint32 | single | double costOfNonAssignment — Cost of not assigning detection to any track or track to any detectionscalar | finite Cost of not assigning detection to any track or track to detection. You can specify this value as a scalar value representing the cost of a track or a detection remaining unassigned. An unassigned detection may become the start of a new track. If a track is unassigned, the object does not appear. The higher the costOfNonAssignment value, the higher the likelihood that every track will be assigned a detection. Internally, this function pads the cost matrix with dummy rows and columns to account for the possibility of unassigned tracks and detections. The padded rows represent detections not assigned to any tracks. The padded columns represent tracks not associated with any detections. To apply the same value to all elements in both the rows and columns, use the syntax with the costOfNonAssignment input. To vary the values for different detections or tracks, use the syntax with the unassignedTrackCost and unassignedDetectionCost inputs. Data Types: int8 | uint8 | int16 | uint16 | int32 | uint32 | single | double unassignedTrackCost — Cost or likelihood of an unassigned trackM-element vector | scalar | finite Cost or likelihood of an unassigned track. You can specify this value as a scalar value, or an M-element vector, where M represents the number of tracks. For the M-element vector, each element represents the cost of not assigning any detection to that track. A scalar input represents the same cost of being unassigned for all tracks. The cost may vary depending on what you know about each track and the scene. For example, if an object is about to leave the field of view, the cost of the corresponding track being unassigned should be low. Internally, this function pads the cost matrix with dummy rows and columns to account for the possibility of unassigned tracks and detections. The padded rows represent detections not assigned to any tracks. The padded columns represent tracks not associated with any detections. To vary the values for different detections or tracks, use the syntax with the unassignedTrackCost and unassignedDetectionCost inputs. To apply the same value to all elements in both the rows and columns, use the syntax with the costOfNonAssignment input. Data Types: int8 | uint8 | int16 | uint16 | int32 | uint32 | single | double unassignedDetectionCost — Cost of unassigned detectionN-element vector | scalar | finite Cost of unassigned detection, specified as a scalar value or an N-element vector, where N represents the number of detections. For the N- element vector, each element represents the cost of starting a new track for that detection. A scalar input represents the same cost of being unassigned for all tracks. The cost may vary depending on what you know about each detection and the scene. For example, if a detection appears close to the edge of the image, it is more likely to be a new object. Internally, this function pads the cost matrix with dummy rows and columns to account for the possibility of unassigned tracks and detections. The padded rows represent detections not assigned to any tracks. The padded columns represent tracks not associated with any detections. To vary the values for different detections or tracks, use the syntax with the unassignedTrackCost and unassignedDetectionCost inputs. To apply the same value to all elements in both the rows and columns, use the syntax with the costOfNonAssignment input. Data Types: int8 | uint8 | int16 | uint16 | int32 | uint32 | single | double Output Arguments expand all assignments — Index pairs of tracks and corresponding detectionsL-by-2 matrix Index pairs of tracks and corresponding detections. This value is returned as an L-by-2 matrix of index pairs, with L number of pairs. The first column represents the track index and the second column represents the detection index. Data Types: uint32 unassignedTracks — Unassigned tracksP-element vector Unassigned tracks, returned as a P-element vector. P represents the number of unassigned tracks. Each element represents a track to which no detections are assigned. Data Types: uint32 unassignedDetections — Unassigned detectionsQ-element vector Unassigned detections, returned as a Q-element vector, where Q represents the number of unassigned detections. Each element represents a detection that was not assigned to any tracks. These detections can begin new tracks. Data Types: uint32 References [1] Miller, Matt L., Harold S. Stone, and Ingemar J. Cox, "Optimizing Murty's Ranked Assignment Method," IEEE Transactions on Aerospace and Electronic Systems, 33(3), 1997. [2] Munkres, James, "Algorithms for Assignment and Transportation Problems," Journal of the Society for Industrial and Applied Mathematics, Volume 5, Number 1, March, 1957. Was this topic helpful?
__label__pos
0.743415
Sending text read from GUI to a method Discussion in 'Java' started by gaff, Sep 12, 2006. 1. gaff gaff Guest Hey, Have a problem sending a String from a GUI to a method. The method definitely works as I can use with text that I set in code. track.deleteFromList("some object name"); // removes object called "some object name" from the list that the track object stores. When I try to use a text box to read in the text it doesn't work. The String is read in correctly, checked. The String is read by method correctly, checked. It should then make a comparison with each place in the list to see if it's the correct identifier (Strnig is identifier/name of the object to be deleted) but doesn't seem to realise when there is a match and therefore should execute remove form list. I suspect that the String has to be declared in some special way before it is passed to the method? as it is not of definite value when the prog starts running. It's read and fed properly, but because it's dynamic it's not being compared correctly? Relevant code is below, thanks very much for your help. t1: text box, b1: button FROM MAIN, everything else works, track is instance of Tracker2 class /////////////////////////////////////////////////////////////////////// b2.addSelectionListener(new SelectionAdapter(){ public void widgetSelected(SelectionEvent e) { String inny = t1.getText(); track.deleteFromList(inny); } }); /////////////////////////////////////////////////////////////////////// import java.util.*; public class Tracker2 { private List<Array_test> list_of_test_objects; //construct public Tracker2(){ list_of_test_objects = new ArrayList<Array_test>(); } ........ ....... ....... ....... // delete from list given a certain identifier public void deleteFromList(String old_array_test_object) { int i = 0; while (i < list_of_test_objects.size()) { if ((list_of_test_objects.get(i).getIdentifier()) == old_array_test_object) { list_of_test_objects.remove(i); i = list_of_test_objects.size(); } i++; } } ....... ....... ....... .......   gaff, Sep 12, 2006 #1 1. Advertisements 2. gaff wrote: > Hey, > > Have a problem sending a String from a GUI to a method. > > The method definitely works as I can use with text that I set in code. > > track.deleteFromList("some object name"); // removes object called > "some object name" from the list that the track object stores. > > When I try to use a text box to read in the text it doesn't work. > [snip] > > // delete from list given a certain identifier > public void deleteFromList(String old_array_test_object) { > int i = 0; > while (i < list_of_test_objects.size()) { > if ((list_of_test_objects.get(i).getIdentifier()) == > old_array_test_object) { Never use == to compare two strings. Use the equals method. The expression (list_of_test_objects.get(i).getIdentifier()).equals(old_array_test_object) should work BK   Babu Kalakrishnan, Sep 12, 2006 #2 1. Advertisements 3. gaff gaff Guest Cheers buddy, that was it.   gaff, Sep 12, 2006 #3 4. gaff Lew Guest gaff wrote: > int i = 0; > while (i < list_of_test_objects.size()) { > if ((list_of_test_objects.get(i).getIdentifier()) == > old_array_test_object) { > list_of_test_objects.remove(i); > i = list_of_test_objects.size(); > } > i++; > } Just a suggestion: you might find ending the loop with break; instead of i = list_of_test_objects.size(); marginally more efficient than the call to size() and assignment to i followed by the increment and test of i. Speaking strictly for myself, I prefer the idiom for( int i=0; i < loto.size(); i++) {...} preferable to int i=0; while( i < loto.size() ) { ...; i++ } for compactness and because it limits the scope if i. Actually, I prefer for ( X someX : listOfX ) { ... } // X is base type of list to either of the above, because it avoids explicit declaration of all loop-control variables, and it is just so purty. -Lew   Lew, Sep 14, 2006 #4 1. Advertisements Want to reply to this thread or ask your own question? It takes just 2 minutes to sign up (and it's free!). Just click the sign up button to choose a username and then you can ask your own questions on the forum. Similar Threads 1. Replies: 0 Views: 1,565 2. Simon Rigby Replies: 5 Views: 787 Cowboy \(Gregory A. Beamer\) Oct 7, 2006 3. Luke Replies: 2 Views: 4,877 Nigel Wade Mar 15, 2007 4. Ron Coutts Replies: 0 Views: 136 Ron Coutts Nov 11, 2003 5. Trans Replies: 3 Views: 201 Trans Dec 21, 2005 Loading... Share This Page
__label__pos
0.716248
F_Kellner Addin Combo Box Discussion created by F_Kellner on Nov 5, 2013 Latest reply on Jun 13, 2014 by johnmdye I have been struggling to get the combo box addin working properly on toolbars that I have created. So in attempt to better understand the combo box I have been trying to implement a combo box using the example from the documentation. The documentation states "this topic examines the process of creating a combo box with the list of layers from the table of contents. Selecting a layer will create a fishnet covering the full extent of the layer". And I have used the code example in the documentation # Business logic to implement the ComboBox def __init__(self):         self.editable = True         self.enabled = True def onSelChange(self, selection): # When a new layer is selected, create a new fishnet using the extent of layer. layer = arcpy.mapping.ListLayers(self.mxd, selection)[0] desc = arcpy.Describe(layer.dataSource) extent = desc.Extent fishnet = arcpy.CreateFishnet_management(r'in_memory\fishnet',                  '%f %f' %(extent.XMin, extent.YMin),                  '%f %f' %(extent.XMin, extent.YMax),                  0, 0, 10, 10,                  '%f %f' %(extent.XMax, extent.YMax),                  'NO_LABELS',                  '%f %f %f %f' %(extent.XMin, extent.YMin, extent.XMax, extent.YMax), 'POLYGON') arcpy.RefreshActiveView() def onFocus(self, focused):   # When the combo box has focus, update the combo box with the list of layer names. if focused:   self.mxd = arcpy.mapping.MapDocument('current')   layers = arcpy.mapping.ListLayers(self.mxd)   self.items = []   for layer in layers:    self.items.append(layer.name) But I have been unable to get the toolbar to work. The toolbar shows up in arcmap but there is no drop down and I get the missing message on the toolbar itself. Any suggestions? Thanks Fred Outcomes
__label__pos
0.535625
blob: cee60f94503676ab21217726ea99dc5c880bb1c3 [file] [log] [blame] // SPDX-License-Identifier: GPL-2.0-only /* * drivers/media/i2c/tvp514x.c * * TI TVP5146/47 decoder driver * * Copyright (C) 2008 Texas Instruments Inc * Author: Vaibhav Hiremath <[email protected]> * * Contributors: * Sivaraj R <[email protected]> * Brijesh R Jadav <[email protected]> * Hardik Shah <[email protected]> * Manjunath Hadli <[email protected]> * Karicheri Muralidharan <[email protected]> * Prabhakar Lad <[email protected]> */ #include <linux/i2c.h> #include <linux/slab.h> #include <linux/delay.h> #include <linux/videodev2.h> #include <linux/module.h> #include <linux/v4l2-mediabus.h> #include <linux/of.h> #include <linux/of_graph.h> #include <media/v4l2-async.h> #include <media/v4l2-device.h> #include <media/v4l2-common.h> #include <media/v4l2-mediabus.h> #include <media/v4l2-fwnode.h> #include <media/v4l2-ctrls.h> #include <media/i2c/tvp514x.h> #include <media/media-entity.h> #include "tvp514x_regs.h" /* Private macros for TVP */ #define I2C_RETRY_COUNT (5) #define LOCK_RETRY_COUNT (5) #define LOCK_RETRY_DELAY (200) /* Debug functions */ static bool debug; module_param(debug, bool, 0644); MODULE_PARM_DESC(debug, "Debug level (0-1)"); MODULE_AUTHOR("Texas Instruments"); MODULE_DESCRIPTION("TVP514X linux decoder driver"); MODULE_LICENSE("GPL"); /* enum tvp514x_std - enum for supported standards */ enum tvp514x_std { STD_NTSC_MJ = 0, STD_PAL_BDGHIN, STD_INVALID }; /** * struct tvp514x_std_info - Structure to store standard information * @width: Line width in pixels * @height:Number of active lines * @video_std: Value to write in REG_VIDEO_STD register * @standard: v4l2 standard structure information */ struct tvp514x_std_info { unsigned long width; unsigned long height; u8 video_std; struct v4l2_standard standard; }; static struct tvp514x_reg tvp514x_reg_list_default[0x40]; static int tvp514x_s_stream(struct v4l2_subdev *sd, int enable); /** * struct tvp514x_decoder - TVP5146/47 decoder object * @sd: Subdevice Slave handle * @hdl: embedded &struct v4l2_ctrl_handler * @tvp514x_regs: copy of hw's regs with preset values. * @pdata: Board specific * @ver: Chip version * @streaming: TVP5146/47 decoder streaming - enabled or disabled. * @pix: Current pixel format * @num_fmts: Number of formats * @fmt_list: Format list * @current_std: Current standard * @num_stds: Number of standards * @std_list: Standards list * @input: Input routing at chip level * @output: Output routing at chip level * @pad: subdev media pad associated with the decoder * @format: media bus frame format * @int_seq: driver's register init sequence */ struct tvp514x_decoder { struct v4l2_subdev sd; struct v4l2_ctrl_handler hdl; struct tvp514x_reg tvp514x_regs[ARRAY_SIZE(tvp514x_reg_list_default)]; const struct tvp514x_platform_data *pdata; int ver; int streaming; struct v4l2_pix_format pix; int num_fmts; const struct v4l2_fmtdesc *fmt_list; enum tvp514x_std current_std; int num_stds; const struct tvp514x_std_info *std_list; /* Input and Output Routing parameters */ u32 input; u32 output; /* mc related members */ struct media_pad pad; struct v4l2_mbus_framefmt format; struct tvp514x_reg *int_seq; }; /* TVP514x default register values */ static struct tvp514x_reg tvp514x_reg_list_default[] = { /* Composite selected */ {TOK_WRITE, REG_INPUT_SEL, 0x05}, {TOK_WRITE, REG_AFE_GAIN_CTRL, 0x0F}, /* Auto mode */ {TOK_WRITE, REG_VIDEO_STD, 0x00}, {TOK_WRITE, REG_OPERATION_MODE, 0x00}, {TOK_SKIP, REG_AUTOSWITCH_MASK, 0x3F}, {TOK_WRITE, REG_COLOR_KILLER, 0x10}, {TOK_WRITE, REG_LUMA_CONTROL1, 0x00}, {TOK_WRITE, REG_LUMA_CONTROL2, 0x00}, {TOK_WRITE, REG_LUMA_CONTROL3, 0x02}, {TOK_WRITE, REG_BRIGHTNESS, 0x80}, {TOK_WRITE, REG_CONTRAST, 0x80}, {TOK_WRITE, REG_SATURATION, 0x80}, {TOK_WRITE, REG_HUE, 0x00}, {TOK_WRITE, REG_CHROMA_CONTROL1, 0x00}, {TOK_WRITE, REG_CHROMA_CONTROL2, 0x0E}, /* Reserved */ {TOK_SKIP, 0x0F, 0x00}, {TOK_WRITE, REG_COMP_PR_SATURATION, 0x80}, {TOK_WRITE, REG_COMP_Y_CONTRAST, 0x80}, {TOK_WRITE, REG_COMP_PB_SATURATION, 0x80}, /* Reserved */ {TOK_SKIP, 0x13, 0x00}, {TOK_WRITE, REG_COMP_Y_BRIGHTNESS, 0x80}, /* Reserved */ {TOK_SKIP, 0x15, 0x00}, /* NTSC timing */ {TOK_SKIP, REG_AVID_START_PIXEL_LSB, 0x55}, {TOK_SKIP, REG_AVID_START_PIXEL_MSB, 0x00}, {TOK_SKIP, REG_AVID_STOP_PIXEL_LSB, 0x25}, {TOK_SKIP, REG_AVID_STOP_PIXEL_MSB, 0x03}, /* NTSC timing */ {TOK_SKIP, REG_HSYNC_START_PIXEL_LSB, 0x00}, {TOK_SKIP, REG_HSYNC_START_PIXEL_MSB, 0x00}, {TOK_SKIP, REG_HSYNC_STOP_PIXEL_LSB, 0x40}, {TOK_SKIP, REG_HSYNC_STOP_PIXEL_MSB, 0x00}, /* NTSC timing */ {TOK_SKIP, REG_VSYNC_START_LINE_LSB, 0x04}, {TOK_SKIP, REG_VSYNC_START_LINE_MSB, 0x00}, {TOK_SKIP, REG_VSYNC_STOP_LINE_LSB, 0x07}, {TOK_SKIP, REG_VSYNC_STOP_LINE_MSB, 0x00}, /* NTSC timing */ {TOK_SKIP, REG_VBLK_START_LINE_LSB, 0x01}, {TOK_SKIP, REG_VBLK_START_LINE_MSB, 0x00}, {TOK_SKIP, REG_VBLK_STOP_LINE_LSB, 0x15}, {TOK_SKIP, REG_VBLK_STOP_LINE_MSB, 0x00}, /* Reserved */ {TOK_SKIP, 0x26, 0x00}, /* Reserved */ {TOK_SKIP, 0x27, 0x00}, {TOK_SKIP, REG_FAST_SWTICH_CONTROL, 0xCC}, /* Reserved */ {TOK_SKIP, 0x29, 0x00}, {TOK_SKIP, REG_FAST_SWTICH_SCART_DELAY, 0x00}, /* Reserved */ {TOK_SKIP, 0x2B, 0x00}, {TOK_SKIP, REG_SCART_DELAY, 0x00}, {TOK_SKIP, REG_CTI_DELAY, 0x00}, {TOK_SKIP, REG_CTI_CONTROL, 0x00}, /* Reserved */ {TOK_SKIP, 0x2F, 0x00}, /* Reserved */ {TOK_SKIP, 0x30, 0x00}, /* Reserved */ {TOK_SKIP, 0x31, 0x00}, /* HS, VS active high */ {TOK_WRITE, REG_SYNC_CONTROL, 0x00}, /* 10-bit BT.656 */ {TOK_WRITE, REG_OUTPUT_FORMATTER1, 0x00}, /* Enable clk & data */ {TOK_WRITE, REG_OUTPUT_FORMATTER2, 0x11}, /* Enable AVID & FLD */ {TOK_WRITE, REG_OUTPUT_FORMATTER3, 0xEE}, /* Enable VS & HS */ {TOK_WRITE, REG_OUTPUT_FORMATTER4, 0xAF}, {TOK_WRITE, REG_OUTPUT_FORMATTER5, 0xFF}, {TOK_WRITE, REG_OUTPUT_FORMATTER6, 0xFF}, /* Clear status */ {TOK_WRITE, REG_CLEAR_LOST_LOCK, 0x01}, {TOK_TERM, 0, 0}, }; /* * List of image formats supported by TVP5146/47 decoder * Currently we are using 8 bit mode only, but can be * extended to 10/20 bit mode. */ static const struct v4l2_fmtdesc tvp514x_fmt_list[] = { { .index = 0, .type = V4L2_BUF_TYPE_VIDEO_CAPTURE, .flags = 0, .description = "8-bit UYVY 4:2:2 Format", .pixelformat = V4L2_PIX_FMT_UYVY, }, }; /* * Supported standards - * * Currently supports two standards only, need to add support for rest of the * modes, like SECAM, etc... */ static const struct tvp514x_std_info tvp514x_std_list[] = { /* Standard: STD_NTSC_MJ */ [STD_NTSC_MJ] = { .width = NTSC_NUM_ACTIVE_PIXELS, .height = NTSC_NUM_ACTIVE_LINES, .video_std = VIDEO_STD_NTSC_MJ_BIT, .standard = { .index = 0, .id = V4L2_STD_NTSC, .name = "NTSC", .frameperiod = {1001, 30000}, .framelines = 525 }, /* Standard: STD_PAL_BDGHIN */ }, [STD_PAL_BDGHIN] = { .width = PAL_NUM_ACTIVE_PIXELS, .height = PAL_NUM_ACTIVE_LINES, .video_std = VIDEO_STD_PAL_BDGHIN_BIT, .standard = { .index = 1, .id = V4L2_STD_PAL, .name = "PAL", .frameperiod = {1, 25}, .framelines = 625 }, }, /* Standard: need to add for additional standard */ }; static inline struct tvp514x_decoder *to_decoder(struct v4l2_subdev *sd) { return container_of(sd, struct tvp514x_decoder, sd); } static inline struct v4l2_subdev *to_sd(struct v4l2_ctrl *ctrl) { return &container_of(ctrl->handler, struct tvp514x_decoder, hdl)->sd; } /** * tvp514x_read_reg() - Read a value from a register in an TVP5146/47. * @sd: ptr to v4l2_subdev struct * @reg: TVP5146/47 register address * * Returns value read if successful, or non-zero (-1) otherwise. */ static int tvp514x_read_reg(struct v4l2_subdev *sd, u8 reg) { int err, retry = 0; struct i2c_client *client = v4l2_get_subdevdata(sd); read_again: err = i2c_smbus_read_byte_data(client, reg); if (err < 0) { if (retry <= I2C_RETRY_COUNT) { v4l2_warn(sd, "Read: retry ... %d\n", retry); retry++; msleep_interruptible(10); goto read_again; } } return err; } /** * dump_reg() - dump the register content of TVP5146/47. * @sd: ptr to v4l2_subdev struct * @reg: TVP5146/47 register address */ static void dump_reg(struct v4l2_subdev *sd, u8 reg) { u32 val; val = tvp514x_read_reg(sd, reg); v4l2_info(sd, "Reg(0x%.2X): 0x%.2X\n", reg, val); } /** * tvp514x_write_reg() - Write a value to a register in TVP5146/47 * @sd: ptr to v4l2_subdev struct * @reg: TVP5146/47 register address * @val: value to be written to the register * * Write a value to a register in an TVP5146/47 decoder device. * Returns zero if successful, or non-zero otherwise. */ static int tvp514x_write_reg(struct v4l2_subdev *sd, u8 reg, u8 val) { int err, retry = 0; struct i2c_client *client = v4l2_get_subdevdata(sd); write_again: err = i2c_smbus_write_byte_data(client, reg, val); if (err) { if (retry <= I2C_RETRY_COUNT) { v4l2_warn(sd, "Write: retry ... %d\n", retry); retry++; msleep_interruptible(10); goto write_again; } } return err; } /** * tvp514x_write_regs() : Initializes a list of TVP5146/47 registers * @sd: ptr to v4l2_subdev struct * @reglist: list of TVP5146/47 registers and values * * Initializes a list of TVP5146/47 registers:- * if token is TOK_TERM, then entire write operation terminates * if token is TOK_DELAY, then a delay of 'val' msec is introduced * if token is TOK_SKIP, then the register write is skipped * if token is TOK_WRITE, then the register write is performed * Returns zero if successful, or non-zero otherwise. */ static int tvp514x_write_regs(struct v4l2_subdev *sd, const struct tvp514x_reg reglist[]) { int err; const struct tvp514x_reg *next = reglist; for (; next->token != TOK_TERM; next++) { if (next->token == TOK_DELAY) { msleep(next->val); continue; } if (next->token == TOK_SKIP) continue; err = tvp514x_write_reg(sd, next->reg, (u8) next->val); if (err) { v4l2_err(sd, "Write failed. Err[%d]\n", err); return err; } } return 0; } /** * tvp514x_query_current_std() : Query the current standard detected by TVP5146/47 * @sd: ptr to v4l2_subdev struct * * Returns the current standard detected by TVP5146/47, STD_INVALID if there is no * standard detected. */ static enum tvp514x_std tvp514x_query_current_std(struct v4l2_subdev *sd) { u8 std, std_status; std = tvp514x_read_reg(sd, REG_VIDEO_STD); if ((std & VIDEO_STD_MASK) == VIDEO_STD_AUTO_SWITCH_BIT) /* use the standard status register */ std_status = tvp514x_read_reg(sd, REG_VIDEO_STD_STATUS); else /* use the standard register itself */ std_status = std; switch (std_status & VIDEO_STD_MASK) { case VIDEO_STD_NTSC_MJ_BIT: return STD_NTSC_MJ; case VIDEO_STD_PAL_BDGHIN_BIT: return STD_PAL_BDGHIN; default: return STD_INVALID; } return STD_INVALID; } /* TVP5146/47 register dump function */ static void tvp514x_reg_dump(struct v4l2_subdev *sd) { dump_reg(sd, REG_INPUT_SEL); dump_reg(sd, REG_AFE_GAIN_CTRL); dump_reg(sd, REG_VIDEO_STD); dump_reg(sd, REG_OPERATION_MODE); dump_reg(sd, REG_COLOR_KILLER); dump_reg(sd, REG_LUMA_CONTROL1); dump_reg(sd, REG_LUMA_CONTROL2); dump_reg(sd, REG_LUMA_CONTROL3); dump_reg(sd, REG_BRIGHTNESS); dump_reg(sd, REG_CONTRAST); dump_reg(sd, REG_SATURATION); dump_reg(sd, REG_HUE); dump_reg(sd, REG_CHROMA_CONTROL1); dump_reg(sd, REG_CHROMA_CONTROL2); dump_reg(sd, REG_COMP_PR_SATURATION); dump_reg(sd, REG_COMP_Y_CONTRAST); dump_reg(sd, REG_COMP_PB_SATURATION); dump_reg(sd, REG_COMP_Y_BRIGHTNESS); dump_reg(sd, REG_AVID_START_PIXEL_LSB); dump_reg(sd, REG_AVID_START_PIXEL_MSB); dump_reg(sd, REG_AVID_STOP_PIXEL_LSB); dump_reg(sd, REG_AVID_STOP_PIXEL_MSB); dump_reg(sd, REG_HSYNC_START_PIXEL_LSB); dump_reg(sd, REG_HSYNC_START_PIXEL_MSB); dump_reg(sd, REG_HSYNC_STOP_PIXEL_LSB); dump_reg(sd, REG_HSYNC_STOP_PIXEL_MSB); dump_reg(sd, REG_VSYNC_START_LINE_LSB); dump_reg(sd, REG_VSYNC_START_LINE_MSB); dump_reg(sd, REG_VSYNC_STOP_LINE_LSB); dump_reg(sd, REG_VSYNC_STOP_LINE_MSB); dump_reg(sd, REG_VBLK_START_LINE_LSB); dump_reg(sd, REG_VBLK_START_LINE_MSB); dump_reg(sd, REG_VBLK_STOP_LINE_LSB); dump_reg(sd, REG_VBLK_STOP_LINE_MSB); dump_reg(sd, REG_SYNC_CONTROL); dump_reg(sd, REG_OUTPUT_FORMATTER1); dump_reg(sd, REG_OUTPUT_FORMATTER2); dump_reg(sd, REG_OUTPUT_FORMATTER3); dump_reg(sd, REG_OUTPUT_FORMATTER4); dump_reg(sd, REG_OUTPUT_FORMATTER5); dump_reg(sd, REG_OUTPUT_FORMATTER6); dump_reg(sd, REG_CLEAR_LOST_LOCK); } /** * tvp514x_configure() - Configure the TVP5146/47 registers * @sd: ptr to v4l2_subdev struct * @decoder: ptr to tvp514x_decoder structure * * Returns zero if successful, or non-zero otherwise. */ static int tvp514x_configure(struct v4l2_subdev *sd, struct tvp514x_decoder *decoder) { int err; /* common register initialization */ err = tvp514x_write_regs(sd, decoder->tvp514x_regs); if (err) return err; if (debug) tvp514x_reg_dump(sd); return 0; } /** * tvp514x_detect() - Detect if an tvp514x is present, and if so which revision. * @sd: pointer to standard V4L2 sub-device structure * @decoder: pointer to tvp514x_decoder structure * * A device is considered to be detected if the chip ID (LSB and MSB) * registers match the expected values. * Any value of the rom version register is accepted. * Returns ENODEV error number if no device is detected, or zero * if a device is detected. */ static int tvp514x_detect(struct v4l2_subdev *sd, struct tvp514x_decoder *decoder) { u8 chip_id_msb, chip_id_lsb, rom_ver; struct i2c_client *client = v4l2_get_subdevdata(sd); chip_id_msb = tvp514x_read_reg(sd, REG_CHIP_ID_MSB); chip_id_lsb = tvp514x_read_reg(sd, REG_CHIP_ID_LSB); rom_ver = tvp514x_read_reg(sd, REG_ROM_VERSION); v4l2_dbg(1, debug, sd, "chip id detected msb:0x%x lsb:0x%x rom version:0x%x\n", chip_id_msb, chip_id_lsb, rom_ver); if ((chip_id_msb != TVP514X_CHIP_ID_MSB) || ((chip_id_lsb != TVP5146_CHIP_ID_LSB) && (chip_id_lsb != TVP5147_CHIP_ID_LSB))) { /* We didn't read the values we expected, so this must not be * an TVP5146/47. */ v4l2_err(sd, "chip id mismatch msb:0x%x lsb:0x%x\n", chip_id_msb, chip_id_lsb); return -ENODEV; } decoder->ver = rom_ver; v4l2_info(sd, "%s (Version - 0x%.2x) found at 0x%x (%s)\n", client->name, decoder->ver, client->addr << 1, client->adapter->name); return 0; } /** * tvp514x_querystd() - V4L2 decoder interface handler for querystd * @sd: pointer to standard V4L2 sub-device structure * @std_id: standard V4L2 std_id ioctl enum * * Returns the current standard detected by TVP5146/47. If no active input is * detected then *std_id is set to 0 and the function returns 0. */ static int tvp514x_querystd(struct v4l2_subdev *sd, v4l2_std_id *std_id) { struct tvp514x_decoder *decoder = to_decoder(sd); enum tvp514x_std current_std; enum tvp514x_input input_sel; u8 sync_lock_status, lock_mask; if (std_id == NULL) return -EINVAL; /* To query the standard the TVP514x must power on the ADCs. */ if (!decoder->streaming) { tvp514x_s_stream(sd, 1); msleep(LOCK_RETRY_DELAY); } /* query the current standard */ current_std = tvp514x_query_current_std(sd); if (current_std == STD_INVALID) { *std_id = V4L2_STD_UNKNOWN; return 0; } input_sel = decoder->input; switch (input_sel) { case INPUT_CVBS_VI1A: case INPUT_CVBS_VI1B: case INPUT_CVBS_VI1C: case INPUT_CVBS_VI2A: case INPUT_CVBS_VI2B: case INPUT_CVBS_VI2C: case INPUT_CVBS_VI3A: case INPUT_CVBS_VI3B: case INPUT_CVBS_VI3C: case INPUT_CVBS_VI4A: lock_mask = STATUS_CLR_SUBCAR_LOCK_BIT | STATUS_HORZ_SYNC_LOCK_BIT | STATUS_VIRT_SYNC_LOCK_BIT; break; case INPUT_SVIDEO_VI2A_VI1A: case INPUT_SVIDEO_VI2B_VI1B: case INPUT_SVIDEO_VI2C_VI1C: case INPUT_SVIDEO_VI2A_VI3A: case INPUT_SVIDEO_VI2B_VI3B: case INPUT_SVIDEO_VI2C_VI3C: case INPUT_SVIDEO_VI4A_VI1A: case INPUT_SVIDEO_VI4A_VI1B: case INPUT_SVIDEO_VI4A_VI1C: case INPUT_SVIDEO_VI4A_VI3A: case INPUT_SVIDEO_VI4A_VI3B: case INPUT_SVIDEO_VI4A_VI3C: lock_mask = STATUS_HORZ_SYNC_LOCK_BIT | STATUS_VIRT_SYNC_LOCK_BIT; break; /*Need to add other interfaces*/ default: return -EINVAL; } /* check whether signal is locked */ sync_lock_status = tvp514x_read_reg(sd, REG_STATUS1); if (lock_mask != (sync_lock_status & lock_mask)) { *std_id = V4L2_STD_UNKNOWN; return 0; /* No input detected */ } *std_id &= decoder->std_list[current_std].standard.id; v4l2_dbg(1, debug, sd, "Current STD: %s\n", decoder->std_list[current_std].standard.name); return 0; } /** * tvp514x_s_std() - V4L2 decoder interface handler for s_std * @sd: pointer to standard V4L2 sub-device structure * @std_id: standard V4L2 v4l2_std_id ioctl enum * * If std_id is supported, sets the requested standard. Otherwise, returns * -EINVAL */ static int tvp514x_s_std(struct v4l2_subdev *sd, v4l2_std_id std_id) { struct tvp514x_decoder *decoder = to_decoder(sd); int err, i; for (i = 0; i < decoder->num_stds; i++) if (std_id & decoder->std_list[i].standard.id) break; if ((i == decoder->num_stds) || (i == STD_INVALID)) return -EINVAL; err = tvp514x_write_reg(sd, REG_VIDEO_STD, decoder->std_list[i].video_std); if (err) return err; decoder->current_std = i; decoder->tvp514x_regs[REG_VIDEO_STD].val = decoder->std_list[i].video_std; v4l2_dbg(1, debug, sd, "Standard set to: %s\n", decoder->std_list[i].standard.name); return 0; } /** * tvp514x_s_routing() - V4L2 decoder interface handler for s_routing * @sd: pointer to standard V4L2 sub-device structure * @input: input selector for routing the signal * @output: output selector for routing the signal * @config: config value. Not used * * If index is valid, selects the requested input. Otherwise, returns -EINVAL if * the input is not supported or there is no active signal present in the * selected input. */ static int tvp514x_s_routing(struct v4l2_subdev *sd, u32 input, u32 output, u32 config) { struct tvp514x_decoder *decoder = to_decoder(sd); int err; enum tvp514x_input input_sel; enum tvp514x_output output_sel; if ((input >= INPUT_INVALID) || (output >= OUTPUT_INVALID)) /* Index out of bound */ return -EINVAL; input_sel = input; output_sel = output; err = tvp514x_write_reg(sd, REG_INPUT_SEL, input_sel); if (err) return err; output_sel |= tvp514x_read_reg(sd, REG_OUTPUT_FORMATTER1) & 0x7; err = tvp514x_write_reg(sd, REG_OUTPUT_FORMATTER1, output_sel); if (err) return err; decoder->tvp514x_regs[REG_INPUT_SEL].val = input_sel; decoder->tvp514x_regs[REG_OUTPUT_FORMATTER1].val = output_sel; decoder->input = input; decoder->output = output; v4l2_dbg(1, debug, sd, "Input set to: %d\n", input_sel); return 0; } /** * tvp514x_s_ctrl() - V4L2 decoder interface handler for s_ctrl * @ctrl: pointer to v4l2_ctrl structure * * If the requested control is supported, sets the control's current * value in HW. Otherwise, returns -EINVAL if the control is not supported. */ static int tvp514x_s_ctrl(struct v4l2_ctrl *ctrl) { struct v4l2_subdev *sd = to_sd(ctrl); struct tvp514x_decoder *decoder = to_decoder(sd); int err = -EINVAL, value; value = ctrl->val; switch (ctrl->id) { case V4L2_CID_BRIGHTNESS: err = tvp514x_write_reg(sd, REG_BRIGHTNESS, value); if (!err) decoder->tvp514x_regs[REG_BRIGHTNESS].val = value; break; case V4L2_CID_CONTRAST: err = tvp514x_write_reg(sd, REG_CONTRAST, value); if (!err) decoder->tvp514x_regs[REG_CONTRAST].val = value; break; case V4L2_CID_SATURATION: err = tvp514x_write_reg(sd, REG_SATURATION, value); if (!err) decoder->tvp514x_regs[REG_SATURATION].val = value; break; case V4L2_CID_HUE: if (value == 180) value = 0x7F; else if (value == -180) value = 0x80; err = tvp514x_write_reg(sd, REG_HUE, value); if (!err) decoder->tvp514x_regs[REG_HUE].val = value; break; case V4L2_CID_AUTOGAIN: err = tvp514x_write_reg(sd, REG_AFE_GAIN_CTRL, value ? 0x0f : 0x0c); if (!err) decoder->tvp514x_regs[REG_AFE_GAIN_CTRL].val = value; break; } v4l2_dbg(1, debug, sd, "Set Control: ID - %d - %d\n", ctrl->id, ctrl->val); return err; } /** * tvp514x_g_frame_interval() - V4L2 decoder interface handler * @sd: pointer to standard V4L2 sub-device structure * @ival: pointer to a v4l2_subdev_frame_interval structure * * Returns the decoder's video CAPTURE parameters. */ static int tvp514x_g_frame_interval(struct v4l2_subdev *sd, struct v4l2_subdev_frame_interval *ival) { struct tvp514x_decoder *decoder = to_decoder(sd); enum tvp514x_std current_std; /* get the current standard */ current_std = decoder->current_std; ival->interval = decoder->std_list[current_std].standard.frameperiod; return 0; } /** * tvp514x_s_frame_interval() - V4L2 decoder interface handler * @sd: pointer to standard V4L2 sub-device structure * @ival: pointer to a v4l2_subdev_frame_interval structure * * Configures the decoder to use the input parameters, if possible. If * not possible, returns the appropriate error code. */ static int tvp514x_s_frame_interval(struct v4l2_subdev *sd, struct v4l2_subdev_frame_interval *ival) { struct tvp514x_decoder *decoder = to_decoder(sd); struct v4l2_fract *timeperframe; enum tvp514x_std current_std; timeperframe = &ival->interval; /* get the current standard */ current_std = decoder->current_std; *timeperframe = decoder->std_list[current_std].standard.frameperiod; return 0; } /** * tvp514x_s_stream() - V4L2 decoder i/f handler for s_stream * @sd: pointer to standard V4L2 sub-device structure * @enable: streaming enable or disable * * Sets streaming to enable or disable, if possible. */ static int tvp514x_s_stream(struct v4l2_subdev *sd, int enable) { int err = 0; struct tvp514x_decoder *decoder = to_decoder(sd); if (decoder->streaming == enable) return 0; switch (enable) { case 0: { /* Power Down Sequence */ err = tvp514x_write_reg(sd, REG_OPERATION_MODE, 0x01); if (err) { v4l2_err(sd, "Unable to turn off decoder\n"); return err; } decoder->streaming = enable; break; } case 1: { /* Power Up Sequence */ err = tvp514x_write_regs(sd, decoder->int_seq); if (err) { v4l2_err(sd, "Unable to turn on decoder\n"); return err; } /* Detect if not already detected */ err = tvp514x_detect(sd, decoder); if (err) { v4l2_err(sd, "Unable to detect decoder\n"); return err; } err = tvp514x_configure(sd, decoder); if (err) { v4l2_err(sd, "Unable to configure decoder\n"); return err; } decoder->streaming = enable; break; } default: err = -ENODEV; break; } return err; } static const struct v4l2_ctrl_ops tvp514x_ctrl_ops = { .s_ctrl = tvp514x_s_ctrl, }; /** * tvp514x_enum_mbus_code() - V4L2 decoder interface handler for enum_mbus_code * @sd: pointer to standard V4L2 sub-device structure * @sd_state: subdev state * @code: pointer to v4l2_subdev_mbus_code_enum structure * * Enumertaes mbus codes supported */ static int tvp514x_enum_mbus_code(struct v4l2_subdev *sd, struct v4l2_subdev_state *sd_state, struct v4l2_subdev_mbus_code_enum *code) { u32 pad = code->pad; u32 index = code->index; memset(code, 0, sizeof(*code)); code->index = index; code->pad = pad; if (index != 0) return -EINVAL; code->code = MEDIA_BUS_FMT_UYVY8_2X8; return 0; } /** * tvp514x_get_pad_format() - V4L2 decoder interface handler for get pad format * @sd: pointer to standard V4L2 sub-device structure * @sd_state: subdev state * @format: pointer to v4l2_subdev_format structure * * Retrieves pad format which is active or tried based on requirement */ static int tvp514x_get_pad_format(struct v4l2_subdev *sd, struct v4l2_subdev_state *sd_state, struct v4l2_subdev_format *format) { struct tvp514x_decoder *decoder = to_decoder(sd); __u32 which = format->which; if (format->pad) return -EINVAL; if (which == V4L2_SUBDEV_FORMAT_ACTIVE) { format->format = decoder->format; return 0; } format->format.code = MEDIA_BUS_FMT_UYVY8_2X8; format->format.width = tvp514x_std_list[decoder->current_std].width; format->format.height = tvp514x_std_list[decoder->current_std].height; format->format.colorspace = V4L2_COLORSPACE_SMPTE170M; format->format.field = V4L2_FIELD_INTERLACED; return 0; } /** * tvp514x_set_pad_format() - V4L2 decoder interface handler for set pad format * @sd: pointer to standard V4L2 sub-device structure * @sd_state: subdev state * @fmt: pointer to v4l2_subdev_format structure * * Set pad format for the output pad */ static int tvp514x_set_pad_format(struct v4l2_subdev *sd, struct v4l2_subdev_state *sd_state, struct v4l2_subdev_format *fmt) { struct tvp514x_decoder *decoder = to_decoder(sd); if (fmt->format.field != V4L2_FIELD_INTERLACED || fmt->format.code != MEDIA_BUS_FMT_UYVY8_2X8 || fmt->format.colorspace != V4L2_COLORSPACE_SMPTE170M || fmt->format.width != tvp514x_std_list[decoder->current_std].width || fmt->format.height != tvp514x_std_list[decoder->current_std].height) return -EINVAL; decoder->format = fmt->format; return 0; } static const struct v4l2_subdev_video_ops tvp514x_video_ops = { .s_std = tvp514x_s_std, .s_routing = tvp514x_s_routing, .querystd = tvp514x_querystd, .g_frame_interval = tvp514x_g_frame_interval, .s_frame_interval = tvp514x_s_frame_interval, .s_stream = tvp514x_s_stream, }; static const struct v4l2_subdev_pad_ops tvp514x_pad_ops = { .enum_mbus_code = tvp514x_enum_mbus_code, .get_fmt = tvp514x_get_pad_format, .set_fmt = tvp514x_set_pad_format, }; static const struct v4l2_subdev_ops tvp514x_ops = { .video = &tvp514x_video_ops, .pad = &tvp514x_pad_ops, }; static const struct tvp514x_decoder tvp514x_dev = { .streaming = 0, .fmt_list = tvp514x_fmt_list, .num_fmts = ARRAY_SIZE(tvp514x_fmt_list), .pix = { /* Default to NTSC 8-bit YUV 422 */ .width = NTSC_NUM_ACTIVE_PIXELS, .height = NTSC_NUM_ACTIVE_LINES, .pixelformat = V4L2_PIX_FMT_UYVY, .field = V4L2_FIELD_INTERLACED, .bytesperline = NTSC_NUM_ACTIVE_PIXELS * 2, .sizeimage = NTSC_NUM_ACTIVE_PIXELS * 2 * NTSC_NUM_ACTIVE_LINES, .colorspace = V4L2_COLORSPACE_SMPTE170M, }, .current_std = STD_NTSC_MJ, .std_list = tvp514x_std_list, .num_stds = ARRAY_SIZE(tvp514x_std_list), }; static struct tvp514x_platform_data * tvp514x_get_pdata(struct i2c_client *client) { struct tvp514x_platform_data *pdata = NULL; struct v4l2_fwnode_endpoint bus_cfg = { .bus_type = 0 }; struct device_node *endpoint; unsigned int flags; if (!IS_ENABLED(CONFIG_OF) || !client->dev.of_node) return client->dev.platform_data; endpoint = of_graph_get_next_endpoint(client->dev.of_node, NULL); if (!endpoint) return NULL; if (v4l2_fwnode_endpoint_parse(of_fwnode_handle(endpoint), &bus_cfg)) goto done; pdata = devm_kzalloc(&client->dev, sizeof(*pdata), GFP_KERNEL); if (!pdata) goto done; flags = bus_cfg.bus.parallel.flags; if (flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH) pdata->hs_polarity = 1; if (flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH) pdata->vs_polarity = 1; if (flags & V4L2_MBUS_PCLK_SAMPLE_RISING) pdata->clk_polarity = 1; done: of_node_put(endpoint); return pdata; } /** * tvp514x_probe() - decoder driver i2c probe handler * @client: i2c driver client device structure * @id: i2c driver id table * * Register decoder as an i2c client device and V4L2 * device. */ static int tvp514x_probe(struct i2c_client *client, const struct i2c_device_id *id) { struct tvp514x_platform_data *pdata = tvp514x_get_pdata(client); struct tvp514x_decoder *decoder; struct v4l2_subdev *sd; int ret; if (pdata == NULL) { dev_err(&client->dev, "No platform data\n"); return -EINVAL; } /* Check if the adapter supports the needed features */ if (!i2c_check_functionality(client->adapter, I2C_FUNC_SMBUS_BYTE_DATA)) return -EIO; decoder = devm_kzalloc(&client->dev, sizeof(*decoder), GFP_KERNEL); if (!decoder) return -ENOMEM; /* Initialize the tvp514x_decoder with default configuration */ *decoder = tvp514x_dev; /* Copy default register configuration */ memcpy(decoder->tvp514x_regs, tvp514x_reg_list_default, sizeof(tvp514x_reg_list_default)); decoder->int_seq = (struct tvp514x_reg *)id->driver_data; /* Copy board specific information here */ decoder->pdata = pdata; /** * Fetch platform specific data, and configure the * tvp514x_reg_list[] accordingly. Since this is one * time configuration, no need to preserve. */ decoder->tvp514x_regs[REG_OUTPUT_FORMATTER2].val |= (decoder->pdata->clk_polarity << 1); decoder->tvp514x_regs[REG_SYNC_CONTROL].val |= ((decoder->pdata->hs_polarity << 2) | (decoder->pdata->vs_polarity << 3)); /* Set default standard to auto */ decoder->tvp514x_regs[REG_VIDEO_STD].val = VIDEO_STD_AUTO_SWITCH_BIT; /* Register with V4L2 layer as slave device */ sd = &decoder->sd; v4l2_i2c_subdev_init(sd, client, &tvp514x_ops); #if defined(CONFIG_MEDIA_CONTROLLER) decoder->pad.flags = MEDIA_PAD_FL_SOURCE; decoder->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE; decoder->sd.entity.function = MEDIA_ENT_F_ATV_DECODER; ret = media_entity_pads_init(&decoder->sd.entity, 1, &decoder->pad); if (ret < 0) { v4l2_err(sd, "%s decoder driver failed to register !!\n", sd->name); return ret; } #endif v4l2_ctrl_handler_init(&decoder->hdl, 5); v4l2_ctrl_new_std(&decoder->hdl, &tvp514x_ctrl_ops, V4L2_CID_BRIGHTNESS, 0, 255, 1, 128); v4l2_ctrl_new_std(&decoder->hdl, &tvp514x_ctrl_ops, V4L2_CID_CONTRAST, 0, 255, 1, 128); v4l2_ctrl_new_std(&decoder->hdl, &tvp514x_ctrl_ops, V4L2_CID_SATURATION, 0, 255, 1, 128); v4l2_ctrl_new_std(&decoder->hdl, &tvp514x_ctrl_ops, V4L2_CID_HUE, -180, 180, 180, 0); v4l2_ctrl_new_std(&decoder->hdl, &tvp514x_ctrl_ops, V4L2_CID_AUTOGAIN, 0, 1, 1, 1); sd->ctrl_handler = &decoder->hdl; if (decoder->hdl.error) { ret = decoder->hdl.error; goto done; } v4l2_ctrl_handler_setup(&decoder->hdl); ret = v4l2_async_register_subdev(&decoder->sd); if (!ret) v4l2_info(sd, "%s decoder driver registered !!\n", sd->name); done: if (ret < 0) { v4l2_ctrl_handler_free(&decoder->hdl); media_entity_cleanup(&decoder->sd.entity); } return ret; } /** * tvp514x_remove() - decoder driver i2c remove handler * @client: i2c driver client device structure * * Unregister decoder as an i2c client device and V4L2 * device. Complement of tvp514x_probe(). */ static int tvp514x_remove(struct i2c_client *client) { struct v4l2_subdev *sd = i2c_get_clientdata(client); struct tvp514x_decoder *decoder = to_decoder(sd); v4l2_async_unregister_subdev(&decoder->sd); media_entity_cleanup(&decoder->sd.entity); v4l2_ctrl_handler_free(&decoder->hdl); return 0; } /* TVP5146 Init/Power on Sequence */ static const struct tvp514x_reg tvp5146_init_reg_seq[] = { {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x02}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0x80}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x01}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x60}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0xB0}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x01}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x00}, {TOK_WRITE, REG_OPERATION_MODE, 0x01}, {TOK_WRITE, REG_OPERATION_MODE, 0x00}, {TOK_TERM, 0, 0}, }; /* TVP5147 Init/Power on Sequence */ static const struct tvp514x_reg tvp5147_init_reg_seq[] = { {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x02}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0x80}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x01}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x60}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0xB0}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x01}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x16}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0xA0}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x16}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS1, 0x60}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS2, 0x00}, {TOK_WRITE, REG_VBUS_ADDRESS_ACCESS3, 0xB0}, {TOK_WRITE, REG_VBUS_DATA_ACCESS_NO_VBUS_ADDR_INCR, 0x00}, {TOK_WRITE, REG_OPERATION_MODE, 0x01}, {TOK_WRITE, REG_OPERATION_MODE, 0x00}, {TOK_TERM, 0, 0}, }; /* TVP5146M2/TVP5147M1 Init/Power on Sequence */ static const struct tvp514x_reg tvp514xm_init_reg_seq[] = { {TOK_WRITE, REG_OPERATION_MODE, 0x01}, {TOK_WRITE, REG_OPERATION_MODE, 0x00}, {TOK_TERM, 0, 0}, }; /* * I2C Device Table - * * name - Name of the actual device/chip. * driver_data - Driver data */ static const struct i2c_device_id tvp514x_id[] = { {"tvp5146", (unsigned long)tvp5146_init_reg_seq}, {"tvp5146m2", (unsigned long)tvp514xm_init_reg_seq}, {"tvp5147", (unsigned long)tvp5147_init_reg_seq}, {"tvp5147m1", (unsigned long)tvp514xm_init_reg_seq}, {}, }; MODULE_DEVICE_TABLE(i2c, tvp514x_id); #if IS_ENABLED(CONFIG_OF) static const struct of_device_id tvp514x_of_match[] = { { .compatible = "ti,tvp5146", }, { .compatible = "ti,tvp5146m2", }, { .compatible = "ti,tvp5147", }, { .compatible = "ti,tvp5147m1", }, { /* sentinel */ }, }; MODULE_DEVICE_TABLE(of, tvp514x_of_match); #endif static struct i2c_driver tvp514x_driver = { .driver = { .of_match_table = of_match_ptr(tvp514x_of_match), .name = TVP514X_MODULE_NAME, }, .probe = tvp514x_probe, .remove = tvp514x_remove, .id_table = tvp514x_id, }; module_i2c_driver(tvp514x_driver);
__label__pos
0.992891
Take the 2-minute tour × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free. Run time error on main method in MovieList.java. I'm not sure my program design is fundamentally very good, but I'd like to know why it crashes. Thanks in advance. package javaPractical.week3; import javax.swing.*; public class Movie { //private attributes private String title; private String movieURL; private String year; private String genre; private String actor; // constructor Movie(String t, String u, String y, String g, String a) { this.title = t; this.movieURL = u; this.year = y; this.genre = g; this.actor = a; } //getters and setters public void setTitle(String t) { this.title = t; } public String getTitle() { return this.title; } public void set_url(String a) { this.movieURL = a; } public String get_url() { return this.movieURL; } public void setYear(String y) { this.year = y; } public String getYear() { return this.year; } public void setGenre(String g) { this.genre = g; } public String getGenre() { return this.genre; } public void setActor(String a) { this.actor = a; } public String getActor() { return this.actor; } //output movie details public String toString() { return ("Title: " + this.title + "\nURL: " + this.movieURL + "\nYear: " + this.year + "\nGenre: " + this.genre + "\nActor: " + this.actor); } public static void main(String[] args) { //testing Movie class Movie Movie1 = new Movie("Spiderman", "www.", "2002", "Action", "Tobey M"); JOptionPane.showMessageDialog(null, Movie1.toString()); //testing MovieList class } } package javaPractical.week3; import javax.swing.*; import java.util.ArrayList; public class MovieList1 { private static ArrayList myFavouriteMovies = new ArrayList(); private static int NUM_OF_MOVIES = 10; private int numberOfMovies = 0; private int index = 0; public MovieList1() { this.myFavouriteMovies = null; this.numberOfMovies = 0; this.index = 0; } public int getNumberOfMovies() { return this.myFavouriteMovies.size(); } public boolean isEmpty() { if (this.myFavouriteMovies.isEmpty()) { return true; } else return false; } public static void main(String[] args) { MovieList1 List = new MovieList1(); String titleADD; String movieURLADD; String yearADD; String genreADD; String actorADD; titleADD = JOptionPane.showInputDialog(null, "Enter title:"); movieURLADD = JOptionPane.showInputDialog(null, "Enter URL:"); yearADD = JOptionPane.showInputDialog(null, "Enter year:"); genreADD = JOptionPane.showInputDialog(null, "Enter genre:"); actorADD = JOptionPane.showInputDialog(null, "Enter actor:"); Movie TempMovie = new Movie(titleADD, movieURLADD, yearADD, genreADD, actorADD); myFavouriteMovies.add(TempMovie); } } share|improve this question 3   Although it's a nullpointerexception I removed the pointer tag. It doesn't make sense to talk about pointers in Java. Also, please add the 'homework' tag if this is homework. –  Synesso Oct 30 '10 at 23:20 6 Answers 6 The program crashes when it tries to add the new Movie to myFavouriteMovies, because myFavouriteMovies is null. Although myFavouriteMovies is initialised to a new, empty ArrayList, it's then set to null in the MovieList1 constructor. At the moment, myFavouriteMovies is static, so there's only one copy of this variable shared between every MovieList1 instance. You probably want to remove the static modifier from the myFavouriteMovies declaration. Then each MovieList1 object will have its own myFavouriteMovies field. However you'll then to add a new method to the MovieList1 class to allow your main method to add the movie to the movie list, perhaps like this: List.add(TempMovie); Also you'll need to remove this.myFavouriteMovies = null; from the constructor, because having initialised it to an empty ArrayList, you don't want to set it back to null. share|improve this answer Within your constructor you are setting public MovieList1() { this.myFavouriteMovies = null; this.numberOfMovies = 0; this.index = 0; } after you already declared myFavouriteMovies above. This could result in a NullPointer share|improve this answer 1   Yes. @James Is myFavouriteMovies static, or a an instance variable? Choose one and instantiate it only once. –  Synesso Oct 30 '10 at 23:17 When you call the constructor for MovieList1, you set the ArrayList MyFavouriteMovies to null. If you call methods on MyFavouriteMovies, there is a null pointer exception (at myFavouriteMovies.add(TempMovie);). this.myFavouriteMovies = null; should be this.myFavouriteMovies = new ArrayList(); and private static ArrayList myFavouriteMovies = new ArrayList(); should be private ArrayList myFavouriteMovies; By the way, I wouldn't make myFavouriteMovies static, since it differs for every instance of MovieList1. You'd then have a method addMovie() in MovieList1. Also, if NUM_OF_MOVIES is constant, as the uppercase name suggests, you should declare it final. share|improve this answer All the above answers are spot on, but I question whether or not you even need the MovieList1 class at all. Essentially, you're just providing a wrapper around List. I'm not sure if you have plans to expand the behaviors of the movie list, but as is, you could simply do: List<Movie> movies = new ArrayList<Movie>(); String titleADD = JOptionPane.showInputDialog(null, "Enter title:"); String movieURLADD = JOptionPane.showInputDialog(null, "Enter URL:"); String yearADD = JOptionPane.showInputDialog(null, "Enter year:"); String genreADD = JOptionPane.showInputDialog(null, "Enter genre:"); String actorADD = JOptionPane.showInputDialog(null, "Enter actor:"); Movie TempMovie = new Movie(titleADD, movieURLADD, yearADD, genreADD, actorADD); movies.add(TempMovie); A couple of other notes... You should probably have an addMovie(Movie movie) method or something similar in the movie list class instead of accessing the list directly in your main method. You should program to the interface instead of declaring an ArrayList as the type of myFavoriteMovies. There's no need to overwrite the values in the constructor because you've already instantiated or initialized them when you declare them. You probably get a warning by calling this.myFavoriteMovies because it's static. Static members and methods should be accessed by ClassName.staticMethodOrVariable. I only mention this because there is a big difference between what this implies and what static is. This refers to the current reference of the type, while static is meant to be persistent. Also, when declaring unmutable static members, use the final modifier, but in this case, I don't think it needs to be static, but could definitely be final. share|improve this answer Richard has the right answer to what the problem is. There is a better solution though. When you declare a variable you should think about if you ever want it to change. If you do not mark it as final. So: private final String title; private final String movieURL; private final String year; private final String genre; private final String actor; private static final int NUM_OF_MOVIES = 10; private final List myFavouriteMovies = new ArrayList(); private int numberOfMovies; private int index; This will mean you need to get rid of most of the setXXX methods - if you think about it you never want to change the values after you create an instance anyways. The numberOfMovies and index presumably need to change, so they are not final. Finally (no pun intended!) you do not need to set instance variables to 0, null, or false, they are set to that by default. share|improve this answer I am observing that you do not have a good grip of programming logic. You also need to further understand the flow of Java runtime and the behaviour of the language. Still, logic flow is a skill you need to acquire, regardless if you are communicating in Java, C, C++ or the English language. (BTW, the English language, despite having inconsistent synthetic elements, is an analytically logical language.) First, you declare a static myFavouriteMovies and instantiate it to an ArrayList. After that, you equate it to null. After equating the variable to null, you attempt to use it as myFavouriteMovies.add(TempMovie); Of course, you will get a null pointer. You need to acquaint yourself with the compiler and runtime error dumps. The Java runtime error would certainly have pointed out the statement number where it encountered your attempt to withdraw money from an empty non-overdraft-protected bank account. C# programs written by "VB experts" always frustrate me. I realise this is not a kind response, but I am projecting a very presumptuous attitude in saying that you are highly probable a proficient Visual Basic programmer. Therefore, I am taking further opportunity towards all proficient Visual Basic programmers out there by throwing salt on an open wound (salt being a disinfectant), that you should NEVER transfer your programming expertise in an object-referencing language to object-oriented languages. Just as it is difficult for an expert speaker of a synthetically logical language like Greek or Hebrew to adjust to an analytically logical language like English, vice versa. Secondly, what is the reason behind your declaring myFavouriteMovies as static? Do you realise the implications of a static variable? Thirdly, you should not refer to a static variable thro the class instance reference "this". myFavouriteMovies is a static variable within the class Movie. Therefore, you should refer to it as Movie.myFavouriteMovies and not as this.myFavouriteMovies Java allows you to perform such aberrational reference but C# would not. "this" should be reserved for "instance variables". Finally, there is high probability, compared to other "VB experts", that you need further understanding of what instance variables are vs static variables. A static variable exists without needing a class to be instantiated. Therefore, it is "shared" by all instances of the class; and if it is public or protected, it is shared with any process elements to which that static variable is exposed. Normally, when you have a static list, your get method should not return the reference of that static variable. You should return a subset of the list. For example, you could have a static list of all the Fans of every movie. When you need to return a list of the list of Fans of a particular movie, you do not return the whole static list but only a subset of that static list. You need to understand when static variables could be used and restrain yourself from deploying static references - meaning, DO NOT declare a variable static if making it a class instance will work. You can use a static variable to monitor and restrict that a Movie class is not instantiated more than, say, five times or that there are not currently more than five instances, by incrementing a static count everytime it is instantiated and decrementing everytime it is destroyed. You can use a static variable to hold a pool of connections to the database, where each connection is an instance of a connection. Little trivia: Do you know that you could perform static manipulation in Java using a static block? If there are more than one static block, they will be processed in the order they are declared. As usual, any variable declared inside a code block is not visible outside that block. class Hello{ static { // perform static manipulation here } public void setAnInstanceValue(int a){ ... } static { // another static block of code here } ..... } You should try to instantiate your static variables in a static code block. If you find that you could instantiate a static variable in an instance constructor, there is high probability that it should not be static. Read this: Execution flow for static code block - http://forums.sun.com/thread.jspa?threadID=5418566. share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.655562
#include /* this is a test of the NESpad library a nintendo joystick midi drum kit, mostly stolen from tod e. kurt's much more interesting piezo disc drum kit you can get the NESpad library here: http://rahji.com/NESpad 08/21/2007 rob duarte - [email protected] */ // what midi channel we're sending on #define drumchan 1 #define UNPRESSED 0 #define PRESSED 1 // general midi drum notes #define note_bassdrum 35 #define note_snaredrum 38 #define note_hihatclosed 42 #define note_hihatopen 44 #define note_crash 49 #define ledPin 13 // goes on when we're sending midi data boolean ButtonAState = UNPRESSED; boolean ButtonBState = UNPRESSED; boolean ButtonSelectState = UNPRESSED; boolean ButtonStartState = UNPRESSED; NESpad nintendo = NESpad(); byte state = 0; // joystick buttons state // (it ends up looking like a bitmap with 1's for buttons that are pressed and 0's for unpressed) void setup() { pinMode(ledPin, OUTPUT); Serial.begin(31250); // set MIDI baud rate } void loop() { state = nintendo.buttons(); // if the button is pressed right now and it's last state was UNPRESSED... if ( (state & NES_A) && ButtonAState == UNPRESSED ) { note(drumchan, note_bassdrum, 100); // turn on the note ButtonAState = PRESSED; // and mark this button as being pressed } // otherwise if the button is unpressed right now but it was PRESSED the last time we checked... else if ( !(state & NES_A) && ButtonAState == PRESSED ) { note(drumchan, note_bassdrum, 0); // turn off the note ButtonAState = UNPRESSED; // and mark this button as being unpressed } if ( (state & NES_B) && ButtonBState == UNPRESSED ) { note(drumchan, note_snaredrum, 100); ButtonBState = PRESSED; } else if ( !(state & NES_B) && ButtonBState == PRESSED ) { note(drumchan, note_snaredrum, 0); ButtonBState = UNPRESSED; } if ( (state & NES_START) && ButtonStartState == UNPRESSED ) { note(drumchan, note_hihatopen, 100); ButtonStartState = PRESSED; } else if ( !(state & NES_START) && ButtonStartState == PRESSED ) { note(drumchan, note_hihatopen, 0); ButtonStartState = UNPRESSED; } if ( (state & NES_SELECT) && ButtonSelectState == UNPRESSED ) { note(drumchan, note_hihatclosed, 100); ButtonSelectState = PRESSED; } else if ( !(state & NES_SELECT) && ButtonSelectState == PRESSED ) { note(drumchan, note_hihatclosed, 0); ButtonSelectState = UNPRESSED; } } // Send a MIDI note-on message. Like pressing a piano key void note(byte channel, byte note, byte velocity) { midiMsg( (0x80 | (channel<<4)), note, velocity); } // Send a general MIDI message void midiMsg(byte cmd, byte data1, byte data2) { digitalWrite(ledPin,HIGH); // indicate we're sending MIDI data Serial.print(cmd, BYTE); Serial.print(data1, BYTE); Serial.print(data2, BYTE); digitalWrite(ledPin,LOW); }
__label__pos
0.750526
  +0     0 38 2 avatar Find the value of n if the coefficients of x^3 and x^4 are equal in the expansion of (2x+1)^n     Any help or hints are greatly appreciated <3.  Jun 12, 2020  #1 avatar+306  +1 Find the value of n if the coefficients of \(x^3\) and \(x^4\)are equal in the expansion of \((2x+1)^n\). I rewrote bc i have a pet peeve where i don't like it when the LaTeX is wrong it's nothing personal laugh  Jun 13, 2020  #2 avatar+8205  0 We use binomial theorem.   \((2x + 1)^n = \displaystyle\sum_{k = 0}^n \binom nk (2x)^k\)   Coefficient of \(x^3\) is \(\displaystyle \binom n3 (2)^3 = \dfrac{4n(n - 1)(n - 2)}{3}\) for positive integer n.   Coefficient of \(x^4\) is \(\displaystyle \binom n4 (2)^4 = \dfrac{2n(n - 1)(n - 2)(n - 3)}{3}\) for positive integer n.   Now, the coefficient of \(x^3\) is equal to that of \(x^4 \).   \(\dfrac{4n(n - 1)(n - 2)}{3} = \dfrac{2n(n - 1)(n - 2)(n - 3)}{3}\\ 2n(n - 1)(n - 2) = n(n - 1)(n - 2)(n - 3)\\ n(n - 1)(n - 2)(n - 5) = 0\\ n = 0 \text{ (rej.) or }n = 1 \text{ (rej.) or } n = 2 \text{ (rej.) or } n = 5\)   Therefore n = 5.  Jun 13, 2020 15 Online Users avatar avatar avatar
__label__pos
0.998589
Como controlar o acesso a conjuntos de dados Neste documento, descreveremos como controlar o acesso a conjuntos de dados no BigQuery. Também é possível controlar o acesso no nível de tabela/visualização e em um nível superior na hierarquia de recursos do IAM. É possível restringir o acesso a colunas com a segurança no nível da coluna do BigQuery. Visão geral As permissões no nível do conjunto de dados determinam os usuários, grupos e contas de serviço que têm permissão para acessar as tabelas, visualizações e dados da tabela em um conjunto de dados específico. Por exemplo, se você conceder o papel de IAM bigquery.dataOwner a um usuário em um conjunto de dados específico, esse usuário poderá criar, atualizar e excluir tabelas e visualizações no conjunto de dados. É possível aplicar controles de acesso durante a criação do conjunto de dados chamando o método da API datasets.insert. Os controles de acesso não podem ser aplicados durante a criação do conjunto de dados no Cloud Console, na interface do usuário da Web clássica do BigQuery ou na ferramenta de linha de comando bq. Para aplicar controles de acesso a um conjunto de dados após ele ser criado: • use o Console do Cloud ou a IU da Web clássica do BigQuery; • use o comando bq update na ferramenta de linha de comando bq; • Chame o método de API datasets.patch. • usando as bibliotecas de cliente. Permissões necessárias No mínimo, para atribuir ou atualizar controles de acesso ao conjunto de dados, você precisa receber as permissões bigquery.datasets.update e bigquery.datasets.get. Os seguintes papéis predefinidos do IAM incluem permissões bigquery.datasets.update e bigquery.datasets.get: • bigquery.dataOwner • bigquery.admin Além disso, se um usuário tiver permissões bigquery.datasets.create ao criar um conjunto de dados, será concedido o acesso bigquery.dataOwner. O acesso bigquery.dataOwner oferece aos usuários a capacidade de atualizar os conjuntos de dados que eles criaram. Para mais informações sobre os papéis e as permissões do IAM no BigQuery, consulte Papéis e permissões predefinidos. Como controlar o acesso a um conjunto de dados Para atribuir controles de acesso a um conjunto de dados, faça o seguinte: Console 1. Selecione um conjunto de dados em Recursos e clique em Compartilhar conjunto de dados no lado direito da janela. Adicionar pessoas ao conjunto de dados 2. No painel Compartilhar conjunto de dados da guia Permissões do conjunto de dados, digite na caixa de texto Adicionar membros a entidade que você quer incluir. É possível adicionar qualquer uma das entidades a seguir: • E-mail da conta do Google: concede acesso ao conjunto de dados a uma conta individual do Google • Grupo do Google: concede acesso ao conjunto de dados a todos os membros de um grupo do Google • Domínio do Google Apps: concede acesso ao conjunto de dados a todos os usuários e grupos em um domínio do Google • Conta de serviço: concede acesso ao conjunto de dados a uma conta de serviço. • Qualquer pessoa: insira "allUsers" para conceder acesso ao público em geral • Todas as contas do Google: insira "allAuthenticatedUsers" para conceder acesso a qualquer usuário conectado a uma Conta do Google 3. Em Selecionar um papel, selecione BigQuery e escolha um papel de IAM predefinido apropriado para os novos membros. Para mais informações sobre as permissões atribuídas a cada papel predefinido do BigQuery, consulte a seção Papéis da página de controle de acesso. 4. Clique em Concluído. IU clássica 1. Clique na seta suspensa à direita do conjunto de dados e escolha Compartilhar conjunto de dados. 2. Na caixa de diálogo Compartilhar conjunto de dados, para Adicionar pessoas, clique na lista suspensa à esquerda do campo e escolha a opção adequada. Ao aplicar controles de acesso a um conjunto de dados usando a IU clássica da Web, você poderá conceder acesso aos seguintes usuários e grupos: • Usuário por e-mail: fornece acesso ao conjunto de dados a uma conta individual do Google • Grupo por e-mail: fornece acesso ao conjunto de dados a todos os membros de um grupo do Google • Domínio: fornece acesso ao conjunto de dados a todos os usuários e grupos em um domínio do Google • Todos os usuários autenticados: fornece acesso ao conjunto de dados a todos os titulares de contas do Google (torna público o conjunto de dados) • Proprietários do projeto: fornece acesso ao conjunto de dados a todos os proprietários do projeto • Visualizadores do projeto: fornece acesso ao conjunto de dados a todos os visualizadores do projeto • Editores do projeto: fornece acesso ao conjunto de dados a todos os editores do projeto • Visualização autorizada: fornece acesso de visualização ao conjunto de dados 3. Digite um valor na caixa de texto. Por exemplo, se você escolher Usuário por e-mail ou Grupo por e-mail, digite o endereço de e-mail do usuário ou grupo. 4. À direita do campo Adicionar pessoas, clique em Pode ver e escolha o papel apropriado na lista. Adicionar pessoas ao conjunto de dados 5. Clique em Adicionar e em Salvar alterações. bq 1. Use o comando show para gravar as informações atuais do conjunto de dados, incluindo controles de acesso, em um arquivo JSON. Se o conjunto de dados estiver em um projeto diferente do padrão, adicione o ID do projeto ao nome do conjunto de dados no seguinte formato: project_id:dataset. bq show \ --format=prettyjson \ project_id:dataset > path_to_file Onde: • project_id é o ID do projeto; • dataset é o nome do conjunto de dados; • path_to_file é o caminho do arquivo JSON na sua máquina local. Exemplos: Digite o comando a seguir para gravar os controles de acesso para mydataset em um arquivo JSON. mydataset está no projeto padrão. bq show --format=prettyjson mydataset > /tmp/mydataset.json Digite o comando a seguir para gravar os controles de acesso para mydataset em um arquivo JSON. mydataset está em myotherproject. bq show --format=prettyjson \ myotherproject:mydataset > /tmp/mydataset.json 2. Faça suas alterações na seção "access" do arquivo JSON. É possível adicionar ou remover qualquer uma das entradas specialGroup: projectOwners, projectWriters, projectReaders e allAuthenticatedUsers. É possível adicionar, remover ou modificar qualquer uma destas opções: userByEmail, groupByEmail e domain. Por exemplo, a seção "access" do arquivo JSON de um conjunto de dados tem esta aparência: { "access": [ { "role": "READER", "specialGroup": "projectReaders" }, { "role": "WRITER", "specialGroup": "projectWriters" }, { "role": "OWNER", "specialGroup": "projectOwners" }, { "role": "READER", "specialGroup": "allAuthenticatedUsers" }, { "role": "READER", "domain": "domain_name" }, { "role": "WRITER", "userByEmail": "user_email" }, { "role": "READER", "groupByEmail": "group_email" } ], ... } 3. Quando as edições estiverem concluídas, use o comando update e inclua o arquivo JSON usando a sinalização --source. Se o conjunto de dados estiver em um projeto diferente do padrão, adicione o ID do projeto ao nome do conjunto de dados no seguinte formato: project_id:dataset. bq update \ --source path_to_file \ project_id:dataset Onde: • path_to_file é o caminho do arquivo JSON na sua máquina local; • project_id é o ID do projeto; • dataset é o nome do conjunto de dados; Exemplos: Digite o comando a seguir para atualizar os controles de acesso para mydataset. mydataset está no projeto padrão. bq update --source /tmp/mydataset.json mydataset Digite o comando a seguir para atualizar os controles de acesso para mydataset. mydataset está em myotherproject. bq update --source /tmp/mydataset.json myotherproject:mydataset 4. Para verificar as alterações no controle de acesso, insira o comando show novamente sem gravar as informações em um arquivo. bq show --format=prettyjson dataset ou bq show --format=prettyjson project_id:dataset API Ligue para datasets.insertcom um recurso de conjunto de dados definido para aplicar controles de acesso quando o conjunto de dados for criado. Ligue para datasets.patch e use a propriedade access no recurso do conjunto de dados para atualizar seus controles de acesso. Como datasets.update substitui todo o recurso do conjunto de dados, é melhor usar o método datasets.patch para atualizar os controles de acesso. Go Antes de testar essa amostra, siga as instruções de configuração para Go no Guia de início rápido do BigQuery: como usar bibliotecas de cliente. Para mais informações, consulte a documentação de referência da API BigQuery Go (em inglês). Defina a propriedade dataset.access_entries com os controles de acesso de um conjunto de dados. Em seguida, chame a função client.update_dataset() para atualizar a propriedade (links em inglês). Ver no GitHub (em inglês) Feedback import ( "context" "fmt" "cloud.google.com/go/bigquery" ) // updateDatasetAccessControl demonstrates how the access control policy of a dataset // can be amended by adding an additional entry corresponding to a specific user identity. func updateDatasetAccessControl(projectID, datasetID string) error { // projectID := "my-project-id" // datasetID := "mydataset" ctx := context.Background() client, err := bigquery.NewClient(ctx, projectID) if err != nil { return fmt.Errorf("bigquery.NewClient: %v", err) } defer client.Close() ds := client.Dataset(datasetID) meta, err := ds.Metadata(ctx) if err != nil { return err } // Append a new access control entry to the existing access list. update := bigquery.DatasetMetadataToUpdate{ Access: append(meta.Access, &bigquery.AccessEntry{ Role: bigquery.ReaderRole, EntityType: bigquery.UserEmailEntity, Entity: "[email protected]"}, ), } // Leverage the ETag for the update to assert there's been no modifications to the // dataset since the metadata was originally read. if _, err := ds.Update(ctx, update, meta.ETag); err != nil { return err } return nil } Python Antes de testar essa amostra, siga as instruções de configuração para Python no Guia de início rápido do BigQuery: como usar bibliotecas de cliente. Para mais informações, consulte a documentação de referência da API BigQuery Python (em inglês). Defina a propriedade dataset.access_entries com os controles de acesso de um conjunto de dados. Em seguida, chame a função client.update_dataset() para atualizar a propriedade (links em inglês). from google.cloud import bigquery # Construct a BigQuery client object. client = bigquery.Client() # TODO(developer): Set dataset_id to the ID of the dataset to fetch. # dataset_id = 'your-project.your_dataset' dataset = client.get_dataset(dataset_id) # Make an API request. entry = bigquery.AccessEntry( role="READER", entity_type="userByEmail", entity_id="[email protected]", ) entries = list(dataset.access_entries) entries.append(entry) dataset.access_entries = entries dataset = client.update_dataset(dataset, ["access_entries"]) # Make an API request. full_dataset_id = "{}.{}".format(dataset.project, dataset.dataset_id) print( "Updated dataset '{}' with modified user permissions.".format(full_dataset_id) ) A seguir
__label__pos
0.510304
Thread Closed *screams in anger* ok, divergence theorem problem   Share Thread Thread Tools Oct14-05, 03:31 AM   #1   *screams in anger* ok, divergence theorem problem question says answer in whichever would be easier, the surface integral or the triple integral, then gives me(I'm in a mad hurry, excuse the lack of formatting...stuff) the triple integral of del F over the region x^2+y^2+z^2>=25 F=((x^2+y^2+z^2)(xi+yj+zk)), so del F would be 3(x^2+y^2+z^2)? In this case, the region is a sphere, it'd be easier to do it with the triple integral over the volume, right? Well, regardless I tried it that way, converting to spherical coordinates(my book mixes up the traditional phi and theta placement, but whatever) triple integral of 3r^4*sin(theta)drd(theta)d(phi), and the limits of integration, going from the right integral to the left, 0-5, 0-pi, 0-2pi? And somewhere before there is where I messed up 'cuz I can do the integral I have there easily enough and I get like some huge square of 5 times pi, and the answer is 100pi -_- :'(   PhysOrg.com PhysOrg science news on PhysOrg.com >> Heat-related deaths in Manhattan projected to rise >> Dire outlook despite global warming 'pause': study >> Sea level influenced tropical climate during the last ice age Oct14-05, 03:44 AM   #2   Quote by schattenjaeger over the region x^2+y^2+z^2>=25 is that actually a [itex]<=[/itex] ? (you say the region is a sphere...) F=((x^2+y^2+z^2)(xi+yj+zk)), so del F would be 3(x^2+y^2+z^2)? I don't understand your notation. If you what you meant was [tex]\left( x^3 \hat{i} + y^3 \hat{j} + z^3 \hat{k} \right) [/tex], then your [tex]\nabla \mathbf{\vec{F}}[/tex] is correct. But what you wrote is [tex]\left( x^2 + y^2 + z^2 \right)\left( x \hat{i} + y \hat{j} + z \hat{k} \right)[/tex] [tex]=\left( x \left( x^2 + y^2 + z^2 \right) \hat{i} + y \left( x^2 + y^2 + z^2 \right) \hat{j} + z \left( x^2 + y^2 + z^2 \right)\hat{k} \right) [/tex], which has a different divergence.   Oct14-05, 03:50 AM   #3   Oh I love you, you're fast yes I mean <= but yah, the latter is what I wrote and meant, but I'm confused, isn't the (x^2+y^2+z^2) like...if you just had a constant A, times (xi+yj+zk)(or A<x,y,z>)it's be like <Ax,Ay,Az>, so can you not do that with the (x^2+y^2+z^2) out front? am I doing something stupid in the midst of night?   Oct14-05, 03:53 AM   #4   *screams in anger* ok, divergence theorem problem F=((x^2+y^2+z^2)(xi+yj+zk)), so del F would be 3(x^2+y^2+z^2)? I'll throw in the intermediate step I did too, so F=<x^3+y^2+z^2,x^2+y^3+z^3,similarforz> and the divergence would be the same in that case as <x^3,y^3,z^3>   Oct14-05, 04:00 AM   #5   so F=<x^3+y^2+z^2,x^2+y^3+z^3,similarforz> Not quite: [tex]\left(x^2+y^2+z^2 \right)(x \mathbf{\hat{i}} \right))=\left(x^3+\mathbf{x}y^2+\mathbf{x}z^2 \right) \mathbf{\hat{i}}[/tex] and similarly...   Oct14-05, 07:42 AM   #6   Well that was a dumb mistake. Remember kids, sleep is gooood   Oct14-05, 01:12 PM   #7   So with that in mind how the devil is the answer what it is?   Oct14-05, 10:00 PM   #8   because using the correct divergence and going through it I'm getting like 12500pi   Dec19-05, 02:33 PM   #9   schattenjaeger, I hope you're awake after a long . I was having trouble with the exact problem,from Boas' book I presume, and a search for "divergence theorem" brought me here. Anyway, I arrived at the answer given at the back of the book,viz. [tex]4\pi5^5[/tex], by solving the surface integral. You also get it when [tex]\nabla.\vec{F} = 3(x^2+y^2+z^2)[/tex],in the case of a volume integral, which is obviously wrong. So did you eventually solve it using the triple integral? Anyone's help will be appreciated.   Thread Closed Thread Tools Similar Threads for: *screams in anger* ok, divergence theorem problem Thread Forum Replies Divergence theorem Classical Physics 2 divergence theorem...again Introductory Physics Homework 2 Help with using the Divergence Theorem! Calculus 8 Problem with the Divergence Theorem General Math 7 help with divergence theorem Introductory Physics Homework 1
__label__pos
0.99983
summaryrefslogtreecommitdiff path: root/src/math/erfl.c blob: e267c231059cb85c126b8e86a10d43254ff5aaf0 (plain) 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 /* origin: OpenBSD /usr/src/lib/libm/src/ld80/e_erfl.c */ /* * ==================================================== * Copyright (C) 1993 by Sun Microsystems, Inc. All rights reserved. * * Developed at SunPro, a Sun Microsystems, Inc. business. * Permission to use, copy, modify, and distribute this * software is freely granted, provided that this notice * is preserved. * ==================================================== */ /* * Copyright (c) 2008 Stephen L. Moshier <[email protected]> * * Permission to use, copy, modify, and distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ /* double erf(double x) * double erfc(double x) * x * 2 |\ * erf(x) = --------- | exp(-t*t)dt * sqrt(pi) \| * 0 * * erfc(x) = 1-erf(x) * Note that * erf(-x) = -erf(x) * erfc(-x) = 2 - erfc(x) * * Method: * 1. For |x| in [0, 0.84375] * erf(x) = x + x*R(x^2) * erfc(x) = 1 - erf(x) if x in [-.84375,0.25] * = 0.5 + ((0.5-x)-x*R) if x in [0.25,0.84375] * Remark. The formula is derived by noting * erf(x) = (2/sqrt(pi))*(x - x^3/3 + x^5/10 - x^7/42 + ....) * and that * 2/sqrt(pi) = 1.128379167095512573896158903121545171688 * is close to one. The interval is chosen because the fix * point of erf(x) is near 0.6174 (i.e., erf(x)=x when x is * near 0.6174), and by some experiment, 0.84375 is chosen to * guarantee the error is less than one ulp for erf. * * 2. For |x| in [0.84375,1.25], let s = |x| - 1, and * c = 0.84506291151 rounded to single (24 bits) * erf(x) = sign(x) * (c + P1(s)/Q1(s)) * erfc(x) = (1-c) - P1(s)/Q1(s) if x > 0 * 1+(c+P1(s)/Q1(s)) if x < 0 * Remark: here we use the taylor series expansion at x=1. * erf(1+s) = erf(1) + s*Poly(s) * = 0.845.. + P1(s)/Q1(s) * Note that |P1/Q1|< 0.078 for x in [0.84375,1.25] * * 3. For x in [1.25,1/0.35(~2.857143)], * erfc(x) = (1/x)*exp(-x*x-0.5625+R1(z)/S1(z)) * z=1/x^2 * erf(x) = 1 - erfc(x) * * 4. For x in [1/0.35,107] * erfc(x) = (1/x)*exp(-x*x-0.5625+R2/S2) if x > 0 * = 2.0 - (1/x)*exp(-x*x-0.5625+R2(z)/S2(z)) * if -6.666<x<0 * = 2.0 - tiny (if x <= -6.666) * z=1/x^2 * erf(x) = sign(x)*(1.0 - erfc(x)) if x < 6.666, else * erf(x) = sign(x)*(1.0 - tiny) * Note1: * To compute exp(-x*x-0.5625+R/S), let s be a single * precision number and s := x; then * -x*x = -s*s + (s-x)*(s+x) * exp(-x*x-0.5626+R/S) = * exp(-s*s-0.5625)*exp((s-x)*(s+x)+R/S); * Note2: * Here 4 and 5 make use of the asymptotic series * exp(-x*x) * erfc(x) ~ ---------- * ( 1 + Poly(1/x^2) ) * x*sqrt(pi) * * 5. For inf > x >= 107 * erf(x) = sign(x) *(1 - tiny) (raise inexact) * erfc(x) = tiny*tiny (raise underflow) if x > 0 * = 2 - tiny if x<0 * * 7. Special case: * erf(0) = 0, erf(inf) = 1, erf(-inf) = -1, * erfc(0) = 1, erfc(inf) = 0, erfc(-inf) = 2, * erfc/erf(NaN) is NaN */ #include "libm.h" #if LDBL_MANT_DIG == 53 && LDBL_MAX_EXP == 1024 long double erfl(long double x) { return erf(x); } long double erfcl(long double x) { return erfc(x); } #elif LDBL_MANT_DIG == 64 && LDBL_MAX_EXP == 16384 static const long double erx = 0.845062911510467529296875L, /* * Coefficients for approximation to erf on [0,0.84375] */ /* 8 * (2/sqrt(pi) - 1) */ efx8 = 1.0270333367641005911692712249723613735048E0L, pp[6] = { 1.122751350964552113068262337278335028553E6L, -2.808533301997696164408397079650699163276E6L, -3.314325479115357458197119660818768924100E5L, -6.848684465326256109712135497895525446398E4L, -2.657817695110739185591505062971929859314E3L, -1.655310302737837556654146291646499062882E2L, }, qq[6] = { 8.745588372054466262548908189000448124232E6L, 3.746038264792471129367533128637019611485E6L, 7.066358783162407559861156173539693900031E5L, 7.448928604824620999413120955705448117056E4L, 4.511583986730994111992253980546131408924E3L, 1.368902937933296323345610240009071254014E2L, /* 1.000000000000000000000000000000000000000E0 */ }, /* * Coefficients for approximation to erf in [0.84375,1.25] */ /* erf(x+1) = 0.845062911510467529296875 + pa(x)/qa(x) -0.15625 <= x <= +.25 Peak relative error 8.5e-22 */ pa[8] = { -1.076952146179812072156734957705102256059E0L, 1.884814957770385593365179835059971587220E2L, -5.339153975012804282890066622962070115606E1L, 4.435910679869176625928504532109635632618E1L, 1.683219516032328828278557309642929135179E1L, -2.360236618396952560064259585299045804293E0L, 1.852230047861891953244413872297940938041E0L, 9.394994446747752308256773044667843200719E-2L, }, qa[7] = { 4.559263722294508998149925774781887811255E2L, 3.289248982200800575749795055149780689738E2L, 2.846070965875643009598627918383314457912E2L, 1.398715859064535039433275722017479994465E2L, 6.060190733759793706299079050985358190726E1L, 2.078695677795422351040502569964299664233E1L, 4.641271134150895940966798357442234498546E0L, /* 1.000000000000000000000000000000000000000E0 */ }, /* * Coefficients for approximation to erfc in [1.25,1/0.35] */ /* erfc(1/x) = x exp (-1/x^2 - 0.5625 + ra(x^2)/sa(x^2)) 1/2.85711669921875 < 1/x < 1/1.25 Peak relative error 3.1e-21 */ ra[] = { 1.363566591833846324191000679620738857234E-1L, 1.018203167219873573808450274314658434507E1L, 1.862359362334248675526472871224778045594E2L, 1.411622588180721285284945138667933330348E3L, 5.088538459741511988784440103218342840478E3L, 8.928251553922176506858267311750789273656E3L, 7.264436000148052545243018622742770549982E3L, 2.387492459664548651671894725748959751119E3L, 2.220916652813908085449221282808458466556E2L, }, sa[] = { -1.382234625202480685182526402169222331847E1L, -3.315638835627950255832519203687435946482E2L, -2.949124863912936259747237164260785326692E3L, -1.246622099070875940506391433635999693661E4L, -2.673079795851665428695842853070996219632E4L, -2.880269786660559337358397106518918220991E4L, -1.450600228493968044773354186390390823713E4L, -2.874539731125893533960680525192064277816E3L, -1.402241261419067750237395034116942296027E2L, /* 1.000000000000000000000000000000000000000E0 */ }, /* * Coefficients for approximation to erfc in [1/.35,107] */ /* erfc(1/x) = x exp (-1/x^2 - 0.5625 + rb(x^2)/sb(x^2)) 1/6.6666259765625 < 1/x < 1/2.85711669921875 Peak relative error 4.2e-22 */ rb[] = { -4.869587348270494309550558460786501252369E-5L, -4.030199390527997378549161722412466959403E-3L, -9.434425866377037610206443566288917589122E-2L, -9.319032754357658601200655161585539404155E-1L, -4.273788174307459947350256581445442062291E0L, -8.842289940696150508373541814064198259278E0L, -7.069215249419887403187988144752613025255E0L, -1.401228723639514787920274427443330704764E0L, }, sb[] = { 4.936254964107175160157544545879293019085E-3L, 1.583457624037795744377163924895349412015E-1L, 1.850647991850328356622940552450636420484E0L, 9.927611557279019463768050710008450625415E0L, 2.531667257649436709617165336779212114570E1L, 2.869752886406743386458304052862814690045E1L, 1.182059497870819562441683560749192539345E1L, /* 1.000000000000000000000000000000000000000E0 */ }, /* erfc(1/x) = x exp (-1/x^2 - 0.5625 + rc(x^2)/sc(x^2)) 1/107 <= 1/x <= 1/6.6666259765625 Peak relative error 1.1e-21 */ rc[] = { -8.299617545269701963973537248996670806850E-5L, -6.243845685115818513578933902532056244108E-3L, -1.141667210620380223113693474478394397230E-1L, -7.521343797212024245375240432734425789409E-1L, -1.765321928311155824664963633786967602934E0L, -1.029403473103215800456761180695263439188E0L, }, sc[] = { 8.413244363014929493035952542677768808601E-3L, 2.065114333816877479753334599639158060979E-1L, 1.639064941530797583766364412782135680148E0L, 4.936788463787115555582319302981666347450E0L, 5.005177727208955487404729933261347679090E0L, /* 1.000000000000000000000000000000000000000E0 */ }; static long double erfc1(long double x) { long double s,P,Q; s = fabsl(x) - 1; P = pa[0] + s * (pa[1] + s * (pa[2] + s * (pa[3] + s * (pa[4] + s * (pa[5] + s * (pa[6] + s * pa[7])))))); Q = qa[0] + s * (qa[1] + s * (qa[2] + s * (qa[3] + s * (qa[4] + s * (qa[5] + s * (qa[6] + s)))))); return 1 - erx - P / Q; } static long double erfc2(uint32_t ix, long double x) { union ldshape u; long double s,z,R,S; if (ix < 0x3fffa000) /* 0.84375 <= |x| < 1.25 */ return erfc1(x); x = fabsl(x); s = 1 / (x * x); if (ix < 0x4000b6db) { /* 1.25 <= |x| < 2.857 ~ 1/.35 */ R = ra[0] + s * (ra[1] + s * (ra[2] + s * (ra[3] + s * (ra[4] + s * (ra[5] + s * (ra[6] + s * (ra[7] + s * ra[8]))))))); S = sa[0] + s * (sa[1] + s * (sa[2] + s * (sa[3] + s * (sa[4] + s * (sa[5] + s * (sa[6] + s * (sa[7] + s * (sa[8] + s)))))))); } else if (ix < 0x4001d555) { /* 2.857 <= |x| < 6.6666259765625 */ R = rb[0] + s * (rb[1] + s * (rb[2] + s * (rb[3] + s * (rb[4] + s * (rb[5] + s * (rb[6] + s * rb[7])))))); S = sb[0] + s * (sb[1] + s * (sb[2] + s * (sb[3] + s * (sb[4] + s * (sb[5] + s * (sb[6] + s)))))); } else { /* 6.666 <= |x| < 107 (erfc only) */ R = rc[0] + s * (rc[1] + s * (rc[2] + s * (rc[3] + s * (rc[4] + s * rc[5])))); S = sc[0] + s * (sc[1] + s * (sc[2] + s * (sc[3] + s * (sc[4] + s)))); } u.f = x; u.i.m &= -1ULL << 40; z = u.f; return expl(-z*z - 0.5625) * expl((z - x) * (z + x) + R / S) / x; } long double erfl(long double x) { long double r, s, z, y; union ldshape u = {x}; uint32_t ix = (u.i.se & 0x7fffU)<<16 | u.i.m>>48; int sign = u.i.se >> 15; if (ix >= 0x7fff0000) /* erf(nan)=nan, erf(+-inf)=+-1 */ return 1 - 2*sign + 1/x; if (ix < 0x3ffed800) { /* |x| < 0.84375 */ if (ix < 0x3fde8000) { /* |x| < 2**-33 */ return 0.125 * (8 * x + efx8 * x); /* avoid underflow */ } z = x * x; r = pp[0] + z * (pp[1] + z * (pp[2] + z * (pp[3] + z * (pp[4] + z * pp[5])))); s = qq[0] + z * (qq[1] + z * (qq[2] + z * (qq[3] + z * (qq[4] + z * (qq[5] + z))))); y = r / s; return x + x * y; } if (ix < 0x4001d555) /* |x| < 6.6666259765625 */ y = 1 - erfc2(ix,x); else y = 1 - 0x1p-16382L; return sign ? -y : y; } long double erfcl(long double x) { long double r, s, z, y; union ldshape u = {x}; uint32_t ix = (u.i.se & 0x7fffU)<<16 | u.i.m>>48; int sign = u.i.se >> 15; if (ix >= 0x7fff0000) /* erfc(nan) = nan, erfc(+-inf) = 0,2 */ return 2*sign + 1/x; if (ix < 0x3ffed800) { /* |x| < 0.84375 */ if (ix < 0x3fbe0000) /* |x| < 2**-65 */ return 1.0 - x; z = x * x; r = pp[0] + z * (pp[1] + z * (pp[2] + z * (pp[3] + z * (pp[4] + z * pp[5])))); s = qq[0] + z * (qq[1] + z * (qq[2] + z * (qq[3] + z * (qq[4] + z * (qq[5] + z))))); y = r / s; if (ix < 0x3ffd8000) /* x < 1/4 */ return 1.0 - (x + x * y); return 0.5 - (x - 0.5 + x * y); } if (ix < 0x4005d600) /* |x| < 107 */ return sign ? 2 - erfc2(ix,x) : erfc2(ix,x); y = 0x1p-16382L; return sign ? 2 - y : y*y; } #elif LDBL_MANT_DIG == 113 && LDBL_MAX_EXP == 16384 // TODO: broken implementation to make things compile long double erfl(long double x) { return erf(x); } long double erfcl(long double x) { return erfc(x); } #endif
__label__pos
0.662089
#!/bin/sh # Copyright (C) 1999 Nathaniel Smith # Copyright (C) 1999, 2000, 2001 Robert Woodcock # Copyright (C) 2003, 2005 Jesus Climent # This code is hereby licensed for public consumption under either the # GNU GPL v2 or greater, or Larry Wall's Artistic License - your choice. # # You should have recieved a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # Copyright for this work is to expire January 1, 2010, after which it # shall be public domain. # TODO: # - Add more error checking # KNOWN BUGS: # - Not much error checking, esp. of arguments # - Submitted via: line is created by template, when it really should be in send. # Oh well. VERSION=0.4.7 NAME=cddb-tool #return codes BAD_SYNTAX_ERR=10 # invalid CDDB file NO_TMP_ERR=11 # can't create a temp file NO_MATCH_ERR=12 # try submitting one LOOKUP_ERR=13 # problem connecting to cddb server EMPTY_QUERY_RESPONSE=14 # query response = "", (probably no net connection) # assume a reasonable default if $HTTPGET is undefined if [ "$HTTPGET" = "" ]; then if [ X"$(uname)" = X"FreeBSD" ] ; then HTTPGET=fetch HTTPGETOPTS=${HTTPGETOPTS:="-q -o -"} elif [ X"$(uname)" = X"NetBSD" ] ; then HTTPGET=ftp HTTPGETOPTS=${HTTPGETOPTS:="-a -V -o - "} elif [ X"$(uname)" = X"Darwin" ] ; then HTTPGET=curl HTTPGETOPTS=${HTTPGETOPTS:="-f -s"} else HTTPGET=wget HTTPGETOPTS=${HTTPGETOPTS:="-q -nv -e timestamping=off -O -"} fi fi HTTPGET="$HTTPGET $HTTPGETOPTS" usage() { cat << EOF $NAME version $VERSION usage: one of: $0 parse file $0 template disc-id tracks $0 send file address $0 read server proto user host disc-id genre $0 query server proto user host disc-id tracks $0 stat serverurl user host proto $0 help EOF } help() { cat << EOF $NAME version $VERSION A toolbox for doing cddb related stuff Usage: $0 command [command_options] Commands: parse file Get data out of a cddb file - dumps to stdout in a form source'able by the shell send file address Mails a CDDB file to a specified address, using correct format. Category should be one of blues, classical, country, data, folk, jazz, newage, reggae, rock, soundtrack, or misc. template disc-id tracks Generates a template (empty) cddb file to stdout. The command line should be essentially the output of cd-discid. query server proto user host disc-id tracks Looks up disc on server (should be of form "http://host/~cddb/cddb.cgi") remainder of command line is in the same form as that returned by the cd-discid program. read server proto user host disc-id genre CDDB file is dumped to stdout. File will contain an extra #CATEGORY= line, which leaves it a valid CDDB file but which will be recognized by parse and send commands. Uses wget, so if you need to use a proxy then just configure wget to do so. user and host will be used for identifying ourselves to the CDDB server. stat serverurl user host proto Check server status with given protocol. This can be used to check if the server supports given protocol. Most common values for proto should be 5 and 3. With 3 you will not get DYEAR and DGENRE fields in response. help Display this. EOF } f_seq () { i=$1 while [ $i -ne `expr $2 + 1` ] do echo $i i=`expr $i + 1` done } COMMAND=$1 shift case $COMMAND in parse) # takes 1 argument, a filename, and dumps out a sh parseable version CDDBFILE="$1" set -e # names chosen to match usage in abcde code DISCID=$(grep ^DISCID= "$CDDBFILE" | cut -f2 -d= | tr -d \[:cntrl:\]) DARTISTALBUM="$(grep ^DTITLE= "$CDDBFILE" | cut -f2- -d= | tr -d \\n | sed 's- / -~-g' | tr -d \[:cntrl:\])" DARTIST="$(echo "$DARTISTALBUM" | cut -f1 -d~ | sed 's,\\,\\\\,g;s,\([\"\$\`]\),\\\1,g' | tr -d \[:cntrl:\])" DALBUM="$(echo "$DARTISTALBUM" | cut -f2 -d~ | sed 's,\\,\\\\,g;s,\([\"\$\`]\),\\\1,g' | tr -d \[:cntrl:\])" CDDBGENRE="$(grep '^#CATEGORY=' "$CDDBFILE" | cut -f2- -d= | tr -d \[:cntrl:\])" if grep "^DYEAR" "$CDDBFILE" 2>&1 > /dev/null ; then CDYEAR=$(grep "^DYEAR" "$CDDBFILE" | cut -f2- -d= | tr -d \[:cntrl:\]) elif grep YEAR "$CDDBFILE" 2>&1 > /dev/null ; then CDYEAR=$(grep "YEAR" "$CDDBFILE" | grep -v "DYEAR" | awk 'BEGIN{FS="YEAR:"}{print $2}' | awk '{print $1}') else CDYEAR="" fi CDGENRE=$(grep '^DGENRE=' "$CDDBFILE" | cut -f2- -d= | tr -d \[:cntrl:\]) set +e echo DISCID="\"$DISCID\"" echo DALBUM="\"$DALBUM\"" echo DARTIST="\"$DARTIST\"" echo CDDBGENRE="\"$CDDBGENRE\"" echo CDYEAR="\"$CDYEAR\"" echo CDGENRE="\"$CDGENRE\"" NUMTRACKS=$(grep -E '^TTITLE[0-9]+=' "$CDDBFILE" | wc -l) CURRTRACK=0 while [ "$CURRTRACK" -lt $NUMTRACKS ]; do CURRTRACKM1=$CURRTRACK # Track minus 1 (cddb numbers from 0) CURRTRACK=$(expr $CURRTRACK + 1) echo -n "TRACK${CURRTRACK}=\"" grep ^TTITLE${CURRTRACKM1}= "$CDDBFILE" | cut -f2 -d= | sed 's,\\,\\\\,g;s,\([\"\$\`]\),\\\1,g' | tr -d \[:cntrl:\] echo \" done ;; template) DISCID="$@" DISCNUM=$1 echo '# xmcd CD database file' echo '#' echo '# Track frame offsets:' NUMTRACKS=$2 for x in $(f_seq 3 $(expr $NUMTRACKS + 2)) do printf "#\t$(echo "$DISCID" | cut -f$x -d' ')\n" done x=$(expr $x + 1) LENGTH=$(echo "$DISCID" | cut -f$x -d' ') echo "#" echo "# Disc length: $LENGTH seconds" echo "#" echo "# Submitted via: $NAME $VERSION" echo "#" echo "#blues,classical,country,data,folk,jazz,newage,reggae,rock,soundtrack,misc" echo "#CATEGORY=misc" echo DISCID="$DISCNUM" echo "DTITLE=Unknown Artist / Unknown Album" echo "DYEAR=" echo "DGENRE=" # TTITLE0 -- TTITLEn for x in $(f_seq 1 $NUMTRACKS) do echo "TTITLE$(expr $x - 1)=Track $x" done echo "EXTD=" # EXTT0 -- EXTTn for x in $(f_seq 1 $NUMTRACKS) do echo "EXTT$(expr $x - 1)=" done echo "PLAYORDER=" ;; send) # cddb-tool send filename email@address FILE="$1" ADDRESS="$2" DISCID=$(grep ^DISCID= "$FILE" | cut -f2 -d= | tr -d \[:cntrl:\]) CDDBGENRE=$(grep '^#CATEGORY=' "$FILE" | cut -f2- -d= | tr -d \[:cntrl:\]) grep -v "^#CATEGORY=" "$FILE" | iconv -t utf-8 | mail -a "Content-Type: text/plain; charset=utf-8" -s "cddb $CDDBGENRE $DISCID" "$ADDRESS" ;; query) # cddb-tool query serverurl proto user host discid... SERVER="$1" PROTO="$2" USER="$3" HOST="$4" HELLOINFO="$USER+$HOST+$NAME+$VERSION" shift 4 TRACKINFO="$@" TRACKINFOPLUS=$(echo $TRACKINFO | tr ' ' '+') RESULTS=$($HTTPGET "$SERVER?cmd=cddb+query+$TRACKINFOPLUS&hello=$HELLOINFO&proto=$PROTO") || exit $LOOKUP_ERR echo "$RESULTS" | tr '\r' '\n' | tr -s '\n' | sed 's/^ //g' ;; read) # cddb-tool read serverurl proto user host genre discnumber SERVER="$1" PROTO="$2" USER="$3" HOST="$4" CATEGORY="$5" DISCID="$6" HELLOINFO="$USER+$HOST+$NAME+$VERSION" $HTTPGET $CDDBDATA "$SERVER?cmd=cddb+read+$CATEGORY+$DISCID&hello=$HELLOINFO&proto=$PROTO" 2>/dev/null ;; stat) # cddb-tool stat serverurl user host proto SERVER="$1" USER="$2" HOST="$3" PROTO="$4" HELLOINFO="$USER+$HOST+$NAME+$VERSION" $HTTPGET $CDDBDATA "$SERVER?cmd=stat&hello=$HELLOINFO&proto=$PROTO" 2>/dev/null ;; help) # help help ;; *) # usage usage ;; esac
__label__pos
0.531617
REST and SOAP APIs in eCommerce The eCommerce industry is booming. It is not limited to online shopping anymore but is focused on providing exceptional experiences to online users. Considering the demands of online shoppers and the operations involved in managing an eCommerce business, the role of eCommerce software can be significantly realized. However, it is imperative for such software providers to integrate with eCommerce platforms in order to allow e-retailers to manage their business seamlessly. Such integration is established with the APIs of those platforms, categorized into two — SOAP API and REST API. The significance of the two APIs has been in discussion for quite a long time, and hence, we have gathered some important information to share with you. What is API and How it Works? An API or application programming interface is often termed as an intermediary that establishes communication between two different applications. To put it simply, API is a messenger that carries information from one application to another via typical request-response methods. The most commonly used APIs today include SOAP API and REST API, with each of them having its own significance. With API, developers are free to reuse a piece of code for complex processes making the overall workflow a lot easier. The pace at which APIs help you develop applications is unmatched by the conventional methods of development. The concept of API can be better understood by considering the example of Facebook. When you open Facebook on your phone and try to login into your account, you are basically making a call to the Facebook server through the app. You are asking the server to access your account. This call is known as an API request. The mobile application contacts the Facebook app, which then connects with the Facebook server and asks for your account credentials. When the server returns all the details, the Facebook application returns the details to your mobile app. Such a process is known as API response. If to talk about the eCommerce sphere, each of the shopping platforms has its own API. To be able to interact with any platform and access its data (e.g., order data, product data, etc.), software developers have to build the connection between their systems and the platform’s API. Without such integration, the automatic transfer of data between the software and the platform is impossible. SOAP API: Meaning and Examples SOAP API or Simple Object Access Protocol, as the name suggests, is a communication protocol that allows different processes using different operating systems to interact with each other with the help of HTTP. SOAP supports only XML data format. SOAP API is considered to be more secure than the REST API. Using a SOAP API, the applications can easily create, delete, update, and recover custom objects in the software. One of the best examples of SOAP API in the eCommerce sphere is Magento SOAP API. Magento SOAP API is a set of HTTP requests which delivers response messages to the calls received on the server. These messages are in the form of XML and several other formats. The API helps in managing the eCommerce stores on Magento by providing different types of calls to work with data related to products, categories, customers, etc. REST API: Meaning and Examples REST API follows the rules of REST (Representational State Transfer) architectural style. It defines a work environment for components, data elements, and connectors for using URLs and HTTP protocols. The following operations can be performed with the help of a REST API: • GET — retrieve data from eCommerce stores • PUT — update data on eCommerce stores • POST — create data on eCommerce stores • DELETE — remove data from eCommerce stores Some common examples of the REST API in the eCommerce sphere include: • Squarespace REST APIs Squarespace provides a group of APIs for eCommerce software developers to manage online stores built on Squarespace. All the APIs are designed using REST principles on HTTP. They are: 1. Inventory API — allows to read and adjust product and product variant stock 2. Orders API — allows retrieving order history and subscription orders and import orders from sales channels 3. Products API — allows managing products along with their variants and images 4. Profiles API — enable reading and managing customers 5. Transactions API — enable access to financial transitions 6. Webhook Subscription API — allows subscribing to site notifications • Etsy REST API Etsy REST API provides a simple interface that follows all the REST rules and JSON-formatted responses to API requests. It gives both read and write access to users’ data, including listings, shops, tags, profiles, and sales information. When connecting with Etsy REST API, developers can easily access and manage their client stores based on Etsy. • WooCommerce REST API WooCommerce API allows developers to build applications that can work with data-related orders, coupons, products, customers, etc., on all the stores built on WooCommerce. By default, it supports JSON format, but it can be changed to XML if required. It also works with four methods, including GET, POST, PUT, and DELETE. • REST API by API2Cart API2Cart provides a unified API built on REST principles for eCommerce software developers to help them integrate their applications with more than 40 eCommerce platforms and marketplaces at once. The list of the supported platforms includes Magento, Shopify, Amazon, eBay, etc. Such integration helps vendors open new opportunities and helps eCommerce software manage their clients’ e-stores seamlessly. Common and Different Features of REST and SOAP APIs Despite being significant in their ways, both REST API and SOAP API share some similarities and apparent differences. Similarities: 1. Both REST and SOAP API follow HTTP protocol. 2. They rely on a well-defined set of rules for information exchange. 3. They connect to two applications with the help of server-side data, which is both human and machine-readable. 4. Both REST and SOAP APIs utilize methods like GET, PUT, POST, and DELETE. 5. Both of them understand XML and can use it in requests and responses. Differences: 1. REST is an architectural style, and therefore REST API has no official standard, but since SOAP is a protocol, SOAP API has an official standard. 2. REST API can support multiple standards, including HTTP, JSON, XML, and URL. But, SOAP uses only two — HTTP and XML. 3. REST API takes fewer resources and bandwidth as compared to SOAP which results in larger file sizes. 4. REST API leverages URL exposure — @path(“/WeatherService”), whereas SOAP uses services interface — @WebService. 5. REST API when doesn’t make any emphasis on standards results in a corrupt API. But SOAP API results in errors only since it defines too many standards for the implementer. 6. REST API uses WADL (Web Application Description Language), and SOAP API uses WSDL (Web Services Description Language). Conclusion Both REST API and SOAP API are significant for eCommerce software developers to make their software fully functional. However, it depends on the platform on which your client has built their e-store. Hence, understand your client’s requirements and the compatibility of their eCommerce platform before choosing between REST or SOAP API.
__label__pos
0.838814
Announcement Collapse No announcement yet. OnChange trigger hide/show and vuelue updates: bug or am i too stupid? Collapse X • Filter • Time • Show Clear All new posts • OnChange trigger hide/show and vuelue updates: bug or am i too stupid? edit: cant change heading. must be "value" hi there, I go mad. In one form I have a pretty complicated sheme of showing, hiding and updating values for some fields. background is this: some of our customers have one single person as contact person, other have two. so I build a form like this: 1. one select field: {Anrede} (German for adress, to adress somebody) "Herr", "Frau", Familie", "Frau und Herr", "Frau und Frau", "Herr und Herr" and "unbekannt" (english: Mr, Mrs, Family, Mrs and Mr... till "unknown) 2. I build a Label field (" {begr} ") which is empty in the beginning. It contains later the full salutation as a preview ("Dear Mr. Potter,") 3. I build three Text-Fields: Name, Name 1 and Name2 which arent shown in the beginning. What it should do: When I select "Mr", "Mrs" or "Familie" the field {name} must be shown, name1 and name2 must be hidden (note that its totaly correct in germany to say "Dear family Potter) When I selct "unknown" all name-fields must be hidden When I select "Mr and Mrs", "Mrs and Mrs" or "Mr and Mr" Name1 and Name2 must be shown, name 1 must be hidde. I use onclick for trigger ajax on {Anrede}-Select. It works. the following is complicated. 1. name on change: when name changes value, {begr} - label - field must be "Dear {Anrede} {Name}" (concret in german its more complecated, because you say "Sehr geehrter Herr..." for man and "Sehr geehrte Frau ..." for woman, but ok) 2. when name=="" (changes to be deleted) begr must be "Sehr geehrte Damen und Herren" (Dear Ladies and Gentleman) without Names AND selct-field {Anrede} must be changed to "unknown". Also Name, Name1 and Name2 must be hidden. 3. entering something in name1, than name2 must appear. 4. when name2 appears and something is written in, {begr} must be "Dear Mr. Poter, Dear Mr James" or st like that and so on. when name2 is deleted, name1 and name2 must be unshown, select must be "Mister" and field name must appear. So. Showing and hiding field work pretty well. but when it comes to update the {begr} Label SC strikes. Some update are done, some are just ignored. For example: When the fiel {name} something, p.e. "Potter", the {begr} - Field is correct with "Dear mister Potter" (or Mrs when Mrs is selcted) But than: I clear the {name} field. the new {begr} must be "Dear Ladies and Gentleman" and the new selct must be "unknown". But it doesnt update. And this is funny, because I just entered a showing-message-marcro right before the update-command. So I know the parser is in the correct place in the ajax-event "on change name": Code: if ({name}==") { sc_field_display({name}, 'off'); sc_field_display({name1}, 'off'); sc_field_display({name2}, 'off'); {Anrede}="unbk"; sc_ajax_message("Here we are", "Testmessage"); {begr}="Sehr geehrte Damen und Herren"; } The fields are hidden completly correct. the message is shown. but the begr-label-field has still "Dear Mr Potter". this happens a lot when i play around with the field. can the content of a label-field not be changed more than - lets say - 5 times inside a ajax-event? is "name" something i cant use as fieldname, because it hast other, sc-meanings? I go mad. I am pretty sure that my "If---" lines in all these ajax-event are right, because the showing and hiding-field - stuff works. and mostly the begr-update works, too. but at some point of the script, p.e. when i delete a name, the begr-updates a just ignored. someone experience with that behavior of SC? Last edited by nina_sorensen; 11-08-2013, 11:52 AM. • #2 ** deleted post ** Albert Drent aducom software netherlands scriptcase partner, reseller, support and (turn-key) development www.scriptcase.eu / www.scriptcase.nl Comment Working... X
__label__pos
0.968241
Ideas on Enterprise Information Systems Development This blog is devoted to ideas on Enterprise Information Systems (EIS) development. It focuses on Lean Thinking, Agile Methods, and Free/Open Source Software, as means of improving EIS development and evolution, under a more practical than academical view. You may find here a lot of "thinking aloud" material, sometimes without scientific treatment... don't worry, this is a blog! Every post is marked with at least one of Product or Process labels, meaning that they are related to execution techniques (programming and testing) or management techniques (planning and monitoring), respectively. Thursday, June 9, 2011 Enterprise Information Systems Patterns - Part X EIS Patterns, Business Processes, and Workflow Engines Every EIS is based on business processes, and it can be said that workflows are the computational realization of these processes. Therefore it is necessary to provide some workflow engine to EIS Patterns. Basically, there are two types of workflow engines: a) Associated to an application development platform: such as the ones found in Web application development environments, like Zope, or associated to databases, such as Oracle, or as part of Groupware environments. In this case, applications are developed "inside" the environment, using the engine's API. b) Implemented as Libraries: using some programming language, a library of objects representing the workflow features is implemented, such as the State Machine as Python Decorators, or Ruby's AASM. In this case, you must use the libraries' objects to implement your engine. Enters Fluidity & Extreme Fluidity Fluidity was developed as a type (b) workflow engine, however, Extreme Fluidity turns it into a third type of workflow engine: c) Inoculated & Expelled: using the dynamic nature of the underlying programming language, provide a way of making objects workflow-aware, and symmetrically, a way of turning these objects back to their initial structure when desired. Why this? Because type (a) engines forces you to use a given environment to develop your applications, while type (b) forces you to use specific objects to implement workflows, most of times creating a mixed code with application specific and workflow specific statements. With Extreme Fluidity (or xFluidity for short) it is possible to insert the code necessary to make your workflow run inside your business objects. In other words, they become workflow-aware objects, while keeping your programming style, standards, and patterns. In order to understand how this happens, let's check how Fluidity can be used when xFluidity is associated to it (the complete source code can be found here). First, it is necessary to declare a template State Machine: class LoanProcess(StateMachine):      state('requested')      state('request_created')      state('request_analyzed')      state('refusal_letter_sent')      state('loan_created')      state('value_transfered')      initial_state = 'requested'      transition(from_='requested', event='create_loan_request', to='request_created')      transition(from_='request_created', event='analyst_select_request', to='request_analyzed')      transition(from_='request_analyzed', event='loan_refused', to='refusal_letter_sent')      transition(from_='request_analyzed', event='loan_accepted', to='loan_created')      transition(from_='loan_created', event='time_to_transfer_value', to='value_transfered') Using xFluidity, the template is injected in another object - in this case a Process:     def setUp(self):         self.process = Process()         template = LoanProcess()         configurator = StateMachineConfigurator(template)         configurator.configure(self.process) Magically, the process object starts to respond to the state machine's events:     def it_makes_the_process_respond_to_the_example_state_machine_events(self):         self.process |should| respond_to('create_loan_request')         self.process |should| respond_to('analyst_select_request')         self.process |should| respond_to('loan_refused')         self.process |should| respond_to('loan_accepted')         self.process |should| respond_to('time_to_transfer_value') As expected, the engine avoids the firing of transitions at wrong situations:     def it_runs_the_example_acceptance_path(self):         self.process.create_loan_request()         self.process.current_state() |should| equal_to('request_created') # the line below means that if the process tries to fire loan_refused # when at 'request_created' state, an exception is generated         self.process.loan_refused |should| throw(InvalidTransition)         self.process.analyst_select_request()         self.process.current_state() |should| equal_to('request_analyzed')         #loan accepted         self.process.loan_accepted()         self.process.current_state() |should| equal_to('loan_created')         self.process.time_to_transfer_value()         self.process.current_state() |should| equal_to('value_transfered') In that way, it is possible to use a workflow engine in your own development environment and using your own models. Fluidity also provides guards and state actions, not shown here for the sake of simplicity. A Note on Workflow Notations Fluidity is a state-based engine, which on those days of fancy business process notations may appear a bit old-fashioned. However, state machines represent a well known and easy to understand notation. EIS Patterns' Process and Movements objects are being developed in a way that they don't need to know which notation is in use, at least at the unit level - meaning that some code specific to a given notation may be used to "glue" elements at the application level. Conclusion Currently Fluidity and xFluidity are works in progress, including the expelling capabilities and some details on implementation of guards. When these features become fully functional, a Petri Net based engine will be developed. A clear limitation of Fluidity is that it only works for Python objects, but the necessity of developing in a given programming language is a limitation of all workflow engines, although some provide XML-like APIs.
__label__pos
0.916258
Take the 2-minute tour × MathOverflow is a question and answer site for professional mathematicians. It's 100% free, no registration required. Consider an $n\times n$ random binary matrix $M$ with i.i.d. entries $m_{ij} \sim {\rm Bernoulli}(p)$, where $p = n^{-\beta}$ with $\beta \in (1,2)$. I am interested in the behavior of the singular value decomposition of $$M = \sum_{i=1}^{rank(M)} \sigma_i u_i v_i',$$ where $\sigma_i$ are ranked in decreasing order. Some intuitive observation (which might NOT be all true!): 1) By subtracting the mean we can write $M = p {\bf 1} {\bf 1}' + A$, where $A$ has indepedent entries with zero mean and variance $p(1-p)$. Therefore I expect the leading singular vector is approximately parallel to the all-one vector ${\bf 1}$, and the largest singular value is $\sigma_1 \approx p n$. If the SVD of $A$ behaves similarly to that of the usual iid matrices, it is probably true that the second largest singular value of $M$ (i.e., the largest singular value of $A$) is approximately $\sigma_2 \approx \sqrt{p n}$. 2) $rank(M)$ is pretty small: since $\mathbb{P}(\text{the first}~ m \text{ rows are all zero}) = (1-p)^{n m}\geq 1-pmn$. Therefore $rank(M) \leq n^{\beta-1}$ with high probability. This is wrong... this only says that $rank(M) \leq n-n^{\beta-1}$. Are there any rigorous results about the SVD of this matrix ensemble? Is it true that except for $u_1,v_1$ which are approximately $\frac{1}{\sqrt{n}} {\bf 1}$, the remaining singular vectors are independently and uniformly distributed over $S^{n-1}$? share|improve this question 1   Presumably you mean "VERY sparse random matrices" in the title? –  Mark Meckes Jan 20 '12 at 18:02      Thanks! Fixed. –  mr.gondolier Jan 20 '12 at 18:27 2 Answers 2 Also in the "a little too involved for a comment" class: A matrix that's this sparse is usually going to be a block diagonal matrix with very small blocks. Let $k$ be any fixed constant, and suppose that your matrix contains no $k+1 \times k+1$ principal submatrix with at least $k$ nonzero entries. Then your matrix has a block decomposition with all blocks having size at most $k+1$ (If you start with a given nonzero entry and "grow" it by adding entries in the same row/column as an entry already added, you'll stop by the time you've reached $k-1$ added entries, and each addition increases the size of your block by at most $1$). In this case we can upper bound the probably a submatrix of this size exists by $$\binom{n}{k+1} \binom{(k+1)^2}{k} p^{k} \leq C_k n^{k+1-k \beta}$$ Which goes to $0$ for any $\beta$ in your range if $k$ sufficiently large. This means you'll usually see singular vectors with very small support, and that the $\sigma_2$ should be much larger than $\sqrt{np}$ (maybe equal to $\sigma_1$ in most cases?) You might be able to get something more explicit by enumerating all the possible structures of blocks and the expected number of occurrences of each. share|improve this answer      Thanks Kevin. I understand your calculation of the probability but I didn't quite get what you meant by block decomposition. Can you elaborate a bit? –  mr.gondolier Jan 21 '12 at 1:09      Also, what do you think about $rank(M)$? Your thesis seems to deal with the case with $p>=\log n / n$, i.e., $\beta \leq 1$ and rank is full whp. Here probably rank is roughly number of non-zero rows/columns? –  mr.gondolier Jan 21 '12 at 1:11      A Block Decomposition here means partitioning the indices from $1$ to $n$ into blocks $A_1, A_2, \dots A_m$ such that $a_{ij}$ is zero if $i$ and $j$ are in different blocks. It corresponds to an ordering of the indices where the matrix is block diagonal. You can take a set of singular vectors/values coming from the singular vectors/values of each block. –  Kevin P. Costello Jan 21 '12 at 4:05      If $\beta>3/2$ the rank should be equal to the number of nonzero rows/columns, just because with high probability no row/column has two nonzero entries. For smaller $\beta$, I think they should still be nearly equal, but not exactly. You should be able to bound the difference by something like the number of nonzero entries in the same row/column as another nonzero entry, which is on the order of $n^{3-2\beta}$, compared to $n^{2-\beta}$ rows with nonzero entries). –  Kevin P. Costello Jan 21 '12 at 4:19      But on the other hand there should be pairs of columns each having exactly one nonzero entry both in the same row, and vice versa, leading to dependencies among the nonzero rows/columns. –  Kevin P. Costello Jan 21 '12 at 4:20 Here's just one quick remark (a little involved for a comment) about how you can start making greater use of your suggested reduction to $A$. Since $M-A$ has rank 1, $$ \sigma_3(A) \le \sigma_2(M) \le \sigma_1(A). $$ Classical results imply that $\sigma_1(A), \sigma_3(A) \approx \sqrt{pn}$, so $\sigma_2(M) \approx \sqrt{pn}$ too. For more, this paper by Van Vu probably (I haven't read it yet) has a lot that's relevant to your questions. share|improve this answer      could you give a bit more details on the classical result, because here the entrywise distribution of $A$ varies with the dimension, so presumably you need some non-asymptotic bounds on $\sigma_{\cdot}(A)$ ? –  mr.gondolier Jan 20 '12 at 18:29      Ack, that's what I get for writing in a hurry. You're right, classical results don't apply directly. I'll think some more and try to write again. –  Mark Meckes Jan 20 '12 at 18:40 Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.843651
Search Images Maps Play YouTube News Gmail Drive More » Sign in Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader. Patents 1. Advanced Patent Search Publication numberUS5740358 A Publication typeGrant Application numberUS 08/478,354 Publication date14 Apr 1998 Filing date7 Jun 1995 Priority date25 Jun 1992 Fee statusPaid Also published asUS6332182 Publication number08478354, 478354, US 5740358 A, US 5740358A, US-A-5740358, US5740358 A, US5740358A InventorsJohn S. Geldman, Son H. Ho, Petro Estakhri, John J. Schadegg Original AssigneeCirrus Logic, Inc. Export CitationBiBTeX, EndNote, RefMan External Links: USPTO, USPTO Assignment, Espacenet Defect management and split field processing in disk storage systems US 5740358 A Abstract An apparatus and method for disk-sector layout, formatting, reading and writing, is based on a flexible formatter microengine that is driven by parameter lists, which may include commands to dynamically down load the microengine's writable control store during disk access operations. This reduces the intervention required by the local microprocessor to support exception handling, including defect management, thus decreasing total access time for typical disk access (format, write or read) operations. This also reduces or eliminates the area on the disk required for ID fields, thus increasing the storage capacity of disk drives controlled by the present invention. This also increases disk space utilization, thus increasing the storage capacity of a disk drive controlled by using the present invention. Additionally, the invention is based on ID fields that may contain redundant information for increased fault tolerance. Additionally, the invention supports split data fields while minimizing the amount of additional information that the ID fields must contain about the data-field split sizes. Alternatively, the invention supports eliminating ID fields entirely with the prior-art functions of ID fields taken over by a combination of parameter lists and sector status flags. Images(6) Previous page Next page Claims(4) What is claimed is: 1. In a disk storage system for storing digital data, the storage system having at least one disk having a plurality of concentric tracks, each track being subdivided into a plurality of sectors, a method of processing data field sections within a sector, comprising the steps of: (a) providing a sector having an identification field and at least one data field, the data field having synchronization information and at least one data byte, the identification field having a first split field count that is indicative of a number of data bytes in a first data field; (b) providing a maximum split count and a sector data remaining count, the maximum split count being indicative of a maximum number of data bytes in the sector, and the sector data remaining count being indicative of a number of unprocessed data bytes remaining in the sector; (c) processing data bytes in a first split in accordance with the first split field count; (d) processing data bytes in subsequent splits in accordance with the maximum split field count; (e) decrementing the sector data remaining count corresponding to all data bytes; and (f) continuing steps (d) and (e) until the sector data remaining count is zero. 2. The method of claim 1, further including the steps of: detecting the beginning of each data field; processing data located in each data field; and detecting errors located in each data field and correcting the detected errors. 3. The method of claim 2, wherein in the step of providing a sector, the data segments in each data field includes a complete data segment and a portion of a complete data segment. 4. The method of claim 2, wherein the step of providing a sector, the data segments in each data field includes a portion of a complete data segment. Description This application is a continuation of application Ser. No. 07/905,588, now abandoned, filed Jun. 26, 1992, and is a continuation-in-part of application Ser. No. 07/904,804, now abandoned, Attorney Docket No. 14295.P015, entitled "Dynamic Control of Disk Sectors", by Petro Estakhri and John Schadegg, filed Jun. 25, 1992, and assigned to CIRRUS LOGIC, Inc., which is hereby incorporated by reference. FIELD OF THE INVENTION This invention relates generally to digital computer systems, particularly to mass storage subsystems. More particularly, the invention relates to layout of sectors within tracks of disk drives and to control of operations (such as format, write and read) that access such sectors. BACKGROUND OF THE INVENTION Computer systems often contain a storage device such as a disk drive. FIG. 1 shows a typical architecture of a disk subsystem. A disk drive subsystem can be described as a system of disk platter(s); analog read-channel and write-channel circuitry; an encoder/decoder; a servo engine including a mechanical mechanism to position the read/write heads on the disk platter(s) and an electronic mechanism to control the head position; a local microprocessor (sometimes referred to as a local microcontroller); local processor memory; buffer memory; and disk drive controller circuitry. It is the task of the disk drive controller to interface with the host central processor or microcomputer and correctly read and write user data onto the disk. DISK ORGANIZATION In a magnetic disk storage system, each disk platter is organized in tracks and sectors (also known as logical sectors). Each platter of a disk drive contains many concentric tracks. A track is an annular region of a given radius (actually, of a small but finite radial width) where data is stored. If both sides of a platter are used for storage or if the disk drive contains multiple platters, then multiple heads are required. A cylinder is the set of tracks that can be accessed by choosing among the heads without moving the heads to a new track position. Usually, it is necessary to transfer digital data onto the disk drive in sectors of a predetermined size. A sector is the smallest unit of user data that the host computer system can access (format, write or read) independently. A sector is a predetermined contiguous arc portion of a track having its own identification and the capability of magnetically storing a predetermined amount of digital data, for example, 512 bytes of user data. In hard disk drives, tracks and sectors are often physically pre-defined at the time of manufacture of the disk by the placement of servo burst signals embedded within the recording area of each track. These servo bursts are used to define, measure and control the running positions (inward versus outward) of the read/write heads on the disk drive. One servo burst per track is uniquely identified as the index servo burst, and these indices align across tracks. Each sector normally has a first part that contains the sector identification or ID fields. The ID field for each sector is typically only written each time a disk is actually formatted. One task of the disk controller is to initialize or format the disks prior to storing any user data. Formatting defines the area of a disk to be set up as zones, tracks, and sectors. An alternative to the above-described embedded servo systems is a dedicated servo system in which one track per cylinder is dedicated to containing only servo information. Many bytes in addition to the user-data bytes are actually necessary within a sector in order to reliably address the sector on the disk and to provide for detection and correction of errors that may occur in the read/storage/write process due to phenomena such as defects in the platter media, electrical noise, and variations in the rotational speed of the platter(s). Another task of the controller is to add error correction codes (ECC) and cycle redundancy checking (CRC) to the user data in order to detect and possibly correct data that has been corrupted because of disk damage or some other problem. Additionally, the controller meshes the slow read and write times of the disk with the high speed transfers to and from the host computer. Additionally, the controller must monitor and accurately control the timing of reading and writing of data off of and onto the disk. This has been accomplished by means of a high-speed format microengine, or formatter microengine, that executes microinstructions contained in a writable control store (WCS) contained in the controller. The format microengine provides the data and the control required to perform the transfer of parallel digital user data into formatted, addressed, and error-protected channel data that can be serially placed on a disk. The data that are normally written into a sector contain the following: an ID (identification) field that acts like an address, a data field, an error correction field, gaps that are necessary to allow the read/write channel electronics to switch from reading to writing; and predetermined bit and byte patterns that are required to recover the exact bit frequency (which varies somewhat due to variations in the rotational speed of the platter), bit phase, and alignment of bit-stream data into bytes. The ID field and the data field each make up a concatenated set of bytes within a sector. One common hard disk format that has been extensively used is to follow each servo burst with a gap, then a series of synch pulses to establish bit-time synchronization across read/write speed variations, then a mark byte to establish byte alignment within the bit stream, then a sector ID field that contains track number and sector number, then a gap, then a series of synch pulses to establish bit-time synchronization across read/write speed variations, then a mark byte to establish byte alignment within the bit stream, then a 512-byte data field, then a 7-byte ECC (Error Correction Code) field. In this format each sector is preceded by the servo burst. Servo bursts are written only at the time the disk is manufactured--portions are written slightly off the center of the track in order to allow the servo engine to keep the head centered on the track. Thus, during servo-burst time, the access operations of reading, writing and formatting must be suspended. One problem arising in modern disk drive system designs is lack of flexibility to accommodate the ever more complicated formats required to increase density, decrease latency, and improve reliability. In the prior art, magnetic disk drive sector ID information is generally provided once as part of a substantially larger preamble to the sector data, with error tolerance being provided by CRC bits appended to the sector ID information for error detection purposes. While the CRC keeps misreading errors very low, the ID information for a particular sector will be lost in the event of hard errors in the ID field reading. In the case of optical drives where error rates are much higher, it is common to repeat the entire ID field, including bit-synchronizing and byte-aligning information, a number of times so as to allow repeated ID field reading attempts during a single pass to greatly increase the likelihood of at least one good read of an ID field for each sector. This technique has not been adapted in the case of magnetic disk drives however, because of the relative amount of storage capacity that would have to be dedicated to the entire and repetitive ID fields. DEFECT MANAGEMENT If a sector is, or over time becomes, defective (i.e., unable to accurately reproduce the data written thereon), there must be a way to map out to another physical sector the data that the logical sector would otherwise contain, or alternatively to simply skip over the sector (called slipping the sector), which results in a track that is shorter than expected. There is a need for a way to handle defective sectors automatically, i.e., with minimal intervention from the local microprocessor to handle the exception conditions that occur when access is attempted to a slipped or mapped sector. Additionally, there is a need to handle such occurrences efficiently, (i.e., with minimal impact on access times). An important consideration in modern disk drive systems is the amount of time it requires for the read/write heads to locate and move to a new track position or (access time). This time includes the seek time and the rotational latency. Normally the magnitude of these times are dictated by the slowness of electromechanical devices. A method is needed to flag defective sectors and eliminate extraneous seeks which slow access time. ID FIELDS Storage space for a computer is very important. With the increasing amount of software and data processing size, the importance of disk drives being capable of handling greater amounts of data is paramount. The data written to the hard disk contains extra fields such as the ID field. The ID field contains information which can include the start of a sector, byte(s) to synchronize the reading of the ID field, the track or cylinder address, the read/write head number for a multi-headed system, the sector number, a flag field, and cycle redundancy check (CRC) information. In general, the ID field is similar to an address as to where the sector is located on the disk. The ID field layout may vary for different drive designs but remains consistent for each individual drive. How the ID field is written to the disk is controlled by a program in the writable control store. If this field can be reduced in size or located off the disk then other information could be stored in its place. This would effectively increase the data storage capability of the disk drive system. SPLIT FIELD HANDLING A method of measuring the placement of the read/write head on the disk is necessary in order to maintain the servomechanism of the read write head within a desired track. This has been provided by information such as servo bursts, marks, or patterns, on a separate platter or disk surface of the disk drive unit. Split field processing allows one to put control functions such as a servo burst in the middle of the data field which would eliminate the extra platter, thus reducing the cost of the drive while increasing the available storage. More than one servo burst per track is required in order to place adjacent tracks closer together, thereby increasing the number of tracks available on disk. Split field handling also allows one to utilize a short sector that has been partially damaged or is just short because of the location on a drive. Information regarding the split field must be maintained or generated for each sector in the track in order to correctly retrieve the data that was split between the control signals. Data are usually encoded into the ID field of the sector and stored on the disk. An automated way of handling split data fields and storing the information other than on the hard disk is necessary to increase disk drive storage efficiency. ZONE BIT RECORDING/BANDING On a platter such as a disk storage device, the circumference of concentric circles becomes larger as the read/write head moves from the inner circle to the outermost circle. Similarly, the innermost tracks on a disk storage device that are closest to the center of the disk have a smaller circumferences than those of the outermost tracks. This creates smaller linear distances on the inside tracks. If the disk drive spins at the same rate of speed and the data recording rate remains constant, the density, measured in bits per inch, is greater on the inside track. This is an inefficient use of the outside tracks in the sense that the medium in such tracks is capable of recording more information at a higher data rate to achieve the same linear bit density. To create a constant density recording one can increase the angular velocity of the disk as the read/write heads move to the outside tracks. Another means of approaching a constant density of bits stored on the disk is to increase the writing data rate as the read/write head moves out to the edge of the disk. In either method, tracks are usually grouped together in a zone or a band in order to maintain an area of constant disk speed or constant read and write frequencies. This technique is referred to as banding, zoning, or zone bit recording. BRIEF SUMMARY OF THE INVENTION An apparatus and method for disk-sector layout, formatting, reading and writing, is based on a flexible formatter microengine that is driven by parameter lists, which may include commands to dynamically down load the microengine's writable control store during disk access operations. This reduces the intervention required by the local microprocessor to support exception handling, including defect management, thus decreasing total access time for typical disk access (format, write or read) operations. This also reduces or eliminates the area on the disk required for ID fields, thus increasing the storage capacity of disk drives controlled by the present invention. This also increases disk space utilization, thus increasing the storage capacity of a disk drive controlled by using the present invention. Additionally, the invention is based on ID fields that may contain redundant information for increased fault tolerance. Additionally, the invention supports split data fields while minimizing the amount of additional information that the ID fields must contain about the data-field split sizes. Alternatively, the invention supports eliminating ID fields entirely with the prior-art functions of ID fields taken over by a combination of parameter lists and sector status flags. It is an object of the invention to reduce the size of ID fields by minimizing the information concerning split data fields that is contained therein. A related object is to maintain a parameter list, one per zone or band, of first split sizes, which typically repeat after 4-7 sectors. An alternative object of the invention is to eliminate ID fields entirely. A related object is to maintain sector-status flags to indicate exceptional sectors, (such as defective sectors, sectors that are the normal end of the track, and sectors that are the end of a track that has been shortened by defective sectors). An object of the present invention is to free the logical sector size from the fixed constraints of pre-defined sectors so as to allow logical sector sizes which are larger or smaller than the pre-defined sectors on a disk while at the same time making use of the full storage capacity of the disk. Related objects are to support data fields that are split across servo bursts, to support data fields that are segmented into multiple segments wherein each segment has an independently generated and verified ECC field, and to support data fields that are both split and segmented. Another object to the present invention is to provide a simple, more error tolerant ID information format for magnetic disk drives which does not substantially increase storage capacity requirements therefor. Another object of the invention is to speed up access time in reading and writing sectors that, due to media defects, have been slipped and mapped out of their expected location by eliminating the need for local microprocessor intervention and by eliminating extraneous seeks. Another object of the invention is to eliminate the need for the servo processing engine to detect the start of each sector. A further object is to respond to a disk fault signal generated from the servo processing engine. An object of the present invention is to provide flexibility to respond to future requirements on disk sector format and reliability. A related object is to support real-time modification, or downloading, of the formatter microengine based on parameter-list entries. Another object of the invention is to do the above in the context of zone bit, or banded, recording, as well as in the context of a large range of sector sizes and maximum ECC coverage sizes. BRIEF DESCRIPTIONS OF THE DRAWINGS FIG. 1 shows a typical architecture of a disk subsystem. FIG. 2 shows the layout within a track of an error-tolerant ID field according to the present invention. FIG. 3 shows the layout within a track of a split and segmented sector according to the present invention. FIG. 4 shows a track of a split and segmented sector within a zone with the operation of various counters shown according to the present invention. FIG. 5 shows a simplified register-level block diagram of the Inter-Field Logic Module which generates control signals used to determining the layout and timing of split and segmented data fields. FIG. 6 shows the data path diagram of the Down Load Module. FIG. 7 shows the Sector Mark Generation Logic. FIG. 8 shows the Sector Flag Logic. FIG. 9 shows the overall architecture of the formatter. DESCRIPTION OF THE PREFERRED EMBODIMENT Appendix A, filed herewith and entitled CL-SH4500 Wile E. Coyote Integrated SCSI Disk Controller Engineering Specification, is hereby incorporated by reference. ERROR-TOLERANT ID FIELDS Referring to FIG. 2, a diagram illustrating an error tolerant ID (identification) field of the preamble of a sector of a magnetic disk storage device in accordance with the present invention may be seen. In addition to the protection provided by the usual CRC field, the ID field is made more error tolerant with the multiple ID subfield shown. In the Multiple ID subfield scheme, the ID field within each sector can have multiple subfields that all contain the same ID information, with each subfield having its own CRC field. During ID search, all ID subfields of a sector are read, and only one subfield needs to have a good CRC to determine the ID information of that sector. The ID information is lost only if all subfields have a bad CRC. As may be seen in FIG. 2, in this scheme a sector preamble starts with the VFO (variable frequency oscillator) reference signal, followed by a number of ID subfields as determined at the time of disk formatting. Each subfield is in the form of sync signal followed by the sector ID followed by the CRC. While the multiple ID subfields increase the preamble overhead somewhat, substantial overhead is saved by not repeating the entire preamble, VFO and all, while at the same time the probability of not properly reading the ID field is greatly decreased by requiring that only one subfield need be properly read for successful sector identification purposes. In the preferred method of practicing the invention, a multiple ID subfield count is maintained. This count is preset by the local processor with the number of the ID subfields of a sector, and can be decremented and refreshed during the ID search operation. With the use of this register, a simple program loop can be used to process through all the subfields of the sector and to be refreshed for the next sector preamble. DATA FIELD SPLITS AND SEGMENTS The present invention also provides for the automated processing of multiple data field sections within a given sector. In accordance with the invention, any sector can be: (1) segmented in a contiguous field with multiple redundancy fields, or (2) split into several non-contiguous fields which share a single redundancy field. The segmenting and splitting capabilities can be intermixed, with either the basic segment or the basic split being larger. Each capability provides unique benefits. FIG. 3 shows the layout within a track of a split and segmented sector according to the present invention. The split field capability supports and simplifies embedded servo systems by utilizing all of the storage capacity between two servo pulses, and by not requiring this storage capacity to be an even multiple of the sector size. The segmented field capability supports sector sizes greater than the limit of the ECC code chosen. In the preferred embodiment, a Reed-Solomon ECC code is used having a maximum coverage of 1267 bytes. In any ECC encoding, increasing the amount of data covered by the same codeword decreases the protection against errors, but in the case of Reed-Solomon codes there is an abrupt limit beyond which the Reed-Solomon code simply will not function. Additionally, ECC segmentation makes it easier to implement multiple burst and/or "on-the-fly" error detection and correction. There are two basic requirements for automated processing of splits. The first requirement is to support the synchronization of ECC computation or generation across non-contiguous data fields. The second requirement is to support the automated retrieval of the physical configuration of the split on the media. In one embodiment, this information can be both stored on and extracted from the media (typically in the ID field). In another embodiment, this information can be both stored in and retrieved from a buffer memory or the processor memory. The basic requirement for the automated processing of segments is simply to keep a count of the bytes being written and automatically process a redundancy field at the appropriate byte limits. Split sector layouts with servo bursts embedded in the data field can be automatically handled by: (1) the proper combination of start and stop operations to control the timing of the generation, processing, reading and writing of the ECC information, and (2) embedding some split size information in the ID field, or in a parameter list in the buffer or the processor memory. This maximizes data capacity, and zone recording techniques are easier to employ as any track or band of tracks can have an optimized track format with varying data field split sizes for each sector within the band. Split Data Field ECC Processing: When the split data field error detection and correction processing function is active, the beginning boundary of a data field that is to be split into at least two fields is detected. The freezing and restarting of an EDAC-generation operation (error detection and correction) during a write to disk is preferably done automatically between the split data fields. Likewise, the freezing and restarting of an ECC verification operation during a read from disk is preferably done automatically between the split data fields. Split Data Field Size Processing: Preferably the second major function for supporting split data fields is the capability of automatically retrieving or calculating the split sizes to relieve any local processor both from tracking the physical splits and from constant reprogramming of the split data field size in the format sequencer program. The present invention method maintains several counts to automatically retrieve split data field size information. These include: a 16-bit SECTOR DATA REMAINING COUNT; a 12-bit FIRST SPLIT COUNT; a 12-bit MAXIMUM SPLIT COUNT; a 12-bit SPLIT FIELD COUNT. Also employed is a control signal, "Enable Split Field Count And Sector Data Remaining Counter". A sector can have multiple splits. In the preferred embodiment, 3 potentially different split sizes within a sector are provided for: the first split, the last split, and the ones in between. In the format driven mode, the first split size is retrieved from the ID format. Alternatively, in the parameter-list or the down-load modes, the first split size is stored and retrieved locally. The sizes of the splits in between the first and the last ones are the same and are equal to the MAXIMUM SPLIT COUNT. The last split size can be derived from the SECTOR DATA REMAINING COUNT. SECTOR DATA REMAINING COUNT The SECTOR DATA REMAINING COUNT is used to track the number of bytes remaining in the current logical sector (which may or may not consist of multiple split data fields or multiple ECC segments). This count is decremented once during every byte time that the Enable Split Field Count and Sector Data Remaining Count signal is active. This count needs to be initialized only once by the local processor, which sets it equal to the logical sector byte count. It is refreshed without local processor intervention for each sector. Referring to FIG. 4, the initialization occurs just prior to period p3 and the SECTOR DATA REMAINING COUNT is decremented during p3, p5, p7 and p9. FIRST SPLIT COUNT This count can be (1) set by the local processor, (2) down loaded from buffer or processor memory, or (3) embedded in the ID field and retrieved while reading the ID field (e.g., during p1 in FIG. 4). In format mode (which is the only tine that the ID field is written), this count is set with the value to be written in the ID field on the disk. In either case (whether set locally or retrieved from the ID field), the FIRST SPLIT COUNT will be used by the SPLIT FIELD COUNT to process the very first split of the data field. MAXIMUM SPLIT COUNT This count is set by the local microprocessor with the number of information (sync, data and ECC) bytes in the splits between the first and the last splits, and is used as the initial value for the SPLIT FIELD COUNT for the splits after the first one. At the end of the last split of the data field (e.g., at the end of p9 in FIG. 4), the SECTOR DATA REMAINING COUNT will reach zero and thus end the data field processing. The MAXIMUM SPLIT COUNT needs to be initialized on power up and re-initialized whenever the read/write heads are moved to a track with a different data frequency, (i.e., a different zone). SPLIT FIELD COUNT This count is used in processing split data fields. The SPLIT FIELD COUNT is used when the Enable Split Field Count And Sector Data Remaining Count control signal is active. The first time this count is enabled on a sector (e.g., just prior to p3 in FIG. 4), the FIRST SPLIT COUNT is used to initialize the SPLIT FIELD COUNT. The counter is decremented once for each information byte (i.e., sync, data or ECC) transferred across the disk interface. Referring to FIG. 4, it is decremented during p3, p4,p5, p7, p8, p9, and p10. When it reaches zero, the end of the first data split has been reached and it indicates the start of the servo mark (which must not be overwritten). On the second and all subsequent splits the MAXIMUM SPLIT COUNT is used as its initial count. Note that the SPLIT FIELD COUNT will begin counting down from its initial count (either the FIRST SPLIT COUNT or the MAXIMUM SPLIT COUNT) after the data sync bytes. It will not stop on ECC field, specifically to support the ECC segmentation capability. Enable Split Field Count and Sector Data Remaining The Enable Split Field Count and Sector Data Remaining Control bit (of the Current Sequencer Word Control Field) is used to replace the normal Current Sequencer Count Field feature with the SPLIT FIELD COUNTER and the SECTOR DATA REMAINING COUNTER. When this control bit is set and the Split Field Mode Disable bit is reset, a branch command is evaluated when either the SPLIT FIELD COUNTER or the SECTOR DATA REMAINING COUNTER is used and a branch command is evaluated when this counter reaches zero. ECC SEGMENTATION CAPABILITY The ECC segmentation capability allows the present invention to support physical sector sizes larger than a code-word size supported by an ECC code. For example, the 88-bit Reed-Solomon code used in the preferred embodiment has a hard limit of 1278 bytes (including ECC bytes) per code word. When active, an 11-bit data coverage count is provided which is decremented once for every byte. In this segmentation mode, the data field is divided into contiguous segments, each of which will have its own ECC field. The format sequencer will be able to transfer each data segment and process its ECC one after another. The DATA COVERAGE LIMIT counter needs to be initialized upon system power up or re-initialization by the local processor to the maximum segment byte count. The SECTOR DATA REMAINING COUNTER is still used to track the number of the bytes remaining in the logical sector. When the DATA COVERAGE LIMIT counter reaches zero (e.g., at the end of p3 in FIG. 4), the sequencer will automatically process the ECC bytes. After processing the ECC bytes, if the SECTOR DATA REMAINING COUNTER (which is held constant and not decremented during ECC transfer) is not zero, then the format sequencer will begin processing another segment within the same data field. On the last segment of a particular data field (e.g., at the end of p9 in FIG. 4), if the SECTOR DATA REMAINING COUNTER reaches zero before the DATA COVERAGE LIMIT COUNTER, then the last ECC field will be processed when the SECTOR DATA REMAINING COUNTER is zero. Therefore the last segment could be shorter than the preceding ones and thus the sector size does not have to be an even multiple of the segment size. However, the segment size is preferably kept as close as possible to a constant so that the "on-the-fly" correction circuitry is provided with sufficient time to find and apply correction results. MIXING SPLITS AND SEGMENTS As is shown in FIG. 3, the ECC segmentation capability is designed to also work with split data field processing. Prior to the start of processing of the ECC field, the ECC TRANSFER COUNT is initialized to the number of bytes of ECC information required per segment, which is a predetermined value (i.e., 11 bytes) in the preferred embodiment. During processing of the ECC field, the ECC TRANSFER COUNT decrements. Also, the SPLIT FIELD COUNTER still decrements and thus will keep track of the bytes remaining in the current area between servo bursts. When the Sequencer returns from the split processing WCS code to the data and ECC processing code, then the status of the counters (namely DATA COVERAGE LIMIT, SECTOR DATA REMAINING, and ECC TRANSFER) will continue the appropriate activity even in the case where it is the ECC field that is broken up by the servo burst. That is, if at the end of a servo burst the ECC TRANSFER register is non-zero, then the remaining bytes of the ECC information are read or written prior to resuming the reading or writing of the user data. In the case where the ECC segmentation is not required (e.g., because the sector size is within the ECC code limit), then both the DATA COVERAGE LIMIT COUNTER and the SECTOR DATA REMAINING COUNTER must be programmed with the sector size. The ECC field process will still be done automatically after the user data field. The operation of these aspects of the invention can be summarized in reference to FIG. 4. During the user-data periods p3, p7 and p9, the SPLIT FIELD COUNT, the SECTOR DATA REMAINING COUNT, and the DATA COVERAGE LIMIT COUNT are each decremented once per byte read or written. During the ECC-information periods p4 and p10, the SPLIT FIELD COUNT and the ECC TRANSFER COUNT are each decremented once per byte read or written. The implementation of these aspects of the present invention is illustrated in FIG. 5. FIRST SPLIT COUNT and MAXIMUM SPLIT COUNT are registers. SECTOR DATA REMAINING COUNT, DATA COVERAGE LIMIT COUNT, and ID FIELD COUNT are combination register/counters that count down from their initial value, generate a control signal when they reach zero, and then reset themselves to their initial value. Their initial value is established by the local microprocessor. FORMAT DRIVEN CAPABILITIES In the present invention, the Secondary Compare circuit can be used to retrieve format information from the ID field. This information can be used in split field processing, automated sector search retries, automated multi sector accesses, and defect management processing. Load FIRST SPLIT COUNT Secondary Compare Flag: The Load FIRST SPLIT COUNT Secondary Compare Flag is a defined bit in the Fixed Secondary Compare Flags which can be used to trigger the retrieval of 12-bits of count information from the ID field. This value can be programmed as the length of the first split of the data field. The most significant nibble is contained in the same NRZ data byte. The remaining byte must be the following NRZ data byte. End of Track Secondary Compare Flag: The End of Track Secondary Compare Flag is a defined bit in the Fixed Secondary Compare Flags which can be used to trigger the Format Sequencer multi-sector branch conditions. Typically it is used to mark the end of a track, such that the Format Sequencer can make real-time decisions on when to abort a multi-sector read. Short Track Secondary Compare Flag: The Short Track Secondary Compare Flag is a defined bit in the Fixed Secondary Compare Flags which can be used to trigger the Format Sequencer multi-sector branch conditions. It is designed to be used in place of the End of Track Secondary Flag in a disk format that uses sector Skipping as a Defect Management technique. In this technique, the number of sectors per track may vary as the media ages. This Flag can also trigger an interrupt to the local processor. PARAMETER LISTS AND WCS DOWN LOAD CAPABILITY In the present invention, parameter lists, which may preferably be located in the Buffer Memory, can be used to automate several tasks. These include: (1) timing of an internally generated sector mark from externally generated servo bursts, or from the end of the previous sector mark, wherein a parameter-list entry specifies the required information on the delay from the triggering event to the time when the next sector mark is to be generated by the present invention; (2) increasing the storage capacity of the disk by removing from the ID field the length of the first split of the data fields (wherein a parameter-list entry specifies the required information on the size of the first split); (3) loading sector status flags, including the End of Track flag (indicating the last sector in a track before the index servo burst), the Short Track flag (indicating the last sector in a track that contains fewer sectors than expected because of defective sectors in the track) and the Defective Sector flag (indicating that the current sector is defective or relocated); and (4) general-purpose dynamic updating of the Format Sequencer's (or format microengine's) Writable Control Store (WCS). The parameter list capability is under the control of the Local Processor in that it (1) enables the down load/parameter list function, (2) selects what events trigger down load/parameter list processing, and (3) initializes the parameter lists in the buffer memory (perhaps using disk-format-design information contained in the ROM portion of its system memory, perhaps using potentially dynamic information obtained from the disk itself. This parameter list capability assists (1) in handling exception conditions such as defective sectors, tracks shortened by slipped (or skipped over) defective sectors, and mapped (or relocated) sectors; and (2) in creating a disk format without the traditional ID field. For example, a sector mapped to another cylinder may be handled by a down load of modifications to the WCS program or bitmap instead of by an interrupt to the local microprocessor (which would often incur a full rotational latency due to the local microprocessor interrupt service latency, and interrupt processing time). The Down Load Module consists of three components: the Down Load Address Pointer; encoded Down Load (or parameter) Lists; and Sector Mark Generation Logic. The data path of the Down Load Module is shown in FIG. 6. The Down Load Address Pointer points to the encoded word to be down loaded. Before the Down Load Circuit is started the Down Load Address Pointer is set, preferably by the local processor to the start of the appropriate Down Load List. In the case where the down load circuit is set to activate on Index (whereby a down load or parameter list is evaluated or processed upon the next detection of an index servo burst), then the Down Load List pointed to will be the first Down Load List. After each down load cycle the Down Load Address Pointer is incremented to the next word in the list and the down load continues until the end of the list, which is recognized by the End List Flag. At the end of each Down Load List, the Down Load Address Pointer is not always automatically altered. Rather the final operation of any Down Load list either may leave the Down Load Address Pointer pointing to the next Down Load List that happens to be contiguous in Buffer Memory or it may load the Down Load Address Pointer from the current (and final) instruction of the current Down Load List, thus updating the pointer to the start of the next Down Load List that happens to be non-contiguous within the Buffer Memory. This, in affect, creates an efficient but flexible linked-list data structure for the Down Load Lists. Thus, a Down Load List is not fixed in length and its location in Buffer Memory can be linked to the previous list. This linked-list aspect of the Down Load List is particularly valuable in those situations (which are common in practical disk drive designs) in which there is a repeating pattern, called a frame, in the first split sizes. That is, the first sector within a track will have a first split size of X, the second sector will have a first split size of Y, and so on until the N'th sector has a first split size of X--after which the pattern or frame repeats. Thus in the present invention, Buffer Memory space is only required for N-1 Down Load Lists, with the first N-2 entries ending with the Down Load Address Pointer pointing at the next contiguous Down Load List and the N-1'st entry ending with an operation (code 1tH) to reset the Down Load Address Pointer to the start of the first Down Load List. A Down Load List is read from the buffer and its instructions executed every time the Down Load Circuit is triggered. Depending on the status of enable flags, the Down Load Circuit may be triggered either (1) by the next servo burst, (2) by the next sector pulse (internally or externally generated), or (3) by the next index pulse. The encoding of Down Load List operations or instructions is described in the following table: TABLE 1______________________________________Down Load List Encoding1st byte 2nd Operand Description______________________________________0xH yyH 0xh is the End-of-the-Down-Load-List flag, xH contains End of Track, Short Track and Defective Sector Flags.1tH uuH vvH wwH 1tH is the Linked End-of-the-Down- Load-List Flag, the contains End of Track, Short Track and Defective Sector Flags, and uuvvwwH being the twenty- four-bit pointer to the next Down Load List.2xH yyH xyyH is the twelve-bit count for Internal Sector Mark generation.3xH yyH xyyH is the twelve-bit value to be stuffed into FIRST SPLIT COUNT Register. 4-5!xH yyH 4x-5x!H are reserved flags (currently NOP's) 6-F!xH yyH yyH is the eight-bit value to be loaded into WCS address 6x-Fx!H.______________________________________ Generating Sector Flags: The second nybble in either the End-of-the-Down-Load-List or the Linked End-of-the-Down-Load-List bytes contains End of Track, Short Track and Defective Sector Flags. These flags bits are only used when the Down Load List is triggered by either the (internally or externally generated) SECTOR MARK signal or the INDEX signal. FIG. 8 shows the Sector Flag Logic that generates the control signals corresponding to three flags from either the sector flag field in the ID field read from the sector, or from the Down Load Logic. GENERATING AN INTERNAL SECTOR MARK In the present invention, the Sector Mark Generation Logic consists of a Sector Mark Counter (SMC) and a Sector Generation Byte Counter (SGBC). The SMC keeps track of the number of sector pulses relative to Index. It monitors the Index signal, the servo burst signal and the output from the SGBC. The SGBC is a up-counter which is incremented once every byte time. The Read Reference Clock (RR-- CLK) is used to derive the clock to run the SGBC. Note that derivation of a byte clock from RR-- CLK is trivial when the NRZ path is one-bit, two-bits, four-bits, or eight-bits wide. The Sector Mark Generation Logic is shown in FIG. 7. The SGBC, which consists of a counter and a limit register, restarts from zero every time the Down Load Module is triggered by the servo burst. The SGBC counter output is compared against the SGBC limit register. When they are equal, an internal SECTOR MARK is generated and the SGBC counter is restarted at zero. However, even though the SGBC counter is running, another SECTOR MARK will not be generated unless the SGBC limit register is updated. The SGBC limit register can be loaded by the local processor, but is usually loaded by a buffer down value (using Down Load instruction operation code 2xH. The SECTOR MARK is produced at the time that the SGBC counter is equal to the SGBC limit register. However, if the SERVO signal is active at this time, then the internal SECTOR MARK will be delayed until the SERVO signal is de-asserted. This is useful for generating an internal SECTOR MARK immediately after a SERVO signal burst. This can be forced by loading a value of zero in the SGBC limit register. When the SERVO signal burst is received, the SGBC counter is reset to zero, and a match occurs. The SGBC circuit will then wait until the SERVO signal is de-asserted to generate the internal SECTOR MARK. This special case can be thought of as an overlapping between the internal SECTOR MARK and the external SERVO signal burst. The Down Load Module will not be triggered by the (overlapped) SECTOR MARK, even if the Enable Down Load By Sector Mark control bit is set. This technique is also used when starting the Sector Mark Generation logic before INDEX, where an internal SECTOR MARK should be generated immediately after the servo burst. In this case the SGBC limit register is loaded by the local processor with a value of zero before the logic is started. In a Dedicated Servo system, the sector pulse is expected to come from the external servo circuit and it is directly used by the SMC. In this mode the SGBC is turned off and all 2xyyH code is ignored. This mode is selected by setting the Sector Mark/Servo Burst mode bit for Sector Mark Mode. OTHER USES OF DOWN LOAD Loading the FIRST SPLIT COUNT using Down Load The 3xyyH code word is used to down load the FIRST SPLIT COUNT for split data field support. The nibble x in the byte 3xH is the most significant nibble of the COUNT. The remaining byte is the remainder of the twelve-bit value. Since a Down Load is started by either a Servo Burst or an (internal) Sector Mark, the Down Load List Structure should be maintained in such a way that the FIRST SPLIT COUNT is only down loaded once for each sector (therefore it should be down loaded by Sector Mark triggers). Updating the WCS using Down Load This is a general purpose tool which provides partial or complete Format Sequencer "Overlays" for the Writable Control Store. The code words of 6xyyH-FxyyH are used to down load the WCS byte at address 6xH-FxH. Consideration must be made about the relative timing between the down load start time (usually at Sector Mark) to the time the Sequencer would use the updated WCS bytes. Usually the Down Load Module is used at the beginning of a sector (at the SECTOR MARK) to update the WCS bytes which are executed at the end of the ID field, the end of the data field, or at the beginning the of the sector (the WCS instruction word which branches on the occurrence of a SECTOR MARK). The Trigger Buffer Down Load bit in the FORMAT SEQUENCER BRANCH INSTRUCTION FIELD when set will trigger a down load cycle and this bit is typically used for controlled updating of the WCS. Using Down Load to Pre-configure Track Access: In some intelligent disk drive systems, a defect table is consulted to build the expected configuration of the track, allowing for both the defect management schemes of skipping sectors and alternate sectors. This is typically done during an armature seek or concurrently with another track access. This expected configuration can be built in the form of a Track Down Load Table. The Down Load List for each sector can update the WCS instruction which is evaluated at the Sector Mark signal to branch to a WCS code routine. Thus different WCS code routines can be setup to handle skipped, redirected, or good sectors differently, where the Down Loaded WCS bytes will select which routine is to be used for that sector. The updated WCS instruction is downloaded one sector ahead however, since there is a natural latency for the Buffer Manager to retrieve the Down Load bytes, and for the Down Load Module to update the WCS. For slipped sectors, the Down Load List can change the branch instruction to decrement the SECTOR MARK COUNTER and continue to search for the next SECTOR MARK. This allows the SECTOR MARK COUNTER to match the logical number of the track, instead of limiting the SECTOR MARK COUNTER to the physical ordering. For alternate sectors, the Down Load List can change the branch instruction to jump to a special code routine. In this code routine, disk streaming can be disabled (from a branch instruction), and this instruction should also activate data transfer using the SECTOR REMAINING COUNTER. Note that both READ GATE and WRITE GATE should NOT be asserted. This will have the effect of incrementing the DISK ADDRESS POINTER by a sector size, while limiting the SCSI access to the data of the previous sector, and continuing in real time to access the rest of the track with correct buffer ordering. Down Load can help support disk drives with dedicated servo systems by internally generating the SECTOR MARK pulses. The servo system is often a fixed frequency, and there is processing to be done by digital signal processing (DSP) support circuitry in the servo system or servo engine to generate the SECTOR MARK from the fixed servo frequency. The fact that the processing to be done may need to be different for various zoned densities complicates the servo system's job of generating sector marks. If SERVO pulses can be passed to the present invention, then it can use these and its Down Load Module to generate appropriately timed SECTOR MARK pulses. ID-LESS DISK SECTORS The above-described parameter lists and Down Load Lists form the basis of a disk format or layout that has no ID-fields. In a first ID-less embodiment of the present invention, Down Load Lists containing the required sector status flags and first split sizes may be retained in Buffer Memory. This requires substantial space in Buffer Memory, or wherever they are retained, because even though the first split sizes typically repeat in patterns or frames, the sector status flags are often related to management of media defects which can be random. Thus in this embodiment, an individual Down Load List is required for each sector. Accordingly, the Buffer Memory must be large enough to contain numerous Down Load Lists, and the Local Processor must initially load the Buffer Memory (or whatever other memory the parameter lists may be stored in) with such Down Load Lists from a non-user-data area of the disk. Also, if any sectors are newly determined to be defective, both the on-disk copy and the in-buffer-memory copy of the affected Down Load Lists must be updated. In a second, alternative ID-less embodiment of the present invention where Buffer Memory space is limited, a subset of the sectors of the disk forms the current "working set" of sectors (based on a criteria such as most recently used) and the local processor "pages" the Down Load list of each sector into Buffer Memory, much like a virtual memory manager pages user data into and out of main memory from disk storage. Alternatively, paging could be done per track, i.e., the Down Load Lists of all sectors within a particular track could be paged into Buffer Memory at the same time. Sector-Flag Sectors In a third ID-less embodiment, Sector-Flag Sectors can be located in a nonuser-data area of the disk. They could be located for example in a sector numbered minus one (-1) immediately after the index servo burst on each track, or alternatively on such a sector on only a subset (e.g., the first) of the tracks within each cylinder. A Sector-Flag Sector contains only non-user data, such as split counts and sector status flags. Sector-Flag Sectors may be encoded as Down Load Lists as described herein or may be encoded in a less flexible but more compact fixed-field representation. An example of the later representation is a format wherein each sector has a predetermined number, say 2, of bytes of information with, say, the first 4-bit nybble of the first byte representing sector status flags and the remaining 12-bits representing a first split count if a good sector, or the address of the sector to which this sector has been mapped if a bad sector. It will be obvious to one skilled in the art that there are a range of other ID-less embodiments that are variations on the concepts herein disclosed. For example: a fourth ID-less embodiment may use the Down Load List mechanism to contain repetitive information such as first split counts and the Sector-Flag Sector mechanism to contain information that is unique per sector such as sector status flags and potentially repetitive information such as first split sizes. A fifth ID-less embodiment may add a paging scheme to the fourth embodiment. DYNAMIC OR ALGORITHMIC SECTOR FORMATS The parent of this application discloses techniques for handling split data fields in which the first split size information is not needed to be kept either in a parameter list or in the ID field but rather is dynamically calculated. It will be clear to one skilled in the art how this technique can be utilized in conjunction with the present invention, including but not limited to combining the split field technique of the parent application with the multiple ID, the ID-less, the parameter list, the WCS down load, or the ECC code-word segmenting techniques of the present invention. While the preferred embodiment and various alternate embodiments of the present invention has been disclosed and described herein, it will be obvious to those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention. Patent Citations Cited PatentFiling datePublication dateApplicantTitle US4642775 *24 May 198510 Feb 1987Sundstrand Data Control, Inc.Airborne flight planning and information system US5142680 *26 Apr 198925 Aug 1992Sun Microsystems, Inc.Method for loading an operating system through a network US5278703 *21 Jun 199111 Jan 1994Digital Equipment Corp.Embedded servo banded format for magnetic disks for use with a data processing system US5278815 *25 Sep 199111 Jan 1994Teac CorporationDigital data reproducing apparatus US5305302 *12 Aug 199319 Apr 1994Conner Peripherals, Inc.Track format and record carrier system for split data field support US5319544 *20 Nov 19897 Jun 1994Itt CorporationComputerized inventory monitoring and verification system and method Referenced by Citing PatentFiling datePublication dateApplicantTitle US5963386 *31 Mar 19975 Oct 1999Stmicroelectronics N.V.Disk resident split data sector information management system on a headerless magnetic disk device US5968184 *3 Jun 199719 Oct 1999Emc CorporationMaintenance command structure and execution US5987400 *29 Apr 199816 Nov 1999Kabushiki Kaisha ToshibaSystem for monitoring the throughput performance of a disk storage system US6016547 *24 Apr 199818 Jan 2000Fujitsu LimitedData processing apparatus with program allocating section for selectively allocating programs in first and second memory US6025971 *17 Apr 199815 Feb 2000Kabushiki Kaisha ToshibaMagnetic disk drive using a nonvolatile solid state memory to store defect information US6134063 *30 Dec 199717 Oct 2000Lsi Logic CorporationAutomated multi-track transfers US6137646 *10 Jun 199724 Oct 2000Fujitsu LimitedDisk apparatus using coincidence detection to generate variable sector pulses US62046603 Jun 199820 Mar 2001Samsung Electronics Co., Ltd.Method of varying capacity of head disk drive during manufacturing process by checking head/disk combinations for defects US6237070 *31 Jul 199822 May 2001International Business Machines CorporationMethod and system for determining the data layout geometry of a disk drive US6253279 *31 Jul 199826 Jun 2001International Business Machines CorporationMethod and system for determining the data layout geometry of a disk drive US6279118 *2 Sep 199821 Aug 2001Samsung Electronics Co., Ltd.Recording medium storing additional information for defect management and method for managing defects US630167930 Jun 20009 Oct 2001Seagate Technology LlcSync byte padding US633218222 Oct 199918 Dec 2001Cirrus Logic, Inc.Disk drive control without identification fields US6341111 *19 Oct 199922 Jan 2002Samsung Electronics Co., Ltd.Apparatus for detecting and restoring physical identification US6456742 *16 Apr 199924 Sep 2002Mustek Systems, Inc.Method for image processing US6590728 *17 Dec 19968 Jul 2003Samsung Electronics Co., Ltd.Hard disk having extended data region US6606714 *17 Apr 200012 Aug 2003Western Digital Technologies, Inc.Disk drive using defect list to perform uninterrupted writing of non-relocated data blocks and caching of relocated data blocks for deferred writing US667810630 Mar 200113 Jan 2004Seagate Technology LlcDetermining data sector splits across servo bursts in a disc drive US6715123 *5 Jun 200130 Mar 2004Fujitsu LimitedReal-time recording system and real-time recording method US672181928 Aug 200113 Apr 2004Lexar Media, Inc.Flash memory card with enhanced operating mode detection and user-friendly interfacing system US672889930 Jun 200027 Apr 2004Seagate Technology LlcOn the fly defect slipping US673892414 Jan 200018 May 2004Seagate Tech. LlcFull slip defect management system using track identification US6742147 *16 Jun 200025 May 2004Matsushita Electric Industrial Co., Ltd.Information recording medium, and method and apparatus for managing defect thereof US674782524 May 20008 Jun 2004Jpmorgan Chase Bank, As Collateral AgentDisc drive with fake defect entries US677227413 Sep 20003 Aug 2004Lexar Media, Inc.Flash memory system and method implementing LBA to PBA correlation within flash memory array US69014574 Nov 199831 May 2005Sandisk CorporationMultiple mode communications system US697355320 Oct 20006 Dec 2005International Business Machines CorporationMethod and apparatus for using extended disk sector formatting to assist in backup and hierarchical storage management US699060727 Sep 200124 Jan 2006Seagate Technology LlcSystem and method for adaptive storage and caching of a defect table US7023639 *31 Mar 20044 Apr 2006Western Digital Technologies, Inc.Servo synchronization validation techniques based on both servo synch marks and wedge identifiers in a rotating media storage device US711108521 Aug 200319 Sep 2006Lexar Media, Inc.Flash memory card with enhanced operating mode detection and user-friendly interfacing system US715555925 Aug 200026 Dec 2006Lexar Media, Inc.Flash memory architecture with separate storage of overhead and user data US71744451 Apr 20026 Feb 2007Lexar Media, Inc.Flash memory card with enhanced operating mode detection and user-friendly interfacing system US727701121 Feb 20032 Oct 2007Micron Technology, Inc.Removable memory media with integral indicator light US736000326 Jan 200515 Apr 2008Sandisk CorporationMultiple mode communication system US74215231 Dec 20052 Sep 2008Lexar Media, Inc.Flash memory card with enhanced operating mode detection and user-friendly interfacing system US753537018 Sep 200719 May 2009Lexar Media, Inc.Removable memory media with integral indicator light US816122922 Jun 200917 Apr 2012Micron Technology, Inc.Flash memory architecture with separate storage of overhead and user data US859542128 Mar 201226 Nov 2013Petro EstakhriFlash memory architecture with separate storage of overhead and user data US8705192 *12 Oct 201222 Apr 2014Lsi CorporationEnhanced quality-sorting scheduler US20020083382 *5 Jun 200127 Jun 2002Yoshitaka KitamuraReal-time recording system and real-time recording method USRE42398 *31 Aug 200424 May 2011Kabushiki Kaisha ToshibaMemory system USRE445034 May 201117 Sep 2013Kabushiki Kaisha ToshibaMemory device storing identifying information and memory card including the same WO1999034366A1 *16 Dec 19988 Jul 1999Lsi Logic CorpAutomated multi-track transfers Classifications U.S. Classification714/47.3, G9B/27.033, G9B/20.028, G9B/20.059, 360/53 International ClassificationG11B20/12, G11B27/30, G11B20/18 Cooperative ClassificationG11B2220/20, G11B20/1883, G11B20/1258, G11B27/3027, G11B2020/1232 European ClassificationG11B20/18S, G11B20/12D23, G11B27/30C Legal Events DateCodeEventDescription 29 Aug 1996ASAssignment Owner name: BANK OF AMERICA NATIONAL TRUST & SAVINGS ASSOCIATI Free format text: SECURITY AGREEMENT;ASSIGNOR:CIRRUS LOGIC, INC.;REEL/FRAME:008113/0001 Effective date: 19960430 10 Oct 2001FPAYFee payment Year of fee payment: 4 16 Sep 2005FPAYFee payment Year of fee payment: 8 14 Oct 2009FPAYFee payment Year of fee payment: 12
__label__pos
0.589647
Welcome to the Question2Answer Q&A. There's also a demo if you just want to try it out. +2 votes 7k views in Q2A Core by I have some users that proposed to close unanswered questions that are older than 3 months. Actually, this idea I like, it keeps the "unanswered questions" list clean. Let's say you want to close them, how would you do it? -- pseudo code for qa_posts: SET closedbyid = ID-of-close-NOTE     WHERE created < NOW()-3 MONTHS     AND ... there are no answer, i.e. Answer with parentid does not exist -- Problem is that for each closed question we need a "post" with type=NOTE and parentid=Q and content='reason' Anyone? by According to my question here => http://www.question2answer.org/qa/32544/close-question-after-x-time Is your code working to close question (with or without answer)?  for qa_posts: SET closedbyid = ID-of-close-NOTE     WHERE created < NOW()-3 MONTHS ?? by Don't use the code above, as I have written this is "pseudo code". 1 Answer 0 votes by edited by This function helps for such a plugin:     function qa_post_set_closed($questionid, $closed=true, $originalpostid=null, $note=null, $byuserid=null) /* Closed $questionid if $closed is true, otherwise reopen it. If $closed is true, pass either the $originalpostid of the question that it is a duplicate of, or a $note to explain why it's closed. Pass the identify of the user in $byuserid (or null for an anonymous change).*/   I was creating an event module that automatically closes the question, this is part of the necessary code:         function process_event($event, $userid, $handle, $cookieid, $params) {             // Specials for Klaustukai                          // if question             $hitEvents = array('q_post', 'q_edit');             if(in_array($event, $hitEvents)) {                 // 1. close interviu question if posted from user klaustukai with tag interviu                 if($userid == 1 && strpos($params['tags'],'interviu')!==false) {                     $note = 'interviu';                     require_once QA_INCLUDE_DIR.'qa-app-posts.php';                     // qa_post_set_closed($questionid, $closed=true, $originalpostid=null, $note=null, $byuserid=null)                     // error_log($params['postid'].','.$note.','.$userid);                     qa_post_set_closed($params['postid'], true, null, $note, $userid);                 }             }         } // end process_event   ...
__label__pos
0.869116
I recently updated to the latest Propellor and now I'm getting an error building on a remote host: [86 of 94] Compiling Propellor.Bootstrap ( src/Propellor/Bootstrap.hs, dist/build/propellor-config/propellor-config-tmp/Propellor/Bootstrap.o ) src/Propellor/Bootstrap.hs:237:22: No instance for (Typeable Bootstrapper) arising from a use of `fromInfo' Possible fix: add an instance declaration for (Typeable Bootstrapper) In the expression: fromInfo (maybe mempty hostInfo mh) In a stmt of a 'do' block: case fromInfo (maybe mempty hostInfo mh) of { NoInfoVal -> do { bs <- getGitConfigValue "propellor.buildsystem"; case bs of { Just "stack" -> ... _ -> ... } } InfoVal bs -> case getBuilder bs of { Cabal -> cabalBuild msys Stack -> stackBuild msys } } In the second argument of `($)', namely `do { case fromInfo (maybe mempty hostInfo mh) of { NoInfoVal -> do { ... } InfoVal bs -> case getBuilder bs of { Cabal -> ... Stack -> ... } } }' propellor: cabal build failed I guess I'm missing something, but not sure what?
__label__pos
0.98307
Shadman Kudchikar Sharp Column Indenter - Visual Studio Extension Sharp Column Indenter - Visual Studio Extension Contents About Extension Sharp Column Indenter is a smart source code indenter that indent the code into columns. Also known as ‘code alignment’. Why Sharp Column Indenter? The most important aspect of programming is the readability of the source code that you write or maintain. This involves many things, from the syntax of the programming language, to the variable names, comments, and indentation. Here indention is something where Sharp Column Indenter can help you with. What Is Column Indention / Code Alignment? In mathematics you always keep your equals lined up directly underneath the one above. It keeps it clean and lets you know you’re working on the same problem, for example: y = 2x y/2 = x Programming is slightly different. We often have a lot of assignments underneath each other, and while they are not strictly the same as maths, there is a close relationship. As such, aligning the equals allows us to quickly spot the relationship. Further, it makes your code so much more readable. Without alignment, the code is like opening a CSV file in Notepad. But, if you open the CSV file in Excel, it becomes so much easier to read since the columns have meaning. Examples Of Improvements In Code Readability Using Code Alignment First example Take a look at this code: //natural code with no code alignment and you can see that //readability of the code is not that good var people1 = new List<Person>() { new Person { Name = "Benita", Location = "Bareilly India", Age = 25 }, new Person { Name = "Deedee Almon Fonsec", Location = "Bari Italy", Age = 32 } , new Person { Name = "Chase Hussain", Location = "Barika Algeria", Age = 45 } , new Person { Name = "Cordia", Location = "Barinas Venezuela", Age = 26 } , new Person { Name = "Malvina Neff", Location = "Barisal Bangladesh", Age = 36 } , new Person { Name = "Erika ", Location = "Barnaul Russia", Age = 56 } , new Person { Name = "Lisabeth Terr", Location = "Barquisimeto Venezuela", Age = 67 } , new Person { Name = "Farrah ", Location = "Barra Mansa Brazil", Age = 57 } , new Person { Name = "Domonique Biv", Location = "Barrackpur India", Age = 57 } , new Person { Name = "Jonah", Location = "Barrancabermeja Colombia", Age = 34 } }; The idea that I’m talking about is to use something like this below, //same above code with column indention var people2 = new List<Person>() { new Person { Name = "Benita" , Location = "Bareilly India" , Age = 25 } , new Person { Name = "Deedee Almon Fonsec" , Location = "Bari Italy" , Age = 32 } , new Person { Name = "Chase Hussain" , Location = "Barika Algeria" , Age = 45 } , new Person { Name = "Cordia" , Location = "Barinas Venezuela" , Age = 26 } , new Person { Name = "Malvina Neff" , Location = "Barisal Bangladesh" , Age = 36 } , new Person { Name = "Erika " , Location = "Barnaul Russia" , Age = 56 } , new Person { Name = "Lisabeth Terr" , Location = "Barquisimeto Venezuela" , Age = 67 } , new Person { Name = "Farrah " , Location = "Barra Mansa Brazil" , Age = 57 } , new Person { Name = "Domonique Biv" , Location = "Barrackpur India" , Age = 57 } , new Person { Name = "Jonah" , Location = "Barrancabermeja Colombia" , Age = 34 } }; The Sharp Column Indenter extension allows you to align by more than just the equals. As you start to see the benefits of alignment, you see that there is so much more to align with. Compare these: var benita = new Person() { Name = "Benita" }; var deedeeAlmon = new Person() { Name = "Deedee Almon Fonsec" }; var chaseHussain = new Person() { Name = "Chase Hussain" }; var cordia = new Person() { Name = "Cordia" }; benita.Age = 35; deedeeAlmon.Age = 12; chaseHussain.Age = 24; cordia.Age = 22; same code with column indention, var benita = new Person ( ) { Name = "Benita" } ; var deedeeAlmon = new Person ( ) { Name = "Deedee Almon Fonsec" } ; var chaseHussain = new Person ( ) { Name = "Chase Hussain" } ; var cordia = new Person ( ) { Name = "Cordia" } ; benita . Age = 35 ; deedeeAlmon . Age = 12 ; chaseHussain . Age = 24 ; cordia . Age = 22 ; By aligning by the dot we can clearly see that we are setting the same property on each variable, and the thing that changes is the variable name. This might seem a bit crazy now, but once you start aligning things, it’s addictive. Steps To Use Sharp Column Indenter Step 1: Select text you want to align. Select text you want to align Step 2: Select Apply Column Indention command in Edit menu. Select Apply Column Indention command in Edit menu Thats it! I turned Sharp Column Indenter project into a Github repo so you can, you know, contribute to it by making pull requests. If you have constructive criticism, or know of other tools that do the same thing, please leave a comment. comments powered by Disqus
__label__pos
0.999949
15 $\begingroup$ Compiler verification efforts often come down to proving the compiler fully abstract: that it preserves and reflects (contextual) equivalences. Instead of providing full abstraction proofs, some recent (categorical based) compiler verification work by Hasegawa [1,2] and Egger et. al. [3] prove the full completeness of various CPS translations. Question: What is the difference between full completeness and full abstraction? To me, completeness just looks like equivalence reflection for a translation and fullness appears to be a consequence of equivalence preservation. Note: Both Curien [7] and Abramsky [8] explore the relationship between definability, full abstraction, and to some extent full completeness. I suspect these resources may have the answer to my question, but after a surface read I have yet to confirm that. Some Background: The term "full completeness" was coined by Abramsky and Jagadeesan [4] to characterize the correctness of a game-semantic model of Multiplicative Linear Logic. Blute [5] provides the following definition: Let $\mathcal{F}$ be a free category. We say that a categorical model $\mathcal{M}$ is fully complete for $\mathcal{F}$ or that we have full completeness of $\mathcal{F}$ with respect to $\mathcal{M}$ if, with respect to some interpretation of the generators, the unique free functor $[\![ - ]\!] : \mathcal{F} \rightarrow \mathcal{M}$ is full. As far as I can tell, Hasegawa in [6] is the first to adapt full completeness to describe a program translation instead of a categorical semantic model. In this case, the Girard translation from simply typed lambda calculus to the linear lambda calculus. Later, in [1], he defines full completeness of the CPS translation $(\cdot)^\circ$ as: If $\Gamma^{\circ};\emptyset \vdash N : (\sigma^\circ \rightarrow o) \multimap o$ is derivable in the linear lambda calculus, then there exists $\Gamma \vdash M : \sigma$ in the computational lambda calculus such that $\Gamma^{\circ};\emptyset \vdash M^\circ = N : (\sigma^\circ \rightarrow o) \multimap o$ holds in the linear lambda calculus. (where $o$ is a base type in the linear lambda calculus (target language), but not in the computational lambda calculus (source language).) To me, Hasegawa's definition seems like a fullness and should really be combined with completeness to get full completeness. Egger et. al. [3] define full completeness of a CPS translation $(\cdot)^v$ as a combination of (1) completeness and (2) fullness: (1): If $\Theta \vdash M,N : \tau$ and $\Theta^v | - \vdash M^v =_{\beta\eta} N^v : !\tau^v$ then $\Theta \vdash M =_{\lambda_c} N : \tau$ (2): If $\Theta^v | - \vdash t : !\tau^v$ then there exists a term $\Theta \vdash M : \tau$ such that $\Theta^v | - \vdash M^v =_{\beta\eta} t : !\tau^v$ (where $=_{\lambda_c}$ is Moggi's computational equational theory) [1] "Linearly Used Effects: Monadic and CPS Transformations into Linear Lambda Calculus", Hasegawa 2002 [2] "Semantics of Linear Continuation-Passing in Call-by-Name", Hasegawa 2004 [3] "Linear-use CPS Translations in the Enriched Effect Calculus", Egger et. al. 2012 [4] "Games and Full Completeness for Multiplicative Linear Logic", Abramsky and Jagadeesan 1992 [5] "Category Theory for Linear Logicians", Blute 2003 [6] "Girard Translation and Logical Predicates", Hasegawa 2000 [7] "Definability and full abstraction", Curien 2007 [8] "Axioms for Definability and Full Completeness", Abramsky 1999 $\endgroup$ 12 $\begingroup$ Unfortunately, there are too many things are going on here. So, it is easy to mix things up. The use of "full" in "full completeness" and "full abstraction" refer to completely different ideas of fullness. But, there is also some vague connection between them. So, this is going to be a complicated answer. Full completeness: "Sound and complete" is a property you want for a traditional logic to have with respect to its semantics. Soundness means that whatever you can prove in the logic is true in the semantic model. Completeness means that whatever is true in the semantic model is provable in the logic. We say that a logic is sound and complete for a particular semantic model. When we come to constructive logic, such as Martin-Lof type theory or linear logic, we care not only about whether formulas provable, but also what their proofs are. A provable formula may have many proofs and a constructive logic wants to keep them apart. So, a semantics for a constructive logic involves specifying not only whether a formula is true, but also some abstract semantic notion of "proof" ("evidence") for its truth. Abramsky and colleagues coined the term "full completeness" to mean that the proofs in the logic can express all the semantic proofs in the model. So, "full" refers to proofs here. A "complete" logic can prove everything it needs to. A "fully complete" logic has all the proofs that it needs to have. So "full completeness" means "constructive completeness" or "proof completeness". This has nothing to do with full abstraction. Full abstraction: "Adequate and fully abstract" is a property you want for the semantic model of a programming language. (Note the first difference: we are now dealing with the properties of the semantic model, not the properties of the language!) Adequacy means that, whenever two terms have the same meaning in the semantic model, they are observationally equivalent in the programming language (with respect to some notion of execution). Full abstraction means that, if two terms are observationally equivalent, they have the same meaning in the semantic model. These ideas can be related to soundness and completeness, but in a somewhat contrived way. If we think of the semantic model of a programming language as a "logic" or a "proof method" to talk about observational equivalence, then adequacy means that this proof method is sound; full abstraction means that this proof method is complete. There is no notion of "full completeness" here because nobody has yet produced a semantic model that represents a constructive proof method. (But, such a thing is theoretically possible, and one of these days somebody might do so.) In your case, you are interested in translations rather than semantic models. The properties of adequacy and full abstraction can be extended to deal with translations as follows. You think of the target language as your "semantic model", i.e., a formalism that you fully understand somehow. If so, you have some notion of equivalence for it. Then, we say that the translation is adequate if, whenever the translations of two source programs are equivalent in the target language, they are observationally equivalent in the source language. We say that it is fully abstract if, whenever two source programs are observationally equivalent in the source language, their translations are equivalent in the target language. In reality, I don't know of any target languages that we really fully "understand". All we know is some other notion of observational equivalence for the target language. In that case, the translation is adequate if the observational equivalence of the translations in the target language implies observational equivalence in the source language. $$ \tau(M) \cong \tau(N) \Longrightarrow M \cong N$$ The translation is fully abstract if the observational equivalence of the terms in the source language implies the observational equivalence of the translations in the target language. $$M \cong N \Longrightarrow \tau(M) \cong \tau(N)$$ Some authors take "fully abstract translation" to mean the combination of these two properties: $$M \cong N \iff \tau(M) \cong \tau(N)$$ Egger et al seem to be similarly extending the idea of full completeness to translations. In their set-up, formulas are types and proofs are terms. Their translation translates types as well as terms. They call their translation fully complete if the translation of a type $A$ has only those terms that are obtained by translating the original terms of type $A$. $$\forall N : \tau(A).\; \exists M : A.\, \tau(M) = N$$ Now for the vague connection between full completeness and full abstraction. Proving that a semantic model or a translation is fully abstract often involves some of definability. This is because our languages are generally higher-order. So, if the semantic model or the target language has too many "contexts" then it will be able to poke our terms or semantic meanings in undesirable ways and spoil their equivalence. "Undesirable ways" means in ways that the programming language itself cannot poke them. So, to get full abstraction, we need to ensure that the "contexts" available in the semantic model or the target language do come from those in the source language in some form. Note that this relates to the full completeness property. Why do we want such properties? It has nothing to do with compilers! We want these properties in order to claim that the source language embeds into the target language. If we are happy with a particular target language (as being clean, understandable, somehow fundamental or God-given) then, if the source language embeds into it, then we can claim that there is nothing new in the source language. It is just a fragment of the target language that we know and love. It is just syntactic sugar. So, fully abstract translations are given by people to establish that particular target languages are great. They are also sometimes given by people who have a big or complicated language to deal with. So, instead of defining a semantics for it directly, they translate it to some core language and then give semantics to the core language. For instance, the Haskell report does this. But the full abstraction of these translations is rarely ever proved because the source languages are big and complicated. People take it on faith that the translation is good. Once again, this has nothing to do with compilers. Compilers are rarely ever adequate or fully abstract. And, they don't need to be! All that a compiler needs to do is to preserve the execution behavior of complete programs. The target language of a compiler is generally huge, which means that it has lots of contexts that can mess up equivalence. So, equivalent programs in the source language are almost never contextually equivalent when compiled. | cite | improve this answer | | $\endgroup$ • $\begingroup$ What do you mean by saying that there aren't any languages that we really fully "understand"? $\endgroup$ – Martin Berger Nov 29 '13 at 11:44 • $\begingroup$ What to you mean by saying that nobody has yet produced a semantic model that represents a constructive proof method? $\endgroup$ – Martin Berger Nov 29 '13 at 11:45 • 1 $\begingroup$ Sorry, but the implications for "translations" seem in the wrong direction to me, compared to your earlier text. Full abstraction for, say, PCF asks for M≅N⟹τ(M)≅τ(N) (with τ being the denotational semantics, and ignoring the need to change symbols): as you say, "Full abstraction means that, if two terms are observationally equivalent, they have the same meaning in the semantic model". But your implication is the other way around (namely, you write τ(M)≅τ(N)⟹M≅N)! Or do translations work differently from denotational semantics? $\endgroup$ – Blaisorblade Mar 26 '14 at 3:39 • 1 $\begingroup$ @Blaisorblade: You are absolutely right! I made a correction to the text of my answer. $\endgroup$ – Uday Reddy Mar 27 '14 at 21:23 • 1 $\begingroup$ Full abstraction is also of interest for language level security, and potentially for cross-language integration. I.e. it's useful to know that nothing in the target language can violate abstractions of the source language. $\endgroup$ – dmbarbour May 28 '14 at 20:46 5 $\begingroup$ Summary: full completeness means that the interpretation function is not just complete, but also surjective on programs. Full abstraction has no requirement for surjectivity. $\newcommand{\semb}[1]{[\![ #1 ]\!]}$ Details: The detailed meaning of full abstraction and full completeness depends on the nature of what/where/how you are interpreting. Here is a rendition for the interpretation of one typed programming language into another. You have an interpretation function $\semb{.}$ which maps three things. • Types $A$ of the source language to types $\semb{A}$ in the target language. • Contexts $\Gamma$ in the source language to contexts $\semb{\Gamma}$ in the target language. • Programs in context $\Gamma \vdash P : \alpha$ to programs in context $\semb{\Gamma} \vdash \semb{P} : \semb{\alpha}$. In a categorical interpretation the first two maps collapse into one. The interpretation function can have various properties, e.g. it can be compositional, or preserve termination or ... Full abstraction is one such property. Recall that full abstraction of $\semb{.}$ means that $$ P \cong^S Q\ \ \ \ \ \text{iff}\ \ \ \ \ \semb{P} \cong^T \semb{Q} $$ for all $P, Q$. Here $\cong^S$ is the chosen notion of typed program equivalence for the source language while $\cong^T$ plays that role for the target language. More precisely, because we are in a typed setting, $$ P \cong^S_{\Gamma, \alpha} Q\ \ \ \ \ \text{iff}\ \ \ \ \ \semb{P} \cong^T_{\semb{\Gamma}, \semb{\alpha}} \semb{Q} $$ for all appropriate $\Gamma, \alpha, P, Q$. Full abstraction implies that $\semb{.}$ is sound and complete.The reason we speak of full abstraction and not just soundness & completeness is that we also want that the target language is 'somehow' non-trivial, e.g. not a term model. But formalising this non-triviality is hard, and we just allude to non-triviality by terminology. Now full abstraction does not imply that $\semb{.}$ is surjective on types, contexts or programs in context. Full completeness means that the map $\semb{.}$ is (complete and) surjective on programs in context for all definable contexts and definable types, i.e. any program $\semb{\Gamma} \vdash Q : \semb{\alpha}$ in the target language is the denotation of some $\Gamma \vdash P : \alpha$ in the source language, i.e. $Q = \semb{P}$. Note that this does not require $\semb{.}$ to be surjective on types and contexts, because that property rarely holds in the interpretations we are typically interested in. | cite | improve this answer | | $\endgroup$ Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.984672
Rename multiple files using Python Prerequisite : OS module in Python In Python3, rename() method is used to rename a file or directory. This method is a part of the os module and comes extremely handy. Syntax for os.rename() : os.rename(src, dst) : src is source address of file to be renamed and dst is destination with the new name. Now say given n images in a folder having random names. For example, consider the image below: Now the requirement is to rename them in ordered fashion like hostel1, hostel2, …and so on. Doing this manually would be a tedious task but this target can be achieved using the rename() and listdir() methods in os module. The listdir method lists out all the content of a given directory. Syntax for listdir() : list = os.listdir(‘Src’) : Where Src is the source to be listed out. The following code will do the job for us. It traverses through the lists of all the images in xyz folder, defines the destination (dst) and source (src) addresses and renames using rename module.   Below is the implementation : filter_none edit close play_arrow link brightness_4 code # Pythono3 code to rename multiple  # files in a directory or folder    # importing os module import os    # Function to rename multiple files def main():        for count, filename in enumerate(os.listdir("xyz")):         dst ="Hostel" + str(count) + ".jpg"         src ='xyz'+ filename         dst ='xyz'+ dst                    # rename() function will         # rename all the files         os.rename(src, dst)    # Driver Code if __name__ == '__main__':            # Calling main() function     main() chevron_right Output : The output of this code will look something like this – Note : This code may not run in online IDE, since it use external image file directory. My Personal Notes arrow_drop_up Check out this Author's contributed articles. If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to [email protected]. See your article appearing on the GeeksforGeeks main page and help other Geeks. Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below. Improved By : ChristopherGarland Article Tags : 7 Please write to us at [email protected] to report any issue with the above content.
__label__pos
0.717419
2 $\begingroup$ Let $(G,+)$,$(G^\prime,+)$ two abelian groups and $H, H^\prime$ two proper subgroups of $G$ and $G^\prime$ respectively. Let $\text{Hom}(G,G^\prime) = \{f:G \rightarrow G^\prime : f\text{ homomorphism}\}$. Prove that $\text{Hom}(G,G^\prime)$ is a group endowed with the operation definded by: $$(f+g)(x) = f(x)+g(x), \forall x\in G$$ I thought it would be easily proven by showing that the group axioms hold. But trying this: $i)$ Closure: Let $f,g\in \text{Hom}(G,G^\prime)$. Then: $$(f+g)(x) = f(x) + g(x)$$ But I don't know how to show that $f(x)+g(x)$ could be again in $\text{Hom}(G,G^\prime)$. Do I have to somehow use the fact that $G^\prime$ is commutative? $ii)$ Associativity: Let $f,g,h\in \text{Hom}(G,G^\prime)$. Then: $$(f+g)(x)+h(x) = f(x)+g(x)+h(x)=f(x)+(g+h)(x)$$ $iii)$ Identity element: Let $f\in \text{Hom}(G,G^\prime)$. $$(f+0)(x) = f(x) + 0(x) = 0(x)+f(x) = f(x)$$ $iv)$ Inverse element: Let $f \in \text{Hom}(G,G^\prime)$. Then: $$0 = \dots?$$ I'm kind of stuck and don't know how to approach the last. Any help or advise would be appreciated. Thanks. $\endgroup$ 1 • $\begingroup$ The inverse element of $f$ is simply $-f$. $\endgroup$ – abhi01nat Sep 22 '20 at 12:28 1 $\begingroup$ $i)$ Closure: Let $f,g\in Hom(G,G^\prime)$. Then: $$(f+g)(x) = f(x) + g(x)$$ But I don't know how to show that $f(x)+g(x)$ could be again in $Hom(G,G^\prime)$. Do I have to somehow use the fact that $G^\prime$ is commutative? Indeed. So we have two group homomorphisms $f,g\in Hom(G,G^\prime)$ and we already know how to add them. So lets check that the result is a homomorphism: $$(f+g)(x+y)=f(x+y)+g(x+y)=f(x)+f(y)+g(x)+g(y)=$$ $$=f(x)+g(x)+f(y)+g(y)=(f+g)(x)+(f+g)(y)$$ Do you see where we've applied commutativity? $ii)$ Associativity: Let $f,g,h\in Hom(G,G^\prime)$. Then: $$(f+g)(x)+h(x) = f(x)+g(x)+h(x)=f(x)+(g+h)(x)$$ That looks ok. To be formally correct you have to write the initial and final conditions, i.e. you should start with $((f+g)+h)(x)$ and end with $(f+(g+h))(x)$. $iii)$ Identity element: Let $f\in Hom(G,G^\prime)$. $$(f+0)(x) = f(x) + 0(x) = 0(x)+f(x) = f(x)$$ Again, that looks fine. You just missed the definition: $0(x):=0_{G^\prime}$. $iv)$ Inverse element: Let $f \in Hom(G,G^\prime)$. Then: $0 = ...?$ How about $g(x):=-f(x)$? Can you complete the proof? $\endgroup$ 1 • $\begingroup$ Thank you very much for the help! $\endgroup$ – MathematicalMoose Sep 23 '20 at 7:51 1 $\begingroup$ You don't need to show that $f(x)+g(x)$ is an element in $Hom(G,G')$, this is of course not true. You need to show that the function $f+g$ is a function in $Hom(G,G')$, i.e you have to show that $f+g$ is a homomorphism. And indeed: $(f+g)(x+y)=f(x+y)+g(x+y)=f(x)+f(y)+g(x)+g(y)=(f(x)+g(x))+(f(y)+g(y))$ $=(f+g)(x)+(f+g)(y)$ We used commutativity here to change the order of summation. So $f+g$ is indeed a group homomorphism. The identity element is indeed the function which maps every element of $G$ to $0$. As for the inverse: given $f\in Hom(G,G')$ define a function $g:G\to G'$ by $g(x)=-f(x)$. This is a homomorphism because: $g(x+y)=-f(x+y)=-(f(x)+f(y))=-f(x)-f(y)=g(x)+g(y)$ And hence $g\in Hom(G,G')$. Also, for every $x\in G$ we have: $(f+g)(x)=f(x)+g(x)=f(x)-f(x)=0=0(x)$. And similarly $(g+f)(x)=0(x)$. So $g$ is the inverse of $f$. $\endgroup$ Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.995594
Take the tour × Mathematics Stack Exchange is a question and answer site for people studying math at any level and professionals in related fields. It's 100% free, no registration required. Given a $\sigma$-algebra on a Cartesian product of a collection of sets, do there always exist a $\sigma$-algebra on each set, so that their product $\sigma$-algebra on their Cartesian product will be the same as the given $\sigma$-algebra? If yes, how to construct the $\sigma$-algebras on the sets from the $\sigma$-algebra on their Cartesian product? If no, when they exist, how to construct the component $\sigma$-algebras? Thanks and regards! Any reference is appreciated as well. share|improve this question   You may need to specify if the collection is finite or infinite. –  Arturo Magidin Feb 21 '11 at 2:02   @Arturo: the collection is arbitrary. But if the treatment is different for finite and for infinite cases, I would like to hear for both. –  Tim Feb 21 '11 at 2:03   @Tim: I don't know off the top of my head, I'd need to try to remember what the $\sigma$-algebra on an infinite product is defined to be. Remember, for example, the issues with infinite products in topology: the product topology in the infinite product is not the infinite product of the topologies. –  Arturo Magidin Feb 21 '11 at 2:04 1   @Tim: Yes, according to planetmath (planetmath.org/encyclopedia/InfiniteProductMeasure.html) you define the product $\sigma$-algebra much like you do in topology: the generating sets have to have almost all components equal to the entire thing. So it's not the "box" $\sigma$-algebra. –  Arturo Magidin Feb 21 '11 at 2:07   @Arturo: Thanks! I understood how to generate product $\sigma$-algebra from component ones. Here I would like to know the construction in the reverse direction. –  Tim Feb 21 '11 at 2:13 show 1 more comment 2 Answers up vote 3 down vote accepted Let $A=\{0,1\}$, $B=\{0,1\}$, $\mathbb{S}=\{\emptyset, \{(0,0),(1,1)\}, \{(1,0),(0,1)\}, A\times B\}$. Then $\mathbb{S}$ is a $\sigma$-algebra on $A\times B$. Are there $\sigma$-algebras on $A$ and on $B$ whose product is $\mathbb{S}$? Well, there aren't that many $\sigma$-algebras on the two-element set. There's the total $\sigma$-algebra, $\mathcal{P}(\{0,1\})$, and the $\sigma$-algebra $\{\emptyset,\{0,1\}\}$, and that's it (if it contains any proper subset, then it contains its complement, and it is the entire thing). Call these $\mathbb{S}_1$ and $\mathbb{S}_2$, respectively. Is $\mathbb{S}$ the product of two of these? No. It's not equal to $\mathbb{S}_1\times\mathbb{S}_i$, because this $\sigma$-algebra contains $\{1\}\times \{0,1\}$, but $\mathbb{S}$ does not. Symmetrically, it's not equal to $\mathbb{S}_i\times \mathbb{S}_1$. So the only possibility would be that it is equal to $$\mathbb{S}_2\times\mathbb{S}_2 = \Bigl\{\emptyset,\{0,1\}\Bigr\} \times \Bigl\{\emptyset,\{0,1\}\Bigr\} = \Bigl\{ \emptyset, A\times B\Bigr\}$$ which is not equal to $\mathbb{S}$. So $\mathbb{S}$ is a $\sigma$-algebra on $A\times B$ which is not a product of a $\sigma$-algebra on $A$ and a $\sigma$-algebra on $B$. If it doesn't work for two, it has no hope of working for an infinite number of factors, so the potential distinction I suggested in the comments is in fact irrelevant. Added. To answer the added question: suppose that there do exist $\sigma$-algebras $\mathbb{S}_i$ on $X_i$ such that $\mathbb{S}=\prod\mathbb{S}_i$. Any $A_i\in\mathbb{S}_i$ will necessarily satisfy that $$A\times\prod_{j\neq i} X_j \in\mathbb{S}.$$ Conversely, suppose that $A\subseteq X_i$ is such that $A\times \prod_{j\neq i}X_j\in\mathbb{S}$. Looking at the projection from $\prod X_i$ to $X_i$, you get that $A\in\mathbb{S}_i$. The projection does not work; I'm fairly sure it is the case that such an $A$ will lie in $\mathbb{S}_i$, but I find myself unable to prove it right now. Edit: This gap was filled by Morning in his answer to this question. Assuming this is the case, then $\mathbb{S}_i$ would consist exactly of the subsets of $X_i$ such that $A\times\prod_{j\neq i}X_j\in\mathbb{S}$. You can see how this fails in the example above. The collection of all $A\subseteq \{0,1\}$ such that $A\times\{0,1\}\in\mathbb{S}$ is just $\{\emptyset, \{0,1\}\}$, but the product of these does not exhaust the $\sigma$-algebra. If you define the $\mathbb{S}_i$ this way, then the product $\sigma$-algebra $\prod\mathbb{S}_i$ is always contained in $\mathbb{S}$, but need not be equal to it. share|improve this answer   @Arturo: Thanks! When there exist, how to construct the component $\sigma$-algebras? –  Tim Feb 21 '11 at 2:45   @Arturo: is there some necessary and/or sufficient condition to tell if a sigma algebra on a Cartesian product can be decomposed this way? –  Tim Feb 21 '11 at 2:53   @Tim: I was adding some stuff. There is basically one possible candidate; if it works, then that's what you want, if it doesn't work, then it's not a product $\sigma$-algebra. –  Arturo Magidin Feb 21 '11 at 3:00   @Arturo: Thanks! (1) I was wondering how you got the "Conversely" in your added part. The projection is a measurable mapping, can it really always map a measurable set to another measurable set? (2) In your last paragraph, by defining $\mathbb{S}_i$ did you mean that $\mathbb{S}_i$ is the sigma algebra generated from the those $A$ of $X_i$ in the conversely part? –  Tim Feb 21 '11 at 4:05   @Tim: Maybe I'm off on that part. Let me think about it. For the last paragraph: if I'm not mistaken, the collection of all such $A$ is already a $\sigma$ algebra: it is closed under countable unions, contains the empty set and all of $X_i$, and it is closed under complements because the complement of $A\times\prod_{j\neq i}X_j$ is $(X_i-A)\times\prod_{j\neq i}X_j$. –  Arturo Magidin Feb 21 '11 at 4:13 show 2 more comments Suppose $A$ and $B$ are each 2-element sets. It's easy to create a measure on the 4-element set $A\times B$ which is not the product of measures on $A$ and $B$. Or perhaps I misunderstood you? share|improve this answer   Never mind -- yes, I did misunderstand you. You're talking about $\sigma$-algebras, not about measures. Sorry. –  Carl Offner Feb 21 '11 at 2:21   Well, I suppose you could simply consider this: let $A = \{a_0, a_1\}$ and $B = \{b_0, b_1\}$. Consider the $\sigma$-algebra on $A\times B$ consisting of the empty set, the whole set, and the two sets $\{(a_0, b_1), (a_1, b_0)\}$ and $\{(a_0, b_0), (a_1, b_1)\}$. –  Carl Offner Feb 21 '11 at 2:28   Yes, you can. You were right in the first place. Just replace "measure" with "$\sigma$-algebra". –  Arturo Magidin Feb 21 '11 at 2:32   Thanks! If we are now talking about measure space instead of sigma algebra, could you explain why a given measure on the product sigma algebra may not be decomposable into component measures so that their product is the given measure? –  Tim Feb 22 '11 at 15:21   @Tim: Sorry, I didn't see your question until just now. I won't try to typeset this, but given the 4-point product of two 2-point spaces (as above, but now saying that every set is measurable), you could define a measure on the four points by giving them values 1, 1, 1, and 2 (in any order, really). You should be able to convince yourself that no measures on the 2-point spaces could generate this. –  Carl Offner Feb 23 '11 at 23:01 add comment Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.983209
Home > There Was > There Was An Error Parsing The Query Webmatrix There Was An Error Parsing The Query Webmatrix If a question is poorly phrased then either ask for Has an SRB been considered for use 1 Num. Browse other questions tagged mysql sql sql-server-cenot having their Middle East? there having this error. Why don't miners get boiled Solution 2 Accept Solution Reject Solution Actually the system doesn't seems was apparently with line 15. an I just begin asp.net classed and this error in assress Is parsing query possible? Is it unethical of me and can I get in trouble if How do you enforce handwriting standards a probability represent anything? Not the answer webmatrix It shows- "There was in orbit to launch to escape velocity? I created a SQL All-Star 179480 Points 26098 Posts ModeratorMVP Re: There was an error parsingline break after the select. There Was An Error Parsing The Query Sql Ce Browse other questions tagged sql razor parsing per Byte of an Algorithm? https://forums.iis.net/t/next/1172242 to get started.need your password?I've found out that the following query works has expected```select count(*)from added a new item "SQL server compact 4.0 local database". Parsing sql queries Parse error: parse parsing into any Linux machine through grub2 secure? There Was An Error Parsing The Query. Token Line Number = 1 Token Line Offset = 5 Err. talkers used during WW2? I have multipleinsert statements, one per line, in a script error Mikesdotnett...webmatrix-2 razor-2 or ask your own question.Paradox of the wavefunction collapse into un error changed both.Kristian Lagaffe 19-Jan-14 2:52am JCahyaatnttearjee----- Solution didn't work! 2 solutions Top @mhcodner: But that's an entirely different issue, isn't it. While trying to insert data, i get an involving MultinormalDistribution?Members Sign in to your account ...or Join us Download, Vote, Comment, Publish. When answering a question you use a verbatim string literal, i.e. there Server compact 4.0 local database. file which which created the insert statements OK but didn't load any data. admin passwords if only trusted users are allowed on the network? parsing Does the reciprocal of row enabled an insert OK. Dozens of earthworms came on my terrace and died an questions SQL questions discussionsforums All Message Boards...What do you call a black eye. Why is international first class much decrease when dissolved in water and increase when burnt? http://stackoverflow.com/questions/8286685/there-was-an-error-parsing-the-query-token-line-number-1-token-line-offset who's solution guided me. the I wrote the whole Query in a single Line... an Help!!! Thanks to JCahyaatnttearjee a wizard early a good idea? traffic from one client to another travel via the access point?Kristian Lagaffe 19-Jan-14 1:51am parsing low speeds with flaps, slats extended?All the error ?? Par. : the immigrants more likely to commit crimes?How is being able to breakcommunities Sign up or log in to customize your list.I don't know how to5 Err.Java beginner exercise : Write a class "Air Plane" Why Treat my content as plain text, not as HTML Preview 0 … Existing http://enhtech.com/there-was/repair-sql-there-was-an-error-parsing-the-query.php Wanted to input from csv and used the created script from my csvstore to the next level?Disproving Euler proposition by brute force in C Does Wi-Fi email is in use. 1 Num. Par. : Is Certificate validation done completely local? Has an SRB been considered for usetable where i am trying to input theses values.Optional Password I have read and agree to the Terms of a black eye. With the space theerror for the last 3days and cant fixe it. Could you share a create table I have the Where ismore expensive than international economy class? 1 Num. Another Solution that I found by myself was thatinto any Linux machine through grub2 secure? How do I respond to the Query in C# Having error while connecting with .sdf file?Understand that English isn't everyone's first languageDisks on a Modern computer? John Let's work to help developers, an custom extension? Has an SRB been considered for useany flight search engines? error Seem to be working fineKind regardsAlbert SwartComments: ** Comment from web user: for homework assignments as a TA? Then i am tryin to insert these per Byte of an Algorithm? So i of westerners such that it doesn't appear to be yucky? Your Email This Par. : the message? error parsing the query up vote 3 down vote favorite I need help.... Stainless Steel Fasteners Does the mass of sulfur really [ Token line number = 1,Token line offset = 62,Token in er...
__label__pos
0.938629
Over a million developers have joined DZone. {{announcement.body}} {{announcement.title}} Investigating Memory Leaks Part 1 - Writing Leaky Code DZone's Guide to Investigating Memory Leaks Part 1 - Writing Leaky Code · Java Zone Free Resource Build vs Buy a Data Quality Solution: Which is Best for You? Gain insights on a hybrid approach. Download white paper now! I found this little problem the other day: there’s this server that runs for a while and then falls over. It’s then restarted by its startup script and the whole process repeats itself. This doesn't sound that bad as it isn't business critical although there is a significant loss of data, so I decided to take a closer look and to find out exactly what's going wrong. The first thing to note is that the server passes all it's unit tests and a whole bunch of integration tests. It runs well in all test environments using test data, so what's going wrong in production? It's easy to guess that in production it's probably under a heavier load than test, or than had been allowed for in its design, and therefore it's running out of resources, but what resources and where? That's the tricky question. In order to demonstrate how to investigate this problem, the first thing to do is to write some leaky sample code and I'm going to use the Producer Consumer pattern to do this because I can demonstrate the big problem with it. To demonstrate leaky code1 I need, as usual, I need a highly contrived scenario and in this scenario imagine that you work for a stockbroker on a system that records their sales of stocks and shares in a database. Orders are taken and placed in queue by a simple thread. Another thread then picks up the order from the queue and writes it to the database. The Order POJO is very straight forward and looks like this: public class Order { private final int id; private final String code; private final int amount; private final double price; private final long time; private final long[] padding; /**   * @param id   *  The order id   * @param code   *  The stock code   * @param amount   *  the number of shares   * @param price   *  the price of the share   * @param time   *  the transaction time   */public Order(int id, String code, int amount, double price, long time) {super();this.id = id;this.code = code;this.amount = amount;this.price = price;this.time = time;// This just makes the Order object bigger so that   // the example runs out of heap more quickly.this.padding = new long[3000];   Arrays.fill(padding, 0, padding.length - 1, -2);} public int getId() {return id;} public String getCode() {return code;} public int getAmount() {return amount;} public double getPrice() {return price;} public long getTime() {return time;} } The Order POJO is part of a simple Spring application, which has three key abstractions that create a new thread when Spring calls theirstart() methods. The first of these is OrderFeed. It's run() method creates a new dummy order and places it on the queue. It then sleeps for a moment before creating the next order. public class OrderFeed implements Runnable { private static Random rand = new Random(); private static int id = 0; private final BlockingQueue<Order> orderQueue; public OrderFeed(BlockingQueue<Order> orderQueue) {this.orderQueue = orderQueue;} /**   * Called by Spring after loading the context. Start producing orders   */public void start() { Thread thread = new Thread(this, "Order producer");   thread.start();} /** The main run loop */@Overridepublic void run() { while (true) {Order order = createOrder();   orderQueue.add(order);   sleep();}   } private Order createOrder() { final String[] stocks = { "BLND.L", "DGE.L", "MKS.L", "PSON.L", "RIO.L", "PRU.L","LSE.L", "WMH.L" };int next = rand.nextInt(stocks.length);long now = System.currentTimeMillis();   Order order = new Order(++id, stocks[next], next * 100, next * 10, now);return order;} private void sleep() {try {TimeUnit.MILLISECONDS.sleep(100);} catch (InterruptedException e) {e.printStackTrace();}   } } The second class is OrderRecord, which is responsible for taking orders from the queue and writing them to the database. The problem is that it takes significantly longer to write the orders to the database then it does to produce them. This is demonstrated by the long, 1 second, sleep in my recordOrder(…) method. public class OrderRecord implements Runnable { private final BlockingQueue<Order> orderQueue; public OrderRecord(BlockingQueue<Order> orderQueue) {this.orderQueue = orderQueue;} public void start() { Thread thread = new Thread(this, "Order Recorder");   thread.start();} @Overridepublic void run() { while (true) { try {Order order = orderQueue.take();   recordOrder(order);} catch (InterruptedException e) {e.printStackTrace();}   }   } /**   * Record the order in the database   *    * This is a dummy method   *    * @param order   *  The order   * @throws InterruptedException   */public void recordOrder(Order order) throws InterruptedException {TimeUnit.SECONDS.sleep(1);} } The result is obvious: the OrderRecord thread just can't keep up and the queue will get longer and longer until the JVM runs out of heap space and falls over. That's the big problem with the Producer Consumer pattern: the consumer has to be able to keep up with the producer. Just to prove his point I've added a third class, OrderMonitor, which prints the queue size every couple of seconds so that you can see things going wrong. public class OrderQueueMonitor implements Runnable { private final BlockingQueue<Order> orderQueue; public OrderQueueMonitor(BlockingQueue<Order> orderQueue) {this.orderQueue = orderQueue;} public void start() { Thread thread = new Thread(this, "Order Queue Monitor");   thread.start();} @Overridepublic void run() { while (true) { try {TimeUnit.SECONDS.sleep(2);int size = orderQueue.size();   System.out.println("Queue size is:" + size);} catch (InterruptedException e) {e.printStackTrace();}   }   } } Just to complete the line-up I’ve included the Spring context below: <?xml version="1.0" encoding="UTF-8"?> <beans xmlns="http://www.springframework.org/schema/beans" xmlns:p="http://www.springframework.org/schema/p" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:context="http://www.springframework.org/schema/context" xsi:schemaLocation="http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans.xsd http://www.springframework.org/schema/context http://www.springframework.org/schema/context/spring-context-3.1.xsd" default-init-method="start" default-destroy-method="destroy"> <bean id="theQueue" class="java.util.concurrent.LinkedBlockingQueue"/> <bean id="orderProducer" class="com.captaindebug.producerconsumer.problem.OrderRecord"> <constructor-arg ref="theQueue"/> </bean> <bean id="OrderRecorder" class="com.captaindebug.producerconsumer.problem.OrderFeed"> <constructor-arg ref="theQueue"/> </bean> <bean id="QueueMonitor" class="com.captaindebug.producerconsumer.problem.OrderQueueMonitor"> <constructor-arg ref="theQueue"/> </bean> </beans> The next thing to do is to start up the leaky sample code. You can do this by changing to the following directory /<your-path>/git/captaindebug/producer-consumer/target/classes ...and then by typing the following command: java -cp /path-to/spring-beans-3.2.3.RELEASE.jar:/path-to/spring-context-3.2.3.RELEASE.jar:/path-to/spring-core-3.2.3.RELEASE.jar:/path-to/slf4j-api-1.6.1-javadoc.jar:/path-to/commons-logging-1.1.1.jar:/path-to/spring-expression-3.2.3.RELEASE.jar:. com.captaindebug.producerconsumer.problem.Main ...where "path-to" is the path to your jar files The one thing that I really hate about Java is that fact that it's  SO difficult to run any program from the command line. You have to figure out what the classpath is, what options and properties need setting and what the main class is. Surely, it must be possible to think of a way of simply typing Java programName and the JVM figures out where everything is, especially if we start using convention over configuration: how hard can it be? You can also monitor the leaky app by attaching a simple jconsole. If you're running it remotely, then you'll need to add the following options to the command line above (picking your own port number): -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.port=9010 -Dcom.sun.management.jmxremote.local.only=false -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false …and if you take a look at the amount of heap used you’ll see it gradually increasing as the queue gets bigger. If a kilobyte of memory leaks away, then you’ll probably never spot it; if a giga byte of memory leaks the problem will be obvious. So, all that’s left to do for the moment is to sit back as wait for some memory to leak away before moving on to the next stage of the investigation. More on that next time... 1The source code can be found in my  Producer Consumer project on GitHub. Build vs Buy a Data Quality Solution: Which is Best for You? Maintaining high quality data is essential for operational efficiency, meaningful analytics and good long-term customer relationships. But, when dealing with multiple sources of data, data quality becomes complex, so you need to know when you should build a custom data quality tools effort over canned solutions. Download our whitepaper for more insights into a hybrid approach. Topics: Published at DZone with permission of Roger Hughes, DZone MVB. See the original article here. Opinions expressed by DZone contributors are their own. {{ parent.title || parent.header.title}} {{ parent.tldr }} {{ parent.urlSource.name }}
__label__pos
0.867661
Decentralized Artificial Intelligence: Integration of Blockchain and AI Update 14 Apr 2023 • Reading Time 6 Minute Image Decentralized Artificial Intelligence: Integration of Blockchain and AI Reading Time: 6 minutes Artificial intelligence (AI) technology is growing these days very rapidly. It refers to using machines and algorithms for tasks that usually require human intelligence, such as natural language processing, to solve problems. Previously, the development of AI projects were mostly developed exclusively by large companies such as Microsoft with OpenAI and Google with DeepMind. This exclusivity hindered independent researchers and developers from developing similar technologies, leading to the development of blockchain-based AI technology known as decentralized artificial intelligence. So, what is decentralized artificial intelligence, and what are its uses in the AI industry? Let’s read the following article. Article Summary • 🤖 Decentralized artificial intelligence (AI) is an AI system that combines the power of artificial intelligence with blockchain technology. It refers to the development of AI applications on a decentralized blockchain infrastructure. • 🧨 Some issues in the AI industry include centralized AI systems, data security vulnerabilities, and limited computing power to train AI models. • 🔗 Blockchain can address these limitations in AI by decentralizing AI systems, improving data security and computing power, and supporting the automation of AI agents in executing smart contracts. • 👠Some examples of crypto AI include SingularityNET (AGIX), Fetch AI (FET), Render Network (RNDR), Bittensor (TAO), and Alethea (ALI). What is Decentralized Artificial Intelligence (AI)? Decentralized Artificial Intelligence Source: Coinrivet Decentralized AI is a type of AI system that combines the power of artificial intelligence with blockchain technology. It refers to the development of AI applications on a decentralized blockchain infrastructure. Development and decision-making are not determined by a large entity or company but rather by a community of developers who take responsibility through a decentralized autonomous organization (DAO). Overview of Artificial Intelligence (AI) Artificial Intelligence (AI) replicates human intelligence in machines programmed to perform problem-solving, decision-making, perception, and natural language processing tasks. AI algorithms understand patterns and solve problems without the need for human assistance. It analyzes incoming data, learns from it, and uses its knowledge to solve a given task. So, these algorithms work by processing data independently and finding solutions without human assistance. The Problems in the AI Industry • Centralized AI Systems AI models rely on big data to train them to understand patterns and complete tasks. The data in these systems is usually collected from one or more sources and controlled by a single entity. It means that only a few organizations or individuals have access to and benefit from the technology. Thus, it could create a ‘digital divide,’ where those who do not have access to the data or technology are left behind. • Data Privacy and Security Personal information may include sensitive details such as biometrics, health records, financial information, and personal identity. Privacy concerns arise when this information is collected without consent, shared without authorization, or used for purposes other than intended. AI algorithms can potentially expose these vulnerabilities by analyzing and identifying patterns in the data. It can lead to data breaches and identity theft by unauthorized users. Therefore, computer system security needs to be strengthened to keep pace with the rapid development of AI. • Computing Power AI companies need extensive resources and significant computing power to train AI models. AI companies often face the choice of investing in their hardware and sacrificing scalability or choosing a cloud service provider and paying a higher price. The high computation cost for AI creates significant barriers for researchers and organizations pursuing advancements in this field. Based on data from Messari, Stablility.AI, the creator of Stable Diffusion, a text-to-image generator, used 4,000 Nvidia A100 GPUs running in AWS' cloud to train their AI models. It cost them more than $50 million for a month. How Blockchain Addresses the Limitations of AI? The integration of blockchain technology and AI can potentially develop a wide range of applications by utilizing the advantages of both technologies. Here’s how blockchain helps improve AI limitations. Decentralizing the System In AI centralized system, all development and decision-making processes are conducted by a single entity. However, in a decentralized system, many different entities can contribute to developing and improving the AI system. It can lead to a broader perspective and more innovation, as many contributors can participate in developing the AI system. In addition, by using a decentralized system, developers can create and train AI models transparently with less risk of bias. Improving Data Security Combining the two technologies can create a secure and intelligent system. Decentralized AI models can provide greater privacy for individuals. Blockchain technology can encrypt data and ensure only authorized users can access it. Blockchain technology uses encryption to protect data from unauthorized users, making it more secure than centralized AI systems. Authorized users can access the data, and multiple nodes can quickly detect and address any anomalies. This high security makes it more difficult for hackers to steal or destroy sensitive information. Decentralized Computing Power Decentralized computing networks can overcome the need for significant computing power to develop AI models. In this network, individuals or companies can use the ‘idle’ computing resources of the users who provides them. The user that provides the resources will be incentivized in the form of crypto assets. Decentralized computing networks can offer lower prices compared to centralized cloud providers. It is because there is no additional cost for the provider of computing resources, unlike the centralized providers that must purchase and maintain hardware to sell their services. AI Agent Automation The main advantage of integrating these two technologies is that it enables fully automated AI agents to perform transactions and make contracts with each other without the need for human intervention. Users can use a crypto asset to run the AI agent. Furthermore, the AI agent will execute smart contracts autonomously to carry out the user’s tasks. It can accelerate various AI-driven industries, such as tourism, transportation, robotics, e-commerce, etc. 5 AI-based Blockchain Projects Decentralized Artificial Intelligence Source: Turing As technology develops, including artificial intelligence, more and more crypto or blockchain-based projects aim to decentralize AI. Below are examples of some decentralized applications that use artificial intelligence: AI Marketplace – SingularityNET (AGIX) SingularityNet allows independent developers to easily share and sell AI programs using a public website and decentralized marketplace. The platform runs on the Ethereum and Cardano networks, making it one of the most data-secure AI marketplaces. Autonomous Agents – Fetch AI (FET) Fetch AI runs an automation system with Autonomous Economic Agent (AEA) technology, or Digital Twin. Digital Twin can execute tasks independently and share data with other Digital Twins. On March 28, 2023, Fetch AI launched a wallet integrated with GPT-3.5 to provide instant assistance. Users can also ask about wallet functions, project fundamentals, and the broader Fetch AI ecosystem. Decentralized Infrastructure – Render Network (RNDR) Render Network is one of the projects addressing computing power issues in the AI industry. Render Network is a peer-to-peer Graphics Processing Unit or GPU provider that connects users who want to render work with users with ‘idle’ GPUs. Quoted from Messari, rendering is taking raw information (materials, lighting, etc.) from a 2D or 3D scene and processing it on a computer to get the final result. This process is widely used in animation, games, simulators, and special effects in film and television. Decentralized Neural Network – Bittensor (TAO) Bittensor is a decentralized machine learning network. On this platform, independent developers collaboratively train machine learning models and deploy them across a distributed network of nodes. Unlike large entities that control machine learning development behind closed doors, Bittensor aims to create an open-source machine learning network accessible to anyone. Generative AI – Alethea (ALI) Alethea allows users to create avatar characters based on text descriptions. By utilizing AI technology, the avatar can interact with other users. The platform also combines AI technology with iNFT (intelligent NFT), which allows users to create NFTs embedded with AI animation, voice recognition, and generative AI capabilities. Conclusion The combination of blockchain and AI has given rise to decentralized AI, which utilizes the security and transparency of blockchain technology to enhance data processing. In this context, developers can gather data from multiple sources and verify it through the blockchain network, ensuring the accuracy and reliability of AI models. Overall, the integration between blockchain and AI is still in the early stages. But, the potential for creating new decentralized data applications is enormous. With interacting economic and technological factors, the trend of decentralized AI is likely to become more exciting than ever. Buy AI Tokens on Pintu Interested in investing in AI tokens? Take it easy, you can buy various AI tokens such as GRT, FET, ROSE, OCEAN and others safely and easily on Pintu. Furthermore, Pintu has subjected all its crypto assets to a thorough evaluation process, emphasizing the importance of prudence. Pintu is also compatible with popular wallets such as Metamask to facilitate your transactions. Download Pintu app on Play Store and App Store! Your security is guaranteed because Pintu is regulated and supervised by Bappebti and Kominfo. In addition to executing transactions, on Pintu Apps, you can also learn more about crypto through various Pintu Academy articles updated weekly! All Pintu Academy articles are made for knowledge and educational purposes, not as financial advice. References 1. Toshendra Kumar Sharma, The Potential Of Decentralized Artificial Intelligence In The Future, Blockchain Council, accessed 12 April 2023. 2. Patrick Dynamo Defi, The Synergies between AI and Crypto, Twitter, accessed 12 April 2023. 3. Sami Kassab, Decentralizing Machine Learning, Messari, accessed 12 April 2023. 4. AI Work, Decentralized Artificial Intelligence, Medium, accessed 13 April 2023. 5. Turing Team, The Future of AI and Blockchain Technology & How It Complements Each Other? Turing, accessed 13 April 2023. Author:Ginisita Dofany Share
__label__pos
0.963501
use strict; use Test::More tests => 6; BEGIN{use_ok('CGI::Application');} $ENV{CGI_APP_RETURN_ONLY} = 1; { my $app = CGI::Application->new; $app->header_type('none'); my $warn = ''; local $SIG{__WARN__} = sub { $warn = shift; }; $app->header_props(-type => 'banana/ripe'); like( $warn, qr/header_type set to 'none'/, "warn if we set header while header type is none", ); } { my $app = CGI::Application->new; eval { $app->header_props(123); }; like( $@, qr/odd number/i, "croak on odd number of non-ref args to header_props", ); eval { $app->header_add(123); }; like( $@, qr/odd number/i, "croak on odd number of non-ref args to header_add", ); } { my $app = CGI::Application->new; $app->header_props({ -type => 'banana/ripe' }); $app->header_add({ -expires => '1d' }); like( $app->run, qr{Content-type: banana/ripe}i, "headed added via hashref arg to header_props", ); like( $app->run, qr{^Expires: }im, "headed added via hashref arg to header_add", ); }
__label__pos
0.998641
Serializing Floats Writing and reading floating point numbers to disk (or network) in a cross-platform manner is surprisingly non-trivial. The IEE754 floating point standard doesn't specify which byte should go where in the memory. Also, the standard C library lacks functions to access various components of a floating point number. When you simply store a float somewhere in memory, the bytes in the floating point number can be laid out in little-endian or big-endian fashion. If you need to communicate this value to some other computer, then the layout should be converted to a standard one so that the other party can read it properly. Since the C platform lacks the necessary tools, I have to make my own. First of all, although very likely, little-endian and big-endian may not be the only options. For example, a machine could store two-byte sequences in reverse, creating a NUXI problem. I just can't know without trying it out, because standard libraries don't give me any information. My end goal is to read and write floats in little-endian format. The x64 family of processors already do this. I don't know how ARM based tablet and phone processors do it. Ideally, the tests should be run before compilation and the results should be used to activate relevant pieces of code. However, running a test program on the target might be too difficult. Since phones and tablets are pretty much walled garden environments, you can't simply upload a.out and run it. You need to package it as an application, adapt your code to run within the wierd environment and get results over the network somehow. This isn't going to be worth it for a small test like this. However, a framework which simply downloads some code from a secure server and runs it can be a good project. In any case, I'm going to go with a run-time test. The end result will work like this: First, I'll run the tests to figure out the floating point layout of the CPU. Based on the results, I'll compute two permutations, one for converting from native to little-endian and one for the other direction. When I want to write a float, I'll write it in native format to a properly aligned area. From there, I'll permute it out into the output buffer. Reading will be done in a similar fashion. The Permutations A CPU is free to store a floating point in memory in any layout. This is the CPU's private business. However, if it uses the IEEE754 standard, the stored 32-bit and 64-bit values should have the following format: 32-bit: SIGN(bit 31)-EXPONENT(bits 30-23)-FRACTION(bits 22-0) 64-bit: SIGN(bit 63)-EXPONENT(bits 62-52)-FRACTION(bits 51-0) The CPU could scramble the bits around, storing third bit of the fraction at the highest byte, putting the sign at the first byte etc. However, I doubt that any sane person would do such a thing. So, I'm going to assume that the fields exponent and fraction are represented continuously in the memory. For a 32-bit value, I should have the following four bytes: bit 76543210 A: SEEEEEEE B: EFFFFFFF C: FFFFFFFF D: FFFFFFFF The order of these bytes may differ from CPU to CPU, but the contents of these should be the same. Otherwise, the algorithm I'm implementing here would fail spectacularly. In order to find which byte goes where, I shall make special floating point numbers and then check which bytes have the 1s. Here is a template for 32 bit floating point numbers: SEEEEEEE EFFFFFFF FFFFFFFF FFFFFFFF bit 76543210 76543210 76543210 76543210 F is fraction, E is exponent and S is sign. For 32-bit numbers, finding out the order of the fraction bytes is sufficient to figure out the overall permutation. The only remaining byte is the upmost byte. For this purpose, I'll use denormalized floating point numbers. These numbers have 0 as exponent, so the only set bits are in the fraction. SEEEEEEE EFFFFFFF FFFFFFFF FFFFFFFF bit 76543210 76543210 76543210 76543210 ----------------------------------------- 00000000 00000000 00000000 00000001 = 1/2^149 = A 00000000 00000000 00000001 00000000 = A * 256 = B 00000000 00000001 00000000 00000000 = B * 256 sign bit check 10000000 00000000 00000000 00000000 = -0.0 Since these are denormalized numbers, they don't necessarily have accurate decimal or hexadecimal representations. The last floating point number is used as a check to verify the position of the sign bit. gcc-4.8.3 has no problems parsing the negative zero "-0.0" into the proper encoding. The 64-bit case is also the same. I just play with the fraction part to find out the encoding. The only byte which doesn't have a fraction bit is the upmost byte. SEEEEEEE EEEEFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF FFFFFFFF 76543210 76543210 76543210 76543210 76543210 76543210 76543210 76543210 ----------------------------------------------------------------------- 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000001 = 1 / 2^1074 = A 00000000 00000000 00000000 00000000 00000000 00000000 00000001 00000000 = A * 256 = B 00000000 00000000 00000000 00000000 00000000 00000001 00000000 00000000 00000000 00000000 00000000 00000000 00000001 00000000 00000000 00000000 00000000 00000000 00000000 00000001 00000000 00000000 00000000 00000000 00000000 00000000 00000001 00000000 00000000 00000000 00000000 00000000 00000000 00000001 00000000 00000000 00000000 00000000 00000000 00000000 Now, in order to find the permutations, I will store these floats somewhere in memory and then search that memory for a set byte. For the first 64-bit example above, the set byte corresponds to the lowest byte. If the index of this set byte is T, then it means that the 0zix little-endian byte should go to Tzix position when converting the float to native layout: T= find_byte(1) toNative[0]= T fromNative[T]= 0 Preliminary Code and Results Below is some preliminary code which does the work. I will maintain the library within alib since this is quite small and suitable for direct injection into another C module. #include <stdio.h> #include <stdlib.h> #include <stdint.h> uint8_t spc[32]; uint8_t *ali16; void align() { uint64_t T; T= (uint64_t) spc; if (T%16) { ali16= spc + (16-T%16); } else { ali16= spc; } } void prt_byte(uint8_t V) { int i; for(i=7;i>=0;i--) printf("%c", (V&(1<<i)) ? '1' : '.'); } void prt_bytes(int N) { int i; for(i=0;i<N;i++) { if (i) printf(" "); prt_byte(ali16[i]); } printf("\n"); } void prt_flo(float *V) { *(float*) ali16= *V; prt_bytes(4); } void prt_dou(double *V) { *(double*) ali16= *V; prt_bytes(8); } static int find_byte(int nb, uint8_t V) { int i; for(i=0;i<nb;i++) if (ali16[i]==V) return i; return 0; } void flo_layout(int dbg,float *F, uint8_t *fromNative,uint8_t *toNative) { char *msg[]= { " 1 / 2^149 = A ", " A * 256 = B ", " B * 256 = C ", }; int i; *F= 1; if (dbg) printf("finding float layout\n"); for(i=0;i<149;i++) *F/= 2; for(i=0;i<3;i++) { *(float*) ali16= *F; toNative[i]= find_byte(4, 1); if (dbg) { printf("%s\n", msg[i]); prt_flo(F); } *F *= 256; } *F= -0.0; *(float*) ali16= *F; if (dbg) { printf("negative zero= %g\n", *F); prt_flo(F); } toNative[i]= find_byte(4, 0x80); for(i=0;i<4;i++) fromNative[toNative[i]]= i; if (dbg) { printf("write: "); for(i=0;i<4;i++) printf(" %d", fromNative[i]); printf("\n"); printf(" read: "); for(i=0;i<4;i++) printf(" %d", toNative[i]); printf("\n"); } } void dou_layout(int dbg,double *F, uint8_t *fromNative,uint8_t *toNative) { char *msg[]= { " 1 / 2^1074 = A ", " A * 256 = B ", " B * 256 = C ", " C * 256 = D ", " D * 256 = E ", " E * 256 = F ", " F * 256 = G ", }; int i; *F= 1; if (dbg) printf("finding double layout\n"); for(i=0;i<1074;i++) *F/= 2; for(i=0;i<7;i++) { *(double*) ali16= *F; toNative[i]= find_byte(8, 1); if (dbg) { printf("%s\n", msg[i]); prt_dou(F); } *F *= 256; } *F= -0.0; *(double*) ali16= *F; if (dbg) { printf("negative zero= %g\n", *F); prt_dou(F); } toNative[i]= find_byte(8, 0x80); for(i=0;i<8;i++) fromNative[toNative[i]]= i; if (dbg) { printf("write: "); for(i=0;i<8;i++) printf(" %d", fromNative[i]); printf("\n"); printf(" read: "); for(i=0;i<8;i++) printf(" %d", toNative[i]); printf("\n"); } } int main() { float F; double D; uint8_t Pwrite[8], Pread[8]; align(); flo_layout(1,&F, Pwrite, Pread); dou_layout(1,&D, Pwrite, Pread); return 0; } Unsurprisingly, on an x64 machine the code results in trivial permutations: finding float layout 1 / 2^149 = A .......1 ........ ........ ........ A * 256 = B ........ .......1 ........ ........ B * 256 = C ........ ........ .......1 ........ negative zero= -0 ........ ........ ........ 1....... write: 0 1 2 3 read: 0 1 2 3 finding double layout 1 / 2^1074 = A .......1 ........ ........ ........ ........ ........ ........ ........ A * 256 = B ........ .......1 ........ ........ ........ ........ ........ ........ B * 256 = C ........ ........ .......1 ........ ........ ........ ........ ........ C * 256 = D ........ ........ ........ .......1 ........ ........ ........ ........ D * 256 = E ........ ........ ........ ........ .......1 ........ ........ ........ E * 256 = F ........ ........ ........ ........ ........ .......1 ........ ........ F * 256 = G ........ ........ ........ ........ ........ ........ .......1 ........ negative zero= -0 ........ ........ ........ ........ ........ ........ ........ 1....... write: 0 1 2 3 4 5 6 7 read: 0 1 2 3 4 5 6 7 Final Thoughts This is probably all make-work. However, I don't want to take chances. Even if all permutations turn out to be trivial on the platforms I use, doing floating point I/O in this way is still beneficial. There are some processors which are picky about the alignment of floating point values. This will at least help with that. Links IEEE An explanation of the IEEE format along with the meanings of different non-numbers. Portable Encoding From the same site, I found this accidentally while poking around. Gives a generalized encoding solution without relying on IEEE standard.
__label__pos
0.980334
Why you should avoid password autofill With all the different passwords you have to manage, it can be tempting to just let your browser store your passwords for convenience. However, this “convenience” can come at a cost: hackers easily accessing your accounts. If you’ve been considering using password autofill, it’s important to understand the risks that come with it and why […] The risks of password autofill Hackers may have found an effective way to track users using a web browser or password manager’s password autofill feature. Here’s how they do it and what you can do to protect yourself. Why password autofill is so dangerous Modern web browsers and password managers have a feature that enables usernames and passwords to be […] How can private browsing protect your online privacy? You may not know it, but some of the websites you visit or apps you download leave a trackable digital footprint. One step you can do to avoid leaving your online behavior exposed to third parties is to use private browsers to surf the internet. What is private browsing? Your web browser — whether it’s […] Which web browser is the most secure? Enterprise cybersecurity is a holistic system that involves employing security practices at every level of use. This includes picking out the most secure web browser. Consider the security features of these popular web browsers when picking yours. Microsoft Edge Microsoft Edge, Windows 10’s current default browser, is an improvement over its predecessor Internet Explorer (IE). […] Can private browsing keep you safe? Privacy is a precious commodity in this era. Every website you visit or app you download leaves a digital footprint that can be tracked by anyone. Fortunately, most — if not all — web browsers offer private browsing features to keep your internet activity a little bit safer from prying eyes. What is private browsing? […] Autocomplete password risks Web browsers come with features to improve user experience. One of the most popular ones is auto-fill passwords. These are designed for users to store and automatically use their account credentials to access websites and other applications. While auto-fill passwords are convenient, they come with security risks. Read more. Why auto-fill passwords are so dangerous […] How to pick the most secure browser The internet is a Wild West of sorts, as one could never know what kind of threat they’ll come across. This is why for businesses, it’s important to identify secure browsers to keep threats at bay. Microsoft Edge Microsoft Edge, Windows’ current default browser, is an improvement over its predecessor Internet Explorer (IE). Edge was […] Invaluable tips for online safety Let’s be honest, surfing the net in the comfort of your home or with the privacy of a small screen feels safe. However, certain sites could be snooping on your online activity by placing digital trackers called “cookies” on your devices. Here’s what happens to your personal details whenever you go online. With the headlines […] Autocomplete password risks Advertisements and “helpful” suggestions based on your internet browsing habits can be troubling. But what’s even more alarming is that hackers have found another way of tracking you via seemingly harmless autocomplete passwords. Here’s what you need to know. Why auto-fill passwords are so dangerous As of December 2018, there are 4.1 billion internet users […]
__label__pos
0.680576
[TASK] Sync CMS Fluid with Flow Fluid 1.1 (part1) [Packages/TYPO3.CMS.git] / typo3 / sysext / fluid / Classes / ViewHelpers / Format / StripTagsViewHelper.php 1 <?php 2 namespace TYPO3\CMS\Fluid\ViewHelpers\Format; 3 4 /* * 5 * This script is backported from the TYPO3 Flow package "TYPO3.Fluid". * 6 * * 7 * It is free software; you can redistribute it and/or modify it under * 8 * the terms of the GNU Lesser General Public License, either version 3 * 9 * of the License, or (at your option) any later version. * 10 * * 11 * The TYPO3 project - inspiring people to share! * 12 * */ 13 14 /** 15 * Removes tags from the given string (applying PHPs strip_tags() function) 16 * 17 * @see http://www.php.net/manual/function.strip-tags.php 18 * 19 * = Examples = 20 * 21 * <code title="default notation"> 22 * <f:format.stripTags>Some Text with <b>Tags</b> and an &Uuml;mlaut.</f:format.stripTags> 23 * </code> 24 * <output> 25 * Some Text with Tags and an &Uuml;mlaut. (strip_tags() applied. Note: encoded entities are not decoded) 26 * </output> 27 * 28 * <code title="inline notation"> 29 * {text -> f:format.stripTags()} 30 * </code> 31 * <output> 32 * Text without tags (strip_tags() applied) 33 * </output> 34 * 35 * @api 36 */ 37 class StripTagsViewHelper extends \TYPO3\CMS\Fluid\Core\ViewHelper\AbstractViewHelper { 38 39 /** 40 * Disable the escaping interceptor because otherwise the child nodes would be escaped before this view helper 41 * can decode the text's entities. 42 * 43 * @var boolean 44 */ 45 protected $escapingInterceptorEnabled = FALSE; 46 47 /** 48 * Escapes special characters with their escaped counterparts as needed using PHPs strip_tags() function. 49 * 50 * @param string $value string to format 51 * @return mixed 52 * @see http://www.php.net/manual/function.strip-tags.php 53 * @api 54 */ 55 public function render($value = NULL) { 56 if ($value === NULL) { 57 $value = $this->renderChildren(); 58 } 59 if (!is_string($value)) { 60 return $value; 61 } 62 return strip_tags($value); 63 } 64 } 65 66 ?>
__label__pos
0.999843
tech-x11 archive [Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index] xvfb not built on vax and x68k? x11/Xserver/Makefile.common and external/mit/xorg/server/xorg-server/Makefile.common specifically set XSERVER_XVFB=no for x68k and vax, but builds it for all other ports. This appears to have been set in the first version of x11/Xserver/Makefile.common. Does anyone have any idea why this is done (apart from possibly to save build time), and more to the point, any objections to making those two ports consistent with all the others so all NetBSD ports build the Xvfb virtual framebuffer X server? Home | Main Index | Thread Index | Old Index
__label__pos
0.91802
How do I change the artwork in a Power Presentation? How to change Vector artwork in a Power Presentation PresentationPro Vector graphics are fully editable in PowerPoint. All of the vector art can be modified inside PowerPoint. Most of the artwork is setup as a grouped set of PowerPoint shapes. Click on it once to select the group. With the group selected, you can resize the entire group or set a style to the entire group. Hold down the shift key while you resize it and the group will keep the aspect ratio and not get stretched out.   Click again, but on a specific piece inside the group to highlight just that shape. Now Right-click on that shape and you can use PowerPoint shape formatting to change things like color.  
__label__pos
0.982747
How Do Academic Disciplines Use PowerPoint? I’m happy to say that my article reviewing PPT files has been published. I used C# to automate the analysis of a 30,000 PowerPoint files from a large academic publisher. It was a fun exercise in “big” data, involving a lot of files and a lot of data cleaning. How Do Academic Disciplines Use PowerPoint? http://link.springer.com/article/10.1007/s10755-016-9381-8? This project analyzed PowerPoint files created by an academic publisher to supplement textbooks. An automated analysis of 30,263 files revealed clear differences by disciplines. Single-paradigm “hard” disciplines used less complex writing but had more words than multi-paradigm “soft” disciplines. The “hard” disciplines also used a greater number of small graphics and fewer large ones. Disciplines identified by students as being more effective users of PowerPoint used larger images and more complex sentences than disciplines identified as being less effective in this regard. This investigation suggests that PowerPoint best practices are not universal and that we need to account for disciplinary differences when creating presentation guidelines. http://link.springer.com/article/10.1007/s10755-016-9381-8?
__label__pos
0.942405
Ekstensi Visual Studio Code untuk memudahkan ngoding Ekstensi Visual Studio Code untuk memudahkan ngoding Hallo, pada kesempatan kali ini kita akan membahas seputar ekstensi yang dapat membantu dan mempermudah kita membuat coding pada Visual Studio Code. Mungkin sebagian dari teman – teman masih ada yang belum tau mengenai salah satu text editor ini. Jadi, Visual Studio Code ini merupakan salah satu text editor seperti sublime, atom, dan text editor […] The post Ekstensi Visual Studio Code untuk memudahkan ngoding first appeared on PUCC. Baca:  Perancangan Aplikasi SmartPhone Android Untuk Pengontrol Lampu Berbasiskan Mikrokontroler (2)
__label__pos
0.980815
XTM Cloud v 13.0 Data – Custom fields – Terminology custom fields 1. Go to Configuration > Data > Custom fields > Terminology custom fields. 2. Select the Add button and enter a field name. 3. Choose one type: 1. Checkbox 2. Date 3. Dropdown: enter a list of values 4. Multiple selection: enter a list of values 5. Number 6. Text field 4. Select Save. Note TBX name refers to the term property when it is exported to a TBX file. The concept or term radio buttons enable you to define where the field should be displayed. The Active checkbox allows you to activate or deactivate this custom field.
__label__pos
0.999916
PercentageCalculator .pro Discount Percentage Fraction to % Decimal to % 3 percent of 800? Percentage Calculator What is 'X' percent of 'W' What is % of Answer:  'P' out of 'W' as a percentage What is out of as a percentage? Answer: % 'P' is 'X' percent of What? is % of what? Answer: Solution for 'What is 3% of 800?' In the following example consider that: Solution Steps The following question is of the type "How much X percent of W", where W is the whole amount and X is the percentage figure or rate". Let's say that you need to find 3 percent of 800. What are the steps? Step 1: first determine the value of the whole amount. We assume that the whole amount is 800. Step 2: determine the percentage, which is 3. Step 3: Convert the percentage 3% to its decimal form by dividing 3 into 100 to get the decimal number 0.03: 3100 = 0.03 Notice that dividing into 100 is the same as moving the decimal point two places to the left. 3.0 → 0.30 → 0.03 Step 4: Finally, find the portion by multiplying the decimal form, found in the previous step, by the whole amount: 0.03 x 800 = 24 (answer). The steps above are expressed by the formula: P = W × X%100 This formula says that: "To find the portion or the part from the whole amount, multiply the whole by the percentage, then divide the result by 100". The symbol % means the percentage expressed in a fraction or multiple of one hundred. Replacing these values in the formula, we get: P = 800 × 3100 = 800 × 0.03 = 24 (answer) Therefore, the answer is 24 is 3 percent of 800. Sample percentage problems See also:
__label__pos
0.99989
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [List Home] Re: [wtp-dev] JSDT facet gone missing in 3.0.4? Looks like it was removed in 3.0.2 through bug 243183, and at least noticed at the latest from bug 21928, but not actually removed from the server tools plug-ins. I regret how poorly this was handled, and am surprised by the number of plugin.xmls that now have to be updated to prevent this error message. Even if the facet is added back into 3.0.4, is/will there be a mechanism for deprecating or automatically migrating it out of 3.1 (by M6 if that's the course of action we decide upon)? Regards, --- Nitin Dahyabhai Eclipse WTP Source Editing IBM Rational
__label__pos
0.624575
Comments on: Sleuthing for Temporary Space Spendthrifts https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/ Love Your Data Fri, 24 Mar 2017 15:13:09 +0000 hourly 1 https://wordpress.org/?v=4.7.3 By: Damir Vadas https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-9184253 Fri, 18 Dec 2015 18:01:17 +0000 http://www.pythian.com/news/?p=38635#comment-9184253 plan shows no paraleleism ]]> By: Ed https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-8060689 Fri, 01 May 2015 01:31:50 +0000 http://www.pythian.com/news/?p=38635#comment-8060689 Hi, Thanks for this very good info. Can you confirm that the SQL that you posted is the updated one with the group by fix on parallel execution? :-) Also, just want to clarify, is it correct to assume that the TEMP usage for the SQL is therefore TEMP_USAGE / TOTAL_EXECS? Can we assume whether a SQL is parallel or not by the value of PX_FLAGS? Any reply much appreciated. Thanks ]]> By: Damir Vadas https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-5031557 Sun, 09 Nov 2014 14:54:56 +0000 http://www.pythian.com/news/?p=38635#comment-5031557 SQL is: [PRE] INSERT INTO RPM_ROLLUP_GTT (ITEM, DIFF_ID, DEPT, CLASS, SUBCLASS, ZONE_NODE_TYPE, LOCATION, TARGET_ITEM, TARGET_DIFF_ID, TARGET_ZONE_NODE_TYPE, TARGET_ZONE_NODE, ROLLUP_TYPE) SELECT INNER.ITEM, INNER.DIFF_ID, INNER.DEPT, INNER.CLASS, INNER.SUBCLASS, INNER.ZONE_NODE_TYPE, INNER.LOCATION, INNER.TARGET_ITEM, INNER.TARGET_DIFF_ID, INNER.TARGET_ZONE_NODE_TYPE, INNER.TARGET_ZONE_NODE, :B14 FROM (SELECT I.ITEM, I.DIFF_ID, I.DEPT, I.CLASS, I.SUBCLASS, I.ZONE_NODE_TYPE, I.LOCATION, I.ZONE_ID, I.TARGET_ITEM, I.TARGET_DIFF_ID, I.TARGET_ZONE_NODE, I.TARGET_ZONE_NODE_TYPE, ROW_NUMBER () OVER (PARTITION BY I.ITEM, I.DIFF_ID, I.LOCATION, I.ZONE_NODE_TYPE ORDER BY I.ACTION_DATE) ROW_CNT, SUM ( DECODE (I.SELLING_RETAIL, NVL (P.SELLING_RETAIL, -999), 0, 1) + DECODE (I.SELLING_RETAIL_CURRENCY, NVL (P.SELLING_RETAIL_CURRENCY, -999), 0, 1) + DECODE (I.SELLING_UOM, NVL (P.SELLING_UOM, -999), 0, 1) + DECODE (NVL (I.MULTI_UNITS, -999), NVL (P.MULTI_UNITS, -999), 0, 1) + DECODE (NVL (I.MULTI_UNIT_RETAIL, -999), NVL (P.MULTI_UNIT_RETAIL, -999), 0, 1) + DECODE (NVL (I.MULTI_UNIT_RETAIL_CURRENCY, -999), NVL (P.MULTI_UNIT_RETAIL_CURRENCY, -999), 0, 1) + DECODE (NVL (I.MULTI_SELLING_UOM, -999), NVL (P.MULTI_SELLING_UOM, -999), 0, 1) + DECODE (NVL (I.CLEAR_RETAIL, -999), NVL (P.CLEAR_RETAIL, -999), 0, 1) + DECODE (NVL (I.CLEAR_RETAIL_CURRENCY, -999), NVL (P.CLEAR_RETAIL_CURRENCY, -999), 0, 1) + DECODE (NVL (I.CLEAR_UOM, -999), NVL (P.CLEAR_UOM, -999), 0, 1) + DECODE (NVL (I.SIMPLE_PROMO_RETAIL, -999), NVL (P.SIMPLE_PROMO_RETAIL, -999), 0, 1) + DECODE (NVL (I.SIMPLE_PROMO_RETAIL_CURRENCY, -999), NVL (P.SIMPLE_PROMO_RETAIL_CURRENCY, -999), 0, 1) + DECODE (NVL (I.SIMPLE_PROMO_UOM, -999), NVL (P.SIMPLE_PROMO_UOM, -999), 0, 1) + DECODE (I.ON_SIMPLE_PROMO_IND, NVL (P.ON_SIMPLE_PROMO_IND, -999), 0, 1) + DECODE (NVL (I.COMPLEX_PROMO_RETAIL, -999), NVL (P.COMPLEX_PROMO_RETAIL, -999), 0, 1) + DECODE ( NVL (I.COMPLEX_PROMO_RETAIL_CURRENCY, -999), NVL (P.COMPLEX_PROMO_RETAIL_CURRENCY, -999), 0, 1) + DECODE (NVL (I.COMPLEX_PROMO_UOM, -999), NVL (P.COMPLEX_PROMO_UOM, -999), 0, 1) + DECODE (I.ON_COMPLEX_PROMO_IND, NVL (P.ON_COMPLEX_PROMO_IND, -999), 0, 1) + DECODE (NVL (I.PRICE_CHANGE_ID, -999), NVL (P.PRICE_CHANGE_ID, -999), 0, 1) + DECODE (NVL (I.PRICE_CHANGE_DISPLAY_ID, -999), NVL (P.PRICE_CHANGE_DISPLAY_ID, -999), 0, 1) + DECODE (NVL (I.PC_EXCEPTION_PARENT_ID, -999), NVL (P.PC_EXCEPTION_PARENT_ID, -999), 0, 1) + DECODE (NVL (I.PC_CHANGE_TYPE, -999), NVL (P.PC_CHANGE_TYPE, -999), 0, 1) + DECODE (NVL (I.PC_CHANGE_AMOUNT, -999), NVL (P.PC_CHANGE_AMOUNT, -999), 0, 1) + DECODE (NVL (I.PC_CHANGE_CURRENCY, -999), NVL (P.PC_CHANGE_CURRENCY, -999), 0, 1) + DECODE (NVL (I.PC_CHANGE_PERCENT, -999), NVL (P.PC_CHANGE_PERCENT, -999), 0, 1) + DECODE (NVL (I.PC_CHANGE_SELLING_UOM, -999), NVL (P.PC_CHANGE_SELLING_UOM, -999), 0, 1) + DECODE (NVL (I.PC_NULL_MULTI_IND, -999), NVL (P.PC_NULL_MULTI_IND, -999), 0, 1) + DECODE (NVL (I.PC_MULTI_UNITS, -999), NVL (P.PC_MULTI_UNITS, -999), 0, 1) + DECODE (NVL (I.PC_MULTI_UNIT_RETAIL, -999), NVL (P.PC_MULTI_UNIT_RETAIL, -999), 0, 1) + DECODE ( NVL (I.PC_MULTI_UNIT_RETAIL_CURRENCY, -999), NVL (P.PC_MULTI_UNIT_RETAIL_CURRENCY, -999), 0, 1) + DECODE (NVL (I.PC_MULTI_SELLING_UOM, -999), NVL (P.PC_MULTI_SELLING_UOM, -999), 0, 1) + DECODE (NVL (I.PC_PRICE_GUIDE_ID, -999), NVL (P.PC_PRICE_GUIDE_ID, -999), 0, 1) + DECODE (NVL (I.CLEAR_EXCEPTION_PARENT_ID, -999), NVL (P.CLEAR_EXCEPTION_PARENT_ID, -999), 0, 1) + DECODE (NVL (I.CLEARANCE_ID, -999), NVL (P.CLEARANCE_ID, -999), 0, 1) + DECODE (NVL (I.CLEARANCE_DISPLAY_ID, -999), NVL (P.CLEARANCE_DISPLAY_ID, -999), 0, 1) + DECODE (NVL (I.CLEAR_MKDN_INDEX, -999), NVL (P.CLEAR_MKDN_INDEX, -999), 0, 1) + DECODE (NVL (I.CLEAR_START_IND, -999), NVL (P.CLEAR_START_IND, -999), 0, 1) + DECODE (NVL (I.CLEAR_CHANGE_TYPE, -999), NVL (P.CLEAR_CHANGE_TYPE, -999), 0, 1) + DECODE (NVL (I.CLEAR_CHANGE_AMOUNT, -999), NVL (P.CLEAR_CHANGE_AMOUNT, -999), 0, 1) + DECODE (NVL (I.CLEAR_CHANGE_CURRENCY, -999), NVL (P.CLEAR_CHANGE_CURRENCY, -999), 0, 1) + DECODE (NVL (I.CLEAR_CHANGE_PERCENT, -999), NVL (P.CLEAR_CHANGE_PERCENT, -999), 0, 1) + DECODE (NVL (I.CLEAR_CHANGE_SELLING_UOM, -999), NVL (P.CLEAR_CHANGE_SELLING_UOM, -999), 0, 1) + DECODE (NVL (I.CLEAR_PRICE_GUIDE_ID, -999), NVL (P.CLEAR_PRICE_GUIDE_ID, -999), 0, 1) + DECODE (NVL (I.LOC_MOVE_FROM_ZONE_ID, -999), NVL (P.LOC_MOVE_FROM_ZONE_ID, -999), 0, 1) + DECODE (NVL (I.LOC_MOVE_TO_ZONE_ID, -999), NVL (P.LOC_MOVE_TO_ZONE_ID, -999), 0, 1) + DECODE (NVL (I.LOCATION_MOVE_ID, -999), NVL (P.LOCATION_MOVE_ID, -999), 0, 1)) OVER (PARTITION BY I.ITEM, I.DIFF_ID, I.LOCATION, I.ZONE_NODE_TYPE) VARIATION_COUNT FROM RPM_FUTURE_RETAIL_GTT P, (SELECT R.*, DECODE ( :B14, :B13, R.ITEM, :B12, R.ITEM_PARENT, :B11, R.ITEM_PARENT, :B10, R.ITEM, :B9, R.ITEM_PARENT, :B8, R.ITEM_PARENT, :B7, R.ITEM_PARENT, :B6, R.ITEM, :B5, R.ITEM, :B4, R.ITEM, :B3, R.ITEM, :B2, R.ITEM_PARENT) AS TARGET_ITEM, DECODE ( :B14, :B13, NVL (R.DIFF_ID, ‘-999’), :B12, ‘-999’, :B11, ‘-999’, :B10, ‘-999’, :B9, ‘-999’, :B8, R.DIFF_ID, :B7, R.DIFF_ID, :B6, R.DIFF_ID, :B5, ‘-999’, :B4, ‘-999’, :B3, ‘-999’, :B2, R.DIFF_ID) AS TARGET_DIFF_ID, DECODE ( :B14, :B13, 1, :B12, R.ZONE_NODE_TYPE, :B11, R.ZONE_NODE_TYPE, :B10, 1, :B9, 1, :B8, R.ZONE_NODE_TYPE, :B7, R.ZONE_NODE_TYPE, :B6, 1, :B5, R.ZONE_NODE_TYPE, :B4, R.ZONE_NODE_TYPE, :B3, 1, :B2, 1) AS TARGET_ZONE_NODE_TYPE, DECODE ( :B14, :B13, R.ZONE_ID, :B12, R.LOCATION, :B11, R.LOCATION, :B10, R.ZONE_ID, :B9, R.ZONE_ID, :B8, R.LOCATION, :B7, R.LOCATION, :B6, R.ZONE_ID, :B5, R.LOCATION, :B4, R.LOCATION, :B3, R.ZONE_ID, :B2, R.ZONE_ID) AS TARGET_ZONE_NODE FROM RPM_FUTURE_RETAIL_GTT R WHERE R.CUR_HIER_LEVEL IN (SELECT /*+ no_unnest CARDINALITY(cr, 1) */ VALUE (CR) FROM TABLE ( CAST ( :B15 AS OBJ_VARCHAR_ID_TABLE)) CR) AND R.MAX_HIER_LEVEL IN (SELECT /*+ no_unnest CARDINALITY(mx, 10) */ VALUE (MX) FROM TABLE ( CAST ( :B16 AS OBJ_VARCHAR_ID_TABLE)) MX) AND R.DEPT = :B1 AND NVL (R.TIMELINE_SEQ, 0) != -999 AND ROWNUM > 0) I WHERE I.DEPT = P.DEPT(+) AND I.TARGET_ITEM = P.ITEM(+) AND I.TARGET_DIFF_ID = NVL (P.DIFF_ID(+), ‘-999’) AND I.TARGET_ZONE_NODE = P.LOCATION(+) AND I.TARGET_ZONE_NODE_TYPE = P.ZONE_NODE_TYPE(+) AND I.ACTION_DATE = P.ACTION_DATE(+) AND I.MAX_HIER_LEVEL = P.MAX_HIER_LEVEL(+) AND P.CUR_HIER_LEVEL(+) = :B17) INNER WHERE INNER.VARIATION_COUNT = 0 AND INNER.ROW_CNT = 1 [/PRE] ]]> By: Damir Vadas https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-5031511 Sun, 09 Nov 2014 14:53:27 +0000 http://www.pythian.com/news/?p=38635#comment-5031511 And SQL is insert into GTT table ——————————————————————————————————————— | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | ——————————————————————————————————————— | 0 | INSERT STATEMENT | | | | 14M(100)| | | 1 | LOAD TABLE CONVENTIONAL | | | | | | |* 2 | VIEW | | 1 | 159 | 14M (1)| 00:06:18 | | 3 | WINDOW SORT | | 1 | 1096 | 14M (1)| 00:06:18 | | 4 | NESTED LOOPS OUTER | | 1 | 1096 | 14M (1)| 00:06:18 | | 5 | VIEW | | 1 | 583 | 14M (1)| 00:06:18 | | 6 | COUNT | | | | | | |* 7 | FILTER | | | | | | |* 8 | TABLE ACCESS FULL | RPM_FUTURE_RETAIL_GTT | 512K| 283M| 217K (2)| 00:00:06 | |* 9 | COLLECTION ITERATOR PICKLER FETCH| | 1 | 2 | 29 (0)| 00:00:01 | |* 10 | COLLECTION ITERATOR PICKLER FETCH| | 10 | 20 | 4 (0)| 00:00:01 | |* 11 | TABLE ACCESS BY INDEX ROWID | RPM_FUTURE_RETAIL_GTT | 1 | 513 | 5 (0)| 00:00:01 | |* 12 | INDEX SKIP SCAN | RPM_FUTURE_RETAIL_GTT_I5 | 63 | | 4 (0)| 00:00:01 | ——————————————————————————————————————— ]]> By: Damir Vadas https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-5031459 Sun, 09 Nov 2014 14:50:45 +0000 http://www.pythian.com/news/?p=38635#comment-5031459 Your query, shows TEMP usage that I do not have on mine TEMP_GROUP: POSITION SQL_ID SQL_PLAN_HASH_VALUE SQL_PLAN_OPERATION SQL_PLAN_LINE_ID NR_EXECS TEMP_MB ———- ————- ——————- ———————- —————- ———- ———- 1 aj5tu2zhpnqdr 3447736756 TABLE ACCESS 8 13 178131 Mine TEMP_GROUP has 120 GB and result shows 178 GB use. Any explanation? Thx Damir Vadas ]]> By: Randolf Geist https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-1388055 Tue, 29 Jan 2013 10:31:48 +0000 http://www.pythian.com/news/?p=38635#comment-1388055 Hi Marc, while working on a completely different matter I came again across this PGA/TEMP ASH stuff, and I have to correct my previous statement, because I mixed things up. The accuraccy of the PGA/TEMP figures in the ASH data seems to be fine. The problem I’ve stumbled upon regarding PGA/TEMP is the fact that I tried to come up with a PGA/TEMP usage figure over time, and due to the sampling this figure for Parallel Execution isn’t that consistent, as not always every PX Slave involved gets sampled every time, so a sum(PGA/TEMP) per sample time jumps a bit up and down, depending on which PX Slaves were sampled or not. This doesn’t mean however that the PGA/TEMP figures shown in the existing samples are incorrect/unreliable, so my previous statement about the reliability needs to be corrected. However, and that is what triggered me to mention this unreliability, the delta figures regarding I/O, that are also part of the ASH data since 11.2, are actually sometimes quite incorrect and fall short of the actual number of I/Os and volume of I/O performed. This I can show quite easily and if I find the time I will blog about it. The same inconsistency can be seen in the Real-Time SQL Monitoring report (“I/O requests” / “I/O Throughput” in the “Metrics” tab), which uses all these figures to populate the “Metrics” tab in the “EM/Active” version of the report. I’ve mixed these two problems, and thought that the PGA/TEMP figures are also unreliable, but that doesn’t seem to be the case – the delta I/O figures seem to be the ones that are often unreliable. Randolf ]]> By: Randolf Geist https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-1312263 Thu, 03 Jan 2013 19:46:21 +0000 http://www.pythian.com/news/?p=38635#comment-1312263 Hi Marc, thanks for taking up my comments and Happy New Year! As I already tried to explain, I don’t think that the ASH TEMP and PGA information can be broken down on SQL execution plan line information, since it seems to reflect the overall PGA / TEMP allocation of the process / session at the time of sampling across all workareas active, so I still believe that your example query showing PGA/TEMP on execution plan line level is a bit misleading. Regarding the accuracy: I still have to come up with good, reproducible examples, but I regularly run my XPLAN_ASH tool on real-life client queries and I can see quite often that the ASH TEMP information does not correspond to the real TEMP allocation as shown in Real-Time SQL Monitoring. One particular issue seems to come from BUFFERED operations that are only there in case of Parallel Execution (like HASH JOIN BUFFERED or additional BUFFER SORTs specific to PX), so may be serial executions are OK, and only Parallel Executions using BUFFERED operations might be affected by the problem. I haven’t had the time yet to investigate further, but I’ve already witnessed numerous inconsistencies. Randolf ]]> By: Marc Fielding https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-1300305 Mon, 31 Dec 2012 21:53:21 +0000 http://www.pythian.com/news/?p=38635#comment-1300305 Hello Randolf, Thanks for the insightful comments. I went back to my original working version of the query, and I had a “PARTITION BY sql_id, sql_exec_id, session_id” clause in it, and somehow lost the session_id part in the final version. Like you say, the temp space counts are independent for each PX process, and I’ve corrected the query. It also now shows the usage in hashing and sorting operations, which does make a lot more sense than PX receive. As far as the accuracy of the data, I would expect that it would be driven off the session statistic “temp space allocated (bytes)”. A test on one (otherwise idle) 11gR2 system: — Temp space consumer select count(distinct id) from big_table where rownum < 10000000; -- While running, in another session select value/1024/1024 from v$sesstat where statistic# = (select statistic# from v$statname where name = 'temp space allocated (bytes)') and sid = 804; 121 select temp_space_allocated/1024/1024 from v$active_session_history where session_id = 804 and sample_time > sysdate-1/(24*60*60); 121 select used_blocks*8192/1024/1024 from v$sort_segment; 121 So, in my simple test at least, the numbers from ASH, v$sort_segment, and v$sesstat do line up at 121m. Are you saying that they’re all inaccurate? Thanks again, and happy new year! Marc ]]> By: Randolf Geist https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-1274579 Tue, 25 Dec 2012 13:55:34 +0000 http://www.pythian.com/news/?p=38635#comment-1274579 Hi Marc, it’s good to see someone writing about this new ASH feature. A few comments: 1. I think your query using the LAG function lacks a bit of precision: In case of a Parallel Execution there will be multiple samples per SAMPLE_ID (one per sampled Parallel Slave), so your ORDER BY SAMPLE_ID of the LAG function is not deterministic for Parallel Executions I think. 2. This also makes the approach of building a delta between samples questionable I think, at least for Parallel Execution. After all it is based on sampling – what if the particular session being sampled doesn’t show up afterwards, or only 100 samples later having released the previous TEMP allocation but allocating TEMP for a completely different part of the execution plan? The question here is probably, what kind of problem do you want to address using your query? 3. I don’t think that you can break down the TEMP_SPACE_ALLOCATED information from ASH on SQL plan line operation level. You can see this from your example where “PX RECEIVE” gets TEMP space allocated. But a “PX RECEIVE” operation doesn’t have a workarea assigned, so why should it show up? I believe the PGA and TEMP allocation in ASH data is not on operation level, but on session level, so represents the total aggregate for all active workareas of that SQL execution at the moment of sampling. This is different from Real-Time SQL Monitoring and Rowsource Statistics where you get the allocation data on workarea level. It looks more like the ASH data is taken from V$PROCESS for PGA and V$TEMPSEG_USAGE for TEMP. 4. I do similar analysis in my XPLAN_ASH tool and I can show there that the TEMP space information from ASH is unreliable and quite often falls short of the actual TEMP space allocation reported by Rowsource Statistics or Real-Time SQL Monitoring, so the information has to be treated carefully and might be misleading (or the way I treat the data in XPLAN_ASH is misleading, but at present I don’t think so). Randolf ]]> By: David Mann https://www.pythian.com/blog/sleuthing-for-temporary-space-spendthrifts/#comment-1263295 Sat, 22 Dec 2012 20:05:42 +0000 http://www.pythian.com/news/?p=38635#comment-1263295 I always wanted to explore an AFTER SERVERERROR trigger to see if there was anything useful I could glean when a ORA-01555 fires. I like your approach though, I will give it a shot. ]]>
__label__pos
0.834453
Tuesday, March 12, 2019 Docker Copy Files When building a Docker image using the Dockerfile, if you need to copy some files from the Docker Engine machine to the Docker Image, you need to use the "correct" source path where the files are located. Let's say you have the following files: /mydocker/Dockerfile /mydocker/file1.sh /mydocker/file2.sh you can't specify the following command within your Dockerfile COPY /mydocker/file1.sh /etc/ COPY /mydocker/file2.sh /etc/ instead, you need to do the following mkdir /mydocker/resources mv /mydocker/file1.sh /mydocker/resources mv /mydocker/file2.sh /mydocker/resources then, in your Dockerfile, specify the following COPY command COPY resources /etc/ No comments:
__label__pos
0.811848
What is CSV? Comma separated values (CSV) is a file format that is used to exchange data between programs. It uses commas to separate values and requires equally many values on each line. Example For a game, we want to save the best player’s score, its name and the date on which the score was achieved. The format we choose therefore is: [player’s name], [score], [date] Bob,12,2013-01-02 This means that Bob scored 12 points at 2013-01-02. We can easily store multiple scores by writing them one below another. Alice,15,2013-03-04 Bob,12,2013-01-02 Charlie,9,2013-05-06 Observation:  CSV files can hold one to many records or lines with values. Example We have a sensor that saves the measured temperature in a file. The first value of a line describes the hour, the second the minute, and the third the temperature. The format is: [hour], [minute], [temperature] 7,38,12 7,39,12 7,40,13 7,41,13 7,42,13 For instance, the first line means that at 7:38 a temperature of 12 were measured. Observation: It can be difficult to understand a CSV file if you don’t know what every value stands for – what the format is. Implementation Read CSV To load a CSV file, a while-loop reads line for line. Within the loop the subroutine ExtractValues() is called to extract all values from the line. After the loop is finished, all values are written to TextWindow by ShowValues(). filename = "data.csv" line_number = 1   '----------------------------------------------------------------------------- '   reads line for line, saves values to csv and shows them '----------------------------------------------------------------------------- line = File.ReadLine(filename, line_number)   While line <> ""   ExtractValues()       line_number = line_number + 1   line = File.ReadLine(filename, line_number) EndWhile   ShowValues()       '----------------------------------------------------------------------------- '   ShowValues - writes all values of csv to the TextWindow '----------------------------------------------------------------------------- Sub ShowValues   For line_number = 1 to Array.GetItemCount(csv)        For value_number = 1 to Array.GetItemCount(csv[line_number])       TextWindow.Write(csv[line_number][value_number] + " ")     EndFor     TextWindow.WriteLine("")   EndFor EndSub The ExtractValues() subroutine works on the current line. It extracts all values and saves them into the CSV array at the according position. '----------------------------------------------------------------------------- '   ExtractValues - saves all values of the current line to csv '----------------------------------------------------------------------------- Sub ExtractValues    'repeat as long as there is a comma   While Text.GetIndexOf(line,",") > 0     comma_position = Text.GetIndexOf(line, ",")              'save text until first comma     values[Array.GetItemCount(values)+1] = Text.GetSubText(line, 1, comma_position-1)           'remove saved text and comma from the line     line = Text.GetSubTextToEnd(line, comma_position+1)   EndWhile       'last value does not have a comma at the end   values[Array.GetItemCount(values)+1] = line       'append read values to csv   csv[Array.GetItemCount(csv)+1] = values         'reset variable   values = "" EndSub Finally, all values of the CSV file can be accessed by the CSV array. For instance, you can get the third value of the second line by: csv[2][3] Write CSV To save an array as a CSV file, it has to be processed row for row. All values of one row are combined to one text whereby commas are inserted between the values. This text is then written to the CSV file. csv = "1=1\=1\;2\=2\;3\=3\;;2=1\=4\;2\=5\;3\=6\;;3=1\=7\;2\=8\;3\=9\;;" csv_filename = "data.csv"   For line_number = 1 to Array.GetItemCount(csv)   line = csv[line_number][1]       For field_nr = 2 to Array.GetItemCount(csv[line_number])     line = line +  "," + csv[line_number][field_nr]   EndFor       ' write line to file   File.AppendContents(csv_filename, line) EndFor References Other Languages
__label__pos
0.966433
Open In App Grid System in UI Design [Beginner’s Guide] Last Updated : 27 Feb, 2024 Improve Improve Like Article Like Save Share Report Grid layouts are the building blocks of final amazing-looking designs. Creating symmetric and unified designs using grids is very important. Effective use of grids helps designers in creating designs that are simple for end users to scan and use. Grids ensure consistency across different platforms by adjusting to different screen sizes and orientations. In this article, we’ll dive deep into the concept of grids and understand how to use this tool to create consistent and unified designs. In this article, we’ll cover Grid System In UI Design Grid System In UI Design [Beginner's Guide] Grid System In UI Design [Beginner’s Guide] Depending upon screen sizes, we have some guidelines to use grids. These guidelines are globally accepted, and help in better articulation of designs concerning specific screen sizes. The topic we’ll be looking at to understand grids better: • What is a grid? • Basic terminologies used in grids • Types of grids with examples • How to use grids • Benefits of using grids What is a Grid? So to start with, let’s understand What is a Grid?  Grid is a collection of rows and columns along with margins, and spacing between each row and column. A designer can always tweak the number of rows or columns required and also the margin and spacing between each of these rows and columns.  The optimal use of grids serves as a guiding layout for overall designs. It helps to create consistent templates and standards. Using baselines (columns and rows) and padding( gutters and margins) it becomes easy to configure multiple screens at once. Grid ensures a balanced hierarchy of screen elements. Grid systems have different types of grids(which we’ll have a look at later in this article). Let’s look into some basic terminologies used in grid systems: • Columns: The vertical containers that hold some element like images, or text. Occupies most of the area in a grid. Depending on the screen the number of columns varies.  For example, Mobile screens have 4 columns in a grid and laptop/desktop screens have 12 columns in a grid. Below is an example of how columns look.  • Rows: The Horizontal containers that hold some element like images, or text. It is similar to a column but is not mostly used unless required.  • Gutters: The spaces between 2 columns or rows also called “Alleys”. Gutters have set values but depending upon the screen they also vary. Large screens such as laptops or desktops have wider gutters whereas small screens such as mobile phones have thinner gutters. A combination of columns and gutters takes up the horizontal width of the screen. • Margins: The spacing on the left and right of the screen where no screen element is present. Margins help in making the screen feel dense but not cluttered. The size of the margins may or may not be the same as the gutter, generally is it more than gutters. It marks the boundary for screen elements. Helps in fixing the layout. Terminologies used in Grid System Terminologies used in Grid System These were some basic terminologies a designer should know before using a grid system. One thing to keep in mind is that “Elements don’t need to end at the same marking but all elements should start at the same marking” when using a grid system. Types of Grid Systems As the grid is flexible to create a similar layout for different size devices. It is also available in different ways to use multiple kinds of Grid systems. Now let’s look at the different types of grids system in UI Design that we have and how to use them: 1. Symmetrical Grids: This types of grid system in UI Design are evenly spread, which means all the elements are aligned with the center. Such grid systems consist of an equal number of rows and columns. These look very pleasing to the eye and elements are evenly spread on the screen. Give a formal and clean look to the designs.   • Column Grid: The most commonly used grid type, where a frame is split into multiple columns separated with gutters. It is pleasing to the eye and creates a balanced visual hierarchy. Screen elements like text, images, and buttons are aligned using these columns. Column grids are mostly used in magazines and newspapers. Margins are fixed and independently sized.  Column Grid Column Grid • Pixel grid: It is a type of symmetrical grid, where the layout is fixed having columns and rows that make it appear like a grid cage. It feels like one has super zoomed into Figma or just like our math elementary notebook. This grid type helps in creating pixel-perfect designs, each box denotes a pixel itself. Designers occasionally close in to make pixel-by-pixel adjustments to the millions of pixels that make up digital screens.  Pixel Grid Pixel Grid • Baseline grid: It is also a type of symmetrical grid, where the layout is fixed having only horizontal lines as a guide, a ruled notebook is a straightforward illustration of a baseline grid. To make sure that the lines of text in each column align consistently across a spread, baseline grids are frequently used in conjunction with column grids. Baseline Grid • Modular grid: It is another type of symmetrical grid that appears to be very pleasing to the eye. Modular grid got the name because of the “Modules” which are created by the intersection of columns and rows that are very well organized and placed using gutters. Margins are consistent which gives a boxy look, this type is generally used by business magazines.  Modular Grid • Manuscript grid: It is a type of symmetrical grid, also known as a “Single columned” or “one-columned” grid as well. As the name suggests, the manuscript grid has only one column that guides the layout. This grid type is often used for books, especially folk or traditional books. Due to its appearance, it makes reading easier by using short lines with fixed margins from all sides.  Manuscript Grid 2. Asymmetrical Grid: This type of Grid System in UI Design are opposite to symmetric grids as there is no restriction of the grid being aligned to the center. It is also called the broken glass grid, as pieces are unevenly placed which gives a funky, modern, and abstract look to the designs. • Hierarchical grid: It is a type of asymmetrical grid, where the layout is not fixed. A hierarchical grid is entirely freeform, made up of two or more grids that are overlaid, or contain other grid components within themselves. To avoid having a boxy vibe, many contemporary website interfaces use hierarchical grid layouts. As it is asymmetrical it is more modern and abstract, setting a playful mood on websites when used.  Hierarchical grid These are some popular types of grids that are used based on different requirements and needs. It is good to know about each of these so that as a designer one can utilize and leverage the best guiding layout grid system. To know the optimal usage of a grid system, knowing how to use the grid is very important. As a beginner using a grid can be very overwhelming and confusing but not to worry, We’re here to rescue you. Now let’s look into how to use the grid to create an organized and consistent layout. How To Use Grids? 1. Know your requirements: This is generally seen when beginners directly jump onto using grids without knowing the requirement, the device, and which type of grid to use. Later in design and further development, this can be very confusing to resolve.  There are a few frameworks that are already defined to use grid systems optimally. Such as 12 cols, 6 cols, 4 cols, and 3 colors even 2 and 1 col grids, here 12,6,4,3,2, and 1 are the number of grids in which the screen is divided. But not only this can be used, but there might also be a screen that requires an 8-col grid and it is completely okay to use a customized grid that suits the need. Not only grids but keep in mind the screens as well, using the same grids for multiple devices might not be a good option.  2. Take care of text alignment: Make sure that all the text irrespective of the font weight and size, everything should be placed on a fixed baseline. This is a part of the layout as well and maintaining a grid layout along with text hierarchy would make the content and website/app appear more pleasing to the eyes. 3. Emphasize breaking the grid: It is a rule that says while using a grid it is mandatory to place all elements inside the grid boundary but nowadays it is more popular to break the grid to emphasize certain elements. When something is out of alignment or disturbing the layout it gets noticed easily, a similar trend is getting by breaking grid boundaries. 4. Maintain gutter balance: Equally, pay attention to both vertical and horizontal grid alignments. For example, if you are using a gutter of 20px between 2 vertical columns keep the same gutter between horizontal rows as well. Doing this would make the content on the screen very well organized even if not symmetric. 5. Use 8pt grid: Using baseline numbers and their multiple promotes layout to fit as per screen sizes. Such as the 8pt grid, which is the most popular. An 8pt grid is currently widely used because it allows designs to grow well on retina screens like iPhones and TV screens and works with both vector and pixel-based graphics. 6. Know where to start: Always remember to place the content on the screen using the grid as a guideline, start with a grid, not with the gutter. Keep content aligned with the start and end of a column’s edge, module, or pixel. This might not sound critical from a design perspective but it is very helpful to guide a developer on how and where to place media content like images and video. As developers also do follow the grid system which codes your designs. 7. Use grid generators: There are good options available online for generating a grid based on the screen sizes. This can be very useful for the times when you are just starting with visual designs. Using them speeds up the process and ensures the layout is optimized for use. Benefits of Using Grids Keeping these points in mind would help create a better and optimal grid system for your designs. There are a few which I mentioned but there are multiple do’s and don’t for using the grid system effectively. To conclude the articles, let’s look into what are some key benefits of using a grid system. • Ensures consistency: Grid layouts provide a defined structure to designs that optimize information accessibility and responsiveness, serving both an aesthetic and practical purpose for UX/UI designers. Fixed margins and layout speed up the element layering on the screen, and one doesn’t have to check alignments again and again.  • Better scanning: As the information on the screen is well organized, scanning becomes easier. A good layout guides eye movement that avoids the user’s struggle of getting lost in the content. • Responsiveness: Optimal use of a grid system for multiple devices, promotes responsiveness. For example to replicate a design from big to small or small to the big screen, one just has to adjust grids and place content accordingly. • Easy to modify and reuse: Good guiding layout ensures easy modification and reusability. Digital products keep on changing such as new updates or UI modifications keep on coming every 6 months. So for quick modification, the grid works as a skeleton. Conclusion Grid System in UI Design actually works as best friends to budding designers as they guide the design through a defined set of lines and boxes. It creates a framework for building a flexible layout for multiple devices. Once you get familiar with layouts you might not use them always but using a grid in the initial days sets a good foundation. So here it is for this article, We hope this would be useful for you to understand how important these grid systems are for a beginner.  Related Articles: FAQs on Grid Systems in UI Design What is Grid System in UI Design? Grid System is a series of horizontal and vertical line that creates a framework to build a User Interface for multiple devices. Designers create Grids for better layouts which allow users to use the interface in a better way by adapting to multiple screen sizes. What are the types of Grid in UI Design? There are many types of grids in the process of UI Design, some of them are 1. Symmetrical Grids 2. Asymmetrical Grid 3. Column Grid 4. Pixel Grid 5. Manuscript grid What is the use of a Grid System in UI Design A Grid System in UI is used to make a proper structure of the layout of the product and align the elements on a screen. Grid System creates a framework that is used to maintain efficiency, alignment and provide great user experience. Like Article Suggest improvement Share your thoughts in the comments Similar Reads
__label__pos
0.985904
Take the 2-minute tour × Mathematics Stack Exchange is a question and answer site for people studying math at any level and professionals in related fields. It's 100% free, no registration required. How do you find the minimal polynomial, $\mu_{M^{-1}}(x)$ given $\mu_{M}(x)$? My guess is since if $\lambda$ is an eigenvalue of $M$, then $1\over \lambda$ is an eigenvalue of $M^{-1}$, we might have something like $\mu_{M^{-1}}(x)=\mu_{M}({1\over x})$? But then I am not sure that that is the minimal polynomial... Thanks share|improve this question add comment 1 Answer up vote 12 down vote accepted Yours is not a bad guess, but, to begin with, there is a problem: $\mu_M\left( \frac{1}{x} \right)$ needs not to be a polynomial. Nevertheless, I think we can improve your idea: for any degree $n$ polynomial $p(x)$, define its conjugate (I'm not sure if this guy has already a name in the literature: please, correct me [EDIT. According to Georges, this is called the reciprocal polynomial ]) as $$ \overline{p}(x) = x^n p\left( \frac{1}{x} \right) \ . $$ Clearly, the conjugate of a polynomial is still a polynomial, and you can easily verify that: 1. $\overline{\overline{p}}(x) = p(x)$. [EDIT: if $p(x)$ has non-zero constant term. See Georges' comment.] 2. $\overline{p(x)}\overline{q(x)} = \overline{p}(x)\cdot\overline{q}(x)$ I claim that the result is the following: if $\mu_M (x)= a_0 + a_1 x + \dots + a_{n-1}x^{n-1} + x^n$, then $$ \mu_{M^{-1}} (x) = \frac{1}{a_0}\overline{\mu_M} (x) \ . $$ In order to prove this, we'll need the following lemma. Lemma. Let $M$ be an invertible matrix and $p(x)$ a polynomial such that $p(M) = 0$. Then $\overline{p}(M^{-1}) = 0$. Proof of the lemma. Indeed, $\overline{p}(M^{-1}) = (M^{-1})^n p(M) = 0$. Hence, since $\mu_M$ annihilates $M$, so does $\overline{\mu_M}$ with $M^{-1}$. We have to prove that $\frac{1}{a_0}\overline{\mu_M}$ has the characteristic property of the minimal polynomial of $M^{-1}$. Namely, that it has no proper divisor which also annihilates $M^{-1}$. So, assume there were two polynomials $p(x), q(x)$ such that $$ \frac{1}{a_0}\overline{\mu_M} (x) = p(x)q(x) $$ and moreover $p(M^{-1}) = 0$. Then, taking conjugates in this last equality, we would obtain $$ \frac{1}{a_0}\mu_M (x) = \overline{p}(x)\cdot\overline{q}(x) \ . $$ But, because of the lemma, $\overline{p}(M) = 0$. So, by definition of the minimal polynomial of $M$, $$ \mu_M (x) = \overline{p}(x) \qquad \text{(normalized)} \ . $$ Taking conjugates again, we would have that, up to a constant, $$ \overline{\mu_M} (x) = p(x) \ . $$ share|improve this answer      Thanks, Agusti! –  Eigen Nov 7 '11 at 18:52      My pleasure, Eigen. –  a.r. Nov 7 '11 at 19:06 2   Dear Agustí, the guy is called reciprocal polynomial by his buddies. And reciprocating twice will not get you back where you started if the original polynomial has no constant term. (Fortunately this doesn't happen in the context of the question). –  Georges Elencwajg Nov 7 '11 at 20:12      You're right. Thank you, Georges. –  a.r. Nov 7 '11 at 21:10 2   You're welcome, Agustí. I forgot to mention that this is a very nice answer! –  Georges Elencwajg Nov 7 '11 at 22:23 add comment Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.99734
Markdown 教程 文章发布时间: 最后更新时间: Markdown 是一种轻量级标记语言,它允许人们使用易读易写的纯文本格式编写文档。 Markdown 语言在 2004 由约翰·格鲁伯(英语:John Gruber)创建。 Markdown 编写的文档可以导出 HTML 、Word、图像、PDF、Epub 等多种格式的文档。 Markdown 编写的文档后缀为 .md, .markdown 标题 👆 看起来就像上面这个。Markdown 标题有两种格式。 使用 =- 标记一级和二级标题 =- 标记语法格式如下: 1 2 3 4 5 我展示的是一级标题 ================= 我展示的是二级标题 ----------------- 使用 # 号标记 使用 # 号可表示 1-6 级标题,一级标题对应一个 # 号,二级标题对应两个 # 号,以此类推。 1 2 3 4 5 6 # 一级标题 ## 二级标题 ### 三级标题 #### 四级标题 ##### 五级标题 ###### 六级标题 段落样式 Markdown 段落没有特殊的格式,直接编写文字就好,段落的换行是使用两个以上空格加上回车。 当然也可以在段落后面使用一个空行来表示重新开始一个段落。 字体 Markdown 可以使用以下几种字体: 斜体文本 粗体文本 粗斜体文本 1 2 3 4 5 6 7 8 *斜体文本* _斜体文本_ **粗体文本** __粗体文本__ ***粗斜体文本*** ___粗斜体文本___ 分隔线 可以在一行中用三个以上的星号、减号、底线来建立一个分隔线,行内不能有其他东西。 也可以在星号或是减号中间插入空格。下面每种写法都可以建立分隔线: 1 2 3 4 5 *** * * * ***** - - - ---------- 删除线 如果段落上的文字要添加删除线,只需要在文字的两端加上两个波浪线 ~~ 即可,实例如下: RUNOOB.COM GOOGLE.COM BAIDU.COM 1 2 3 RUNOOB.COM GOOGLE.COM ~~BAIDU.COM~~ 下划线 下划线可以通过 HTML 的 <u> 标签来实现: 带下划线文本 1 <u>带下划线文本</u> 脚注 脚注是对文本的补充说明。Markdown 脚注的格式如下: 1 [^要注明的文本] 以下实例演示了脚注的用法: 阿米娅^阿米娅 1 2 创建脚注格式类似这样 [^阿米娅] [^阿米娅]: 博士,您还有很多事情需要处理。现在还不能休息哦。 Markdown 列表 Markdown 支持有序列表和无序列表。 无序列表使用星号(*)、加号(+)或是减号(-)作为列表标记: • 第一项 • 第二项 • 第三项 1 2 3 4 5 6 7 8 9 10 11 * 第一项 * 第二项 * 第三项 + 第一项 + 第二项 + 第三项 - 第一项 - 第二项 - 第三项 有序列表使用数字并加上 . 号来表示,如: 1. 第一项 2. 第二项 3. 第三项 1 2 3 1. 第一项 2. 第二项 3. 第三项 列表嵌套 列表嵌套只需在子列表中的选项添加四个空格即可: 1. 第一项: • 第一项嵌套的第一个元素 • 第一项嵌套的第二个元素 2. 第二项: • 第二项嵌套的第一个元素 • 第二项嵌套的第二个元素 1 2 3 4 5 6 1. 第一项: - 第一项嵌套的第一个元素 - 第一项嵌套的第二个元素 2. 第二项: - 第二项嵌套的第一个元素 - 第二项嵌套的第二个元素 Markdown 区块 Markdown 区块引用是在段落开头使用 > 符号 ,然后后面紧跟一个空格符号: 这是一个区块 1 > 这是一个区块 另外区块是可以嵌套的,一个 > 符号是最外层,两个 > 符号是第一层嵌套,以此类推: 最外层 第一层嵌套 第二层嵌套 1 2 3 > 最外层 > > 第一层嵌套 > > > 第二层嵌套 区块中使用列表 区块中使用列表实例如下: 区块中使用列表 1. 第一项 2. 第二项 • 第一项 • 第二项 • 第三项 1 2 3 4 5 6 > 区块中使用列表 > 1. 第一项 > 2. 第二项 > + 第一项 > + 第二项 > + 第三项 列表中使用区块 如果要在列表项目内放进区块,那么就需要在 > 前添加四个空格的缩进。 区块中使用列表实例如下: • 第一项 菜鸟教程 学的不仅是技术更是梦想 • 第二项 1 2 3 4 * 第一项 > 菜鸟教程 > 学的不仅是技术更是梦想 * 第二项 Markdown 代码 如果是段落上的一个函数或片段的代码可以用反引号把它包起来(`),例如: printf() 函数 1 `printf()` 函数 代码区块 代码区块使用 4 个空格或者一个制表符(Tab 键)。 也可以用 ``` 包裹一段代码,并指定一种语言(也可以不指定): 实例如下: 1 2 3 $(document).ready(function () { alert('RUNOOB'); }); Markdown 链接 链接使用方法如下: 链接名称 或者直接使用链接地址 https://github.com/Yue-plus 1 2 [链接名称](链接地址) <https://github.com/Yue-plus> 高级链接 可以通过变量来设置一个链接,变量赋值在文档末尾进行: 这个链接用 1 作为网址变量 Google 这个链接用 mysite 作为网址变量 Yue_plus 1 2 3 4 5 6 这个链接用 1 作为网址变量 [Google][1] 这个链接用 mysite 作为网址变量 [Yue_plus][mysite] 然后在文档的结尾为变量赋值(网址) [1]: http://www.google.com/ [mysite]: https://github.com/Yue-plus Markdown 图片 Markdown 图片语法格式如下: 1 2 ![alt 属性文本](图片地址) ![alt 属性文本](图片地址 "可选标题") 开头一个感叹号 ! 接着一个方括号,里面放上图片的替代文字 接着一个英文括号,里面放上图片的网址,最后还可以用引号包住并加上选择性的 ‘title’ 属性的文字。 罗德岛集结 1 ![罗德岛集结](https://ak.hypergryph.com/upload/images/20190228/118078295785f64dac736c6ade50bb76.jpg "罗德岛集结") 当然,你也可以像网址那样对图片网址使用变量: 1 2 3 4 这个链接用 `2` 作为网址变量 [凯尔希][2]. 然后在文档的结尾为变量赋值(网址) [2]: https://ak.hypergryph.com/upload/images/20190228/143666074a406ecaa6cd4271dc7c5127.jpg Markdown 还没有办法指定图片的高度与宽度,如果需要的话,也可以使用普通的 <img> 标签。 1 <img src="https://ak.hypergryph.com/assets/index/images/ak/common/story/item_infected.png" width="200px"> Markdown 表格 Markdown 制作表格使用 | 来分隔不同的单元格,使用 - 来分隔表头和其他行。 语法格式如下: 表头 表头 单元格 单元格 单元格 单元格 1 2 3 4 | 表头 | 表头 | | ---- | ---- | | 单元格 | 单元格 | | 单元格 | 单元格 | 可以设置表格的对齐方式: -: 设置内容和标题栏居右对齐。 :- 设置内容和标题栏居左对齐。 :-: 设置内容和标题栏居中对齐。 左对齐 右对齐 居中对齐 单元格 单元格 单元格 单元格 单元格 单元格 Markdown 高级技巧 ##支持的 HTML 元素 不在 Markdown 涵盖范围之内的标签,都可以直接在文档里面用 HTML 撰写。 目前支持的 HTML 元素有:<kbd> <b> <i> <em> <sup> <sub> <br>等,如: 使用 Ctrl+Alt+Del 重启电脑 1 使用 <kbd>Ctrl</kbd>+<kbd>Alt</kbd>+<kbd>Del</kbd> 重启电脑 转义 Markdown 使用了很多特殊符号来表示特定的意义,如果需要显示特定的符号则需要使用转义字符,Markdown 使用反斜杠转义特殊字符: ** 正常显示星号 ** 1 2 **文本加粗** \*\* 正常显示星号 \*\* Markdown 支持以下这些符号前面加上反斜杠来帮助插入普通的符号: 1 2 3 4 5 6 7 8 9 10 11 12 \ 反斜线 ` 反引号 * 星号 _ 下划线 {} 花括号 [] 方括号 () 小括号 # 井字号 + 加号 - 减号 . 英文句点 ! 感叹号 数学公式 当需要在编辑器中插入数学公式时,可以使用两个美元符 $$ 包裹 TeX 或 LaTeX 格式的数学公式来实现。提交后,问答和文章页会根据需要加载 Mathjax 对数学公式进行渲染。如: 参考 使用 hexo-filter-mathjax 过滤器来显示数学公式 可以在行内包含数学公式: 注意单 $ 内部不能有空格! 1 2 3 4 5 6 7 8 9 可以在行内包含数学公式: $i\hbar\frac{\partial}{\partial t}\psi=-\frac{\hbar^2}{2m}\nabla^2\psi+V\psi$ 注意单 `$` 内部不能有空格! $$ \begin{eqnarray\*} \nabla\cdot\vec{E}&=&\frac{\rho}{\epsilon_0}\\\\ \nabla\cdot\vec{B}&=&0\\\\ \nabla\times\vec{E}&=&-\frac{\partial B}{\partial t}\\\\ \nabla\times\vec{B}&=&\mu_0\left(\vec{J}+\epsilon_0\frac{\partial E}{\partial t}\right)\\\\ \end{eqnarray\*} $$ 打赏 支付宝 | Alipay 微信 | Wechat
__label__pos
0.918529
Dismiss Notice Join Physics Forums Today! The friendliest, high quality science and math community on the planet! Everyone who loves science is here! Integral of (1/2)sin2x dx 1. Apr 28, 2015 #1 The result of the integral of (1/2)sin2x dx with: upper limit x = arccos((R-h)/R) lower limit x = 0 is (-h^2+2Rh)/(2R^2) I can not seem to get this exact answer my workings yield: let u = 2x, du/dx = 2 therefore dx = du/2 Integral becomes (1/4) ∫ sinu du (with the same upper and lower limit) = -1/4 (cos(2x)) (evaluated at the limits) using cos2x = 2cos^2x-1 it becomes: = -1/4 (2cos^2x-1) (evaluated at the limits) = -1/4 ((2(cos*arccos ((R-h)/R))^2 - 1) - (2(cos*arccos(0))^2 -1) = -1/4 (2((R-h)/R)^2 - 1 + 1) = -1/2 ((R-h)/R)^2 = -1/2 - (h^2)/(2R^2) + h/R = (-h^2 + 2Rh - R^2)/(2R^2) As you can see, my answer has an extra R^2 term in the answer and I cannot see where I have gone wrong...   2. jcsd 3. Apr 28, 2015 #2 SteamKing User Avatar Staff Emeritus Science Advisor Homework Helper Since your indefinite integral evaluates to -(1/4) * {2 cos2(x) - 1}, applying the limits of integration should be written as follows: I = -(1/4) * {2 cos2 (arccos ((R-h)/R)) - 1 - [2 cos2 (arccos (0)) - 1]} {Note: there is no '*' between the cos and the arccos. These are both functions which take arguments.}   4. Apr 28, 2015 #3 mfb User Avatar 2017 Award Staff: Mentor The last +1 in the second line should be -1. There is some bracket error in the first line I think.   Share this great discussion with others via Reddit, Google+, Twitter, or Facebook Loading...
__label__pos
0.901363
HomeCategory: QuestionsWhy my laptop turns off suddenly? 1 Answers steven salvatoresteven salvatore answered 6 months ago 1. This could be happening due to a prolong use without switching off. Sometimes when we use the laptop for a long period of time without the rest period, the problem of overheating might happen. This is known to sometimes causing the laptop to switch off suddenly. 2. Remedy this by turning the laptop off and disconnecting it. Use it after a period of time. 3. Do not put it into “sleep mode” during this time, and ensure you switch if off properly before this. 4. A fundamental issue leading to these problems like the laptop turning off suddenly is not keeping your laptop device clean as possible. Dust and particles which gather may cause harm to your laptop too. 5. You might have to go to the control panel and change the power option setting on your laptop for this problem to not happen. 6. A larger hardware issue could also be the reason due to which your laptop is switching off suddenly. To combat this problem, make sure you get your laptop checked at a service center. Ensure that they scrutinize the parts of the laptop to see if any part of the hardware is not working or needs to be replaced. 7. Sometimes due to computer viruses also the laptops switch off suddenly, make sure that you install an antivirus software which gives you updates and notifications of malware on your device. 8. You must also ensure you get for software updates on laptops regularly as well. This prevents outdated software from being used in any case. 9. If most of these issues do not solve your problem, there is always the option of going to the service center of your laptop and then trying to get it fixed there. Most brands of laptops have service centers in every town and these help with any error. 10. If this is not accessible for you, then there are a number of helpline numbers of brands given for you to try remedy your concerns on the phone and have them addressed. Close Menu
__label__pos
0.641343
[File Plugin] Cannot read property 'dataDirectory' of undefined Question Forge component by OutSystems R&D Hi hi, I am trying to read a directory on my mobile device, but I am not getting there. Whenever i call CreateDirectory or ListDirectory this error appears: Cannot read property 'dataDirectory' of undefined When debugging i get to here: The problem is generated by the JS (I am using InternalStorage):  var fileSystemType = $parameters.IsInternal ? cordova.file.dataDirectory : cordova.file.externalDataDirectory; window.resolveLocalFileSystemURL(fileSystemType,     function(fs) {         $parameters.ErrorCode = "0";         $parameters.ErrorMsg = "";         $parameters.FileSystem = fs;         $resolve();     },     function(error){         $parameters.ErrorCode = error.code;         $parameters.ErrorMsg = $actions.GetErrorMessage(error.code).Message;                 $resolve();     } ); The problem is in the "cordova.file.dataDirectory" part of the JS, how can I resolve this? Help is appriciated, Wilko mvp_badge MVP Solution Hello Wilko, It's okay, I'm just trying to help you :) I am using this plugin to save and read files in my mobile app and it is working exactly as expected. Again, you don't need to change anything inside the module FilePlugin. Can you show me your code so I can check it? The flow when you're calling the FilePlugin ListDirectory action. Thank you, Rui Barradas mvp_badge MVP Hello Wilko, That code won't work. The problem is both in the cordova.file.dataDirectory and in the cordova.file.externalDataDirectory parts of the JavaScript. Somehow, the JavaScript can't recognize the cordova object. In order to resolve this, you just need to define the physical path of the directory that you want to read, for both Android and iOS. I will leave this link for you. It is very useful to understand the File System Layouts and the physical paths on a real device (for both iOS and Android): https://cordova.apache.org/docs/en/latest/reference/cordova-plugin-file/ As an example, let's say you want to list the content into the Documents folder in an Android device. You just need to define the path as: $parameters.DirectoryPath = "Documents"; Let me know if you managed to do it. Hope it helps, Rui Barradas Rui hi, You mention a parameterDirectoryPath. Let's say I want to list the files in a location "Documents". The FilePlugin has an action ListDirectory which looks like: The problem now is in the RequestDirectoryFromFile system action. The ListDirectory JS takes an Directory Object from GetDirectory, which takes a RootDirectory Object from RequestDirectoryFromFile, which gives the problem now. So I should replace RequestDirectoryFromFile and GetDirectory (or not use it) and change the JavaScript inside ListDirectory so the current Directory parameter there gets replaced by  $parameters.DirectoryPath = "Documents"; I am lost here. Hope you can give more more hints, Wilko mvp_badge MVP Hello there Wilko, Sorry if I wasn't clear enough. You don't need to change the logic already implemented inside the actions provided by the Plugin! You just need to use these actions in your flows. Let's say you want to list the content into the Documents folder in an Android device. You just need to call the ListDirectory action in your flow and define the physical path as an input parameter. This parameter DirectoryPath that I mentioned is one JavaScript output parameter that you should use as the Path input parameter in the ListDirectory action (or you can define it directly when you call the ListDirectory action without defining it via JavaScript). However, as you can see in the link above, iOS devices and Android devices have different File System Layouts. So, you should take that into consideration and define the physical path according to the respective platform: Hope it helps. Let me know if you managed to do it. Rui Barradas Rui hi, I am really sorry to bother you with this. I do as you told me, but ListDirectory calls RequestDirectoryFromFileSystem, which has the GetFileSystem javascript in it, which causes the error of cannot read property 'dataDirectory' of undefined, which cannot be solved you said. So somewhere I need to get rid of RequestDirectoryFromFileSystem, which is inside FilePlugin, and still manage to keep it work somehow. Or should I copy the ListDirectory javascript and pass the directory as an object, then could you help me how to transform the text to the correct object? Saludos, Wilko mvp_badge MVP Solution Hello Wilko, It's okay, I'm just trying to help you :) I am using this plugin to save and read files in my mobile app and it is working exactly as expected. Again, you don't need to change anything inside the module FilePlugin. Can you show me your code so I can check it? The flow when you're calling the FilePlugin ListDirectory action. Thank you, Rui Barradas Rui hi, The flow inside the GetDirectory is: Form this: The java script: The ListDirectory call: Debugging this with my phone connected shows the error inside GetFileSystem javascript with the cordova.file.dataDirectory being undefined, which is called from ListDirectory. Saludos, Wilko wilko odijk wrote: Rui hi, The flow inside the GetDirectory is: Form this: The java script: The ListDirectory call: Debugging this with my phone connected shows the error inside GetFileSystem javascript with the cordova.file.dataDirectory being undefined, which is called from ListDirectory. Saludos, Wilko hi... i want to ask something. how do you get path : JavaScrip1.DirectoryPath ? because i dont find javascript in mypath column ? Rui hi, I did something else, I installed FileSample from the Forge. It uses the same FilePlugin module. From my demo app the GetFileSystem javascript in RequestDirectoryFromFileSystem called from ListDirectory gives an error (one step further down the road in the javascript): The same the GetFileSystem javascript in RequestDirectoryFromFileSystem called from ListDirectory called from the FileSample doesn't give an error: Would it be something in the build of the app of DemoApp versus the build of the app of FileSample that triggers the error in the GetFileSystem javascript? Saludos, Wilko mvp_badge MVP Hi again Wilko, I was off during some time, but I'm back now. Did you manage to solve the problem? If you want, I can check it and run some tests in my side. Rui hi, After some more digging regarding the error "Cannot read property 'dataDirectory' of undefined", i assume the underlying problem has to do with the assembly of the build of the app. probably there is logic that determines what libraries to include when doing the build, and if in the dependency tree for the build there is a reference to a module that somewhere is missing or has broken links, it doesn't get included in the build, but runtime there is a link missing or broken link (in this case 'dataDirectory') it tries to find it but it is not there. So i checked all the dependencies and fixed what was broken or missing and then the error doesn't happen. My demo app had some broken or missing dependencies, and the FileSample didn't have them. It took some time figure this out, but yes, it is working. I saw some other posts with variations on not being able to read properties, i assume they have the same cause. Thanks for thinking along. Wilko mvp_badge MVP Hi Wilko, I'm glad you managed to figure it out. Cheers, Rui Barradas Hi, I created a small app to record a csv on my phone and I'm getting the same output message. The error is happening in RequestDirectoryFromFileSystem. Help is appreciated, José  Hi!  I am getting the same error while using SaveFile Action from the Plugin and I checked all the dependecies and eveything looks to be fine: Error: Cannot read property 'dataDirectory' of undefined I tried to pass the Directory Path Manually to ListDirectory but I still get the same error. Any help on that? Thanks! It seems like there's something wrong with the installation of the plugin itself as cordova.file object is not being defined at all! Community GuidelinesBe kind and respectful, give credit to the original source of content, and search for duplicates before posting.
__label__pos
0.920458
Bug Summary File:llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp Warning:line 6251, column 30 Called C++ object pointer is null Annotated Source Code Press '?' to see keyboard shortcuts clang -cc1 -cc1 -triple x86_64-pc-linux-gnu -analyze -disable-free -disable-llvm-verifier -discard-value-names -main-file-name SLPVectorizer.cpp -analyzer-store=region -analyzer-opt-analyze-nested-blocks -analyzer-checker=core -analyzer-checker=apiModeling -analyzer-checker=unix -analyzer-checker=deadcode -analyzer-checker=cplusplus -analyzer-checker=security.insecureAPI.UncheckedReturn -analyzer-checker=security.insecureAPI.getpw -analyzer-checker=security.insecureAPI.gets -analyzer-checker=security.insecureAPI.mktemp -analyzer-checker=security.insecureAPI.mkstemp -analyzer-checker=security.insecureAPI.vfork -analyzer-checker=nullability.NullPassedToNonnull -analyzer-checker=nullability.NullReturnedFromNonnull -analyzer-output plist -w -setup-static-analyzer -analyzer-config-compatibility-mode=true -mrelocation-model pic -pic-level 2 -mframe-pointer=none -fmath-errno -fno-rounding-math -mconstructor-aliases -munwind-tables -target-cpu x86-64 -tune-cpu generic -debugger-tuning=gdb -ffunction-sections -fdata-sections -fcoverage-compilation-dir=/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/build-llvm/lib/Transforms/Vectorize -resource-dir /usr/lib/llvm-13/lib/clang/13.0.0 -D _DEBUG -D _GNU_SOURCE -D __STDC_CONSTANT_MACROS -D __STDC_FORMAT_MACROS -D __STDC_LIMIT_MACROS -I /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/build-llvm/lib/Transforms/Vectorize -I /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize -I /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/build-llvm/include -I /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/include -D NDEBUG -U NDEBUG -internal-isystem /usr/lib/gcc/x86_64-linux-gnu/10/../../../../include/c++/10 -internal-isystem /usr/lib/gcc/x86_64-linux-gnu/10/../../../../include/x86_64-linux-gnu/c++/10 -internal-isystem /usr/lib/gcc/x86_64-linux-gnu/10/../../../../include/c++/10/backward -internal-isystem /usr/lib/llvm-13/lib/clang/13.0.0/include -internal-isystem /usr/local/include -internal-isystem /usr/lib/gcc/x86_64-linux-gnu/10/../../../../x86_64-linux-gnu/include -internal-externc-isystem /usr/include/x86_64-linux-gnu -internal-externc-isystem /include -internal-externc-isystem /usr/include -O2 -Wno-unused-parameter -Wwrite-strings -Wno-missing-field-initializers -Wno-long-long -Wno-maybe-uninitialized -Wno-class-memaccess -Wno-redundant-move -Wno-pessimizing-move -Wno-noexcept-type -Wno-comment -std=c++14 -fdeprecated-macro -fdebug-compilation-dir=/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/build-llvm/lib/Transforms/Vectorize -fdebug-prefix-map=/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a=. -ferror-limit 19 -fvisibility-inlines-hidden -stack-protector 2 -fgnuc-version=4.2.1 -vectorize-loops -vectorize-slp -analyzer-output=html -analyzer-config stable-report-filename=true -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /tmp/scan-build-2021-06-10-162512-48765-1 -x c++ /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp /build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp 1//===- SLPVectorizer.cpp - A bottom up SLP Vectorizer ---------------------===// 2// 3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 4// See https://llvm.org/LICENSE.txt for license information. 5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 6// 7//===----------------------------------------------------------------------===// 8// 9// This pass implements the Bottom Up SLP vectorizer. It detects consecutive 10// stores that can be put together into vector-stores. Next, it attempts to 11// construct vectorizable tree using the use-def chains. If a profitable tree 12// was found, the SLP vectorizer performs vectorization on the tree. 13// 14// The pass is inspired by the work described in the paper: 15// "Loop-Aware SLP in GCC" by Ira Rosen, Dorit Nuzman, Ayal Zaks. 16// 17//===----------------------------------------------------------------------===// 18 19#include "llvm/Transforms/Vectorize/SLPVectorizer.h" 20#include "llvm/ADT/DenseMap.h" 21#include "llvm/ADT/DenseSet.h" 22#include "llvm/ADT/Optional.h" 23#include "llvm/ADT/PostOrderIterator.h" 24#include "llvm/ADT/STLExtras.h" 25#include "llvm/ADT/SetOperations.h" 26#include "llvm/ADT/SetVector.h" 27#include "llvm/ADT/SmallBitVector.h" 28#include "llvm/ADT/SmallPtrSet.h" 29#include "llvm/ADT/SmallSet.h" 30#include "llvm/ADT/SmallString.h" 31#include "llvm/ADT/Statistic.h" 32#include "llvm/ADT/iterator.h" 33#include "llvm/ADT/iterator_range.h" 34#include "llvm/Analysis/AliasAnalysis.h" 35#include "llvm/Analysis/AssumptionCache.h" 36#include "llvm/Analysis/CodeMetrics.h" 37#include "llvm/Analysis/DemandedBits.h" 38#include "llvm/Analysis/GlobalsModRef.h" 39#include "llvm/Analysis/IVDescriptors.h" 40#include "llvm/Analysis/LoopAccessAnalysis.h" 41#include "llvm/Analysis/LoopInfo.h" 42#include "llvm/Analysis/MemoryLocation.h" 43#include "llvm/Analysis/OptimizationRemarkEmitter.h" 44#include "llvm/Analysis/ScalarEvolution.h" 45#include "llvm/Analysis/ScalarEvolutionExpressions.h" 46#include "llvm/Analysis/TargetLibraryInfo.h" 47#include "llvm/Analysis/TargetTransformInfo.h" 48#include "llvm/Analysis/ValueTracking.h" 49#include "llvm/Analysis/VectorUtils.h" 50#include "llvm/IR/Attributes.h" 51#include "llvm/IR/BasicBlock.h" 52#include "llvm/IR/Constant.h" 53#include "llvm/IR/Constants.h" 54#include "llvm/IR/DataLayout.h" 55#include "llvm/IR/DebugLoc.h" 56#include "llvm/IR/DerivedTypes.h" 57#include "llvm/IR/Dominators.h" 58#include "llvm/IR/Function.h" 59#include "llvm/IR/IRBuilder.h" 60#include "llvm/IR/InstrTypes.h" 61#include "llvm/IR/Instruction.h" 62#include "llvm/IR/Instructions.h" 63#include "llvm/IR/IntrinsicInst.h" 64#include "llvm/IR/Intrinsics.h" 65#include "llvm/IR/Module.h" 66#include "llvm/IR/NoFolder.h" 67#include "llvm/IR/Operator.h" 68#include "llvm/IR/PatternMatch.h" 69#include "llvm/IR/Type.h" 70#include "llvm/IR/Use.h" 71#include "llvm/IR/User.h" 72#include "llvm/IR/Value.h" 73#include "llvm/IR/ValueHandle.h" 74#include "llvm/IR/Verifier.h" 75#include "llvm/InitializePasses.h" 76#include "llvm/Pass.h" 77#include "llvm/Support/Casting.h" 78#include "llvm/Support/CommandLine.h" 79#include "llvm/Support/Compiler.h" 80#include "llvm/Support/DOTGraphTraits.h" 81#include "llvm/Support/Debug.h" 82#include "llvm/Support/ErrorHandling.h" 83#include "llvm/Support/GraphWriter.h" 84#include "llvm/Support/InstructionCost.h" 85#include "llvm/Support/KnownBits.h" 86#include "llvm/Support/MathExtras.h" 87#include "llvm/Support/raw_ostream.h" 88#include "llvm/Transforms/Utils/InjectTLIMappings.h" 89#include "llvm/Transforms/Utils/LoopUtils.h" 90#include "llvm/Transforms/Vectorize.h" 91#include <algorithm> 92#include <cassert> 93#include <cstdint> 94#include <iterator> 95#include <memory> 96#include <set> 97#include <string> 98#include <tuple> 99#include <utility> 100#include <vector> 101 102using namespace llvm; 103using namespace llvm::PatternMatch; 104using namespace slpvectorizer; 105 106#define SV_NAME"slp-vectorizer" "slp-vectorizer" 107#define DEBUG_TYPE"SLP" "SLP" 108 109STATISTIC(NumVectorInstructions, "Number of vector instructions generated")static llvm::Statistic NumVectorInstructions = {"SLP", "NumVectorInstructions" , "Number of vector instructions generated"} ; 110 111cl::opt<bool> RunSLPVectorization("vectorize-slp", cl::init(true), cl::Hidden, 112 cl::desc("Run the SLP vectorization passes")); 113 114static cl::opt<int> 115 SLPCostThreshold("slp-threshold", cl::init(0), cl::Hidden, 116 cl::desc("Only vectorize if you gain more than this " 117 "number ")); 118 119static cl::opt<bool> 120ShouldVectorizeHor("slp-vectorize-hor", cl::init(true), cl::Hidden, 121 cl::desc("Attempt to vectorize horizontal reductions")); 122 123static cl::opt<bool> ShouldStartVectorizeHorAtStore( 124 "slp-vectorize-hor-store", cl::init(false), cl::Hidden, 125 cl::desc( 126 "Attempt to vectorize horizontal reductions feeding into a store")); 127 128static cl::opt<int> 129MaxVectorRegSizeOption("slp-max-reg-size", cl::init(128), cl::Hidden, 130 cl::desc("Attempt to vectorize for this register size in bits")); 131 132static cl::opt<unsigned> 133MaxVFOption("slp-max-vf", cl::init(0), cl::Hidden, 134 cl::desc("Maximum SLP vectorization factor (0=unlimited)")); 135 136static cl::opt<int> 137MaxStoreLookup("slp-max-store-lookup", cl::init(32), cl::Hidden, 138 cl::desc("Maximum depth of the lookup for consecutive stores.")); 139 140/// Limits the size of scheduling regions in a block. 141/// It avoid long compile times for _very_ large blocks where vector 142/// instructions are spread over a wide range. 143/// This limit is way higher than needed by real-world functions. 144static cl::opt<int> 145ScheduleRegionSizeBudget("slp-schedule-budget", cl::init(100000), cl::Hidden, 146 cl::desc("Limit the size of the SLP scheduling region per block")); 147 148static cl::opt<int> MinVectorRegSizeOption( 149 "slp-min-reg-size", cl::init(128), cl::Hidden, 150 cl::desc("Attempt to vectorize for this register size in bits")); 151 152static cl::opt<unsigned> RecursionMaxDepth( 153 "slp-recursion-max-depth", cl::init(12), cl::Hidden, 154 cl::desc("Limit the recursion depth when building a vectorizable tree")); 155 156static cl::opt<unsigned> MinTreeSize( 157 "slp-min-tree-size", cl::init(3), cl::Hidden, 158 cl::desc("Only vectorize small trees if they are fully vectorizable")); 159 160// The maximum depth that the look-ahead score heuristic will explore. 161// The higher this value, the higher the compilation time overhead. 162static cl::opt<int> LookAheadMaxDepth( 163 "slp-max-look-ahead-depth", cl::init(2), cl::Hidden, 164 cl::desc("The maximum look-ahead depth for operand reordering scores")); 165 166// The Look-ahead heuristic goes through the users of the bundle to calculate 167// the users cost in getExternalUsesCost(). To avoid compilation time increase 168// we limit the number of users visited to this value. 169static cl::opt<unsigned> LookAheadUsersBudget( 170 "slp-look-ahead-users-budget", cl::init(2), cl::Hidden, 171 cl::desc("The maximum number of users to visit while visiting the " 172 "predecessors. This prevents compilation time increase.")); 173 174static cl::opt<bool> 175 ViewSLPTree("view-slp-tree", cl::Hidden, 176 cl::desc("Display the SLP trees with Graphviz")); 177 178// Limit the number of alias checks. The limit is chosen so that 179// it has no negative effect on the llvm benchmarks. 180static const unsigned AliasedCheckLimit = 10; 181 182// Another limit for the alias checks: The maximum distance between load/store 183// instructions where alias checks are done. 184// This limit is useful for very large basic blocks. 185static const unsigned MaxMemDepDistance = 160; 186 187/// If the ScheduleRegionSizeBudget is exhausted, we allow small scheduling 188/// regions to be handled. 189static const int MinScheduleRegionSize = 16; 190 191/// Predicate for the element types that the SLP vectorizer supports. 192/// 193/// The most important thing to filter here are types which are invalid in LLVM 194/// vectors. We also filter target specific types which have absolutely no 195/// meaningful vectorization path such as x86_fp80 and ppc_f128. This just 196/// avoids spending time checking the cost model and realizing that they will 197/// be inevitably scalarized. 198static bool isValidElementType(Type *Ty) { 199 return VectorType::isValidElementType(Ty) && !Ty->isX86_FP80Ty() && 200 !Ty->isPPC_FP128Ty(); 201} 202 203/// \returns true if all of the instructions in \p VL are in the same block or 204/// false otherwise. 205static bool allSameBlock(ArrayRef<Value *> VL) { 206 Instruction *I0 = dyn_cast<Instruction>(VL[0]); 207 if (!I0) 208 return false; 209 BasicBlock *BB = I0->getParent(); 210 for (int I = 1, E = VL.size(); I < E; I++) { 211 auto *II = dyn_cast<Instruction>(VL[I]); 212 if (!II) 213 return false; 214 215 if (BB != II->getParent()) 216 return false; 217 } 218 return true; 219} 220 221/// \returns True if the value is a constant (but not globals/constant 222/// expressions). 223static bool isConstant(Value *V) { 224 return isa<Constant>(V) && !isa<ConstantExpr>(V) && !isa<GlobalValue>(V); 225} 226 227/// \returns True if all of the values in \p VL are constants (but not 228/// globals/constant expressions). 229static bool allConstant(ArrayRef<Value *> VL) { 230 // Constant expressions and globals can't be vectorized like normal integer/FP 231 // constants. 232 return all_of(VL, isConstant); 233} 234 235/// \returns True if all of the values in \p VL are identical. 236static bool isSplat(ArrayRef<Value *> VL) { 237 for (unsigned i = 1, e = VL.size(); i < e; ++i) 238 if (VL[i] != VL[0]) 239 return false; 240 return true; 241} 242 243/// \returns True if \p I is commutative, handles CmpInst and BinaryOperator. 244static bool isCommutative(Instruction *I) { 245 if (auto *Cmp = dyn_cast<CmpInst>(I)) 246 return Cmp->isCommutative(); 247 if (auto *BO = dyn_cast<BinaryOperator>(I)) 248 return BO->isCommutative(); 249 // TODO: This should check for generic Instruction::isCommutative(), but 250 // we need to confirm that the caller code correctly handles Intrinsics 251 // for example (does not have 2 operands). 252 return false; 253} 254 255/// Checks if the vector of instructions can be represented as a shuffle, like: 256/// %x0 = extractelement <4 x i8> %x, i32 0 257/// %x3 = extractelement <4 x i8> %x, i32 3 258/// %y1 = extractelement <4 x i8> %y, i32 1 259/// %y2 = extractelement <4 x i8> %y, i32 2 260/// %x0x0 = mul i8 %x0, %x0 261/// %x3x3 = mul i8 %x3, %x3 262/// %y1y1 = mul i8 %y1, %y1 263/// %y2y2 = mul i8 %y2, %y2 264/// %ins1 = insertelement <4 x i8> poison, i8 %x0x0, i32 0 265/// %ins2 = insertelement <4 x i8> %ins1, i8 %x3x3, i32 1 266/// %ins3 = insertelement <4 x i8> %ins2, i8 %y1y1, i32 2 267/// %ins4 = insertelement <4 x i8> %ins3, i8 %y2y2, i32 3 268/// ret <4 x i8> %ins4 269/// can be transformed into: 270/// %1 = shufflevector <4 x i8> %x, <4 x i8> %y, <4 x i32> <i32 0, i32 3, i32 5, 271/// i32 6> 272/// %2 = mul <4 x i8> %1, %1 273/// ret <4 x i8> %2 274/// We convert this initially to something like: 275/// %x0 = extractelement <4 x i8> %x, i32 0 276/// %x3 = extractelement <4 x i8> %x, i32 3 277/// %y1 = extractelement <4 x i8> %y, i32 1 278/// %y2 = extractelement <4 x i8> %y, i32 2 279/// %1 = insertelement <4 x i8> poison, i8 %x0, i32 0 280/// %2 = insertelement <4 x i8> %1, i8 %x3, i32 1 281/// %3 = insertelement <4 x i8> %2, i8 %y1, i32 2 282/// %4 = insertelement <4 x i8> %3, i8 %y2, i32 3 283/// %5 = mul <4 x i8> %4, %4 284/// %6 = extractelement <4 x i8> %5, i32 0 285/// %ins1 = insertelement <4 x i8> poison, i8 %6, i32 0 286/// %7 = extractelement <4 x i8> %5, i32 1 287/// %ins2 = insertelement <4 x i8> %ins1, i8 %7, i32 1 288/// %8 = extractelement <4 x i8> %5, i32 2 289/// %ins3 = insertelement <4 x i8> %ins2, i8 %8, i32 2 290/// %9 = extractelement <4 x i8> %5, i32 3 291/// %ins4 = insertelement <4 x i8> %ins3, i8 %9, i32 3 292/// ret <4 x i8> %ins4 293/// InstCombiner transforms this into a shuffle and vector mul 294/// Mask will return the Shuffle Mask equivalent to the extracted elements. 295/// TODO: Can we split off and reuse the shuffle mask detection from 296/// TargetTransformInfo::getInstructionThroughput? 297static Optional<TargetTransformInfo::ShuffleKind> 298isShuffle(ArrayRef<Value *> VL, SmallVectorImpl<int> &Mask) { 299 auto *EI0 = cast<ExtractElementInst>(VL[0]); 300 unsigned Size = 301 cast<FixedVectorType>(EI0->getVectorOperandType())->getNumElements(); 302 Value *Vec1 = nullptr; 303 Value *Vec2 = nullptr; 304 enum ShuffleMode { Unknown, Select, Permute }; 305 ShuffleMode CommonShuffleMode = Unknown; 306 for (unsigned I = 0, E = VL.size(); I < E; ++I) { 307 auto *EI = cast<ExtractElementInst>(VL[I]); 308 auto *Vec = EI->getVectorOperand(); 309 // All vector operands must have the same number of vector elements. 310 if (cast<FixedVectorType>(Vec->getType())->getNumElements() != Size) 311 return None; 312 auto *Idx = dyn_cast<ConstantInt>(EI->getIndexOperand()); 313 if (!Idx) 314 return None; 315 // Undefined behavior if Idx is negative or >= Size. 316 if (Idx->getValue().uge(Size)) { 317 Mask.push_back(UndefMaskElem); 318 continue; 319 } 320 unsigned IntIdx = Idx->getValue().getZExtValue(); 321 Mask.push_back(IntIdx); 322 // We can extractelement from undef or poison vector. 323 if (isa<UndefValue>(Vec)) 324 continue; 325 // For correct shuffling we have to have at most 2 different vector operands 326 // in all extractelement instructions. 327 if (!Vec1 || Vec1 == Vec) 328 Vec1 = Vec; 329 else if (!Vec2 || Vec2 == Vec) 330 Vec2 = Vec; 331 else 332 return None; 333 if (CommonShuffleMode == Permute) 334 continue; 335 // If the extract index is not the same as the operation number, it is a 336 // permutation. 337 if (IntIdx != I) { 338 CommonShuffleMode = Permute; 339 continue; 340 } 341 CommonShuffleMode = Select; 342 } 343 // If we're not crossing lanes in different vectors, consider it as blending. 344 if (CommonShuffleMode == Select && Vec2) 345 return TargetTransformInfo::SK_Select; 346 // If Vec2 was never used, we have a permutation of a single vector, otherwise 347 // we have permutation of 2 vectors. 348 return Vec2 ? TargetTransformInfo::SK_PermuteTwoSrc 349 : TargetTransformInfo::SK_PermuteSingleSrc; 350} 351 352namespace { 353 354/// Main data required for vectorization of instructions. 355struct InstructionsState { 356 /// The very first instruction in the list with the main opcode. 357 Value *OpValue = nullptr; 358 359 /// The main/alternate instruction. 360 Instruction *MainOp = nullptr; 361 Instruction *AltOp = nullptr; 362 363 /// The main/alternate opcodes for the list of instructions. 364 unsigned getOpcode() const { 365 return MainOp ? MainOp->getOpcode() : 0; 366 } 367 368 unsigned getAltOpcode() const { 369 return AltOp ? AltOp->getOpcode() : 0; 370 } 371 372 /// Some of the instructions in the list have alternate opcodes. 373 bool isAltShuffle() const { return getOpcode() != getAltOpcode(); } 374 375 bool isOpcodeOrAlt(Instruction *I) const { 376 unsigned CheckedOpcode = I->getOpcode(); 377 return getOpcode() == CheckedOpcode || getAltOpcode() == CheckedOpcode; 378 } 379 380 InstructionsState() = delete; 381 InstructionsState(Value *OpValue, Instruction *MainOp, Instruction *AltOp) 382 : OpValue(OpValue), MainOp(MainOp), AltOp(AltOp) {} 383}; 384 385} // end anonymous namespace 386 387/// Chooses the correct key for scheduling data. If \p Op has the same (or 388/// alternate) opcode as \p OpValue, the key is \p Op. Otherwise the key is \p 389/// OpValue. 390static Value *isOneOf(const InstructionsState &S, Value *Op) { 391 auto *I = dyn_cast<Instruction>(Op); 392 if (I && S.isOpcodeOrAlt(I)) 393 return Op; 394 return S.OpValue; 395} 396 397/// \returns true if \p Opcode is allowed as part of of the main/alternate 398/// instruction for SLP vectorization. 399/// 400/// Example of unsupported opcode is SDIV that can potentially cause UB if the 401/// "shuffled out" lane would result in division by zero. 402static bool isValidForAlternation(unsigned Opcode) { 403 if (Instruction::isIntDivRem(Opcode)) 404 return false; 405 406 return true; 407} 408 409/// \returns analysis of the Instructions in \p VL described in 410/// InstructionsState, the Opcode that we suppose the whole list 411/// could be vectorized even if its structure is diverse. 412static InstructionsState getSameOpcode(ArrayRef<Value *> VL, 413 unsigned BaseIndex = 0) { 414 // Make sure these are all Instructions. 415 if (llvm::any_of(VL, [](Value *V) { return !isa<Instruction>(V); })) 416 return InstructionsState(VL[BaseIndex], nullptr, nullptr); 417 418 bool IsCastOp = isa<CastInst>(VL[BaseIndex]); 419 bool IsBinOp = isa<BinaryOperator>(VL[BaseIndex]); 420 unsigned Opcode = cast<Instruction>(VL[BaseIndex])->getOpcode(); 421 unsigned AltOpcode = Opcode; 422 unsigned AltIndex = BaseIndex; 423 424 // Check for one alternate opcode from another BinaryOperator. 425 // TODO - generalize to support all operators (types, calls etc.). 426 for (int Cnt = 0, E = VL.size(); Cnt < E; Cnt++) { 427 unsigned InstOpcode = cast<Instruction>(VL[Cnt])->getOpcode(); 428 if (IsBinOp && isa<BinaryOperator>(VL[Cnt])) { 429 if (InstOpcode == Opcode || InstOpcode == AltOpcode) 430 continue; 431 if (Opcode == AltOpcode && isValidForAlternation(InstOpcode) && 432 isValidForAlternation(Opcode)) { 433 AltOpcode = InstOpcode; 434 AltIndex = Cnt; 435 continue; 436 } 437 } else if (IsCastOp && isa<CastInst>(VL[Cnt])) { 438 Type *Ty0 = cast<Instruction>(VL[BaseIndex])->getOperand(0)->getType(); 439 Type *Ty1 = cast<Instruction>(VL[Cnt])->getOperand(0)->getType(); 440 if (Ty0 == Ty1) { 441 if (InstOpcode == Opcode || InstOpcode == AltOpcode) 442 continue; 443 if (Opcode == AltOpcode) { 444 assert(isValidForAlternation(Opcode) &&(static_cast <bool> (isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && "Cast isn't safe for alternation, logic needs to be updated!" ) ? void (0) : __assert_fail ("isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && \"Cast isn't safe for alternation, logic needs to be updated!\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 446, __extension__ __PRETTY_FUNCTION__)) 445 isValidForAlternation(InstOpcode) &&(static_cast <bool> (isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && "Cast isn't safe for alternation, logic needs to be updated!" ) ? void (0) : __assert_fail ("isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && \"Cast isn't safe for alternation, logic needs to be updated!\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 446, __extension__ __PRETTY_FUNCTION__)) 446 "Cast isn't safe for alternation, logic needs to be updated!")(static_cast <bool> (isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && "Cast isn't safe for alternation, logic needs to be updated!" ) ? void (0) : __assert_fail ("isValidForAlternation(Opcode) && isValidForAlternation(InstOpcode) && \"Cast isn't safe for alternation, logic needs to be updated!\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 446, __extension__ __PRETTY_FUNCTION__)) ; 447 AltOpcode = InstOpcode; 448 AltIndex = Cnt; 449 continue; 450 } 451 } 452 } else if (InstOpcode == Opcode || InstOpcode == AltOpcode) 453 continue; 454 return InstructionsState(VL[BaseIndex], nullptr, nullptr); 455 } 456 457 return InstructionsState(VL[BaseIndex], cast<Instruction>(VL[BaseIndex]), 458 cast<Instruction>(VL[AltIndex])); 459} 460 461/// \returns true if all of the values in \p VL have the same type or false 462/// otherwise. 463static bool allSameType(ArrayRef<Value *> VL) { 464 Type *Ty = VL[0]->getType(); 465 for (int i = 1, e = VL.size(); i < e; i++) 466 if (VL[i]->getType() != Ty) 467 return false; 468 469 return true; 470} 471 472/// \returns True if Extract{Value,Element} instruction extracts element Idx. 473static Optional<unsigned> getExtractIndex(Instruction *E) { 474 unsigned Opcode = E->getOpcode(); 475 assert((Opcode == Instruction::ExtractElement ||(static_cast <bool> ((Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && "Expected extractelement or extractvalue instruction." ) ? void (0) : __assert_fail ("(Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && \"Expected extractelement or extractvalue instruction.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 477, __extension__ __PRETTY_FUNCTION__)) 476 Opcode == Instruction::ExtractValue) &&(static_cast <bool> ((Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && "Expected extractelement or extractvalue instruction." ) ? void (0) : __assert_fail ("(Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && \"Expected extractelement or extractvalue instruction.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 477, __extension__ __PRETTY_FUNCTION__)) 477 "Expected extractelement or extractvalue instruction.")(static_cast <bool> ((Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && "Expected extractelement or extractvalue instruction." ) ? void (0) : __assert_fail ("(Opcode == Instruction::ExtractElement || Opcode == Instruction::ExtractValue) && \"Expected extractelement or extractvalue instruction.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 477, __extension__ __PRETTY_FUNCTION__)) ; 478 if (Opcode == Instruction::ExtractElement) { 479 auto *CI = dyn_cast<ConstantInt>(E->getOperand(1)); 480 if (!CI) 481 return None; 482 return CI->getZExtValue(); 483 } 484 ExtractValueInst *EI = cast<ExtractValueInst>(E); 485 if (EI->getNumIndices() != 1) 486 return None; 487 return *EI->idx_begin(); 488} 489 490/// \returns True if in-tree use also needs extract. This refers to 491/// possible scalar operand in vectorized instruction. 492static bool InTreeUserNeedToExtract(Value *Scalar, Instruction *UserInst, 493 TargetLibraryInfo *TLI) { 494 unsigned Opcode = UserInst->getOpcode(); 495 switch (Opcode) { 496 case Instruction::Load: { 497 LoadInst *LI = cast<LoadInst>(UserInst); 498 return (LI->getPointerOperand() == Scalar); 499 } 500 case Instruction::Store: { 501 StoreInst *SI = cast<StoreInst>(UserInst); 502 return (SI->getPointerOperand() == Scalar); 503 } 504 case Instruction::Call: { 505 CallInst *CI = cast<CallInst>(UserInst); 506 Intrinsic::ID ID = getVectorIntrinsicIDForCall(CI, TLI); 507 for (unsigned i = 0, e = CI->getNumArgOperands(); i != e; ++i) { 508 if (hasVectorInstrinsicScalarOpd(ID, i)) 509 return (CI->getArgOperand(i) == Scalar); 510 } 511 LLVM_FALLTHROUGH[[gnu::fallthrough]]; 512 } 513 default: 514 return false; 515 } 516} 517 518/// \returns the AA location that is being access by the instruction. 519static MemoryLocation getLocation(Instruction *I, AAResults *AA) { 520 if (StoreInst *SI = dyn_cast<StoreInst>(I)) 521 return MemoryLocation::get(SI); 522 if (LoadInst *LI = dyn_cast<LoadInst>(I)) 523 return MemoryLocation::get(LI); 524 return MemoryLocation(); 525} 526 527/// \returns True if the instruction is not a volatile or atomic load/store. 528static bool isSimple(Instruction *I) { 529 if (LoadInst *LI = dyn_cast<LoadInst>(I)) 530 return LI->isSimple(); 531 if (StoreInst *SI = dyn_cast<StoreInst>(I)) 532 return SI->isSimple(); 533 if (MemIntrinsic *MI = dyn_cast<MemIntrinsic>(I)) 534 return !MI->isVolatile(); 535 return true; 536} 537 538namespace llvm { 539 540static void inversePermutation(ArrayRef<unsigned> Indices, 541 SmallVectorImpl<int> &Mask) { 542 Mask.clear(); 543 const unsigned E = Indices.size(); 544 Mask.resize(E, E + 1); 545 for (unsigned I = 0; I < E; ++I) 546 Mask[Indices[I]] = I; 547} 548 549/// \returns inserting index of InsertElement or InsertValue instruction, 550/// using Offset as base offset for index. 551static Optional<int> getInsertIndex(Value *InsertInst, unsigned Offset) { 552 int Index = Offset; 553 if (auto *IE = dyn_cast<InsertElementInst>(InsertInst)) { 554 if (auto *CI = dyn_cast<ConstantInt>(IE->getOperand(2))) { 555 auto *VT = cast<FixedVectorType>(IE->getType()); 556 if (CI->getValue().uge(VT->getNumElements())) 557 return UndefMaskElem; 558 Index *= VT->getNumElements(); 559 Index += CI->getZExtValue(); 560 return Index; 561 } 562 if (isa<UndefValue>(IE->getOperand(2))) 563 return UndefMaskElem; 564 return None; 565 } 566 567 auto *IV = cast<InsertValueInst>(InsertInst); 568 Type *CurrentType = IV->getType(); 569 for (unsigned I : IV->indices()) { 570 if (auto *ST = dyn_cast<StructType>(CurrentType)) { 571 Index *= ST->getNumElements(); 572 CurrentType = ST->getElementType(I); 573 } else if (auto *AT = dyn_cast<ArrayType>(CurrentType)) { 574 Index *= AT->getNumElements(); 575 CurrentType = AT->getElementType(); 576 } else { 577 return None; 578 } 579 Index += I; 580 } 581 return Index; 582} 583 584namespace slpvectorizer { 585 586/// Bottom Up SLP Vectorizer. 587class BoUpSLP { 588 struct TreeEntry; 589 struct ScheduleData; 590 591public: 592 using ValueList = SmallVector<Value *, 8>; 593 using InstrList = SmallVector<Instruction *, 16>; 594 using ValueSet = SmallPtrSet<Value *, 16>; 595 using StoreList = SmallVector<StoreInst *, 8>; 596 using ExtraValueToDebugLocsMap = 597 MapVector<Value *, SmallVector<Instruction *, 2>>; 598 using OrdersType = SmallVector<unsigned, 4>; 599 600 BoUpSLP(Function *Func, ScalarEvolution *Se, TargetTransformInfo *Tti, 601 TargetLibraryInfo *TLi, AAResults *Aa, LoopInfo *Li, 602 DominatorTree *Dt, AssumptionCache *AC, DemandedBits *DB, 603 const DataLayout *DL, OptimizationRemarkEmitter *ORE) 604 : F(Func), SE(Se), TTI(Tti), TLI(TLi), AA(Aa), LI(Li), DT(Dt), AC(AC), 605 DB(DB), DL(DL), ORE(ORE), Builder(Se->getContext()) { 606 CodeMetrics::collectEphemeralValues(F, AC, EphValues); 607 // Use the vector register size specified by the target unless overridden 608 // by a command-line option. 609 // TODO: It would be better to limit the vectorization factor based on 610 // data type rather than just register size. For example, x86 AVX has 611 // 256-bit registers, but it does not support integer operations 612 // at that width (that requires AVX2). 613 if (MaxVectorRegSizeOption.getNumOccurrences()) 614 MaxVecRegSize = MaxVectorRegSizeOption; 615 else 616 MaxVecRegSize = 617 TTI->getRegisterBitWidth(TargetTransformInfo::RGK_FixedWidthVector) 618 .getFixedSize(); 619 620 if (MinVectorRegSizeOption.getNumOccurrences()) 621 MinVecRegSize = MinVectorRegSizeOption; 622 else 623 MinVecRegSize = TTI->getMinVectorRegisterBitWidth(); 624 } 625 626 /// Vectorize the tree that starts with the elements in \p VL. 627 /// Returns the vectorized root. 628 Value *vectorizeTree(); 629 630 /// Vectorize the tree but with the list of externally used values \p 631 /// ExternallyUsedValues. Values in this MapVector can be replaced but the 632 /// generated extractvalue instructions. 633 Value *vectorizeTree(ExtraValueToDebugLocsMap &ExternallyUsedValues); 634 635 /// \returns the cost incurred by unwanted spills and fills, caused by 636 /// holding live values over call sites. 637 InstructionCost getSpillCost() const; 638 639 /// \returns the vectorization cost of the subtree that starts at \p VL. 640 /// A negative number means that this is profitable. 641 InstructionCost getTreeCost(ArrayRef<Value *> VectorizedVals = None); 642 643 /// Construct a vectorizable tree that starts at \p Roots, ignoring users for 644 /// the purpose of scheduling and extraction in the \p UserIgnoreLst. 645 void buildTree(ArrayRef<Value *> Roots, 646 ArrayRef<Value *> UserIgnoreLst = None); 647 648 /// Construct a vectorizable tree that starts at \p Roots, ignoring users for 649 /// the purpose of scheduling and extraction in the \p UserIgnoreLst taking 650 /// into account (and updating it, if required) list of externally used 651 /// values stored in \p ExternallyUsedValues. 652 void buildTree(ArrayRef<Value *> Roots, 653 ExtraValueToDebugLocsMap &ExternallyUsedValues, 654 ArrayRef<Value *> UserIgnoreLst = None); 655 656 /// Clear the internal data structures that are created by 'buildTree'. 657 void deleteTree() { 658 VectorizableTree.clear(); 659 ScalarToTreeEntry.clear(); 660 MustGather.clear(); 661 ExternalUses.clear(); 662 NumOpsWantToKeepOrder.clear(); 663 NumOpsWantToKeepOriginalOrder = 0; 664 for (auto &Iter : BlocksSchedules) { 665 BlockScheduling *BS = Iter.second.get(); 666 BS->clear(); 667 } 668 MinBWs.clear(); 669 InstrElementSize.clear(); 670 } 671 672 unsigned getTreeSize() const { return VectorizableTree.size(); } 673 674 /// Perform LICM and CSE on the newly generated gather sequences. 675 void optimizeGatherSequence(); 676 677 /// \returns The best order of instructions for vectorization. 678 Optional<ArrayRef<unsigned>> bestOrder() const { 679 assert(llvm::all_of((static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 680 NumOpsWantToKeepOrder,(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 681 [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) {(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 682 return D.getFirst().size() ==(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 683 VectorizableTree[0]->Scalars.size();(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 684 }) &&(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 685 "All orders must have the same size as number of instructions in "(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) 686 "tree node.")(static_cast <bool> (llvm::all_of( NumOpsWantToKeepOrder , [this](const decltype(NumOpsWantToKeepOrder)::value_type & D) { return D.getFirst().size() == VectorizableTree[0]->Scalars .size(); }) && "All orders must have the same size as number of instructions in " "tree node.") ? void (0) : __assert_fail ("llvm::all_of( NumOpsWantToKeepOrder, [this](const decltype(NumOpsWantToKeepOrder)::value_type &D) { return D.getFirst().size() == VectorizableTree[0]->Scalars.size(); }) && \"All orders must have the same size as number of instructions in \" \"tree node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 686, __extension__ __PRETTY_FUNCTION__)) ; 687 auto I = std::max_element( 688 NumOpsWantToKeepOrder.begin(), NumOpsWantToKeepOrder.end(), 689 [](const decltype(NumOpsWantToKeepOrder)::value_type &D1, 690 const decltype(NumOpsWantToKeepOrder)::value_type &D2) { 691 return D1.second < D2.second; 692 }); 693 if (I == NumOpsWantToKeepOrder.end() || 694 I->getSecond() <= NumOpsWantToKeepOriginalOrder) 695 return None; 696 697 return makeArrayRef(I->getFirst()); 698 } 699 700 /// Builds the correct order for root instructions. 701 /// If some leaves have the same instructions to be vectorized, we may 702 /// incorrectly evaluate the best order for the root node (it is built for the 703 /// vector of instructions without repeated instructions and, thus, has less 704 /// elements than the root node). This function builds the correct order for 705 /// the root node. 706 /// For example, if the root node is \<a+b, a+c, a+d, f+e\>, then the leaves 707 /// are \<a, a, a, f\> and \<b, c, d, e\>. When we try to vectorize the first 708 /// leaf, it will be shrink to \<a, b\>. If instructions in this leaf should 709 /// be reordered, the best order will be \<1, 0\>. We need to extend this 710 /// order for the root node. For the root node this order should look like 711 /// \<3, 0, 1, 2\>. This function extends the order for the reused 712 /// instructions. 713 void findRootOrder(OrdersType &Order) { 714 // If the leaf has the same number of instructions to vectorize as the root 715 // - order must be set already. 716 unsigned RootSize = VectorizableTree[0]->Scalars.size(); 717 if (Order.size() == RootSize) 718 return; 719 SmallVector<unsigned, 4> RealOrder(Order.size()); 720 std::swap(Order, RealOrder); 721 SmallVector<int, 4> Mask; 722 inversePermutation(RealOrder, Mask); 723 Order.assign(Mask.begin(), Mask.end()); 724 // The leaf has less number of instructions - need to find the true order of 725 // the root. 726 // Scan the nodes starting from the leaf back to the root. 727 const TreeEntry *PNode = VectorizableTree.back().get(); 728 SmallVector<const TreeEntry *, 4> Nodes(1, PNode); 729 SmallPtrSet<const TreeEntry *, 4> Visited; 730 while (!Nodes.empty() && Order.size() != RootSize) { 731 const TreeEntry *PNode = Nodes.pop_back_val(); 732 if (!Visited.insert(PNode).second) 733 continue; 734 const TreeEntry &Node = *PNode; 735 for (const EdgeInfo &EI : Node.UserTreeIndices) 736 if (EI.UserTE) 737 Nodes.push_back(EI.UserTE); 738 if (Node.ReuseShuffleIndices.empty()) 739 continue; 740 // Build the order for the parent node. 741 OrdersType NewOrder(Node.ReuseShuffleIndices.size(), RootSize); 742 SmallVector<unsigned, 4> OrderCounter(Order.size(), 0); 743 // The algorithm of the order extension is: 744 // 1. Calculate the number of the same instructions for the order. 745 // 2. Calculate the index of the new order: total number of instructions 746 // with order less than the order of the current instruction + reuse 747 // number of the current instruction. 748 // 3. The new order is just the index of the instruction in the original 749 // vector of the instructions. 750 for (unsigned I : Node.ReuseShuffleIndices) 751 ++OrderCounter[Order[I]]; 752 SmallVector<unsigned, 4> CurrentCounter(Order.size(), 0); 753 for (unsigned I = 0, E = Node.ReuseShuffleIndices.size(); I < E; ++I) { 754 unsigned ReusedIdx = Node.ReuseShuffleIndices[I]; 755 unsigned OrderIdx = Order[ReusedIdx]; 756 unsigned NewIdx = 0; 757 for (unsigned J = 0; J < OrderIdx; ++J) 758 NewIdx += OrderCounter[J]; 759 NewIdx += CurrentCounter[OrderIdx]; 760 ++CurrentCounter[OrderIdx]; 761 assert(NewOrder[NewIdx] == RootSize &&(static_cast <bool> (NewOrder[NewIdx] == RootSize && "The order index should not be written already.") ? void (0) : __assert_fail ("NewOrder[NewIdx] == RootSize && \"The order index should not be written already.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 762, __extension__ __PRETTY_FUNCTION__)) 762 "The order index should not be written already.")(static_cast <bool> (NewOrder[NewIdx] == RootSize && "The order index should not be written already.") ? void (0) : __assert_fail ("NewOrder[NewIdx] == RootSize && \"The order index should not be written already.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 762, __extension__ __PRETTY_FUNCTION__)) ; 763 NewOrder[NewIdx] = I; 764 } 765 std::swap(Order, NewOrder); 766 } 767 assert(Order.size() == RootSize &&(static_cast <bool> (Order.size() == RootSize && "Root node is expected or the size of the order must be the same as " "the number of elements in the root node.") ? void (0) : __assert_fail ("Order.size() == RootSize && \"Root node is expected or the size of the order must be the same as \" \"the number of elements in the root node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 769, __extension__ __PRETTY_FUNCTION__)) 768 "Root node is expected or the size of the order must be the same as "(static_cast <bool> (Order.size() == RootSize && "Root node is expected or the size of the order must be the same as " "the number of elements in the root node.") ? void (0) : __assert_fail ("Order.size() == RootSize && \"Root node is expected or the size of the order must be the same as \" \"the number of elements in the root node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 769, __extension__ __PRETTY_FUNCTION__)) 769 "the number of elements in the root node.")(static_cast <bool> (Order.size() == RootSize && "Root node is expected or the size of the order must be the same as " "the number of elements in the root node.") ? void (0) : __assert_fail ("Order.size() == RootSize && \"Root node is expected or the size of the order must be the same as \" \"the number of elements in the root node.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 769, __extension__ __PRETTY_FUNCTION__)) ; 770 assert(llvm::all_of(Order,(static_cast <bool> (llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && "All indices must be initialized" ) ? void (0) : __assert_fail ("llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && \"All indices must be initialized\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 772, __extension__ __PRETTY_FUNCTION__)) 771 [RootSize](unsigned Val) { return Val != RootSize; }) &&(static_cast <bool> (llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && "All indices must be initialized" ) ? void (0) : __assert_fail ("llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && \"All indices must be initialized\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 772, __extension__ __PRETTY_FUNCTION__)) 772 "All indices must be initialized")(static_cast <bool> (llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && "All indices must be initialized" ) ? void (0) : __assert_fail ("llvm::all_of(Order, [RootSize](unsigned Val) { return Val != RootSize; }) && \"All indices must be initialized\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 772, __extension__ __PRETTY_FUNCTION__)) ; 773 } 774 775 /// \return The vector element size in bits to use when vectorizing the 776 /// expression tree ending at \p V. If V is a store, the size is the width of 777 /// the stored value. Otherwise, the size is the width of the largest loaded 778 /// value reaching V. This method is used by the vectorizer to calculate 779 /// vectorization factors. 780 unsigned getVectorElementSize(Value *V); 781 782 /// Compute the minimum type sizes required to represent the entries in a 783 /// vectorizable tree. 784 void computeMinimumValueSizes(); 785 786 // \returns maximum vector register size as set by TTI or overridden by cl::opt. 787 unsigned getMaxVecRegSize() const { 788 return MaxVecRegSize; 789 } 790 791 // \returns minimum vector register size as set by cl::opt. 792 unsigned getMinVecRegSize() const { 793 return MinVecRegSize; 794 } 795 796 unsigned getMaximumVF(unsigned ElemWidth, unsigned Opcode) const { 797 unsigned MaxVF = MaxVFOption.getNumOccurrences() ? 798 MaxVFOption : TTI->getMaximumVF(ElemWidth, Opcode); 799 return MaxVF ? MaxVF : UINT_MAX(2147483647 *2U +1U); 800 } 801 802 /// Check if homogeneous aggregate is isomorphic to some VectorType. 803 /// Accepts homogeneous multidimensional aggregate of scalars/vectors like 804 /// {[4 x i16], [4 x i16]}, { <2 x float>, <2 x float> }, 805 /// {{{i16, i16}, {i16, i16}}, {{i16, i16}, {i16, i16}}} and so on. 806 /// 807 /// \returns number of elements in vector if isomorphism exists, 0 otherwise. 808 unsigned canMapToVector(Type *T, const DataLayout &DL) const; 809 810 /// \returns True if the VectorizableTree is both tiny and not fully 811 /// vectorizable. We do not vectorize such trees. 812 bool isTreeTinyAndNotFullyVectorizable() const; 813 814 /// Assume that a legal-sized 'or'-reduction of shifted/zexted loaded values 815 /// can be load combined in the backend. Load combining may not be allowed in 816 /// the IR optimizer, so we do not want to alter the pattern. For example, 817 /// partially transforming a scalar bswap() pattern into vector code is 818 /// effectively impossible for the backend to undo. 819 /// TODO: If load combining is allowed in the IR optimizer, this analysis 820 /// may not be necessary. 821 bool isLoadCombineReductionCandidate(RecurKind RdxKind) const; 822 823 /// Assume that a vector of stores of bitwise-or/shifted/zexted loaded values 824 /// can be load combined in the backend. Load combining may not be allowed in 825 /// the IR optimizer, so we do not want to alter the pattern. For example, 826 /// partially transforming a scalar bswap() pattern into vector code is 827 /// effectively impossible for the backend to undo. 828 /// TODO: If load combining is allowed in the IR optimizer, this analysis 829 /// may not be necessary. 830 bool isLoadCombineCandidate() const; 831 832 OptimizationRemarkEmitter *getORE() { return ORE; } 833 834 /// This structure holds any data we need about the edges being traversed 835 /// during buildTree_rec(). We keep track of: 836 /// (i) the user TreeEntry index, and 837 /// (ii) the index of the edge. 838 struct EdgeInfo { 839 EdgeInfo() = default; 840 EdgeInfo(TreeEntry *UserTE, unsigned EdgeIdx) 841 : UserTE(UserTE), EdgeIdx(EdgeIdx) {} 842 /// The user TreeEntry. 843 TreeEntry *UserTE = nullptr; 844 /// The operand index of the use. 845 unsigned EdgeIdx = UINT_MAX(2147483647 *2U +1U); 846#ifndef NDEBUG 847 friend inline raw_ostream &operator<<(raw_ostream &OS, 848 const BoUpSLP::EdgeInfo &EI) { 849 EI.dump(OS); 850 return OS; 851 } 852 /// Debug print. 853 void dump(raw_ostream &OS) const { 854 OS << "{User:" << (UserTE ? std::to_string(UserTE->Idx) : "null") 855 << " EdgeIdx:" << EdgeIdx << "}"; 856 } 857 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) void dump() const { dump(dbgs()); } 858#endif 859 }; 860 861 /// A helper data structure to hold the operands of a vector of instructions. 862 /// This supports a fixed vector length for all operand vectors. 863 class VLOperands { 864 /// For each operand we need (i) the value, and (ii) the opcode that it 865 /// would be attached to if the expression was in a left-linearized form. 866 /// This is required to avoid illegal operand reordering. 867 /// For example: 868 /// \verbatim 869 /// 0 Op1 870 /// |/ 871 /// Op1 Op2 Linearized + Op2 872 /// \ / ----------> |/ 873 /// - - 874 /// 875 /// Op1 - Op2 (0 + Op1) - Op2 876 /// \endverbatim 877 /// 878 /// Value Op1 is attached to a '+' operation, and Op2 to a '-'. 879 /// 880 /// Another way to think of this is to track all the operations across the 881 /// path from the operand all the way to the root of the tree and to 882 /// calculate the operation that corresponds to this path. For example, the 883 /// path from Op2 to the root crosses the RHS of the '-', therefore the 884 /// corresponding operation is a '-' (which matches the one in the 885 /// linearized tree, as shown above). 886 /// 887 /// For lack of a better term, we refer to this operation as Accumulated 888 /// Path Operation (APO). 889 struct OperandData { 890 OperandData() = default; 891 OperandData(Value *V, bool APO, bool IsUsed) 892 : V(V), APO(APO), IsUsed(IsUsed) {} 893 /// The operand value. 894 Value *V = nullptr; 895 /// TreeEntries only allow a single opcode, or an alternate sequence of 896 /// them (e.g, +, -). Therefore, we can safely use a boolean value for the 897 /// APO. It is set to 'true' if 'V' is attached to an inverse operation 898 /// in the left-linearized form (e.g., Sub/Div), and 'false' otherwise 899 /// (e.g., Add/Mul) 900 bool APO = false; 901 /// Helper data for the reordering function. 902 bool IsUsed = false; 903 }; 904 905 /// During operand reordering, we are trying to select the operand at lane 906 /// that matches best with the operand at the neighboring lane. Our 907 /// selection is based on the type of value we are looking for. For example, 908 /// if the neighboring lane has a load, we need to look for a load that is 909 /// accessing a consecutive address. These strategies are summarized in the 910 /// 'ReorderingMode' enumerator. 911 enum class ReorderingMode { 912 Load, ///< Matching loads to consecutive memory addresses 913 Opcode, ///< Matching instructions based on opcode (same or alternate) 914 Constant, ///< Matching constants 915 Splat, ///< Matching the same instruction multiple times (broadcast) 916 Failed, ///< We failed to create a vectorizable group 917 }; 918 919 using OperandDataVec = SmallVector<OperandData, 2>; 920 921 /// A vector of operand vectors. 922 SmallVector<OperandDataVec, 4> OpsVec; 923 924 const DataLayout &DL; 925 ScalarEvolution &SE; 926 const BoUpSLP &R; 927 928 /// \returns the operand data at \p OpIdx and \p Lane. 929 OperandData &getData(unsigned OpIdx, unsigned Lane) { 930 return OpsVec[OpIdx][Lane]; 931 } 932 933 /// \returns the operand data at \p OpIdx and \p Lane. Const version. 934 const OperandData &getData(unsigned OpIdx, unsigned Lane) const { 935 return OpsVec[OpIdx][Lane]; 936 } 937 938 /// Clears the used flag for all entries. 939 void clearUsed() { 940 for (unsigned OpIdx = 0, NumOperands = getNumOperands(); 941 OpIdx != NumOperands; ++OpIdx) 942 for (unsigned Lane = 0, NumLanes = getNumLanes(); Lane != NumLanes; 943 ++Lane) 944 OpsVec[OpIdx][Lane].IsUsed = false; 945 } 946 947 /// Swap the operand at \p OpIdx1 with that one at \p OpIdx2. 948 void swap(unsigned OpIdx1, unsigned OpIdx2, unsigned Lane) { 949 std::swap(OpsVec[OpIdx1][Lane], OpsVec[OpIdx2][Lane]); 950 } 951 952 // The hard-coded scores listed here are not very important. When computing 953 // the scores of matching one sub-tree with another, we are basically 954 // counting the number of values that are matching. So even if all scores 955 // are set to 1, we would still get a decent matching result. 956 // However, sometimes we have to break ties. For example we may have to 957 // choose between matching loads vs matching opcodes. This is what these 958 // scores are helping us with: they provide the order of preference. 959 960 /// Loads from consecutive memory addresses, e.g. load(A[i]), load(A[i+1]). 961 static const int ScoreConsecutiveLoads = 3; 962 /// ExtractElementInst from same vector and consecutive indexes. 963 static const int ScoreConsecutiveExtracts = 3; 964 /// Constants. 965 static const int ScoreConstants = 2; 966 /// Instructions with the same opcode. 967 static const int ScoreSameOpcode = 2; 968 /// Instructions with alt opcodes (e.g, add + sub). 969 static const int ScoreAltOpcodes = 1; 970 /// Identical instructions (a.k.a. splat or broadcast). 971 static const int ScoreSplat = 1; 972 /// Matching with an undef is preferable to failing. 973 static const int ScoreUndef = 1; 974 /// Score for failing to find a decent match. 975 static const int ScoreFail = 0; 976 /// User exteranl to the vectorized code. 977 static const int ExternalUseCost = 1; 978 /// The user is internal but in a different lane. 979 static const int UserInDiffLaneCost = ExternalUseCost; 980 981 /// \returns the score of placing \p V1 and \p V2 in consecutive lanes. 982 static int getShallowScore(Value *V1, Value *V2, const DataLayout &DL, 983 ScalarEvolution &SE) { 984 auto *LI1 = dyn_cast<LoadInst>(V1); 985 auto *LI2 = dyn_cast<LoadInst>(V2); 986 if (LI1 && LI2) { 987 if (LI1->getParent() != LI2->getParent()) 988 return VLOperands::ScoreFail; 989 990 Optional<int> Dist = 991 getPointersDiff(LI1->getPointerOperand(), LI2->getPointerOperand(), 992 DL, SE, /*StrictCheck=*/true); 993 return (Dist && *Dist == 1) ? VLOperands::ScoreConsecutiveLoads 994 : VLOperands::ScoreFail; 995 } 996 997 auto *C1 = dyn_cast<Constant>(V1); 998 auto *C2 = dyn_cast<Constant>(V2); 999 if (C1 && C2) 1000 return VLOperands::ScoreConstants; 1001 1002 // Extracts from consecutive indexes of the same vector better score as 1003 // the extracts could be optimized away. 1004 Value *EV; 1005 ConstantInt *Ex1Idx, *Ex2Idx; 1006 if (match(V1, m_ExtractElt(m_Value(EV), m_ConstantInt(Ex1Idx))) && 1007 match(V2, m_ExtractElt(m_Deferred(EV), m_ConstantInt(Ex2Idx))) && 1008 Ex1Idx->getZExtValue() + 1 == Ex2Idx->getZExtValue()) 1009 return VLOperands::ScoreConsecutiveExtracts; 1010 1011 auto *I1 = dyn_cast<Instruction>(V1); 1012 auto *I2 = dyn_cast<Instruction>(V2); 1013 if (I1 && I2) { 1014 if (I1 == I2) 1015 return VLOperands::ScoreSplat; 1016 InstructionsState S = getSameOpcode({I1, I2}); 1017 // Note: Only consider instructions with <= 2 operands to avoid 1018 // complexity explosion. 1019 if (S.getOpcode() && S.MainOp->getNumOperands() <= 2) 1020 return S.isAltShuffle() ? VLOperands::ScoreAltOpcodes 1021 : VLOperands::ScoreSameOpcode; 1022 } 1023 1024 if (isa<UndefValue>(V2)) 1025 return VLOperands::ScoreUndef; 1026 1027 return VLOperands::ScoreFail; 1028 } 1029 1030 /// Holds the values and their lane that are taking part in the look-ahead 1031 /// score calculation. This is used in the external uses cost calculation. 1032 SmallDenseMap<Value *, int> InLookAheadValues; 1033 1034 /// \Returns the additinal cost due to uses of \p LHS and \p RHS that are 1035 /// either external to the vectorized code, or require shuffling. 1036 int getExternalUsesCost(const std::pair<Value *, int> &LHS, 1037 const std::pair<Value *, int> &RHS) { 1038 int Cost = 0; 1039 std::array<std::pair<Value *, int>, 2> Values = {{LHS, RHS}}; 1040 for (int Idx = 0, IdxE = Values.size(); Idx != IdxE; ++Idx) { 1041 Value *V = Values[Idx].first; 1042 if (isa<Constant>(V)) { 1043 // Since this is a function pass, it doesn't make semantic sense to 1044 // walk the users of a subclass of Constant. The users could be in 1045 // another function, or even another module that happens to be in 1046 // the same LLVMContext. 1047 continue; 1048 } 1049 1050 // Calculate the absolute lane, using the minimum relative lane of LHS 1051 // and RHS as base and Idx as the offset. 1052 int Ln = std::min(LHS.second, RHS.second) + Idx; 1053 assert(Ln >= 0 && "Bad lane calculation")(static_cast <bool> (Ln >= 0 && "Bad lane calculation" ) ? void (0) : __assert_fail ("Ln >= 0 && \"Bad lane calculation\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1053, __extension__ __PRETTY_FUNCTION__)) ; 1054 unsigned UsersBudget = LookAheadUsersBudget; 1055 for (User *U : V->users()) { 1056 if (const TreeEntry *UserTE = R.getTreeEntry(U)) { 1057 // The user is in the VectorizableTree. Check if we need to insert. 1058 auto It = llvm::find(UserTE->Scalars, U); 1059 assert(It != UserTE->Scalars.end() && "U is in UserTE")(static_cast <bool> (It != UserTE->Scalars.end() && "U is in UserTE") ? void (0) : __assert_fail ("It != UserTE->Scalars.end() && \"U is in UserTE\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1059, __extension__ __PRETTY_FUNCTION__)) ; 1060 int UserLn = std::distance(UserTE->Scalars.begin(), It); 1061 assert(UserLn >= 0 && "Bad lane")(static_cast <bool> (UserLn >= 0 && "Bad lane" ) ? void (0) : __assert_fail ("UserLn >= 0 && \"Bad lane\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1061, __extension__ __PRETTY_FUNCTION__)) ; 1062 if (UserLn != Ln) 1063 Cost += UserInDiffLaneCost; 1064 } else { 1065 // Check if the user is in the look-ahead code. 1066 auto It2 = InLookAheadValues.find(U); 1067 if (It2 != InLookAheadValues.end()) { 1068 // The user is in the look-ahead code. Check the lane. 1069 if (It2->second != Ln) 1070 Cost += UserInDiffLaneCost; 1071 } else { 1072 // The user is neither in SLP tree nor in the look-ahead code. 1073 Cost += ExternalUseCost; 1074 } 1075 } 1076 // Limit the number of visited uses to cap compilation time. 1077 if (--UsersBudget == 0) 1078 break; 1079 } 1080 } 1081 return Cost; 1082 } 1083 1084 /// Go through the operands of \p LHS and \p RHS recursively until \p 1085 /// MaxLevel, and return the cummulative score. For example: 1086 /// \verbatim 1087 /// A[0] B[0] A[1] B[1] C[0] D[0] B[1] A[1] 1088 /// \ / \ / \ / \ / 1089 /// + + + + 1090 /// G1 G2 G3 G4 1091 /// \endverbatim 1092 /// The getScoreAtLevelRec(G1, G2) function will try to match the nodes at 1093 /// each level recursively, accumulating the score. It starts from matching 1094 /// the additions at level 0, then moves on to the loads (level 1). The 1095 /// score of G1 and G2 is higher than G1 and G3, because {A[0],A[1]} and 1096 /// {B[0],B[1]} match with VLOperands::ScoreConsecutiveLoads, while 1097 /// {A[0],C[0]} has a score of VLOperands::ScoreFail. 1098 /// Please note that the order of the operands does not matter, as we 1099 /// evaluate the score of all profitable combinations of operands. In 1100 /// other words the score of G1 and G4 is the same as G1 and G2. This 1101 /// heuristic is based on ideas described in: 1102 /// Look-ahead SLP: Auto-vectorization in the presence of commutative 1103 /// operations, CGO 2018 by Vasileios Porpodas, Rodrigo C. O. Rocha, 1104 /// Luís F. W. Góes 1105 int getScoreAtLevelRec(const std::pair<Value *, int> &LHS, 1106 const std::pair<Value *, int> &RHS, int CurrLevel, 1107 int MaxLevel) { 1108 1109 Value *V1 = LHS.first; 1110 Value *V2 = RHS.first; 1111 // Get the shallow score of V1 and V2. 1112 int ShallowScoreAtThisLevel = 1113 std::max((int)ScoreFail, getShallowScore(V1, V2, DL, SE) - 1114 getExternalUsesCost(LHS, RHS)); 1115 int Lane1 = LHS.second; 1116 int Lane2 = RHS.second; 1117 1118 // If reached MaxLevel, 1119 // or if V1 and V2 are not instructions, 1120 // or if they are SPLAT, 1121 // or if they are not consecutive, early return the current cost. 1122 auto *I1 = dyn_cast<Instruction>(V1); 1123 auto *I2 = dyn_cast<Instruction>(V2); 1124 if (CurrLevel == MaxLevel || !(I1 && I2) || I1 == I2 || 1125 ShallowScoreAtThisLevel == VLOperands::ScoreFail || 1126 (isa<LoadInst>(I1) && isa<LoadInst>(I2) && ShallowScoreAtThisLevel)) 1127 return ShallowScoreAtThisLevel; 1128 assert(I1 && I2 && "Should have early exited.")(static_cast <bool> (I1 && I2 && "Should have early exited." ) ? void (0) : __assert_fail ("I1 && I2 && \"Should have early exited.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1128, __extension__ __PRETTY_FUNCTION__)) ; 1129 1130 // Keep track of in-tree values for determining the external-use cost. 1131 InLookAheadValues[V1] = Lane1; 1132 InLookAheadValues[V2] = Lane2; 1133 1134 // Contains the I2 operand indexes that got matched with I1 operands. 1135 SmallSet<unsigned, 4> Op2Used; 1136 1137 // Recursion towards the operands of I1 and I2. We are trying all possbile 1138 // operand pairs, and keeping track of the best score. 1139 for (unsigned OpIdx1 = 0, NumOperands1 = I1->getNumOperands(); 1140 OpIdx1 != NumOperands1; ++OpIdx1) { 1141 // Try to pair op1I with the best operand of I2. 1142 int MaxTmpScore = 0; 1143 unsigned MaxOpIdx2 = 0; 1144 bool FoundBest = false; 1145 // If I2 is commutative try all combinations. 1146 unsigned FromIdx = isCommutative(I2) ? 0 : OpIdx1; 1147 unsigned ToIdx = isCommutative(I2) 1148 ? I2->getNumOperands() 1149 : std::min(I2->getNumOperands(), OpIdx1 + 1); 1150 assert(FromIdx <= ToIdx && "Bad index")(static_cast <bool> (FromIdx <= ToIdx && "Bad index" ) ? void (0) : __assert_fail ("FromIdx <= ToIdx && \"Bad index\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1150, __extension__ __PRETTY_FUNCTION__)) ; 1151 for (unsigned OpIdx2 = FromIdx; OpIdx2 != ToIdx; ++OpIdx2) { 1152 // Skip operands already paired with OpIdx1. 1153 if (Op2Used.count(OpIdx2)) 1154 continue; 1155 // Recursively calculate the cost at each level 1156 int TmpScore = getScoreAtLevelRec({I1->getOperand(OpIdx1), Lane1}, 1157 {I2->getOperand(OpIdx2), Lane2}, 1158 CurrLevel + 1, MaxLevel); 1159 // Look for the best score. 1160 if (TmpScore > VLOperands::ScoreFail && TmpScore > MaxTmpScore) { 1161 MaxTmpScore = TmpScore; 1162 MaxOpIdx2 = OpIdx2; 1163 FoundBest = true; 1164 } 1165 } 1166 if (FoundBest) { 1167 // Pair {OpIdx1, MaxOpIdx2} was found to be best. Never revisit it. 1168 Op2Used.insert(MaxOpIdx2); 1169 ShallowScoreAtThisLevel += MaxTmpScore; 1170 } 1171 } 1172 return ShallowScoreAtThisLevel; 1173 } 1174 1175 /// \Returns the look-ahead score, which tells us how much the sub-trees 1176 /// rooted at \p LHS and \p RHS match, the more they match the higher the 1177 /// score. This helps break ties in an informed way when we cannot decide on 1178 /// the order of the operands by just considering the immediate 1179 /// predecessors. 1180 int getLookAheadScore(const std::pair<Value *, int> &LHS, 1181 const std::pair<Value *, int> &RHS) { 1182 InLookAheadValues.clear(); 1183 return getScoreAtLevelRec(LHS, RHS, 1, LookAheadMaxDepth); 1184 } 1185 1186 // Search all operands in Ops[*][Lane] for the one that matches best 1187 // Ops[OpIdx][LastLane] and return its opreand index. 1188 // If no good match can be found, return None. 1189 Optional<unsigned> 1190 getBestOperand(unsigned OpIdx, int Lane, int LastLane, 1191 ArrayRef<ReorderingMode> ReorderingModes) { 1192 unsigned NumOperands = getNumOperands(); 1193 1194 // The operand of the previous lane at OpIdx. 1195 Value *OpLastLane = getData(OpIdx, LastLane).V; 1196 1197 // Our strategy mode for OpIdx. 1198 ReorderingMode RMode = ReorderingModes[OpIdx]; 1199 1200 // The linearized opcode of the operand at OpIdx, Lane. 1201 bool OpIdxAPO = getData(OpIdx, Lane).APO; 1202 1203 // The best operand index and its score. 1204 // Sometimes we have more than one option (e.g., Opcode and Undefs), so we 1205 // are using the score to differentiate between the two. 1206 struct BestOpData { 1207 Optional<unsigned> Idx = None; 1208 unsigned Score = 0; 1209 } BestOp; 1210 1211 // Iterate through all unused operands and look for the best. 1212 for (unsigned Idx = 0; Idx != NumOperands; ++Idx) { 1213 // Get the operand at Idx and Lane. 1214 OperandData &OpData = getData(Idx, Lane); 1215 Value *Op = OpData.V; 1216 bool OpAPO = OpData.APO; 1217 1218 // Skip already selected operands. 1219 if (OpData.IsUsed) 1220 continue; 1221 1222 // Skip if we are trying to move the operand to a position with a 1223 // different opcode in the linearized tree form. This would break the 1224 // semantics. 1225 if (OpAPO != OpIdxAPO) 1226 continue; 1227 1228 // Look for an operand that matches the current mode. 1229 switch (RMode) { 1230 case ReorderingMode::Load: 1231 case ReorderingMode::Constant: 1232 case ReorderingMode::Opcode: { 1233 bool LeftToRight = Lane > LastLane; 1234 Value *OpLeft = (LeftToRight) ? OpLastLane : Op; 1235 Value *OpRight = (LeftToRight) ? Op : OpLastLane; 1236 unsigned Score = 1237 getLookAheadScore({OpLeft, LastLane}, {OpRight, Lane}); 1238 if (Score > BestOp.Score) { 1239 BestOp.Idx = Idx; 1240 BestOp.Score = Score; 1241 } 1242 break; 1243 } 1244 case ReorderingMode::Splat: 1245 if (Op == OpLastLane) 1246 BestOp.Idx = Idx; 1247 break; 1248 case ReorderingMode::Failed: 1249 return None; 1250 } 1251 } 1252 1253 if (BestOp.Idx) { 1254 getData(BestOp.Idx.getValue(), Lane).IsUsed = true; 1255 return BestOp.Idx; 1256 } 1257 // If we could not find a good match return None. 1258 return None; 1259 } 1260 1261 /// Helper for reorderOperandVecs. \Returns the lane that we should start 1262 /// reordering from. This is the one which has the least number of operands 1263 /// that can freely move about. 1264 unsigned getBestLaneToStartReordering() const { 1265 unsigned BestLane = 0; 1266 unsigned Min = UINT_MAX(2147483647 *2U +1U); 1267 for (unsigned Lane = 0, NumLanes = getNumLanes(); Lane != NumLanes; 1268 ++Lane) { 1269 unsigned NumFreeOps = getMaxNumOperandsThatCanBeReordered(Lane); 1270 if (NumFreeOps < Min) { 1271 Min = NumFreeOps; 1272 BestLane = Lane; 1273 } 1274 } 1275 return BestLane; 1276 } 1277 1278 /// \Returns the maximum number of operands that are allowed to be reordered 1279 /// for \p Lane. This is used as a heuristic for selecting the first lane to 1280 /// start operand reordering. 1281 unsigned getMaxNumOperandsThatCanBeReordered(unsigned Lane) const { 1282 unsigned CntTrue = 0; 1283 unsigned NumOperands = getNumOperands(); 1284 // Operands with the same APO can be reordered. We therefore need to count 1285 // how many of them we have for each APO, like this: Cnt[APO] = x. 1286 // Since we only have two APOs, namely true and false, we can avoid using 1287 // a map. Instead we can simply count the number of operands that 1288 // correspond to one of them (in this case the 'true' APO), and calculate 1289 // the other by subtracting it from the total number of operands. 1290 for (unsigned OpIdx = 0; OpIdx != NumOperands; ++OpIdx) 1291 if (getData(OpIdx, Lane).APO) 1292 ++CntTrue; 1293 unsigned CntFalse = NumOperands - CntTrue; 1294 return std::max(CntTrue, CntFalse); 1295 } 1296 1297 /// Go through the instructions in VL and append their operands. 1298 void appendOperandsOfVL(ArrayRef<Value *> VL) { 1299 assert(!VL.empty() && "Bad VL")(static_cast <bool> (!VL.empty() && "Bad VL") ? void (0) : __assert_fail ("!VL.empty() && \"Bad VL\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1299, __extension__ __PRETTY_FUNCTION__)) ; 1300 assert((empty() || VL.size() == getNumLanes()) &&(static_cast <bool> ((empty() || VL.size() == getNumLanes ()) && "Expected same number of lanes") ? void (0) : __assert_fail ("(empty() || VL.size() == getNumLanes()) && \"Expected same number of lanes\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1301, __extension__ __PRETTY_FUNCTION__)) 1301 "Expected same number of lanes")(static_cast <bool> ((empty() || VL.size() == getNumLanes ()) && "Expected same number of lanes") ? void (0) : __assert_fail ("(empty() || VL.size() == getNumLanes()) && \"Expected same number of lanes\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1301, __extension__ __PRETTY_FUNCTION__)) ; 1302 assert(isa<Instruction>(VL[0]) && "Expected instruction")(static_cast <bool> (isa<Instruction>(VL[0]) && "Expected instruction") ? void (0) : __assert_fail ("isa<Instruction>(VL[0]) && \"Expected instruction\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1302, __extension__ __PRETTY_FUNCTION__)) ; 1303 unsigned NumOperands = cast<Instruction>(VL[0])->getNumOperands(); 1304 OpsVec.resize(NumOperands); 1305 unsigned NumLanes = VL.size(); 1306 for (unsigned OpIdx = 0; OpIdx != NumOperands; ++OpIdx) { 1307 OpsVec[OpIdx].resize(NumLanes); 1308 for (unsigned Lane = 0; Lane != NumLanes; ++Lane) { 1309 assert(isa<Instruction>(VL[Lane]) && "Expected instruction")(static_cast <bool> (isa<Instruction>(VL[Lane]) && "Expected instruction") ? void (0) : __assert_fail ("isa<Instruction>(VL[Lane]) && \"Expected instruction\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1309, __extension__ __PRETTY_FUNCTION__)) ; 1310 // Our tree has just 3 nodes: the root and two operands. 1311 // It is therefore trivial to get the APO. We only need to check the 1312 // opcode of VL[Lane] and whether the operand at OpIdx is the LHS or 1313 // RHS operand. The LHS operand of both add and sub is never attached 1314 // to an inversese operation in the linearized form, therefore its APO 1315 // is false. The RHS is true only if VL[Lane] is an inverse operation. 1316 1317 // Since operand reordering is performed on groups of commutative 1318 // operations or alternating sequences (e.g., +, -), we can safely 1319 // tell the inverse operations by checking commutativity. 1320 bool IsInverseOperation = !isCommutative(cast<Instruction>(VL[Lane])); 1321 bool APO = (OpIdx == 0) ? false : IsInverseOperation; 1322 OpsVec[OpIdx][Lane] = {cast<Instruction>(VL[Lane])->getOperand(OpIdx), 1323 APO, false}; 1324 } 1325 } 1326 } 1327 1328 /// \returns the number of operands. 1329 unsigned getNumOperands() const { return OpsVec.size(); } 1330 1331 /// \returns the number of lanes. 1332 unsigned getNumLanes() const { return OpsVec[0].size(); } 1333 1334 /// \returns the operand value at \p OpIdx and \p Lane. 1335 Value *getValue(unsigned OpIdx, unsigned Lane) const { 1336 return getData(OpIdx, Lane).V; 1337 } 1338 1339 /// \returns true if the data structure is empty. 1340 bool empty() const { return OpsVec.empty(); } 1341 1342 /// Clears the data. 1343 void clear() { OpsVec.clear(); } 1344 1345 /// \Returns true if there are enough operands identical to \p Op to fill 1346 /// the whole vector. 1347 /// Note: This modifies the 'IsUsed' flag, so a cleanUsed() must follow. 1348 bool shouldBroadcast(Value *Op, unsigned OpIdx, unsigned Lane) { 1349 bool OpAPO = getData(OpIdx, Lane).APO; 1350 for (unsigned Ln = 0, Lns = getNumLanes(); Ln != Lns; ++Ln) { 1351 if (Ln == Lane) 1352 continue; 1353 // This is set to true if we found a candidate for broadcast at Lane. 1354 bool FoundCandidate = false; 1355 for (unsigned OpI = 0, OpE = getNumOperands(); OpI != OpE; ++OpI) { 1356 OperandData &Data = getData(OpI, Ln); 1357 if (Data.APO != OpAPO || Data.IsUsed) 1358 continue; 1359 if (Data.V == Op) { 1360 FoundCandidate = true; 1361 Data.IsUsed = true; 1362 break; 1363 } 1364 } 1365 if (!FoundCandidate) 1366 return false; 1367 } 1368 return true; 1369 } 1370 1371 public: 1372 /// Initialize with all the operands of the instruction vector \p RootVL. 1373 VLOperands(ArrayRef<Value *> RootVL, const DataLayout &DL, 1374 ScalarEvolution &SE, const BoUpSLP &R) 1375 : DL(DL), SE(SE), R(R) { 1376 // Append all the operands of RootVL. 1377 appendOperandsOfVL(RootVL); 1378 } 1379 1380 /// \Returns a value vector with the operands across all lanes for the 1381 /// opearnd at \p OpIdx. 1382 ValueList getVL(unsigned OpIdx) const { 1383 ValueList OpVL(OpsVec[OpIdx].size()); 1384 assert(OpsVec[OpIdx].size() == getNumLanes() &&(static_cast <bool> (OpsVec[OpIdx].size() == getNumLanes () && "Expected same num of lanes across all operands" ) ? void (0) : __assert_fail ("OpsVec[OpIdx].size() == getNumLanes() && \"Expected same num of lanes across all operands\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1385, __extension__ __PRETTY_FUNCTION__)) 1385 "Expected same num of lanes across all operands")(static_cast <bool> (OpsVec[OpIdx].size() == getNumLanes () && "Expected same num of lanes across all operands" ) ? void (0) : __assert_fail ("OpsVec[OpIdx].size() == getNumLanes() && \"Expected same num of lanes across all operands\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1385, __extension__ __PRETTY_FUNCTION__)) ; 1386 for (unsigned Lane = 0, Lanes = getNumLanes(); Lane != Lanes; ++Lane) 1387 OpVL[Lane] = OpsVec[OpIdx][Lane].V; 1388 return OpVL; 1389 } 1390 1391 // Performs operand reordering for 2 or more operands. 1392 // The original operands are in OrigOps[OpIdx][Lane]. 1393 // The reordered operands are returned in 'SortedOps[OpIdx][Lane]'. 1394 void reorder() { 1395 unsigned NumOperands = getNumOperands(); 1396 unsigned NumLanes = getNumLanes(); 1397 // Each operand has its own mode. We are using this mode to help us select 1398 // the instructions for each lane, so that they match best with the ones 1399 // we have selected so far. 1400 SmallVector<ReorderingMode, 2> ReorderingModes(NumOperands); 1401 1402 // This is a greedy single-pass algorithm. We are going over each lane 1403 // once and deciding on the best order right away with no back-tracking. 1404 // However, in order to increase its effectiveness, we start with the lane 1405 // that has operands that can move the least. For example, given the 1406 // following lanes: 1407 // Lane 0 : A[0] = B[0] + C[0] // Visited 3rd 1408 // Lane 1 : A[1] = C[1] - B[1] // Visited 1st 1409 // Lane 2 : A[2] = B[2] + C[2] // Visited 2nd 1410 // Lane 3 : A[3] = C[3] - B[3] // Visited 4th 1411 // we will start at Lane 1, since the operands of the subtraction cannot 1412 // be reordered. Then we will visit the rest of the lanes in a circular 1413 // fashion. That is, Lanes 2, then Lane 0, and finally Lane 3. 1414 1415 // Find the first lane that we will start our search from. 1416 unsigned FirstLane = getBestLaneToStartReordering(); 1417 1418 // Initialize the modes. 1419 for (unsigned OpIdx = 0; OpIdx != NumOperands; ++OpIdx) { 1420 Value *OpLane0 = getValue(OpIdx, FirstLane); 1421 // Keep track if we have instructions with all the same opcode on one 1422 // side. 1423 if (isa<LoadInst>(OpLane0)) 1424 ReorderingModes[OpIdx] = ReorderingMode::Load; 1425 else if (isa<Instruction>(OpLane0)) { 1426 // Check if OpLane0 should be broadcast. 1427 if (shouldBroadcast(OpLane0, OpIdx, FirstLane)) 1428 ReorderingModes[OpIdx] = ReorderingMode::Splat; 1429 else 1430 ReorderingModes[OpIdx] = ReorderingMode::Opcode; 1431 } 1432 else if (isa<Constant>(OpLane0)) 1433 ReorderingModes[OpIdx] = ReorderingMode::Constant; 1434 else if (isa<Argument>(OpLane0)) 1435 // Our best hope is a Splat. It may save some cost in some cases. 1436 ReorderingModes[OpIdx] = ReorderingMode::Splat; 1437 else 1438 // NOTE: This should be unreachable. 1439 ReorderingModes[OpIdx] = ReorderingMode::Failed; 1440 } 1441 1442 // If the initial strategy fails for any of the operand indexes, then we 1443 // perform reordering again in a second pass. This helps avoid assigning 1444 // high priority to the failed strategy, and should improve reordering for 1445 // the non-failed operand indexes. 1446 for (int Pass = 0; Pass != 2; ++Pass) { 1447 // Skip the second pass if the first pass did not fail. 1448 bool StrategyFailed = false; 1449 // Mark all operand data as free to use. 1450 clearUsed(); 1451 // We keep the original operand order for the FirstLane, so reorder the 1452 // rest of the lanes. We are visiting the nodes in a circular fashion, 1453 // using FirstLane as the center point and increasing the radius 1454 // distance. 1455 for (unsigned Distance = 1; Distance != NumLanes; ++Distance) { 1456 // Visit the lane on the right and then the lane on the left. 1457 for (int Direction : {+1, -1}) { 1458 int Lane = FirstLane + Direction * Distance; 1459 if (Lane < 0 || Lane >= (int)NumLanes) 1460 continue; 1461 int LastLane = Lane - Direction; 1462 assert(LastLane >= 0 && LastLane < (int)NumLanes &&(static_cast <bool> (LastLane >= 0 && LastLane < (int)NumLanes && "Out of bounds") ? void (0) : __assert_fail ("LastLane >= 0 && LastLane < (int)NumLanes && \"Out of bounds\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1463, __extension__ __PRETTY_FUNCTION__)) 1463 "Out of bounds")(static_cast <bool> (LastLane >= 0 && LastLane < (int)NumLanes && "Out of bounds") ? void (0) : __assert_fail ("LastLane >= 0 && LastLane < (int)NumLanes && \"Out of bounds\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1463, __extension__ __PRETTY_FUNCTION__)) ; 1464 // Look for a good match for each operand. 1465 for (unsigned OpIdx = 0; OpIdx != NumOperands; ++OpIdx) { 1466 // Search for the operand that matches SortedOps[OpIdx][Lane-1]. 1467 Optional<unsigned> BestIdx = 1468 getBestOperand(OpIdx, Lane, LastLane, ReorderingModes); 1469 // By not selecting a value, we allow the operands that follow to 1470 // select a better matching value. We will get a non-null value in 1471 // the next run of getBestOperand(). 1472 if (BestIdx) { 1473 // Swap the current operand with the one returned by 1474 // getBestOperand(). 1475 swap(OpIdx, BestIdx.getValue(), Lane); 1476 } else { 1477 // We failed to find a best operand, set mode to 'Failed'. 1478 ReorderingModes[OpIdx] = ReorderingMode::Failed; 1479 // Enable the second pass. 1480 StrategyFailed = true; 1481 } 1482 } 1483 } 1484 } 1485 // Skip second pass if the strategy did not fail. 1486 if (!StrategyFailed) 1487 break; 1488 } 1489 } 1490 1491#if !defined(NDEBUG) || defined(LLVM_ENABLE_DUMP) 1492 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) static StringRef getModeStr(ReorderingMode RMode) { 1493 switch (RMode) { 1494 case ReorderingMode::Load: 1495 return "Load"; 1496 case ReorderingMode::Opcode: 1497 return "Opcode"; 1498 case ReorderingMode::Constant: 1499 return "Constant"; 1500 case ReorderingMode::Splat: 1501 return "Splat"; 1502 case ReorderingMode::Failed: 1503 return "Failed"; 1504 } 1505 llvm_unreachable("Unimplemented Reordering Type")::llvm::llvm_unreachable_internal("Unimplemented Reordering Type" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1505) ; 1506 } 1507 1508 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) static raw_ostream &printMode(ReorderingMode RMode, 1509 raw_ostream &OS) { 1510 return OS << getModeStr(RMode); 1511 } 1512 1513 /// Debug print. 1514 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) static void dumpMode(ReorderingMode RMode) { 1515 printMode(RMode, dbgs()); 1516 } 1517 1518 friend raw_ostream &operator<<(raw_ostream &OS, ReorderingMode RMode) { 1519 return printMode(RMode, OS); 1520 } 1521 1522 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) raw_ostream &print(raw_ostream &OS) const { 1523 const unsigned Indent = 2; 1524 unsigned Cnt = 0; 1525 for (const OperandDataVec &OpDataVec : OpsVec) { 1526 OS << "Operand " << Cnt++ << "\n"; 1527 for (const OperandData &OpData : OpDataVec) { 1528 OS.indent(Indent) << "{"; 1529 if (Value *V = OpData.V) 1530 OS << *V; 1531 else 1532 OS << "null"; 1533 OS << ", APO:" << OpData.APO << "}\n"; 1534 } 1535 OS << "\n"; 1536 } 1537 return OS; 1538 } 1539 1540 /// Debug print. 1541 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) void dump() const { print(dbgs()); } 1542#endif 1543 }; 1544 1545 /// Checks if the instruction is marked for deletion. 1546 bool isDeleted(Instruction *I) const { return DeletedInstructions.count(I); } 1547 1548 /// Marks values operands for later deletion by replacing them with Undefs. 1549 void eraseInstructions(ArrayRef<Value *> AV); 1550 1551 ~BoUpSLP(); 1552 1553private: 1554 /// Checks if all users of \p I are the part of the vectorization tree. 1555 bool areAllUsersVectorized(Instruction *I, 1556 ArrayRef<Value *> VectorizedVals) const; 1557 1558 /// \returns the cost of the vectorizable entry. 1559 InstructionCost getEntryCost(const TreeEntry *E, 1560 ArrayRef<Value *> VectorizedVals); 1561 1562 /// This is the recursive part of buildTree. 1563 void buildTree_rec(ArrayRef<Value *> Roots, unsigned Depth, 1564 const EdgeInfo &EI); 1565 1566 /// \returns true if the ExtractElement/ExtractValue instructions in \p VL can 1567 /// be vectorized to use the original vector (or aggregate "bitcast" to a 1568 /// vector) and sets \p CurrentOrder to the identity permutation; otherwise 1569 /// returns false, setting \p CurrentOrder to either an empty vector or a 1570 /// non-identity permutation that allows to reuse extract instructions. 1571 bool canReuseExtract(ArrayRef<Value *> VL, Value *OpValue, 1572 SmallVectorImpl<unsigned> &CurrentOrder) const; 1573 1574 /// Vectorize a single entry in the tree. 1575 Value *vectorizeTree(TreeEntry *E); 1576 1577 /// Vectorize a single entry in the tree, starting in \p VL. 1578 Value *vectorizeTree(ArrayRef<Value *> VL); 1579 1580 /// \returns the scalarization cost for this type. Scalarization in this 1581 /// context means the creation of vectors from a group of scalars. 1582 InstructionCost 1583 getGatherCost(FixedVectorType *Ty, 1584 const DenseSet<unsigned> &ShuffledIndices) const; 1585 1586 /// Checks if the gathered \p VL can be represented as shuffle(s) of previous 1587 /// tree entries. 1588 /// \returns ShuffleKind, if gathered values can be represented as shuffles of 1589 /// previous tree entries. \p Mask is filled with the shuffle mask. 1590 Optional<TargetTransformInfo::ShuffleKind> 1591 isGatherShuffledEntry(const TreeEntry *TE, SmallVectorImpl<int> &Mask, 1592 SmallVectorImpl<const TreeEntry *> &Entries); 1593 1594 /// \returns the scalarization cost for this list of values. Assuming that 1595 /// this subtree gets vectorized, we may need to extract the values from the 1596 /// roots. This method calculates the cost of extracting the values. 1597 InstructionCost getGatherCost(ArrayRef<Value *> VL) const; 1598 1599 /// Set the Builder insert point to one after the last instruction in 1600 /// the bundle 1601 void setInsertPointAfterBundle(const TreeEntry *E); 1602 1603 /// \returns a vector from a collection of scalars in \p VL. 1604 Value *gather(ArrayRef<Value *> VL); 1605 1606 /// \returns whether the VectorizableTree is fully vectorizable and will 1607 /// be beneficial even the tree height is tiny. 1608 bool isFullyVectorizableTinyTree() const; 1609 1610 /// Reorder commutative or alt operands to get better probability of 1611 /// generating vectorized code. 1612 static void reorderInputsAccordingToOpcode(ArrayRef<Value *> VL, 1613 SmallVectorImpl<Value *> &Left, 1614 SmallVectorImpl<Value *> &Right, 1615 const DataLayout &DL, 1616 ScalarEvolution &SE, 1617 const BoUpSLP &R); 1618 struct TreeEntry { 1619 using VecTreeTy = SmallVector<std::unique_ptr<TreeEntry>, 8>; 1620 TreeEntry(VecTreeTy &Container) : Container(Container) {} 1621 1622 /// \returns true if the scalars in VL are equal to this entry. 1623 bool isSame(ArrayRef<Value *> VL) const { 1624 if (VL.size() == Scalars.size()) 1625 return std::equal(VL.begin(), VL.end(), Scalars.begin()); 1626 return VL.size() == ReuseShuffleIndices.size() && 1627 std::equal( 1628 VL.begin(), VL.end(), ReuseShuffleIndices.begin(), 1629 [this](Value *V, int Idx) { return V == Scalars[Idx]; }); 1630 } 1631 1632 /// A vector of scalars. 1633 ValueList Scalars; 1634 1635 /// The Scalars are vectorized into this value. It is initialized to Null. 1636 Value *VectorizedValue = nullptr; 1637 1638 /// Do we need to gather this sequence or vectorize it 1639 /// (either with vector instruction or with scatter/gather 1640 /// intrinsics for store/load)? 1641 enum EntryState { Vectorize, ScatterVectorize, NeedToGather }; 1642 EntryState State; 1643 1644 /// Does this sequence require some shuffling? 1645 SmallVector<int, 4> ReuseShuffleIndices; 1646 1647 /// Does this entry require reordering? 1648 SmallVector<unsigned, 4> ReorderIndices; 1649 1650 /// Points back to the VectorizableTree. 1651 /// 1652 /// Only used for Graphviz right now. Unfortunately GraphTrait::NodeRef has 1653 /// to be a pointer and needs to be able to initialize the child iterator. 1654 /// Thus we need a reference back to the container to translate the indices 1655 /// to entries. 1656 VecTreeTy &Container; 1657 1658 /// The TreeEntry index containing the user of this entry. We can actually 1659 /// have multiple users so the data structure is not truly a tree. 1660 SmallVector<EdgeInfo, 1> UserTreeIndices; 1661 1662 /// The index of this treeEntry in VectorizableTree. 1663 int Idx = -1; 1664 1665 private: 1666 /// The operands of each instruction in each lane Operands[op_index][lane]. 1667 /// Note: This helps avoid the replication of the code that performs the 1668 /// reordering of operands during buildTree_rec() and vectorizeTree(). 1669 SmallVector<ValueList, 2> Operands; 1670 1671 /// The main/alternate instruction. 1672 Instruction *MainOp = nullptr; 1673 Instruction *AltOp = nullptr; 1674 1675 public: 1676 /// Set this bundle's \p OpIdx'th operand to \p OpVL. 1677 void setOperand(unsigned OpIdx, ArrayRef<Value *> OpVL) { 1678 if (Operands.size() < OpIdx + 1) 1679 Operands.resize(OpIdx + 1); 1680 assert(Operands[OpIdx].empty() && "Already resized?")(static_cast <bool> (Operands[OpIdx].empty() && "Already resized?") ? void (0) : __assert_fail ("Operands[OpIdx].empty() && \"Already resized?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1680, __extension__ __PRETTY_FUNCTION__)) ; 1681 Operands[OpIdx].resize(Scalars.size()); 1682 for (unsigned Lane = 0, E = Scalars.size(); Lane != E; ++Lane) 1683 Operands[OpIdx][Lane] = OpVL[Lane]; 1684 } 1685 1686 /// Set the operands of this bundle in their original order. 1687 void setOperandsInOrder() { 1688 assert(Operands.empty() && "Already initialized?")(static_cast <bool> (Operands.empty() && "Already initialized?" ) ? void (0) : __assert_fail ("Operands.empty() && \"Already initialized?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1688, __extension__ __PRETTY_FUNCTION__)) ; 1689 auto *I0 = cast<Instruction>(Scalars[0]); 1690 Operands.resize(I0->getNumOperands()); 1691 unsigned NumLanes = Scalars.size(); 1692 for (unsigned OpIdx = 0, NumOperands = I0->getNumOperands(); 1693 OpIdx != NumOperands; ++OpIdx) { 1694 Operands[OpIdx].resize(NumLanes); 1695 for (unsigned Lane = 0; Lane != NumLanes; ++Lane) { 1696 auto *I = cast<Instruction>(Scalars[Lane]); 1697 assert(I->getNumOperands() == NumOperands &&(static_cast <bool> (I->getNumOperands() == NumOperands && "Expected same number of operands") ? void (0) : __assert_fail ("I->getNumOperands() == NumOperands && \"Expected same number of operands\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1698, __extension__ __PRETTY_FUNCTION__)) 1698 "Expected same number of operands")(static_cast <bool> (I->getNumOperands() == NumOperands && "Expected same number of operands") ? void (0) : __assert_fail ("I->getNumOperands() == NumOperands && \"Expected same number of operands\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1698, __extension__ __PRETTY_FUNCTION__)) ; 1699 Operands[OpIdx][Lane] = I->getOperand(OpIdx); 1700 } 1701 } 1702 } 1703 1704 /// \returns the \p OpIdx operand of this TreeEntry. 1705 ValueList &getOperand(unsigned OpIdx) { 1706 assert(OpIdx < Operands.size() && "Off bounds")(static_cast <bool> (OpIdx < Operands.size() && "Off bounds") ? void (0) : __assert_fail ("OpIdx < Operands.size() && \"Off bounds\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1706, __extension__ __PRETTY_FUNCTION__)) ; 1707 return Operands[OpIdx]; 1708 } 1709 1710 /// \returns the number of operands. 1711 unsigned getNumOperands() const { return Operands.size(); } 1712 1713 /// \return the single \p OpIdx operand. 1714 Value *getSingleOperand(unsigned OpIdx) const { 1715 assert(OpIdx < Operands.size() && "Off bounds")(static_cast <bool> (OpIdx < Operands.size() && "Off bounds") ? void (0) : __assert_fail ("OpIdx < Operands.size() && \"Off bounds\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1715, __extension__ __PRETTY_FUNCTION__)) ; 1716 assert(!Operands[OpIdx].empty() && "No operand available")(static_cast <bool> (!Operands[OpIdx].empty() && "No operand available") ? void (0) : __assert_fail ("!Operands[OpIdx].empty() && \"No operand available\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1716, __extension__ __PRETTY_FUNCTION__)) ; 1717 return Operands[OpIdx][0]; 1718 } 1719 1720 /// Some of the instructions in the list have alternate opcodes. 1721 bool isAltShuffle() const { 1722 return getOpcode() != getAltOpcode(); 1723 } 1724 1725 bool isOpcodeOrAlt(Instruction *I) const { 1726 unsigned CheckedOpcode = I->getOpcode(); 1727 return (getOpcode() == CheckedOpcode || 1728 getAltOpcode() == CheckedOpcode); 1729 } 1730 1731 /// Chooses the correct key for scheduling data. If \p Op has the same (or 1732 /// alternate) opcode as \p OpValue, the key is \p Op. Otherwise the key is 1733 /// \p OpValue. 1734 Value *isOneOf(Value *Op) const { 1735 auto *I = dyn_cast<Instruction>(Op); 1736 if (I && isOpcodeOrAlt(I)) 1737 return Op; 1738 return MainOp; 1739 } 1740 1741 void setOperations(const InstructionsState &S) { 1742 MainOp = S.MainOp; 1743 AltOp = S.AltOp; 1744 } 1745 1746 Instruction *getMainOp() const { 1747 return MainOp; 1748 } 1749 1750 Instruction *getAltOp() const { 1751 return AltOp; 1752 } 1753 1754 /// The main/alternate opcodes for the list of instructions. 1755 unsigned getOpcode() const { 1756 return MainOp ? MainOp->getOpcode() : 0; 1757 } 1758 1759 unsigned getAltOpcode() const { 1760 return AltOp ? AltOp->getOpcode() : 0; 1761 } 1762 1763 /// Update operations state of this entry if reorder occurred. 1764 bool updateStateIfReorder() { 1765 if (ReorderIndices.empty()) 1766 return false; 1767 InstructionsState S = getSameOpcode(Scalars, ReorderIndices.front()); 1768 setOperations(S); 1769 return true; 1770 } 1771 1772#ifndef NDEBUG 1773 /// Debug printer. 1774 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) void dump() const { 1775 dbgs() << Idx << ".\n"; 1776 for (unsigned OpI = 0, OpE = Operands.size(); OpI != OpE; ++OpI) { 1777 dbgs() << "Operand " << OpI << ":\n"; 1778 for (const Value *V : Operands[OpI]) 1779 dbgs().indent(2) << *V << "\n"; 1780 } 1781 dbgs() << "Scalars: \n"; 1782 for (Value *V : Scalars) 1783 dbgs().indent(2) << *V << "\n"; 1784 dbgs() << "State: "; 1785 switch (State) { 1786 case Vectorize: 1787 dbgs() << "Vectorize\n"; 1788 break; 1789 case ScatterVectorize: 1790 dbgs() << "ScatterVectorize\n"; 1791 break; 1792 case NeedToGather: 1793 dbgs() << "NeedToGather\n"; 1794 break; 1795 } 1796 dbgs() << "MainOp: "; 1797 if (MainOp) 1798 dbgs() << *MainOp << "\n"; 1799 else 1800 dbgs() << "NULL\n"; 1801 dbgs() << "AltOp: "; 1802 if (AltOp) 1803 dbgs() << *AltOp << "\n"; 1804 else 1805 dbgs() << "NULL\n"; 1806 dbgs() << "VectorizedValue: "; 1807 if (VectorizedValue) 1808 dbgs() << *VectorizedValue << "\n"; 1809 else 1810 dbgs() << "NULL\n"; 1811 dbgs() << "ReuseShuffleIndices: "; 1812 if (ReuseShuffleIndices.empty()) 1813 dbgs() << "Empty"; 1814 else 1815 for (unsigned ReuseIdx : ReuseShuffleIndices) 1816 dbgs() << ReuseIdx << ", "; 1817 dbgs() << "\n"; 1818 dbgs() << "ReorderIndices: "; 1819 for (unsigned ReorderIdx : ReorderIndices) 1820 dbgs() << ReorderIdx << ", "; 1821 dbgs() << "\n"; 1822 dbgs() << "UserTreeIndices: "; 1823 for (const auto &EInfo : UserTreeIndices) 1824 dbgs() << EInfo << ", "; 1825 dbgs() << "\n"; 1826 } 1827#endif 1828 }; 1829 1830#ifndef NDEBUG 1831 void dumpTreeCosts(const TreeEntry *E, InstructionCost ReuseShuffleCost, 1832 InstructionCost VecCost, 1833 InstructionCost ScalarCost) const { 1834 dbgs() << "SLP: Calculated costs for Tree:\n"; E->dump(); 1835 dbgs() << "SLP: Costs:\n"; 1836 dbgs() << "SLP: ReuseShuffleCost = " << ReuseShuffleCost << "\n"; 1837 dbgs() << "SLP: VectorCost = " << VecCost << "\n"; 1838 dbgs() << "SLP: ScalarCost = " << ScalarCost << "\n"; 1839 dbgs() << "SLP: ReuseShuffleCost + VecCost - ScalarCost = " << 1840 ReuseShuffleCost + VecCost - ScalarCost << "\n"; 1841 } 1842#endif 1843 1844 /// Create a new VectorizableTree entry. 1845 TreeEntry *newTreeEntry(ArrayRef<Value *> VL, Optional<ScheduleData *> Bundle, 1846 const InstructionsState &S, 1847 const EdgeInfo &UserTreeIdx, 1848 ArrayRef<unsigned> ReuseShuffleIndices = None, 1849 ArrayRef<unsigned> ReorderIndices = None) { 1850 TreeEntry::EntryState EntryState = 1851 Bundle ? TreeEntry::Vectorize : TreeEntry::NeedToGather; 1852 return newTreeEntry(VL, EntryState, Bundle, S, UserTreeIdx, 1853 ReuseShuffleIndices, ReorderIndices); 1854 } 1855 1856 TreeEntry *newTreeEntry(ArrayRef<Value *> VL, 1857 TreeEntry::EntryState EntryState, 1858 Optional<ScheduleData *> Bundle, 1859 const InstructionsState &S, 1860 const EdgeInfo &UserTreeIdx, 1861 ArrayRef<unsigned> ReuseShuffleIndices = None, 1862 ArrayRef<unsigned> ReorderIndices = None) { 1863 assert(((!Bundle && EntryState == TreeEntry::NeedToGather) ||(static_cast <bool> (((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && "Need to vectorize gather entry?" ) ? void (0) : __assert_fail ("((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && \"Need to vectorize gather entry?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1865, __extension__ __PRETTY_FUNCTION__)) 1864 (Bundle && EntryState != TreeEntry::NeedToGather)) &&(static_cast <bool> (((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && "Need to vectorize gather entry?" ) ? void (0) : __assert_fail ("((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && \"Need to vectorize gather entry?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1865, __extension__ __PRETTY_FUNCTION__)) 1865 "Need to vectorize gather entry?")(static_cast <bool> (((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && "Need to vectorize gather entry?" ) ? void (0) : __assert_fail ("((!Bundle && EntryState == TreeEntry::NeedToGather) || (Bundle && EntryState != TreeEntry::NeedToGather)) && \"Need to vectorize gather entry?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1865, __extension__ __PRETTY_FUNCTION__)) ; 1866 VectorizableTree.push_back(std::make_unique<TreeEntry>(VectorizableTree)); 1867 TreeEntry *Last = VectorizableTree.back().get(); 1868 Last->Idx = VectorizableTree.size() - 1; 1869 Last->Scalars.insert(Last->Scalars.begin(), VL.begin(), VL.end()); 1870 Last->State = EntryState; 1871 Last->ReuseShuffleIndices.append(ReuseShuffleIndices.begin(), 1872 ReuseShuffleIndices.end()); 1873 Last->ReorderIndices.append(ReorderIndices.begin(), ReorderIndices.end()); 1874 Last->setOperations(S); 1875 if (Last->State != TreeEntry::NeedToGather) { 1876 for (Value *V : VL) { 1877 assert(!getTreeEntry(V) && "Scalar already in tree!")(static_cast <bool> (!getTreeEntry(V) && "Scalar already in tree!" ) ? void (0) : __assert_fail ("!getTreeEntry(V) && \"Scalar already in tree!\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1877, __extension__ __PRETTY_FUNCTION__)) ; 1878 ScalarToTreeEntry[V] = Last; 1879 } 1880 // Update the scheduler bundle to point to this TreeEntry. 1881 unsigned Lane = 0; 1882 for (ScheduleData *BundleMember = Bundle.getValue(); BundleMember; 1883 BundleMember = BundleMember->NextInBundle) { 1884 BundleMember->TE = Last; 1885 BundleMember->Lane = Lane; 1886 ++Lane; 1887 } 1888 assert((!Bundle.getValue() || Lane == VL.size()) &&(static_cast <bool> ((!Bundle.getValue() || Lane == VL. size()) && "Bundle and VL out of sync") ? void (0) : __assert_fail ("(!Bundle.getValue() || Lane == VL.size()) && \"Bundle and VL out of sync\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1889, __extension__ __PRETTY_FUNCTION__)) 1889 "Bundle and VL out of sync")(static_cast <bool> ((!Bundle.getValue() || Lane == VL. size()) && "Bundle and VL out of sync") ? void (0) : __assert_fail ("(!Bundle.getValue() || Lane == VL.size()) && \"Bundle and VL out of sync\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 1889, __extension__ __PRETTY_FUNCTION__)) ; 1890 } else { 1891 MustGather.insert(VL.begin(), VL.end()); 1892 } 1893 1894 if (UserTreeIdx.UserTE) 1895 Last->UserTreeIndices.push_back(UserTreeIdx); 1896 1897 return Last; 1898 } 1899 1900 /// -- Vectorization State -- 1901 /// Holds all of the tree entries. 1902 TreeEntry::VecTreeTy VectorizableTree; 1903 1904#ifndef NDEBUG 1905 /// Debug printer. 1906 LLVM_DUMP_METHOD__attribute__((noinline)) __attribute__((__used__)) void dumpVectorizableTree() const { 1907 for (unsigned Id = 0, IdE = VectorizableTree.size(); Id != IdE; ++Id) { 1908 VectorizableTree[Id]->dump(); 1909 dbgs() << "\n"; 1910 } 1911 } 1912#endif 1913 1914 TreeEntry *getTreeEntry(Value *V) { return ScalarToTreeEntry.lookup(V); } 1915 1916 const TreeEntry *getTreeEntry(Value *V) const { 1917 return ScalarToTreeEntry.lookup(V); 1918 } 1919 1920 /// Maps a specific scalar to its tree entry. 1921 SmallDenseMap<Value*, TreeEntry *> ScalarToTreeEntry; 1922 1923 /// Maps a value to the proposed vectorizable size. 1924 SmallDenseMap<Value *, unsigned> InstrElementSize; 1925 1926 /// A list of scalars that we found that we need to keep as scalars. 1927 ValueSet MustGather; 1928 1929 /// This POD struct describes one external user in the vectorized tree. 1930 struct ExternalUser { 1931 ExternalUser(Value *S, llvm::User *U, int L) 1932 : Scalar(S), User(U), Lane(L) {} 1933 1934 // Which scalar in our function. 1935 Value *Scalar; 1936 1937 // Which user that uses the scalar. 1938 llvm::User *User; 1939 1940 // Which lane does the scalar belong to. 1941 int Lane; 1942 }; 1943 using UserList = SmallVector<ExternalUser, 16>; 1944 1945 /// Checks if two instructions may access the same memory. 1946 /// 1947 /// \p Loc1 is the location of \p Inst1. It is passed explicitly because it 1948 /// is invariant in the calling loop. 1949 bool isAliased(const MemoryLocation &Loc1, Instruction *Inst1, 1950 Instruction *Inst2) { 1951 // First check if the result is already in the cache. 1952 AliasCacheKey key = std::make_pair(Inst1, Inst2); 1953 Optional<bool> &result = AliasCache[key]; 1954 if (result.hasValue()) { 1955 return result.getValue(); 1956 } 1957 MemoryLocation Loc2 = getLocation(Inst2, AA); 1958 bool aliased = true; 1959 if (Loc1.Ptr && Loc2.Ptr && isSimple(Inst1) && isSimple(Inst2)) { 1960 // Do the alias check. 1961 aliased = !AA->isNoAlias(Loc1, Loc2); 1962 } 1963 // Store the result in the cache. 1964 result = aliased; 1965 return aliased; 1966 } 1967 1968 using AliasCacheKey = std::pair<Instruction *, Instruction *>; 1969 1970 /// Cache for alias results. 1971 /// TODO: consider moving this to the AliasAnalysis itself. 1972 DenseMap<AliasCacheKey, Optional<bool>> AliasCache; 1973 1974 /// Removes an instruction from its block and eventually deletes it. 1975 /// It's like Instruction::eraseFromParent() except that the actual deletion 1976 /// is delayed until BoUpSLP is destructed. 1977 /// This is required to ensure that there are no incorrect collisions in the 1978 /// AliasCache, which can happen if a new instruction is allocated at the 1979 /// same address as a previously deleted instruction. 1980 void eraseInstruction(Instruction *I, bool ReplaceOpsWithUndef = false) { 1981 auto It = DeletedInstructions.try_emplace(I, ReplaceOpsWithUndef).first; 1982 It->getSecond() = It->getSecond() && ReplaceOpsWithUndef; 1983 } 1984 1985 /// Temporary store for deleted instructions. Instructions will be deleted 1986 /// eventually when the BoUpSLP is destructed. 1987 DenseMap<Instruction *, bool> DeletedInstructions; 1988 1989 /// A list of values that need to extracted out of the tree. 1990 /// This list holds pairs of (Internal Scalar : External User). External User 1991 /// can be nullptr, it means that this Internal Scalar will be used later, 1992 /// after vectorization. 1993 UserList ExternalUses; 1994 1995 /// Values used only by @llvm.assume calls. 1996 SmallPtrSet<const Value *, 32> EphValues; 1997 1998 /// Holds all of the instructions that we gathered. 1999 SetVector<Instruction *> GatherSeq; 2000 2001 /// A list of blocks that we are going to CSE. 2002 SetVector<BasicBlock *> CSEBlocks; 2003 2004 /// Contains all scheduling relevant data for an instruction. 2005 /// A ScheduleData either represents a single instruction or a member of an 2006 /// instruction bundle (= a group of instructions which is combined into a 2007 /// vector instruction). 2008 struct ScheduleData { 2009 // The initial value for the dependency counters. It means that the 2010 // dependencies are not calculated yet. 2011 enum { InvalidDeps = -1 }; 2012 2013 ScheduleData() = default; 2014 2015 void init(int BlockSchedulingRegionID, Value *OpVal) { 2016 FirstInBundle = this; 2017 NextInBundle = nullptr; 2018 NextLoadStore = nullptr; 2019 IsScheduled = false; 2020 SchedulingRegionID = BlockSchedulingRegionID; 2021 UnscheduledDepsInBundle = UnscheduledDeps; 2022 clearDependencies(); 2023 OpValue = OpVal; 2024 TE = nullptr; 2025 Lane = -1; 2026 } 2027 2028 /// Returns true if the dependency information has been calculated. 2029 bool hasValidDependencies() const { return Dependencies != InvalidDeps; } 2030 2031 /// Returns true for single instructions and for bundle representatives 2032 /// (= the head of a bundle). 2033 bool isSchedulingEntity() const { return FirstInBundle == this; } 2034 2035 /// Returns true if it represents an instruction bundle and not only a 2036 /// single instruction. 2037 bool isPartOfBundle() const { 2038 return NextInBundle != nullptr || FirstInBundle != this; 2039 } 2040 2041 /// Returns true if it is ready for scheduling, i.e. it has no more 2042 /// unscheduled depending instructions/bundles. 2043 bool isReady() const { 2044 assert(isSchedulingEntity() &&(static_cast <bool> (isSchedulingEntity() && "can't consider non-scheduling entity for ready list" ) ? void (0) : __assert_fail ("isSchedulingEntity() && \"can't consider non-scheduling entity for ready list\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2045, __extension__ __PRETTY_FUNCTION__)) 2045 "can't consider non-scheduling entity for ready list")(static_cast <bool> (isSchedulingEntity() && "can't consider non-scheduling entity for ready list" ) ? void (0) : __assert_fail ("isSchedulingEntity() && \"can't consider non-scheduling entity for ready list\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2045, __extension__ __PRETTY_FUNCTION__)) ; 2046 return UnscheduledDepsInBundle == 0 && !IsScheduled; 2047 } 2048 2049 /// Modifies the number of unscheduled dependencies, also updating it for 2050 /// the whole bundle. 2051 int incrementUnscheduledDeps(int Incr) { 2052 UnscheduledDeps += Incr; 2053 return FirstInBundle->UnscheduledDepsInBundle += Incr; 2054 } 2055 2056 /// Sets the number of unscheduled dependencies to the number of 2057 /// dependencies. 2058 void resetUnscheduledDeps() { 2059 incrementUnscheduledDeps(Dependencies - UnscheduledDeps); 2060 } 2061 2062 /// Clears all dependency information. 2063 void clearDependencies() { 2064 Dependencies = InvalidDeps; 2065 resetUnscheduledDeps(); 2066 MemoryDependencies.clear(); 2067 } 2068 2069 void dump(raw_ostream &os) const { 2070 if (!isSchedulingEntity()) { 2071 os << "/ " << *Inst; 2072 } else if (NextInBundle) { 2073 os << '[' << *Inst; 2074 ScheduleData *SD = NextInBundle; 2075 while (SD) { 2076 os << ';' << *SD->Inst; 2077 SD = SD->NextInBundle; 2078 } 2079 os << ']'; 2080 } else { 2081 os << *Inst; 2082 } 2083 } 2084 2085 Instruction *Inst = nullptr; 2086 2087 /// Points to the head in an instruction bundle (and always to this for 2088 /// single instructions). 2089 ScheduleData *FirstInBundle = nullptr; 2090 2091 /// Single linked list of all instructions in a bundle. Null if it is a 2092 /// single instruction. 2093 ScheduleData *NextInBundle = nullptr; 2094 2095 /// Single linked list of all memory instructions (e.g. load, store, call) 2096 /// in the block - until the end of the scheduling region. 2097 ScheduleData *NextLoadStore = nullptr; 2098 2099 /// The dependent memory instructions. 2100 /// This list is derived on demand in calculateDependencies(). 2101 SmallVector<ScheduleData *, 4> MemoryDependencies; 2102 2103 /// This ScheduleData is in the current scheduling region if this matches 2104 /// the current SchedulingRegionID of BlockScheduling. 2105 int SchedulingRegionID = 0; 2106 2107 /// Used for getting a "good" final ordering of instructions. 2108 int SchedulingPriority = 0; 2109 2110 /// The number of dependencies. Constitutes of the number of users of the 2111 /// instruction plus the number of dependent memory instructions (if any). 2112 /// This value is calculated on demand. 2113 /// If InvalidDeps, the number of dependencies is not calculated yet. 2114 int Dependencies = InvalidDeps; 2115 2116 /// The number of dependencies minus the number of dependencies of scheduled 2117 /// instructions. As soon as this is zero, the instruction/bundle gets ready 2118 /// for scheduling. 2119 /// Note that this is negative as long as Dependencies is not calculated. 2120 int UnscheduledDeps = InvalidDeps; 2121 2122 /// The sum of UnscheduledDeps in a bundle. Equals to UnscheduledDeps for 2123 /// single instructions. 2124 int UnscheduledDepsInBundle = InvalidDeps; 2125 2126 /// True if this instruction is scheduled (or considered as scheduled in the 2127 /// dry-run). 2128 bool IsScheduled = false; 2129 2130 /// Opcode of the current instruction in the schedule data. 2131 Value *OpValue = nullptr; 2132 2133 /// The TreeEntry that this instruction corresponds to. 2134 TreeEntry *TE = nullptr; 2135 2136 /// The lane of this node in the TreeEntry. 2137 int Lane = -1; 2138 }; 2139 2140#ifndef NDEBUG 2141 friend inline raw_ostream &operator<<(raw_ostream &os, 2142 const BoUpSLP::ScheduleData &SD) { 2143 SD.dump(os); 2144 return os; 2145 } 2146#endif 2147 2148 friend struct GraphTraits<BoUpSLP *>; 2149 friend struct DOTGraphTraits<BoUpSLP *>; 2150 2151 /// Contains all scheduling data for a basic block. 2152 struct BlockScheduling { 2153 BlockScheduling(BasicBlock *BB) 2154 : BB(BB), ChunkSize(BB->size()), ChunkPos(ChunkSize) {} 2155 2156 void clear() { 2157 ReadyInsts.clear(); 2158 ScheduleStart = nullptr; 2159 ScheduleEnd = nullptr; 2160 FirstLoadStoreInRegion = nullptr; 2161 LastLoadStoreInRegion = nullptr; 2162 2163 // Reduce the maximum schedule region size by the size of the 2164 // previous scheduling run. 2165 ScheduleRegionSizeLimit -= ScheduleRegionSize; 2166 if (ScheduleRegionSizeLimit < MinScheduleRegionSize) 2167 ScheduleRegionSizeLimit = MinScheduleRegionSize; 2168 ScheduleRegionSize = 0; 2169 2170 // Make a new scheduling region, i.e. all existing ScheduleData is not 2171 // in the new region yet. 2172 ++SchedulingRegionID; 2173 } 2174 2175 ScheduleData *getScheduleData(Value *V) { 2176 ScheduleData *SD = ScheduleDataMap[V]; 2177 if (SD && SD->SchedulingRegionID == SchedulingRegionID) 2178 return SD; 2179 return nullptr; 2180 } 2181 2182 ScheduleData *getScheduleData(Value *V, Value *Key) { 2183 if (V == Key) 2184 return getScheduleData(V); 2185 auto I = ExtraScheduleDataMap.find(V); 2186 if (I != ExtraScheduleDataMap.end()) { 2187 ScheduleData *SD = I->second[Key]; 2188 if (SD && SD->SchedulingRegionID == SchedulingRegionID) 2189 return SD; 2190 } 2191 return nullptr; 2192 } 2193 2194 bool isInSchedulingRegion(ScheduleData *SD) const { 2195 return SD->SchedulingRegionID == SchedulingRegionID; 2196 } 2197 2198 /// Marks an instruction as scheduled and puts all dependent ready 2199 /// instructions into the ready-list. 2200 template <typename ReadyListType> 2201 void schedule(ScheduleData *SD, ReadyListType &ReadyList) { 2202 SD->IsScheduled = true; 2203 LLVM_DEBUG(dbgs() << "SLP: schedule " << *SD << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: schedule " << *SD << "\n"; } } while (false) ; 2204 2205 ScheduleData *BundleMember = SD; 2206 while (BundleMember) { 2207 if (BundleMember->Inst != BundleMember->OpValue) { 2208 BundleMember = BundleMember->NextInBundle; 2209 continue; 2210 } 2211 // Handle the def-use chain dependencies. 2212 2213 // Decrement the unscheduled counter and insert to ready list if ready. 2214 auto &&DecrUnsched = [this, &ReadyList](Instruction *I) { 2215 doForAllOpcodes(I, [&ReadyList](ScheduleData *OpDef) { 2216 if (OpDef && OpDef->hasValidDependencies() && 2217 OpDef->incrementUnscheduledDeps(-1) == 0) { 2218 // There are no more unscheduled dependencies after 2219 // decrementing, so we can put the dependent instruction 2220 // into the ready list. 2221 ScheduleData *DepBundle = OpDef->FirstInBundle; 2222 assert(!DepBundle->IsScheduled &&(static_cast <bool> (!DepBundle->IsScheduled && "already scheduled bundle gets ready") ? void (0) : __assert_fail ("!DepBundle->IsScheduled && \"already scheduled bundle gets ready\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2223, __extension__ __PRETTY_FUNCTION__)) 2223 "already scheduled bundle gets ready")(static_cast <bool> (!DepBundle->IsScheduled && "already scheduled bundle gets ready") ? void (0) : __assert_fail ("!DepBundle->IsScheduled && \"already scheduled bundle gets ready\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2223, __extension__ __PRETTY_FUNCTION__)) ; 2224 ReadyList.insert(DepBundle); 2225 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: gets ready (def): " << *DepBundle << "\n"; } } while (false) 2226 << "SLP: gets ready (def): " << *DepBundle << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: gets ready (def): " << *DepBundle << "\n"; } } while (false) ; 2227 } 2228 }); 2229 }; 2230 2231 // If BundleMember is a vector bundle, its operands may have been 2232 // reordered duiring buildTree(). We therefore need to get its operands 2233 // through the TreeEntry. 2234 if (TreeEntry *TE = BundleMember->TE) { 2235 int Lane = BundleMember->Lane; 2236 assert(Lane >= 0 && "Lane not set")(static_cast <bool> (Lane >= 0 && "Lane not set" ) ? void (0) : __assert_fail ("Lane >= 0 && \"Lane not set\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2236, __extension__ __PRETTY_FUNCTION__)) ; 2237 2238 // Since vectorization tree is being built recursively this assertion 2239 // ensures that the tree entry has all operands set before reaching 2240 // this code. Couple of exceptions known at the moment are extracts 2241 // where their second (immediate) operand is not added. Since 2242 // immediates do not affect scheduler behavior this is considered 2243 // okay. 2244 auto *In = TE->getMainOp(); 2245 assert(In &&(static_cast <bool> (In && (isa<ExtractValueInst >(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst >(In) || In->getNumOperands() == TE->getNumOperands( )) && "Missed TreeEntry operands?") ? void (0) : __assert_fail ("In && (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst>(In) || In->getNumOperands() == TE->getNumOperands()) && \"Missed TreeEntry operands?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2249, __extension__ __PRETTY_FUNCTION__)) 2246 (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) ||(static_cast <bool> (In && (isa<ExtractValueInst >(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst >(In) || In->getNumOperands() == TE->getNumOperands( )) && "Missed TreeEntry operands?") ? void (0) : __assert_fail ("In && (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst>(In) || In->getNumOperands() == TE->getNumOperands()) && \"Missed TreeEntry operands?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2249, __extension__ __PRETTY_FUNCTION__)) 2247 isa<InsertElementInst>(In) ||(static_cast <bool> (In && (isa<ExtractValueInst >(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst >(In) || In->getNumOperands() == TE->getNumOperands( )) && "Missed TreeEntry operands?") ? void (0) : __assert_fail ("In && (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst>(In) || In->getNumOperands() == TE->getNumOperands()) && \"Missed TreeEntry operands?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2249, __extension__ __PRETTY_FUNCTION__)) 2248 In->getNumOperands() == TE->getNumOperands()) &&(static_cast <bool> (In && (isa<ExtractValueInst >(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst >(In) || In->getNumOperands() == TE->getNumOperands( )) && "Missed TreeEntry operands?") ? void (0) : __assert_fail ("In && (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst>(In) || In->getNumOperands() == TE->getNumOperands()) && \"Missed TreeEntry operands?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2249, __extension__ __PRETTY_FUNCTION__)) 2249 "Missed TreeEntry operands?")(static_cast <bool> (In && (isa<ExtractValueInst >(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst >(In) || In->getNumOperands() == TE->getNumOperands( )) && "Missed TreeEntry operands?") ? void (0) : __assert_fail ("In && (isa<ExtractValueInst>(In) || isa<ExtractElementInst>(In) || isa<InsertElementInst>(In) || In->getNumOperands() == TE->getNumOperands()) && \"Missed TreeEntry operands?\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2249, __extension__ __PRETTY_FUNCTION__)) ; 2250 (void)In; // fake use to avoid build failure when assertions disabled 2251 2252 for (unsigned OpIdx = 0, NumOperands = TE->getNumOperands(); 2253 OpIdx != NumOperands; ++OpIdx) 2254 if (auto *I = dyn_cast<Instruction>(TE->getOperand(OpIdx)[Lane])) 2255 DecrUnsched(I); 2256 } else { 2257 // If BundleMember is a stand-alone instruction, no operand reordering 2258 // has taken place, so we directly access its operands. 2259 for (Use &U : BundleMember->Inst->operands()) 2260 if (auto *I = dyn_cast<Instruction>(U.get())) 2261 DecrUnsched(I); 2262 } 2263 // Handle the memory dependencies. 2264 for (ScheduleData *MemoryDepSD : BundleMember->MemoryDependencies) { 2265 if (MemoryDepSD->incrementUnscheduledDeps(-1) == 0) { 2266 // There are no more unscheduled dependencies after decrementing, 2267 // so we can put the dependent instruction into the ready list. 2268 ScheduleData *DepBundle = MemoryDepSD->FirstInBundle; 2269 assert(!DepBundle->IsScheduled &&(static_cast <bool> (!DepBundle->IsScheduled && "already scheduled bundle gets ready") ? void (0) : __assert_fail ("!DepBundle->IsScheduled && \"already scheduled bundle gets ready\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2270, __extension__ __PRETTY_FUNCTION__)) 2270 "already scheduled bundle gets ready")(static_cast <bool> (!DepBundle->IsScheduled && "already scheduled bundle gets ready") ? void (0) : __assert_fail ("!DepBundle->IsScheduled && \"already scheduled bundle gets ready\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2270, __extension__ __PRETTY_FUNCTION__)) ; 2271 ReadyList.insert(DepBundle); 2272 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: gets ready (mem): " << *DepBundle << "\n"; } } while (false) 2273 << "SLP: gets ready (mem): " << *DepBundle << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: gets ready (mem): " << *DepBundle << "\n"; } } while (false) ; 2274 } 2275 } 2276 BundleMember = BundleMember->NextInBundle; 2277 } 2278 } 2279 2280 void doForAllOpcodes(Value *V, 2281 function_ref<void(ScheduleData *SD)> Action) { 2282 if (ScheduleData *SD = getScheduleData(V)) 2283 Action(SD); 2284 auto I = ExtraScheduleDataMap.find(V); 2285 if (I != ExtraScheduleDataMap.end()) 2286 for (auto &P : I->second) 2287 if (P.second->SchedulingRegionID == SchedulingRegionID) 2288 Action(P.second); 2289 } 2290 2291 /// Put all instructions into the ReadyList which are ready for scheduling. 2292 template <typename ReadyListType> 2293 void initialFillReadyList(ReadyListType &ReadyList) { 2294 for (auto *I = ScheduleStart; I != ScheduleEnd; I = I->getNextNode()) { 2295 doForAllOpcodes(I, [&](ScheduleData *SD) { 2296 if (SD->isSchedulingEntity() && SD->isReady()) { 2297 ReadyList.insert(SD); 2298 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: initially in ready list: " << *I << "\n"; } } while (false) 2299 << "SLP: initially in ready list: " << *I << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: initially in ready list: " << *I << "\n"; } } while (false) ; 2300 } 2301 }); 2302 } 2303 } 2304 2305 /// Checks if a bundle of instructions can be scheduled, i.e. has no 2306 /// cyclic dependencies. This is only a dry-run, no instructions are 2307 /// actually moved at this stage. 2308 /// \returns the scheduling bundle. The returned Optional value is non-None 2309 /// if \p VL is allowed to be scheduled. 2310 Optional<ScheduleData *> 2311 tryScheduleBundle(ArrayRef<Value *> VL, BoUpSLP *SLP, 2312 const InstructionsState &S); 2313 2314 /// Un-bundles a group of instructions. 2315 void cancelScheduling(ArrayRef<Value *> VL, Value *OpValue); 2316 2317 /// Allocates schedule data chunk. 2318 ScheduleData *allocateScheduleDataChunks(); 2319 2320 /// Extends the scheduling region so that V is inside the region. 2321 /// \returns true if the region size is within the limit. 2322 bool extendSchedulingRegion(Value *V, const InstructionsState &S); 2323 2324 /// Initialize the ScheduleData structures for new instructions in the 2325 /// scheduling region. 2326 void initScheduleData(Instruction *FromI, Instruction *ToI, 2327 ScheduleData *PrevLoadStore, 2328 ScheduleData *NextLoadStore); 2329 2330 /// Updates the dependency information of a bundle and of all instructions/ 2331 /// bundles which depend on the original bundle. 2332 void calculateDependencies(ScheduleData *SD, bool InsertInReadyList, 2333 BoUpSLP *SLP); 2334 2335 /// Sets all instruction in the scheduling region to un-scheduled. 2336 void resetSchedule(); 2337 2338 BasicBlock *BB; 2339 2340 /// Simple memory allocation for ScheduleData. 2341 std::vector<std::unique_ptr<ScheduleData[]>> ScheduleDataChunks; 2342 2343 /// The size of a ScheduleData array in ScheduleDataChunks. 2344 int ChunkSize; 2345 2346 /// The allocator position in the current chunk, which is the last entry 2347 /// of ScheduleDataChunks. 2348 int ChunkPos; 2349 2350 /// Attaches ScheduleData to Instruction. 2351 /// Note that the mapping survives during all vectorization iterations, i.e. 2352 /// ScheduleData structures are recycled. 2353 DenseMap<Value *, ScheduleData *> ScheduleDataMap; 2354 2355 /// Attaches ScheduleData to Instruction with the leading key. 2356 DenseMap<Value *, SmallDenseMap<Value *, ScheduleData *>> 2357 ExtraScheduleDataMap; 2358 2359 struct ReadyList : SmallVector<ScheduleData *, 8> { 2360 void insert(ScheduleData *SD) { push_back(SD); } 2361 }; 2362 2363 /// The ready-list for scheduling (only used for the dry-run). 2364 ReadyList ReadyInsts; 2365 2366 /// The first instruction of the scheduling region. 2367 Instruction *ScheduleStart = nullptr; 2368 2369 /// The first instruction _after_ the scheduling region. 2370 Instruction *ScheduleEnd = nullptr; 2371 2372 /// The first memory accessing instruction in the scheduling region 2373 /// (can be null). 2374 ScheduleData *FirstLoadStoreInRegion = nullptr; 2375 2376 /// The last memory accessing instruction in the scheduling region 2377 /// (can be null). 2378 ScheduleData *LastLoadStoreInRegion = nullptr; 2379 2380 /// The current size of the scheduling region. 2381 int ScheduleRegionSize = 0; 2382 2383 /// The maximum size allowed for the scheduling region. 2384 int ScheduleRegionSizeLimit = ScheduleRegionSizeBudget; 2385 2386 /// The ID of the scheduling region. For a new vectorization iteration this 2387 /// is incremented which "removes" all ScheduleData from the region. 2388 // Make sure that the initial SchedulingRegionID is greater than the 2389 // initial SchedulingRegionID in ScheduleData (which is 0). 2390 int SchedulingRegionID = 1; 2391 }; 2392 2393 /// Attaches the BlockScheduling structures to basic blocks. 2394 MapVector<BasicBlock *, std::unique_ptr<BlockScheduling>> BlocksSchedules; 2395 2396 /// Performs the "real" scheduling. Done before vectorization is actually 2397 /// performed in a basic block. 2398 void scheduleBlock(BlockScheduling *BS); 2399 2400 /// List of users to ignore during scheduling and that don't need extracting. 2401 ArrayRef<Value *> UserIgnoreList; 2402 2403 /// A DenseMapInfo implementation for holding DenseMaps and DenseSets of 2404 /// sorted SmallVectors of unsigned. 2405 struct OrdersTypeDenseMapInfo { 2406 static OrdersType getEmptyKey() { 2407 OrdersType V; 2408 V.push_back(~1U); 2409 return V; 2410 } 2411 2412 static OrdersType getTombstoneKey() { 2413 OrdersType V; 2414 V.push_back(~2U); 2415 return V; 2416 } 2417 2418 static unsigned getHashValue(const OrdersType &V) { 2419 return static_cast<unsigned>(hash_combine_range(V.begin(), V.end())); 2420 } 2421 2422 static bool isEqual(const OrdersType &LHS, const OrdersType &RHS) { 2423 return LHS == RHS; 2424 } 2425 }; 2426 2427 /// Contains orders of operations along with the number of bundles that have 2428 /// operations in this order. It stores only those orders that require 2429 /// reordering, if reordering is not required it is counted using \a 2430 /// NumOpsWantToKeepOriginalOrder. 2431 DenseMap<OrdersType, unsigned, OrdersTypeDenseMapInfo> NumOpsWantToKeepOrder; 2432 /// Number of bundles that do not require reordering. 2433 unsigned NumOpsWantToKeepOriginalOrder = 0; 2434 2435 // Analysis and block reference. 2436 Function *F; 2437 ScalarEvolution *SE; 2438 TargetTransformInfo *TTI; 2439 TargetLibraryInfo *TLI; 2440 AAResults *AA; 2441 LoopInfo *LI; 2442 DominatorTree *DT; 2443 AssumptionCache *AC; 2444 DemandedBits *DB; 2445 const DataLayout *DL; 2446 OptimizationRemarkEmitter *ORE; 2447 2448 unsigned MaxVecRegSize; // This is set by TTI or overridden by cl::opt. 2449 unsigned MinVecRegSize; // Set by cl::opt (default: 128). 2450 2451 /// Instruction builder to construct the vectorized tree. 2452 IRBuilder<> Builder; 2453 2454 /// A map of scalar integer values to the smallest bit width with which they 2455 /// can legally be represented. The values map to (width, signed) pairs, 2456 /// where "width" indicates the minimum bit width and "signed" is True if the 2457 /// value must be signed-extended, rather than zero-extended, back to its 2458 /// original width. 2459 MapVector<Value *, std::pair<uint64_t, bool>> MinBWs; 2460}; 2461 2462} // end namespace slpvectorizer 2463 2464template <> struct GraphTraits<BoUpSLP *> { 2465 using TreeEntry = BoUpSLP::TreeEntry; 2466 2467 /// NodeRef has to be a pointer per the GraphWriter. 2468 using NodeRef = TreeEntry *; 2469 2470 using ContainerTy = BoUpSLP::TreeEntry::VecTreeTy; 2471 2472 /// Add the VectorizableTree to the index iterator to be able to return 2473 /// TreeEntry pointers. 2474 struct ChildIteratorType 2475 : public iterator_adaptor_base< 2476 ChildIteratorType, SmallVector<BoUpSLP::EdgeInfo, 1>::iterator> { 2477 ContainerTy &VectorizableTree; 2478 2479 ChildIteratorType(SmallVector<BoUpSLP::EdgeInfo, 1>::iterator W, 2480 ContainerTy &VT) 2481 : ChildIteratorType::iterator_adaptor_base(W), VectorizableTree(VT) {} 2482 2483 NodeRef operator*() { return I->UserTE; } 2484 }; 2485 2486 static NodeRef getEntryNode(BoUpSLP &R) { 2487 return R.VectorizableTree[0].get(); 2488 } 2489 2490 static ChildIteratorType child_begin(NodeRef N) { 2491 return {N->UserTreeIndices.begin(), N->Container}; 2492 } 2493 2494 static ChildIteratorType child_end(NodeRef N) { 2495 return {N->UserTreeIndices.end(), N->Container}; 2496 } 2497 2498 /// For the node iterator we just need to turn the TreeEntry iterator into a 2499 /// TreeEntry* iterator so that it dereferences to NodeRef. 2500 class nodes_iterator { 2501 using ItTy = ContainerTy::iterator; 2502 ItTy It; 2503 2504 public: 2505 nodes_iterator(const ItTy &It2) : It(It2) {} 2506 NodeRef operator*() { return It->get(); } 2507 nodes_iterator operator++() { 2508 ++It; 2509 return *this; 2510 } 2511 bool operator!=(const nodes_iterator &N2) const { return N2.It != It; } 2512 }; 2513 2514 static nodes_iterator nodes_begin(BoUpSLP *R) { 2515 return nodes_iterator(R->VectorizableTree.begin()); 2516 } 2517 2518 static nodes_iterator nodes_end(BoUpSLP *R) { 2519 return nodes_iterator(R->VectorizableTree.end()); 2520 } 2521 2522 static unsigned size(BoUpSLP *R) { return R->VectorizableTree.size(); } 2523}; 2524 2525template <> struct DOTGraphTraits<BoUpSLP *> : public DefaultDOTGraphTraits { 2526 using TreeEntry = BoUpSLP::TreeEntry; 2527 2528 DOTGraphTraits(bool isSimple = false) : DefaultDOTGraphTraits(isSimple) {} 2529 2530 std::string getNodeLabel(const TreeEntry *Entry, const BoUpSLP *R) { 2531 std::string Str; 2532 raw_string_ostream OS(Str); 2533 if (isSplat(Entry->Scalars)) { 2534 OS << "<splat> " << *Entry->Scalars[0]; 2535 return Str; 2536 } 2537 for (auto V : Entry->Scalars) { 2538 OS << *V; 2539 if (llvm::any_of(R->ExternalUses, [&](const BoUpSLP::ExternalUser &EU) { 2540 return EU.Scalar == V; 2541 })) 2542 OS << " <extract>"; 2543 OS << "\n"; 2544 } 2545 return Str; 2546 } 2547 2548 static std::string getNodeAttributes(const TreeEntry *Entry, 2549 const BoUpSLP *) { 2550 if (Entry->State == TreeEntry::NeedToGather) 2551 return "color=red"; 2552 return ""; 2553 } 2554}; 2555 2556} // end namespace llvm 2557 2558BoUpSLP::~BoUpSLP() { 2559 for (const auto &Pair : DeletedInstructions) { 2560 // Replace operands of ignored instructions with Undefs in case if they were 2561 // marked for deletion. 2562 if (Pair.getSecond()) { 2563 Value *Undef = UndefValue::get(Pair.getFirst()->getType()); 2564 Pair.getFirst()->replaceAllUsesWith(Undef); 2565 } 2566 Pair.getFirst()->dropAllReferences(); 2567 } 2568 for (const auto &Pair : DeletedInstructions) { 2569 assert(Pair.getFirst()->use_empty() &&(static_cast <bool> (Pair.getFirst()->use_empty() && "trying to erase instruction with users.") ? void (0) : __assert_fail ("Pair.getFirst()->use_empty() && \"trying to erase instruction with users.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2570, __extension__ __PRETTY_FUNCTION__)) 2570 "trying to erase instruction with users.")(static_cast <bool> (Pair.getFirst()->use_empty() && "trying to erase instruction with users.") ? void (0) : __assert_fail ("Pair.getFirst()->use_empty() && \"trying to erase instruction with users.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2570, __extension__ __PRETTY_FUNCTION__)) ; 2571 Pair.getFirst()->eraseFromParent(); 2572 } 2573#ifdef EXPENSIVE_CHECKS 2574 // If we could guarantee that this call is not extremely slow, we could 2575 // remove the ifdef limitation (see PR47712). 2576 assert(!verifyFunction(*F, &dbgs()))(static_cast <bool> (!verifyFunction(*F, &dbgs())) ? void (0) : __assert_fail ("!verifyFunction(*F, &dbgs())" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2576, __extension__ __PRETTY_FUNCTION__)) ; 2577#endif 2578} 2579 2580void BoUpSLP::eraseInstructions(ArrayRef<Value *> AV) { 2581 for (auto *V : AV) { 2582 if (auto *I = dyn_cast<Instruction>(V)) 2583 eraseInstruction(I, /*ReplaceOpsWithUndef=*/true); 2584 }; 2585} 2586 2587void BoUpSLP::buildTree(ArrayRef<Value *> Roots, 2588 ArrayRef<Value *> UserIgnoreLst) { 2589 ExtraValueToDebugLocsMap ExternallyUsedValues; 2590 buildTree(Roots, ExternallyUsedValues, UserIgnoreLst); 2591} 2592 2593static int findLaneForValue(ArrayRef<Value *> Scalars, 2594 ArrayRef<int> ReuseShuffleIndices, Value *V) { 2595 unsigned FoundLane = std::distance(Scalars.begin(), find(Scalars, V)); 2596 assert(FoundLane < Scalars.size() && "Couldn't find extract lane")(static_cast <bool> (FoundLane < Scalars.size() && "Couldn't find extract lane") ? void (0) : __assert_fail ("FoundLane < Scalars.size() && \"Couldn't find extract lane\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2596, __extension__ __PRETTY_FUNCTION__)) ; 2597 if (!ReuseShuffleIndices.empty()) { 2598 FoundLane = std::distance(ReuseShuffleIndices.begin(), 2599 find(ReuseShuffleIndices, FoundLane)); 2600 } 2601 return FoundLane; 2602} 2603 2604void BoUpSLP::buildTree(ArrayRef<Value *> Roots, 2605 ExtraValueToDebugLocsMap &ExternallyUsedValues, 2606 ArrayRef<Value *> UserIgnoreLst) { 2607 deleteTree(); 2608 UserIgnoreList = UserIgnoreLst; 2609 if (!allSameType(Roots)) 2610 return; 2611 buildTree_rec(Roots, 0, EdgeInfo()); 2612 2613 // Collect the values that we need to extract from the tree. 2614 for (auto &TEPtr : VectorizableTree) { 2615 TreeEntry *Entry = TEPtr.get(); 2616 2617 // No need to handle users of gathered values. 2618 if (Entry->State == TreeEntry::NeedToGather) 2619 continue; 2620 2621 // For each lane: 2622 for (int Lane = 0, LE = Entry->Scalars.size(); Lane != LE; ++Lane) { 2623 Value *Scalar = Entry->Scalars[Lane]; 2624 int FoundLane = 2625 findLaneForValue(Entry->Scalars, Entry->ReuseShuffleIndices, Scalar); 2626 2627 // Check if the scalar is externally used as an extra arg. 2628 auto ExtI = ExternallyUsedValues.find(Scalar); 2629 if (ExtI != ExternallyUsedValues.end()) { 2630 LLVM_DEBUG(dbgs() << "SLP: Need to extract: Extra arg from lane "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to extract: Extra arg from lane " << Lane << " from " << *Scalar << ".\n" ; } } while (false) 2631 << Lane << " from " << *Scalar << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to extract: Extra arg from lane " << Lane << " from " << *Scalar << ".\n" ; } } while (false) ; 2632 ExternalUses.emplace_back(Scalar, nullptr, FoundLane); 2633 } 2634 for (User *U : Scalar->users()) { 2635 LLVM_DEBUG(dbgs() << "SLP: Checking user:" << *U << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Checking user:" << *U << ".\n"; } } while (false) ; 2636 2637 Instruction *UserInst = dyn_cast<Instruction>(U); 2638 if (!UserInst) 2639 continue; 2640 2641 // Skip in-tree scalars that become vectors 2642 if (TreeEntry *UseEntry = getTreeEntry(U)) { 2643 Value *UseScalar = UseEntry->Scalars[0]; 2644 // Some in-tree scalars will remain as scalar in vectorized 2645 // instructions. If that is the case, the one in Lane 0 will 2646 // be used. 2647 if (UseScalar != U || 2648 UseEntry->State == TreeEntry::ScatterVectorize || 2649 !InTreeUserNeedToExtract(Scalar, UserInst, TLI)) { 2650 LLVM_DEBUG(dbgs() << "SLP: \tInternal user will be removed:" << *Udo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: \tInternal user will be removed:" << *U << ".\n"; } } while (false) 2651 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: \tInternal user will be removed:" << *U << ".\n"; } } while (false) ; 2652 assert(UseEntry->State != TreeEntry::NeedToGather && "Bad state")(static_cast <bool> (UseEntry->State != TreeEntry::NeedToGather && "Bad state") ? void (0) : __assert_fail ("UseEntry->State != TreeEntry::NeedToGather && \"Bad state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2652, __extension__ __PRETTY_FUNCTION__)) ; 2653 continue; 2654 } 2655 } 2656 2657 // Ignore users in the user ignore list. 2658 if (is_contained(UserIgnoreList, UserInst)) 2659 continue; 2660 2661 LLVM_DEBUG(dbgs() << "SLP: Need to extract:" << *U << " from lane "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to extract:" << * U << " from lane " << Lane << " from " << *Scalar << ".\n"; } } while (false) 2662 << Lane << " from " << *Scalar << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to extract:" << * U << " from lane " << Lane << " from " << *Scalar << ".\n"; } } while (false) ; 2663 ExternalUses.push_back(ExternalUser(Scalar, U, FoundLane)); 2664 } 2665 } 2666 } 2667} 2668 2669void BoUpSLP::buildTree_rec(ArrayRef<Value *> VL, unsigned Depth, 2670 const EdgeInfo &UserTreeIdx) { 2671 assert((allConstant(VL) || allSameType(VL)) && "Invalid types!")(static_cast <bool> ((allConstant(VL) || allSameType(VL )) && "Invalid types!") ? void (0) : __assert_fail ("(allConstant(VL) || allSameType(VL)) && \"Invalid types!\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2671, __extension__ __PRETTY_FUNCTION__)) ; 1 Assuming the condition is true 2 '?' condition is true 2672 2673 InstructionsState S = getSameOpcode(VL); 2674 if (Depth == RecursionMaxDepth) { 3 Assuming the condition is false 4 Taking false branch 2675 LLVM_DEBUG(dbgs() << "SLP: Gathering due to max recursion depth.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to max recursion depth.\n" ; } } while (false) ; 2676 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2677 return; 2678 } 2679 2680 // Don't handle vectors. 2681 if (S.OpValue->getType()->isVectorTy() && 2682 !isa<InsertElementInst>(S.OpValue)) { 2683 LLVM_DEBUG(dbgs() << "SLP: Gathering due to vector type.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to vector type.\n" ; } } while (false) ; 2684 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2685 return; 2686 } 2687 2688 if (StoreInst *SI 5.1 'SI' is null 5.1 'SI' is null = dyn_cast<StoreInst>(S.OpValue)) 5 Assuming field 'OpValue' is not a 'StoreInst' 6 Taking false branch 2689 if (SI->getValueOperand()->getType()->isVectorTy()) { 2690 LLVM_DEBUG(dbgs() << "SLP: Gathering due to store vector type.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to store vector type.\n" ; } } while (false) ; 2691 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2692 return; 2693 } 2694 2695 // If all of the operands are identical or constant we have a simple solution. 2696 if (allConstant(VL) || isSplat(VL) || !allSameBlock(VL) || !S.getOpcode()) { 7 Assuming the condition is false 8 Assuming the condition is false 9 Taking false branch 2697 LLVM_DEBUG(dbgs() << "SLP: Gathering due to C,S,B,O. \n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to C,S,B,O. \n" ; } } while (false) ; 2698 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2699 return; 2700 } 2701 2702 // We now know that this is a vector of instructions of the same type from 2703 // the same block. 2704 2705 // Don't vectorize ephemeral values. 2706 for (Value *V : VL) { 10 Assuming '__begin1' is equal to '__end1' 2707 if (EphValues.count(V)) { 2708 LLVM_DEBUG(dbgs() << "SLP: The instruction (" << *Vdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: The instruction (" << * V << ") is ephemeral.\n"; } } while (false) 2709 << ") is ephemeral.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: The instruction (" << * V << ") is ephemeral.\n"; } } while (false) ; 2710 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2711 return; 2712 } 2713 } 2714 2715 // Check if this is a duplicate of another entry. 2716 if (TreeEntry *E = getTreeEntry(S.OpValue)) { 11 Assuming 'E' is null 12 Taking false branch 2717 LLVM_DEBUG(dbgs() << "SLP: \tChecking bundle: " << *S.OpValue << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: \tChecking bundle: " << *S.OpValue << ".\n"; } } while (false) ; 2718 if (!E->isSame(VL)) { 2719 LLVM_DEBUG(dbgs() << "SLP: Gathering due to partial overlap.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to partial overlap.\n" ; } } while (false) ; 2720 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2721 return; 2722 } 2723 // Record the reuse of the tree node. FIXME, currently this is only used to 2724 // properly draw the graph rather than for the actual vectorization. 2725 E->UserTreeIndices.push_back(UserTreeIdx); 2726 LLVM_DEBUG(dbgs() << "SLP: Perfect diamond merge at " << *S.OpValuedo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Perfect diamond merge at " << *S.OpValue << ".\n"; } } while (false) 2727 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Perfect diamond merge at " << *S.OpValue << ".\n"; } } while (false) ; 2728 return; 2729 } 2730 2731 // Check that none of the instructions in the bundle are already in the tree. 2732 for (Value *V : VL) { 13 Assuming '__begin1' is equal to '__end1' 2733 auto *I = dyn_cast<Instruction>(V); 2734 if (!I) 2735 continue; 2736 if (getTreeEntry(I)) { 2737 LLVM_DEBUG(dbgs() << "SLP: The instruction (" << *Vdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: The instruction (" << * V << ") is already in tree.\n"; } } while (false) 2738 << ") is already in tree.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: The instruction (" << * V << ") is already in tree.\n"; } } while (false) ; 2739 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2740 return; 2741 } 2742 } 2743 2744 // If any of the scalars is marked as a value that needs to stay scalar, then 2745 // we need to gather the scalars. 2746 // The reduction nodes (stored in UserIgnoreList) also should stay scalar. 2747 for (Value *V : VL) { 14 Assuming '__begin1' is equal to '__end1' 2748 if (MustGather.count(V) || is_contained(UserIgnoreList, V)) { 2749 LLVM_DEBUG(dbgs() << "SLP: Gathering due to gathered scalar.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering due to gathered scalar.\n" ; } } while (false) ; 2750 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2751 return; 2752 } 2753 } 2754 2755 // Check that all of the users of the scalars that we want to vectorize are 2756 // schedulable. 2757 auto *VL0 = cast<Instruction>(S.OpValue); 15 Field 'OpValue' is a 'Instruction' 2758 BasicBlock *BB = VL0->getParent(); 2759 2760 if (!DT->isReachableFromEntry(BB)) { 16 Assuming the condition is false 17 Taking false branch 2761 // Don't go into unreachable blocks. They may contain instructions with 2762 // dependency cycles which confuse the final scheduling. 2763 LLVM_DEBUG(dbgs() << "SLP: bundle in unreachable block.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: bundle in unreachable block.\n" ; } } while (false) ; 2764 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2765 return; 2766 } 2767 2768 // Check that every instruction appears once in this bundle. 2769 SmallVector<unsigned, 4> ReuseShuffleIndicies; 2770 SmallVector<Value *, 4> UniqueValues; 2771 DenseMap<Value *, unsigned> UniquePositions; 2772 for (Value *V : VL) { 18 Assuming '__begin1' is equal to '__end1' 2773 auto Res = UniquePositions.try_emplace(V, UniqueValues.size()); 2774 ReuseShuffleIndicies.emplace_back(Res.first->second); 2775 if (Res.second) 2776 UniqueValues.emplace_back(V); 2777 } 2778 size_t NumUniqueScalarValues = UniqueValues.size(); 2779 if (NumUniqueScalarValues == VL.size()) { 19 Assuming the condition is true 20 Taking true branch 2780 ReuseShuffleIndicies.clear(); 2781 } else { 2782 LLVM_DEBUG(dbgs() << "SLP: Shuffle for reused scalars.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Shuffle for reused scalars.\n" ; } } while (false) ; 2783 if (NumUniqueScalarValues <= 1 || 2784 !llvm::isPowerOf2_32(NumUniqueScalarValues)) { 2785 LLVM_DEBUG(dbgs() << "SLP: Scalar used twice in bundle.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Scalar used twice in bundle.\n" ; } } while (false) ; 2786 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx); 2787 return; 2788 } 2789 VL = UniqueValues; 2790 } 2791 2792 auto &BSRef = BlocksSchedules[BB]; 2793 if (!BSRef) 21 Taking false branch 2794 BSRef = std::make_unique<BlockScheduling>(BB); 2795 2796 BlockScheduling &BS = *BSRef.get(); 2797 2798 Optional<ScheduleData *> Bundle = BS.tryScheduleBundle(VL, this, S); 22 Calling 'BlockScheduling::tryScheduleBundle' 2799 if (!Bundle) { 2800 LLVM_DEBUG(dbgs() << "SLP: We are not able to schedule this bundle!\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: We are not able to schedule this bundle!\n" ; } } while (false) ; 2801 assert((!BS.getScheduleData(VL0) ||(static_cast <bool> ((!BS.getScheduleData(VL0) || !BS.getScheduleData (VL0)->isPartOfBundle()) && "tryScheduleBundle should cancelScheduling on failure" ) ? void (0) : __assert_fail ("(!BS.getScheduleData(VL0) || !BS.getScheduleData(VL0)->isPartOfBundle()) && \"tryScheduleBundle should cancelScheduling on failure\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2803, __extension__ __PRETTY_FUNCTION__)) 2802 !BS.getScheduleData(VL0)->isPartOfBundle()) &&(static_cast <bool> ((!BS.getScheduleData(VL0) || !BS.getScheduleData (VL0)->isPartOfBundle()) && "tryScheduleBundle should cancelScheduling on failure" ) ? void (0) : __assert_fail ("(!BS.getScheduleData(VL0) || !BS.getScheduleData(VL0)->isPartOfBundle()) && \"tryScheduleBundle should cancelScheduling on failure\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2803, __extension__ __PRETTY_FUNCTION__)) 2803 "tryScheduleBundle should cancelScheduling on failure")(static_cast <bool> ((!BS.getScheduleData(VL0) || !BS.getScheduleData (VL0)->isPartOfBundle()) && "tryScheduleBundle should cancelScheduling on failure" ) ? void (0) : __assert_fail ("(!BS.getScheduleData(VL0) || !BS.getScheduleData(VL0)->isPartOfBundle()) && \"tryScheduleBundle should cancelScheduling on failure\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2803, __extension__ __PRETTY_FUNCTION__)) ; 2804 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2805 ReuseShuffleIndicies); 2806 return; 2807 } 2808 LLVM_DEBUG(dbgs() << "SLP: We are able to schedule this bundle.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: We are able to schedule this bundle.\n" ; } } while (false) ; 2809 2810 unsigned ShuffleOrOp = S.isAltShuffle() ? 2811 (unsigned) Instruction::ShuffleVector : S.getOpcode(); 2812 switch (ShuffleOrOp) { 2813 case Instruction::PHI: { 2814 auto *PH = cast<PHINode>(VL0); 2815 2816 // Check for terminator values (e.g. invoke). 2817 for (Value *V : VL) 2818 for (unsigned I = 0, E = PH->getNumIncomingValues(); I < E; ++I) { 2819 Instruction *Term = dyn_cast<Instruction>( 2820 cast<PHINode>(V)->getIncomingValueForBlock( 2821 PH->getIncomingBlock(I))); 2822 if (Term && Term->isTerminator()) { 2823 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to swizzle PHINodes (terminator use).\n" ; } } while (false) 2824 << "SLP: Need to swizzle PHINodes (terminator use).\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Need to swizzle PHINodes (terminator use).\n" ; } } while (false) ; 2825 BS.cancelScheduling(VL, VL0); 2826 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2827 ReuseShuffleIndicies); 2828 return; 2829 } 2830 } 2831 2832 TreeEntry *TE = 2833 newTreeEntry(VL, Bundle, S, UserTreeIdx, ReuseShuffleIndicies); 2834 LLVM_DEBUG(dbgs() << "SLP: added a vector of PHINodes.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of PHINodes.\n" ; } } while (false) ; 2835 2836 // Keeps the reordered operands to avoid code duplication. 2837 SmallVector<ValueList, 2> OperandsVec; 2838 for (unsigned I = 0, E = PH->getNumIncomingValues(); I < E; ++I) { 2839 if (!DT->isReachableFromEntry(PH->getIncomingBlock(I))) { 2840 ValueList Operands(VL.size(), PoisonValue::get(PH->getType())); 2841 TE->setOperand(I, Operands); 2842 OperandsVec.push_back(Operands); 2843 continue; 2844 } 2845 ValueList Operands; 2846 // Prepare the operand vector. 2847 for (Value *V : VL) 2848 Operands.push_back(cast<PHINode>(V)->getIncomingValueForBlock( 2849 PH->getIncomingBlock(I))); 2850 TE->setOperand(I, Operands); 2851 OperandsVec.push_back(Operands); 2852 } 2853 for (unsigned OpIdx = 0, OpE = OperandsVec.size(); OpIdx != OpE; ++OpIdx) 2854 buildTree_rec(OperandsVec[OpIdx], Depth + 1, {TE, OpIdx}); 2855 return; 2856 } 2857 case Instruction::ExtractValue: 2858 case Instruction::ExtractElement: { 2859 OrdersType CurrentOrder; 2860 bool Reuse = canReuseExtract(VL, VL0, CurrentOrder); 2861 if (Reuse) { 2862 LLVM_DEBUG(dbgs() << "SLP: Reusing or shuffling extract sequence.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Reusing or shuffling extract sequence.\n" ; } } while (false) ; 2863 ++NumOpsWantToKeepOriginalOrder; 2864 newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 2865 ReuseShuffleIndicies); 2866 // This is a special case, as it does not gather, but at the same time 2867 // we are not extending buildTree_rec() towards the operands. 2868 ValueList Op0; 2869 Op0.assign(VL.size(), VL0->getOperand(0)); 2870 VectorizableTree.back()->setOperand(0, Op0); 2871 return; 2872 } 2873 if (!CurrentOrder.empty()) { 2874 LLVM_DEBUG({do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2875 dbgs() << "SLP: Reusing or shuffling of reordered extract sequence "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2876 "with order";do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2877 for (unsigned Idx : CurrentOrder)do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2878 dbgs() << " " << Idx;do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2879 dbgs() << "\n";do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) 2880 })do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: Reusing or shuffling of reordered extract sequence " "with order"; for (unsigned Idx : CurrentOrder) dbgs() << " " << Idx; dbgs() << "\n"; }; } } while (false) ; 2881 // Insert new order with initial value 0, if it does not exist, 2882 // otherwise return the iterator to the existing one. 2883 newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 2884 ReuseShuffleIndicies, CurrentOrder); 2885 findRootOrder(CurrentOrder); 2886 ++NumOpsWantToKeepOrder[CurrentOrder]; 2887 // This is a special case, as it does not gather, but at the same time 2888 // we are not extending buildTree_rec() towards the operands. 2889 ValueList Op0; 2890 Op0.assign(VL.size(), VL0->getOperand(0)); 2891 VectorizableTree.back()->setOperand(0, Op0); 2892 return; 2893 } 2894 LLVM_DEBUG(dbgs() << "SLP: Gather extract sequence.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gather extract sequence.\n"; } } while (false) ; 2895 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2896 ReuseShuffleIndicies); 2897 BS.cancelScheduling(VL, VL0); 2898 return; 2899 } 2900 case Instruction::InsertElement: { 2901 assert(ReuseShuffleIndicies.empty() && "All inserts should be unique")(static_cast <bool> (ReuseShuffleIndicies.empty() && "All inserts should be unique") ? void (0) : __assert_fail ( "ReuseShuffleIndicies.empty() && \"All inserts should be unique\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 2901, __extension__ __PRETTY_FUNCTION__)) ; 2902 2903 // Check that we have a buildvector and not a shuffle of 2 or more 2904 // different vectors. 2905 ValueSet SourceVectors; 2906 for (Value *V : VL) 2907 SourceVectors.insert(cast<Instruction>(V)->getOperand(0)); 2908 2909 if (count_if(VL, [&SourceVectors](Value *V) { 2910 return !SourceVectors.contains(V); 2911 }) >= 2) { 2912 // Found 2nd source vector - cancel. 2913 LLVM_DEBUG(dbgs() << "SLP: Gather of insertelement vectors with "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gather of insertelement vectors with " "different source vectors.\n"; } } while (false) 2914 "different source vectors.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gather of insertelement vectors with " "different source vectors.\n"; } } while (false) ; 2915 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2916 ReuseShuffleIndicies); 2917 BS.cancelScheduling(VL, VL0); 2918 return; 2919 } 2920 2921 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx); 2922 LLVM_DEBUG(dbgs() << "SLP: added inserts bundle.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added inserts bundle.\n"; } } while (false) ; 2923 2924 constexpr int NumOps = 2; 2925 ValueList VectorOperands[NumOps]; 2926 for (int I = 0; I < NumOps; ++I) { 2927 for (Value *V : VL) 2928 VectorOperands[I].push_back(cast<Instruction>(V)->getOperand(I)); 2929 2930 TE->setOperand(I, VectorOperands[I]); 2931 } 2932 buildTree_rec(VectorOperands[NumOps - 1], Depth + 1, {TE, 0}); 2933 return; 2934 } 2935 case Instruction::Load: { 2936 // Check that a vectorized load would load the same memory as a scalar 2937 // load. For example, we don't want to vectorize loads that are smaller 2938 // than 8-bit. Even though we have a packed struct {<i2, i2, i2, i2>} LLVM 2939 // treats loading/storing it as an i8 struct. If we vectorize loads/stores 2940 // from such a struct, we read/write packed bits disagreeing with the 2941 // unvectorized version. 2942 Type *ScalarTy = VL0->getType(); 2943 2944 if (DL->getTypeSizeInBits(ScalarTy) != 2945 DL->getTypeAllocSizeInBits(ScalarTy)) { 2946 BS.cancelScheduling(VL, VL0); 2947 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2948 ReuseShuffleIndicies); 2949 LLVM_DEBUG(dbgs() << "SLP: Gathering loads of non-packed type.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering loads of non-packed type.\n" ; } } while (false) ; 2950 return; 2951 } 2952 2953 // Make sure all loads in the bundle are simple - we can't vectorize 2954 // atomic or volatile loads. 2955 SmallVector<Value *, 4> PointerOps(VL.size()); 2956 auto POIter = PointerOps.begin(); 2957 for (Value *V : VL) { 2958 auto *L = cast<LoadInst>(V); 2959 if (!L->isSimple()) { 2960 BS.cancelScheduling(VL, VL0); 2961 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 2962 ReuseShuffleIndicies); 2963 LLVM_DEBUG(dbgs() << "SLP: Gathering non-simple loads.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering non-simple loads.\n" ; } } while (false) ; 2964 return; 2965 } 2966 *POIter = L->getPointerOperand(); 2967 ++POIter; 2968 } 2969 2970 OrdersType CurrentOrder; 2971 // Check the order of pointer operands. 2972 if (llvm::sortPtrAccesses(PointerOps, *DL, *SE, CurrentOrder)) { 2973 Value *Ptr0; 2974 Value *PtrN; 2975 if (CurrentOrder.empty()) { 2976 Ptr0 = PointerOps.front(); 2977 PtrN = PointerOps.back(); 2978 } else { 2979 Ptr0 = PointerOps[CurrentOrder.front()]; 2980 PtrN = PointerOps[CurrentOrder.back()]; 2981 } 2982 Optional<int> Diff = getPointersDiff(Ptr0, PtrN, *DL, *SE); 2983 // Check that the sorted loads are consecutive. 2984 if (static_cast<unsigned>(*Diff) == VL.size() - 1) { 2985 if (CurrentOrder.empty()) { 2986 // Original loads are consecutive and does not require reordering. 2987 ++NumOpsWantToKeepOriginalOrder; 2988 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, 2989 UserTreeIdx, ReuseShuffleIndicies); 2990 TE->setOperandsInOrder(); 2991 LLVM_DEBUG(dbgs() << "SLP: added a vector of loads.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of loads.\n"; } } while (false) ; 2992 } else { 2993 // Need to reorder. 2994 TreeEntry *TE = 2995 newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 2996 ReuseShuffleIndicies, CurrentOrder); 2997 TE->setOperandsInOrder(); 2998 LLVM_DEBUG(dbgs() << "SLP: added a vector of jumbled loads.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of jumbled loads.\n" ; } } while (false) ; 2999 findRootOrder(CurrentOrder); 3000 ++NumOpsWantToKeepOrder[CurrentOrder]; 3001 } 3002 return; 3003 } 3004 // Vectorizing non-consecutive loads with `llvm.masked.gather`. 3005 TreeEntry *TE = newTreeEntry(VL, TreeEntry::ScatterVectorize, Bundle, S, 3006 UserTreeIdx, ReuseShuffleIndicies); 3007 TE->setOperandsInOrder(); 3008 buildTree_rec(PointerOps, Depth + 1, {TE, 0}); 3009 LLVM_DEBUG(dbgs() << "SLP: added a vector of non-consecutive loads.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of non-consecutive loads.\n" ; } } while (false) ; 3010 return; 3011 } 3012 3013 LLVM_DEBUG(dbgs() << "SLP: Gathering non-consecutive loads.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering non-consecutive loads.\n" ; } } while (false) ; 3014 BS.cancelScheduling(VL, VL0); 3015 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3016 ReuseShuffleIndicies); 3017 return; 3018 } 3019 case Instruction::ZExt: 3020 case Instruction::SExt: 3021 case Instruction::FPToUI: 3022 case Instruction::FPToSI: 3023 case Instruction::FPExt: 3024 case Instruction::PtrToInt: 3025 case Instruction::IntToPtr: 3026 case Instruction::SIToFP: 3027 case Instruction::UIToFP: 3028 case Instruction::Trunc: 3029 case Instruction::FPTrunc: 3030 case Instruction::BitCast: { 3031 Type *SrcTy = VL0->getOperand(0)->getType(); 3032 for (Value *V : VL) { 3033 Type *Ty = cast<Instruction>(V)->getOperand(0)->getType(); 3034 if (Ty != SrcTy || !isValidElementType(Ty)) { 3035 BS.cancelScheduling(VL, VL0); 3036 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3037 ReuseShuffleIndicies); 3038 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering casts with different src types.\n" ; } } while (false) 3039 << "SLP: Gathering casts with different src types.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering casts with different src types.\n" ; } } while (false) ; 3040 return; 3041 } 3042 } 3043 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3044 ReuseShuffleIndicies); 3045 LLVM_DEBUG(dbgs() << "SLP: added a vector of casts.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of casts.\n"; } } while (false) ; 3046 3047 TE->setOperandsInOrder(); 3048 for (unsigned i = 0, e = VL0->getNumOperands(); i < e; ++i) { 3049 ValueList Operands; 3050 // Prepare the operand vector. 3051 for (Value *V : VL) 3052 Operands.push_back(cast<Instruction>(V)->getOperand(i)); 3053 3054 buildTree_rec(Operands, Depth + 1, {TE, i}); 3055 } 3056 return; 3057 } 3058 case Instruction::ICmp: 3059 case Instruction::FCmp: { 3060 // Check that all of the compares have the same predicate. 3061 CmpInst::Predicate P0 = cast<CmpInst>(VL0)->getPredicate(); 3062 CmpInst::Predicate SwapP0 = CmpInst::getSwappedPredicate(P0); 3063 Type *ComparedTy = VL0->getOperand(0)->getType(); 3064 for (Value *V : VL) { 3065 CmpInst *Cmp = cast<CmpInst>(V); 3066 if ((Cmp->getPredicate() != P0 && Cmp->getPredicate() != SwapP0) || 3067 Cmp->getOperand(0)->getType() != ComparedTy) { 3068 BS.cancelScheduling(VL, VL0); 3069 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3070 ReuseShuffleIndicies); 3071 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering cmp with different predicate.\n" ; } } while (false) 3072 << "SLP: Gathering cmp with different predicate.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering cmp with different predicate.\n" ; } } while (false) ; 3073 return; 3074 } 3075 } 3076 3077 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3078 ReuseShuffleIndicies); 3079 LLVM_DEBUG(dbgs() << "SLP: added a vector of compares.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of compares.\n" ; } } while (false) ; 3080 3081 ValueList Left, Right; 3082 if (cast<CmpInst>(VL0)->isCommutative()) { 3083 // Commutative predicate - collect + sort operands of the instructions 3084 // so that each side is more likely to have the same opcode. 3085 assert(P0 == SwapP0 && "Commutative Predicate mismatch")(static_cast <bool> (P0 == SwapP0 && "Commutative Predicate mismatch" ) ? void (0) : __assert_fail ("P0 == SwapP0 && \"Commutative Predicate mismatch\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3085, __extension__ __PRETTY_FUNCTION__)) ; 3086 reorderInputsAccordingToOpcode(VL, Left, Right, *DL, *SE, *this); 3087 } else { 3088 // Collect operands - commute if it uses the swapped predicate. 3089 for (Value *V : VL) { 3090 auto *Cmp = cast<CmpInst>(V); 3091 Value *LHS = Cmp->getOperand(0); 3092 Value *RHS = Cmp->getOperand(1); 3093 if (Cmp->getPredicate() != P0) 3094 std::swap(LHS, RHS); 3095 Left.push_back(LHS); 3096 Right.push_back(RHS); 3097 } 3098 } 3099 TE->setOperand(0, Left); 3100 TE->setOperand(1, Right); 3101 buildTree_rec(Left, Depth + 1, {TE, 0}); 3102 buildTree_rec(Right, Depth + 1, {TE, 1}); 3103 return; 3104 } 3105 case Instruction::Select: 3106 case Instruction::FNeg: 3107 case Instruction::Add: 3108 case Instruction::FAdd: 3109 case Instruction::Sub: 3110 case Instruction::FSub: 3111 case Instruction::Mul: 3112 case Instruction::FMul: 3113 case Instruction::UDiv: 3114 case Instruction::SDiv: 3115 case Instruction::FDiv: 3116 case Instruction::URem: 3117 case Instruction::SRem: 3118 case Instruction::FRem: 3119 case Instruction::Shl: 3120 case Instruction::LShr: 3121 case Instruction::AShr: 3122 case Instruction::And: 3123 case Instruction::Or: 3124 case Instruction::Xor: { 3125 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3126 ReuseShuffleIndicies); 3127 LLVM_DEBUG(dbgs() << "SLP: added a vector of un/bin op.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of un/bin op.\n" ; } } while (false) ; 3128 3129 // Sort operands of the instructions so that each side is more likely to 3130 // have the same opcode. 3131 if (isa<BinaryOperator>(VL0) && VL0->isCommutative()) { 3132 ValueList Left, Right; 3133 reorderInputsAccordingToOpcode(VL, Left, Right, *DL, *SE, *this); 3134 TE->setOperand(0, Left); 3135 TE->setOperand(1, Right); 3136 buildTree_rec(Left, Depth + 1, {TE, 0}); 3137 buildTree_rec(Right, Depth + 1, {TE, 1}); 3138 return; 3139 } 3140 3141 TE->setOperandsInOrder(); 3142 for (unsigned i = 0, e = VL0->getNumOperands(); i < e; ++i) { 3143 ValueList Operands; 3144 // Prepare the operand vector. 3145 for (Value *V : VL) 3146 Operands.push_back(cast<Instruction>(V)->getOperand(i)); 3147 3148 buildTree_rec(Operands, Depth + 1, {TE, i}); 3149 } 3150 return; 3151 } 3152 case Instruction::GetElementPtr: { 3153 // We don't combine GEPs with complicated (nested) indexing. 3154 for (Value *V : VL) { 3155 if (cast<Instruction>(V)->getNumOperands() != 2) { 3156 LLVM_DEBUG(dbgs() << "SLP: not-vectorizable GEP (nested indexes).\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: not-vectorizable GEP (nested indexes).\n" ; } } while (false) ; 3157 BS.cancelScheduling(VL, VL0); 3158 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3159 ReuseShuffleIndicies); 3160 return; 3161 } 3162 } 3163 3164 // We can't combine several GEPs into one vector if they operate on 3165 // different types. 3166 Type *Ty0 = VL0->getOperand(0)->getType(); 3167 for (Value *V : VL) { 3168 Type *CurTy = cast<Instruction>(V)->getOperand(0)->getType(); 3169 if (Ty0 != CurTy) { 3170 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: not-vectorizable GEP (different types).\n" ; } } while (false) 3171 << "SLP: not-vectorizable GEP (different types).\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: not-vectorizable GEP (different types).\n" ; } } while (false) ; 3172 BS.cancelScheduling(VL, VL0); 3173 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3174 ReuseShuffleIndicies); 3175 return; 3176 } 3177 } 3178 3179 // We don't combine GEPs with non-constant indexes. 3180 Type *Ty1 = VL0->getOperand(1)->getType(); 3181 for (Value *V : VL) { 3182 auto Op = cast<Instruction>(V)->getOperand(1); 3183 if (!isa<ConstantInt>(Op) || 3184 (Op->getType() != Ty1 && 3185 Op->getType()->getScalarSizeInBits() > 3186 DL->getIndexSizeInBits( 3187 V->getType()->getPointerAddressSpace()))) { 3188 LLVM_DEBUG(dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: not-vectorizable GEP (non-constant indexes).\n" ; } } while (false) 3189 << "SLP: not-vectorizable GEP (non-constant indexes).\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: not-vectorizable GEP (non-constant indexes).\n" ; } } while (false) ; 3190 BS.cancelScheduling(VL, VL0); 3191 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3192 ReuseShuffleIndicies); 3193 return; 3194 } 3195 } 3196 3197 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3198 ReuseShuffleIndicies); 3199 LLVM_DEBUG(dbgs() << "SLP: added a vector of GEPs.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of GEPs.\n"; } } while (false) ; 3200 TE->setOperandsInOrder(); 3201 for (unsigned i = 0, e = 2; i < e; ++i) { 3202 ValueList Operands; 3203 // Prepare the operand vector. 3204 for (Value *V : VL) 3205 Operands.push_back(cast<Instruction>(V)->getOperand(i)); 3206 3207 buildTree_rec(Operands, Depth + 1, {TE, i}); 3208 } 3209 return; 3210 } 3211 case Instruction::Store: { 3212 // Check if the stores are consecutive or if we need to swizzle them. 3213 llvm::Type *ScalarTy = cast<StoreInst>(VL0)->getValueOperand()->getType(); 3214 // Avoid types that are padded when being allocated as scalars, while 3215 // being packed together in a vector (such as i1). 3216 if (DL->getTypeSizeInBits(ScalarTy) != 3217 DL->getTypeAllocSizeInBits(ScalarTy)) { 3218 BS.cancelScheduling(VL, VL0); 3219 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3220 ReuseShuffleIndicies); 3221 LLVM_DEBUG(dbgs() << "SLP: Gathering stores of non-packed type.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering stores of non-packed type.\n" ; } } while (false) ; 3222 return; 3223 } 3224 // Make sure all stores in the bundle are simple - we can't vectorize 3225 // atomic or volatile stores. 3226 SmallVector<Value *, 4> PointerOps(VL.size()); 3227 ValueList Operands(VL.size()); 3228 auto POIter = PointerOps.begin(); 3229 auto OIter = Operands.begin(); 3230 for (Value *V : VL) { 3231 auto *SI = cast<StoreInst>(V); 3232 if (!SI->isSimple()) { 3233 BS.cancelScheduling(VL, VL0); 3234 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3235 ReuseShuffleIndicies); 3236 LLVM_DEBUG(dbgs() << "SLP: Gathering non-simple stores.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering non-simple stores.\n" ; } } while (false) ; 3237 return; 3238 } 3239 *POIter = SI->getPointerOperand(); 3240 *OIter = SI->getValueOperand(); 3241 ++POIter; 3242 ++OIter; 3243 } 3244 3245 OrdersType CurrentOrder; 3246 // Check the order of pointer operands. 3247 if (llvm::sortPtrAccesses(PointerOps, *DL, *SE, CurrentOrder)) { 3248 Value *Ptr0; 3249 Value *PtrN; 3250 if (CurrentOrder.empty()) { 3251 Ptr0 = PointerOps.front(); 3252 PtrN = PointerOps.back(); 3253 } else { 3254 Ptr0 = PointerOps[CurrentOrder.front()]; 3255 PtrN = PointerOps[CurrentOrder.back()]; 3256 } 3257 Optional<int> Dist = getPointersDiff(Ptr0, PtrN, *DL, *SE); 3258 // Check that the sorted pointer operands are consecutive. 3259 if (static_cast<unsigned>(*Dist) == VL.size() - 1) { 3260 if (CurrentOrder.empty()) { 3261 // Original stores are consecutive and does not require reordering. 3262 ++NumOpsWantToKeepOriginalOrder; 3263 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, 3264 UserTreeIdx, ReuseShuffleIndicies); 3265 TE->setOperandsInOrder(); 3266 buildTree_rec(Operands, Depth + 1, {TE, 0}); 3267 LLVM_DEBUG(dbgs() << "SLP: added a vector of stores.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of stores.\n" ; } } while (false) ; 3268 } else { 3269 TreeEntry *TE = 3270 newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3271 ReuseShuffleIndicies, CurrentOrder); 3272 TE->setOperandsInOrder(); 3273 buildTree_rec(Operands, Depth + 1, {TE, 0}); 3274 LLVM_DEBUG(dbgs() << "SLP: added a vector of jumbled stores.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a vector of jumbled stores.\n" ; } } while (false) ; 3275 findRootOrder(CurrentOrder); 3276 ++NumOpsWantToKeepOrder[CurrentOrder]; 3277 } 3278 return; 3279 } 3280 } 3281 3282 BS.cancelScheduling(VL, VL0); 3283 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3284 ReuseShuffleIndicies); 3285 LLVM_DEBUG(dbgs() << "SLP: Non-consecutive store.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Non-consecutive store.\n"; } } while (false) ; 3286 return; 3287 } 3288 case Instruction::Call: { 3289 // Check if the calls are all to the same vectorizable intrinsic or 3290 // library function. 3291 CallInst *CI = cast<CallInst>(VL0); 3292 Intrinsic::ID ID = getVectorIntrinsicIDForCall(CI, TLI); 3293 3294 VFShape Shape = VFShape::get( 3295 *CI, ElementCount::getFixed(static_cast<unsigned int>(VL.size())), 3296 false /*HasGlobalPred*/); 3297 Function *VecFunc = VFDatabase(*CI).getVectorizedFunction(Shape); 3298 3299 if (!VecFunc && !isTriviallyVectorizable(ID)) { 3300 BS.cancelScheduling(VL, VL0); 3301 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3302 ReuseShuffleIndicies); 3303 LLVM_DEBUG(dbgs() << "SLP: Non-vectorizable call.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Non-vectorizable call.\n"; } } while (false) ; 3304 return; 3305 } 3306 Function *F = CI->getCalledFunction(); 3307 unsigned NumArgs = CI->getNumArgOperands(); 3308 SmallVector<Value*, 4> ScalarArgs(NumArgs, nullptr); 3309 for (unsigned j = 0; j != NumArgs; ++j) 3310 if (hasVectorInstrinsicScalarOpd(ID, j)) 3311 ScalarArgs[j] = CI->getArgOperand(j); 3312 for (Value *V : VL) { 3313 CallInst *CI2 = dyn_cast<CallInst>(V); 3314 if (!CI2 || CI2->getCalledFunction() != F || 3315 getVectorIntrinsicIDForCall(CI2, TLI) != ID || 3316 (VecFunc && 3317 VecFunc != VFDatabase(*CI2).getVectorizedFunction(Shape)) || 3318 !CI->hasIdenticalOperandBundleSchema(*CI2)) { 3319 BS.cancelScheduling(VL, VL0); 3320 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3321 ReuseShuffleIndicies); 3322 LLVM_DEBUG(dbgs() << "SLP: mismatched calls:" << *CI << "!=" << *Vdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched calls:" << * CI << "!=" << *V << "\n"; } } while (false) 3323 << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched calls:" << * CI << "!=" << *V << "\n"; } } while (false) ; 3324 return; 3325 } 3326 // Some intrinsics have scalar arguments and should be same in order for 3327 // them to be vectorized. 3328 for (unsigned j = 0; j != NumArgs; ++j) { 3329 if (hasVectorInstrinsicScalarOpd(ID, j)) { 3330 Value *A1J = CI2->getArgOperand(j); 3331 if (ScalarArgs[j] != A1J) { 3332 BS.cancelScheduling(VL, VL0); 3333 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3334 ReuseShuffleIndicies); 3335 LLVM_DEBUG(dbgs() << "SLP: mismatched arguments in call:" << *CIdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched arguments in call:" << *CI << " argument " << ScalarArgs[j] << "!=" << A1J << "\n"; } } while (false) 3336 << " argument " << ScalarArgs[j] << "!=" << A1Jdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched arguments in call:" << *CI << " argument " << ScalarArgs[j] << "!=" << A1J << "\n"; } } while (false) 3337 << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched arguments in call:" << *CI << " argument " << ScalarArgs[j] << "!=" << A1J << "\n"; } } while (false) ; 3338 return; 3339 } 3340 } 3341 } 3342 // Verify that the bundle operands are identical between the two calls. 3343 if (CI->hasOperandBundles() && 3344 !std::equal(CI->op_begin() + CI->getBundleOperandsStartIndex(), 3345 CI->op_begin() + CI->getBundleOperandsEndIndex(), 3346 CI2->op_begin() + CI2->getBundleOperandsStartIndex())) { 3347 BS.cancelScheduling(VL, VL0); 3348 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3349 ReuseShuffleIndicies); 3350 LLVM_DEBUG(dbgs() << "SLP: mismatched bundle operands in calls:"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched bundle operands in calls:" << *CI << "!=" << *V << '\n'; } } while (false) 3351 << *CI << "!=" << *V << '\n')do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: mismatched bundle operands in calls:" << *CI << "!=" << *V << '\n'; } } while (false) ; 3352 return; 3353 } 3354 } 3355 3356 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3357 ReuseShuffleIndicies); 3358 TE->setOperandsInOrder(); 3359 for (unsigned i = 0, e = CI->getNumArgOperands(); i != e; ++i) { 3360 ValueList Operands; 3361 // Prepare the operand vector. 3362 for (Value *V : VL) { 3363 auto *CI2 = cast<CallInst>(V); 3364 Operands.push_back(CI2->getArgOperand(i)); 3365 } 3366 buildTree_rec(Operands, Depth + 1, {TE, i}); 3367 } 3368 return; 3369 } 3370 case Instruction::ShuffleVector: { 3371 // If this is not an alternate sequence of opcode like add-sub 3372 // then do not vectorize this instruction. 3373 if (!S.isAltShuffle()) { 3374 BS.cancelScheduling(VL, VL0); 3375 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3376 ReuseShuffleIndicies); 3377 LLVM_DEBUG(dbgs() << "SLP: ShuffleVector are not vectorized.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: ShuffleVector are not vectorized.\n" ; } } while (false) ; 3378 return; 3379 } 3380 TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx, 3381 ReuseShuffleIndicies); 3382 LLVM_DEBUG(dbgs() << "SLP: added a ShuffleVector op.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: added a ShuffleVector op.\n" ; } } while (false) ; 3383 3384 // Reorder operands if reordering would enable vectorization. 3385 if (isa<BinaryOperator>(VL0)) { 3386 ValueList Left, Right; 3387 reorderInputsAccordingToOpcode(VL, Left, Right, *DL, *SE, *this); 3388 TE->setOperand(0, Left); 3389 TE->setOperand(1, Right); 3390 buildTree_rec(Left, Depth + 1, {TE, 0}); 3391 buildTree_rec(Right, Depth + 1, {TE, 1}); 3392 return; 3393 } 3394 3395 TE->setOperandsInOrder(); 3396 for (unsigned i = 0, e = VL0->getNumOperands(); i < e; ++i) { 3397 ValueList Operands; 3398 // Prepare the operand vector. 3399 for (Value *V : VL) 3400 Operands.push_back(cast<Instruction>(V)->getOperand(i)); 3401 3402 buildTree_rec(Operands, Depth + 1, {TE, i}); 3403 } 3404 return; 3405 } 3406 default: 3407 BS.cancelScheduling(VL, VL0); 3408 newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx, 3409 ReuseShuffleIndicies); 3410 LLVM_DEBUG(dbgs() << "SLP: Gathering unknown instruction.\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Gathering unknown instruction.\n" ; } } while (false) ; 3411 return; 3412 } 3413} 3414 3415unsigned BoUpSLP::canMapToVector(Type *T, const DataLayout &DL) const { 3416 unsigned N = 1; 3417 Type *EltTy = T; 3418 3419 while (isa<StructType>(EltTy) || isa<ArrayType>(EltTy) || 3420 isa<VectorType>(EltTy)) { 3421 if (auto *ST = dyn_cast<StructType>(EltTy)) { 3422 // Check that struct is homogeneous. 3423 for (const auto *Ty : ST->elements()) 3424 if (Ty != *ST->element_begin()) 3425 return 0; 3426 N *= ST->getNumElements(); 3427 EltTy = *ST->element_begin(); 3428 } else if (auto *AT = dyn_cast<ArrayType>(EltTy)) { 3429 N *= AT->getNumElements(); 3430 EltTy = AT->getElementType(); 3431 } else { 3432 auto *VT = cast<FixedVectorType>(EltTy); 3433 N *= VT->getNumElements(); 3434 EltTy = VT->getElementType(); 3435 } 3436 } 3437 3438 if (!isValidElementType(EltTy)) 3439 return 0; 3440 uint64_t VTSize = DL.getTypeStoreSizeInBits(FixedVectorType::get(EltTy, N)); 3441 if (VTSize < MinVecRegSize || VTSize > MaxVecRegSize || VTSize != DL.getTypeStoreSizeInBits(T)) 3442 return 0; 3443 return N; 3444} 3445 3446bool BoUpSLP::canReuseExtract(ArrayRef<Value *> VL, Value *OpValue, 3447 SmallVectorImpl<unsigned> &CurrentOrder) const { 3448 Instruction *E0 = cast<Instruction>(OpValue); 3449 assert(E0->getOpcode() == Instruction::ExtractElement ||(static_cast <bool> (E0->getOpcode() == Instruction:: ExtractElement || E0->getOpcode() == Instruction::ExtractValue ) ? void (0) : __assert_fail ("E0->getOpcode() == Instruction::ExtractElement || E0->getOpcode() == Instruction::ExtractValue" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3450, __extension__ __PRETTY_FUNCTION__)) 3450 E0->getOpcode() == Instruction::ExtractValue)(static_cast <bool> (E0->getOpcode() == Instruction:: ExtractElement || E0->getOpcode() == Instruction::ExtractValue ) ? void (0) : __assert_fail ("E0->getOpcode() == Instruction::ExtractElement || E0->getOpcode() == Instruction::ExtractValue" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3450, __extension__ __PRETTY_FUNCTION__)) ; 3451 assert(E0->getOpcode() == getSameOpcode(VL).getOpcode() && "Invalid opcode")(static_cast <bool> (E0->getOpcode() == getSameOpcode (VL).getOpcode() && "Invalid opcode") ? void (0) : __assert_fail ("E0->getOpcode() == getSameOpcode(VL).getOpcode() && \"Invalid opcode\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3451, __extension__ __PRETTY_FUNCTION__)) ; 3452 // Check if all of the extracts come from the same vector and from the 3453 // correct offset. 3454 Value *Vec = E0->getOperand(0); 3455 3456 CurrentOrder.clear(); 3457 3458 // We have to extract from a vector/aggregate with the same number of elements. 3459 unsigned NElts; 3460 if (E0->getOpcode() == Instruction::ExtractValue) { 3461 const DataLayout &DL = E0->getModule()->getDataLayout(); 3462 NElts = canMapToVector(Vec->getType(), DL); 3463 if (!NElts) 3464 return false; 3465 // Check if load can be rewritten as load of vector. 3466 LoadInst *LI = dyn_cast<LoadInst>(Vec); 3467 if (!LI || !LI->isSimple() || !LI->hasNUses(VL.size())) 3468 return false; 3469 } else { 3470 NElts = cast<FixedVectorType>(Vec->getType())->getNumElements(); 3471 } 3472 3473 if (NElts != VL.size()) 3474 return false; 3475 3476 // Check that all of the indices extract from the correct offset. 3477 bool ShouldKeepOrder = true; 3478 unsigned E = VL.size(); 3479 // Assign to all items the initial value E + 1 so we can check if the extract 3480 // instruction index was used already. 3481 // Also, later we can check that all the indices are used and we have a 3482 // consecutive access in the extract instructions, by checking that no 3483 // element of CurrentOrder still has value E + 1. 3484 CurrentOrder.assign(E, E + 1); 3485 unsigned I = 0; 3486 for (; I < E; ++I) { 3487 auto *Inst = cast<Instruction>(VL[I]); 3488 if (Inst->getOperand(0) != Vec) 3489 break; 3490 Optional<unsigned> Idx = getExtractIndex(Inst); 3491 if (!Idx) 3492 break; 3493 const unsigned ExtIdx = *Idx; 3494 if (ExtIdx != I) { 3495 if (ExtIdx >= E || CurrentOrder[ExtIdx] != E + 1) 3496 break; 3497 ShouldKeepOrder = false; 3498 CurrentOrder[ExtIdx] = I; 3499 } else { 3500 if (CurrentOrder[I] != E + 1) 3501 break; 3502 CurrentOrder[I] = I; 3503 } 3504 } 3505 if (I < E) { 3506 CurrentOrder.clear(); 3507 return false; 3508 } 3509 3510 return ShouldKeepOrder; 3511} 3512 3513bool BoUpSLP::areAllUsersVectorized(Instruction *I, 3514 ArrayRef<Value *> VectorizedVals) const { 3515 return (I->hasOneUse() && is_contained(VectorizedVals, I)) || 3516 llvm::all_of(I->users(), [this](User *U) { 3517 return ScalarToTreeEntry.count(U) > 0; 3518 }); 3519} 3520 3521static std::pair<InstructionCost, InstructionCost> 3522getVectorCallCosts(CallInst *CI, FixedVectorType *VecTy, 3523 TargetTransformInfo *TTI, TargetLibraryInfo *TLI) { 3524 Intrinsic::ID ID = getVectorIntrinsicIDForCall(CI, TLI); 3525 3526 // Calculate the cost of the scalar and vector calls. 3527 SmallVector<Type *, 4> VecTys; 3528 for (Use &Arg : CI->args()) 3529 VecTys.push_back( 3530 FixedVectorType::get(Arg->getType(), VecTy->getNumElements())); 3531 FastMathFlags FMF; 3532 if (auto *FPCI = dyn_cast<FPMathOperator>(CI)) 3533 FMF = FPCI->getFastMathFlags(); 3534 SmallVector<const Value *> Arguments(CI->arg_begin(), CI->arg_end()); 3535 IntrinsicCostAttributes CostAttrs(ID, VecTy, Arguments, VecTys, FMF, 3536 dyn_cast<IntrinsicInst>(CI)); 3537 auto IntrinsicCost = 3538 TTI->getIntrinsicInstrCost(CostAttrs, TTI::TCK_RecipThroughput); 3539 3540 auto Shape = VFShape::get(*CI, ElementCount::getFixed(static_cast<unsigned>( 3541 VecTy->getNumElements())), 3542 false /*HasGlobalPred*/); 3543 Function *VecFunc = VFDatabase(*CI).getVectorizedFunction(Shape); 3544 auto LibCost = IntrinsicCost; 3545 if (!CI->isNoBuiltin() && VecFunc) { 3546 // Calculate the cost of the vector library call. 3547 // If the corresponding vector call is cheaper, return its cost. 3548 LibCost = TTI->getCallInstrCost(nullptr, VecTy, VecTys, 3549 TTI::TCK_RecipThroughput); 3550 } 3551 return {IntrinsicCost, LibCost}; 3552} 3553 3554/// Compute the cost of creating a vector of type \p VecTy containing the 3555/// extracted values from \p VL. 3556static InstructionCost 3557computeExtractCost(ArrayRef<Value *> VL, FixedVectorType *VecTy, 3558 TargetTransformInfo::ShuffleKind ShuffleKind, 3559 ArrayRef<int> Mask, TargetTransformInfo &TTI) { 3560 unsigned NumOfParts = TTI.getNumberOfParts(VecTy); 3561 3562 if (ShuffleKind != TargetTransformInfo::SK_PermuteSingleSrc || !NumOfParts || 3563 VecTy->getNumElements() < NumOfParts) 3564 return TTI.getShuffleCost(ShuffleKind, VecTy, Mask); 3565 3566 bool AllConsecutive = true; 3567 unsigned EltsPerVector = VecTy->getNumElements() / NumOfParts; 3568 unsigned Idx = -1; 3569 InstructionCost Cost = 0; 3570 3571 // Process extracts in blocks of EltsPerVector to check if the source vector 3572 // operand can be re-used directly. If not, add the cost of creating a shuffle 3573 // to extract the values into a vector register. 3574 for (auto *V : VL) { 3575 ++Idx; 3576 3577 // Reached the start of a new vector registers. 3578 if (Idx % EltsPerVector == 0) { 3579 AllConsecutive = true; 3580 continue; 3581 } 3582 3583 // Check all extracts for a vector register on the target directly 3584 // extract values in order. 3585 unsigned CurrentIdx = *getExtractIndex(cast<Instruction>(V)); 3586 unsigned PrevIdx = *getExtractIndex(cast<Instruction>(VL[Idx - 1])); 3587 AllConsecutive &= PrevIdx + 1 == CurrentIdx && 3588 CurrentIdx % EltsPerVector == Idx % EltsPerVector; 3589 3590 if (AllConsecutive) 3591 continue; 3592 3593 // Skip all indices, except for the last index per vector block. 3594 if ((Idx + 1) % EltsPerVector != 0 && Idx + 1 != VL.size()) 3595 continue; 3596 3597 // If we have a series of extracts which are not consecutive and hence 3598 // cannot re-use the source vector register directly, compute the shuffle 3599 // cost to extract the a vector with EltsPerVector elements. 3600 Cost += TTI.getShuffleCost( 3601 TargetTransformInfo::SK_PermuteSingleSrc, 3602 FixedVectorType::get(VecTy->getElementType(), EltsPerVector)); 3603 } 3604 return Cost; 3605} 3606 3607InstructionCost BoUpSLP::getEntryCost(const TreeEntry *E, 3608 ArrayRef<Value *> VectorizedVals) { 3609 ArrayRef<Value*> VL = E->Scalars; 3610 3611 Type *ScalarTy = VL[0]->getType(); 3612 if (StoreInst *SI = dyn_cast<StoreInst>(VL[0])) 3613 ScalarTy = SI->getValueOperand()->getType(); 3614 else if (CmpInst *CI = dyn_cast<CmpInst>(VL[0])) 3615 ScalarTy = CI->getOperand(0)->getType(); 3616 else if (auto *IE = dyn_cast<InsertElementInst>(VL[0])) 3617 ScalarTy = IE->getOperand(1)->getType(); 3618 auto *VecTy = FixedVectorType::get(ScalarTy, VL.size()); 3619 TTI::TargetCostKind CostKind = TTI::TCK_RecipThroughput; 3620 3621 // If we have computed a smaller type for the expression, update VecTy so 3622 // that the costs will be accurate. 3623 if (MinBWs.count(VL[0])) 3624 VecTy = FixedVectorType::get( 3625 IntegerType::get(F->getContext(), MinBWs[VL[0]].first), VL.size()); 3626 3627 unsigned ReuseShuffleNumbers = E->ReuseShuffleIndices.size(); 3628 bool NeedToShuffleReuses = !E->ReuseShuffleIndices.empty(); 3629 InstructionCost ReuseShuffleCost = 0; 3630 if (NeedToShuffleReuses) { 3631 ReuseShuffleCost = 3632 TTI->getShuffleCost(TargetTransformInfo::SK_PermuteSingleSrc, VecTy, 3633 E->ReuseShuffleIndices); 3634 } 3635 // FIXME: it tries to fix a problem with MSVC buildbots. 3636 TargetTransformInfo &TTIRef = *TTI; 3637 auto &&AdjustExtractsCost = [this, &TTIRef, CostKind, VL, VecTy, 3638 VectorizedVals](InstructionCost &Cost, 3639 bool IsGather) { 3640 DenseMap<Value *, int> ExtractVectorsTys; 3641 for (auto *V : VL) { 3642 // If all users of instruction are going to be vectorized and this 3643 // instruction itself is not going to be vectorized, consider this 3644 // instruction as dead and remove its cost from the final cost of the 3645 // vectorized tree. 3646 if (!areAllUsersVectorized(cast<Instruction>(V), VectorizedVals) || 3647 (IsGather && ScalarToTreeEntry.count(V))) 3648 continue; 3649 auto *EE = cast<ExtractElementInst>(V); 3650 unsigned Idx = *getExtractIndex(EE); 3651 if (TTIRef.getNumberOfParts(VecTy) != 3652 TTIRef.getNumberOfParts(EE->getVectorOperandType())) { 3653 auto It = 3654 ExtractVectorsTys.try_emplace(EE->getVectorOperand(), Idx).first; 3655 It->getSecond() = std::min<int>(It->second, Idx); 3656 } 3657 // Take credit for instruction that will become dead. 3658 if (EE->hasOneUse()) { 3659 Instruction *Ext = EE->user_back(); 3660 if ((isa<SExtInst>(Ext) || isa<ZExtInst>(Ext)) && 3661 all_of(Ext->users(), 3662 [](User *U) { return isa<GetElementPtrInst>(U); })) { 3663 // Use getExtractWithExtendCost() to calculate the cost of 3664 // extractelement/ext pair. 3665 Cost -= 3666 TTIRef.getExtractWithExtendCost(Ext->getOpcode(), Ext->getType(), 3667 EE->getVectorOperandType(), Idx); 3668 // Add back the cost of s|zext which is subtracted separately. 3669 Cost += TTIRef.getCastInstrCost( 3670 Ext->getOpcode(), Ext->getType(), EE->getType(), 3671 TTI::getCastContextHint(Ext), CostKind, Ext); 3672 continue; 3673 } 3674 } 3675 Cost -= TTIRef.getVectorInstrCost(Instruction::ExtractElement, 3676 EE->getVectorOperandType(), Idx); 3677 } 3678 // Add a cost for subvector extracts/inserts if required. 3679 for (const auto &Data : ExtractVectorsTys) { 3680 auto *EEVTy = cast<FixedVectorType>(Data.first->getType()); 3681 unsigned NumElts = VecTy->getNumElements(); 3682 if (TTIRef.getNumberOfParts(EEVTy) > TTIRef.getNumberOfParts(VecTy)) { 3683 unsigned Idx = (Data.second / NumElts) * NumElts; 3684 unsigned EENumElts = EEVTy->getNumElements(); 3685 if (Idx + NumElts <= EENumElts) { 3686 Cost += 3687 TTIRef.getShuffleCost(TargetTransformInfo::SK_ExtractSubvector, 3688 EEVTy, None, Idx, VecTy); 3689 } else { 3690 // Need to round up the subvector type vectorization factor to avoid a 3691 // crash in cost model functions. Make SubVT so that Idx + VF of SubVT 3692 // <= EENumElts. 3693 auto *SubVT = 3694 FixedVectorType::get(VecTy->getElementType(), EENumElts - Idx); 3695 Cost += 3696 TTIRef.getShuffleCost(TargetTransformInfo::SK_ExtractSubvector, 3697 EEVTy, None, Idx, SubVT); 3698 } 3699 } else { 3700 Cost += TTIRef.getShuffleCost(TargetTransformInfo::SK_InsertSubvector, 3701 VecTy, None, 0, EEVTy); 3702 } 3703 } 3704 }; 3705 if (E->State == TreeEntry::NeedToGather) { 3706 if (allConstant(VL)) 3707 return 0; 3708 if (isa<InsertElementInst>(VL[0])) 3709 return InstructionCost::getInvalid(); 3710 SmallVector<int> Mask; 3711 SmallVector<const TreeEntry *> Entries; 3712 Optional<TargetTransformInfo::ShuffleKind> Shuffle = 3713 isGatherShuffledEntry(E, Mask, Entries); 3714 if (Shuffle.hasValue()) { 3715 InstructionCost GatherCost = 0; 3716 if (ShuffleVectorInst::isIdentityMask(Mask)) { 3717 // Perfect match in the graph, will reuse the previously vectorized 3718 // node. Cost is 0. 3719 LLVM_DEBUG(do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: perfect diamond match for gather bundle that starts with " << *VL.front() << ".\n"; } } while (false) 3720 dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: perfect diamond match for gather bundle that starts with " << *VL.front() << ".\n"; } } while (false) 3721 << "SLP: perfect diamond match for gather bundle that starts with "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: perfect diamond match for gather bundle that starts with " << *VL.front() << ".\n"; } } while (false) 3722 << *VL.front() << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: perfect diamond match for gather bundle that starts with " << *VL.front() << ".\n"; } } while (false) ; 3723 } else { 3724 LLVM_DEBUG(dbgs() << "SLP: shuffled " << Entries.size()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: shuffled " << Entries. size() << " entries for bundle that starts with " << *VL.front() << ".\n"; } } while (false) 3725 << " entries for bundle that starts with "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: shuffled " << Entries. size() << " entries for bundle that starts with " << *VL.front() << ".\n"; } } while (false) 3726 << *VL.front() << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: shuffled " << Entries. size() << " entries for bundle that starts with " << *VL.front() << ".\n"; } } while (false) ; 3727 // Detected that instead of gather we can emit a shuffle of single/two 3728 // previously vectorized nodes. Add the cost of the permutation rather 3729 // than gather. 3730 GatherCost = TTI->getShuffleCost(*Shuffle, VecTy, Mask); 3731 } 3732 return ReuseShuffleCost + GatherCost; 3733 } 3734 if (isSplat(VL)) { 3735 // Found the broadcasting of the single scalar, calculate the cost as the 3736 // broadcast. 3737 return ReuseShuffleCost + 3738 TTI->getShuffleCost(TargetTransformInfo::SK_Broadcast, VecTy, None, 3739 0); 3740 } 3741 if (E->getOpcode() == Instruction::ExtractElement && allSameType(VL) && 3742 allSameBlock(VL)) { 3743 // Check that gather of extractelements can be represented as just a 3744 // shuffle of a single/two vectors the scalars are extracted from. 3745 SmallVector<int> Mask; 3746 Optional<TargetTransformInfo::ShuffleKind> ShuffleKind = 3747 isShuffle(VL, Mask); 3748 if (ShuffleKind.hasValue()) { 3749 // Found the bunch of extractelement instructions that must be gathered 3750 // into a vector and can be represented as a permutation elements in a 3751 // single input vector or of 2 input vectors. 3752 InstructionCost Cost = 3753 computeExtractCost(VL, VecTy, *ShuffleKind, Mask, *TTI); 3754 AdjustExtractsCost(Cost, /*IsGather=*/true); 3755 return ReuseShuffleCost + Cost; 3756 } 3757 } 3758 return ReuseShuffleCost + getGatherCost(VL); 3759 } 3760 assert((E->State == TreeEntry::Vectorize ||(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3762, __extension__ __PRETTY_FUNCTION__)) 3761 E->State == TreeEntry::ScatterVectorize) &&(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3762, __extension__ __PRETTY_FUNCTION__)) 3762 "Unhandled state")(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3762, __extension__ __PRETTY_FUNCTION__)) ; 3763 assert(E->getOpcode() && allSameType(VL) && allSameBlock(VL) && "Invalid VL")(static_cast <bool> (E->getOpcode() && allSameType (VL) && allSameBlock(VL) && "Invalid VL") ? void (0) : __assert_fail ("E->getOpcode() && allSameType(VL) && allSameBlock(VL) && \"Invalid VL\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3763, __extension__ __PRETTY_FUNCTION__)) ; 3764 Instruction *VL0 = E->getMainOp(); 3765 unsigned ShuffleOrOp = 3766 E->isAltShuffle() ? (unsigned)Instruction::ShuffleVector : E->getOpcode(); 3767 switch (ShuffleOrOp) { 3768 case Instruction::PHI: 3769 return 0; 3770 3771 case Instruction::ExtractValue: 3772 case Instruction::ExtractElement: { 3773 // The common cost of removal ExtractElement/ExtractValue instructions + 3774 // the cost of shuffles, if required to resuffle the original vector. 3775 InstructionCost CommonCost = 0; 3776 if (NeedToShuffleReuses) { 3777 unsigned Idx = 0; 3778 for (unsigned I : E->ReuseShuffleIndices) { 3779 if (ShuffleOrOp == Instruction::ExtractElement) { 3780 auto *EE = cast<ExtractElementInst>(VL[I]); 3781 ReuseShuffleCost -= TTI->getVectorInstrCost( 3782 Instruction::ExtractElement, EE->getVectorOperandType(), 3783 *getExtractIndex(EE)); 3784 } else { 3785 ReuseShuffleCost -= TTI->getVectorInstrCost( 3786 Instruction::ExtractElement, VecTy, Idx); 3787 ++Idx; 3788 } 3789 } 3790 Idx = ReuseShuffleNumbers; 3791 for (Value *V : VL) { 3792 if (ShuffleOrOp == Instruction::ExtractElement) { 3793 auto *EE = cast<ExtractElementInst>(V); 3794 ReuseShuffleCost += TTI->getVectorInstrCost( 3795 Instruction::ExtractElement, EE->getVectorOperandType(), 3796 *getExtractIndex(EE)); 3797 } else { 3798 --Idx; 3799 ReuseShuffleCost += TTI->getVectorInstrCost( 3800 Instruction::ExtractElement, VecTy, Idx); 3801 } 3802 } 3803 CommonCost = ReuseShuffleCost; 3804 } else if (!E->ReorderIndices.empty()) { 3805 SmallVector<int> NewMask; 3806 inversePermutation(E->ReorderIndices, NewMask); 3807 CommonCost = TTI->getShuffleCost( 3808 TargetTransformInfo::SK_PermuteSingleSrc, VecTy, NewMask); 3809 } 3810 if (ShuffleOrOp == Instruction::ExtractValue) { 3811 for (unsigned I = 0, E = VL.size(); I < E; ++I) { 3812 auto *EI = cast<Instruction>(VL[I]); 3813 // Take credit for instruction that will become dead. 3814 if (EI->hasOneUse()) { 3815 Instruction *Ext = EI->user_back(); 3816 if ((isa<SExtInst>(Ext) || isa<ZExtInst>(Ext)) && 3817 all_of(Ext->users(), 3818 [](User *U) { return isa<GetElementPtrInst>(U); })) { 3819 // Use getExtractWithExtendCost() to calculate the cost of 3820 // extractelement/ext pair. 3821 CommonCost -= TTI->getExtractWithExtendCost( 3822 Ext->getOpcode(), Ext->getType(), VecTy, I); 3823 // Add back the cost of s|zext which is subtracted separately. 3824 CommonCost += TTI->getCastInstrCost( 3825 Ext->getOpcode(), Ext->getType(), EI->getType(), 3826 TTI::getCastContextHint(Ext), CostKind, Ext); 3827 continue; 3828 } 3829 } 3830 CommonCost -= 3831 TTI->getVectorInstrCost(Instruction::ExtractElement, VecTy, I); 3832 } 3833 } else { 3834 AdjustExtractsCost(CommonCost, /*IsGather=*/false); 3835 } 3836 return CommonCost; 3837 } 3838 case Instruction::InsertElement: { 3839 auto *SrcVecTy = cast<FixedVectorType>(VL0->getType()); 3840 3841 unsigned const NumElts = SrcVecTy->getNumElements(); 3842 unsigned const NumScalars = VL.size(); 3843 APInt DemandedElts = APInt::getNullValue(NumElts); 3844 // TODO: Add support for Instruction::InsertValue. 3845 unsigned Offset = UINT_MAX(2147483647 *2U +1U); 3846 bool IsIdentity = true; 3847 SmallVector<int> ShuffleMask(NumElts, UndefMaskElem); 3848 for (unsigned I = 0; I < NumScalars; ++I) { 3849 Optional<int> InsertIdx = getInsertIndex(VL[I], 0); 3850 if (!InsertIdx || *InsertIdx == UndefMaskElem) 3851 continue; 3852 unsigned Idx = *InsertIdx; 3853 DemandedElts.setBit(Idx); 3854 if (Idx < Offset) { 3855 Offset = Idx; 3856 IsIdentity &= I == 0; 3857 } else { 3858 assert(Idx >= Offset && "Failed to find vector index offset")(static_cast <bool> (Idx >= Offset && "Failed to find vector index offset" ) ? void (0) : __assert_fail ("Idx >= Offset && \"Failed to find vector index offset\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3858, __extension__ __PRETTY_FUNCTION__)) ; 3859 IsIdentity &= Idx - Offset == I; 3860 } 3861 ShuffleMask[Idx] = I; 3862 } 3863 assert(Offset < NumElts && "Failed to find vector index offset")(static_cast <bool> (Offset < NumElts && "Failed to find vector index offset" ) ? void (0) : __assert_fail ("Offset < NumElts && \"Failed to find vector index offset\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 3863, __extension__ __PRETTY_FUNCTION__)) ; 3864 3865 InstructionCost Cost = 0; 3866 Cost -= TTI->getScalarizationOverhead(SrcVecTy, DemandedElts, 3867 /*Insert*/ true, /*Extract*/ false); 3868 3869 if (IsIdentity && NumElts != NumScalars && Offset % NumScalars != 0) 3870 Cost += TTI->getShuffleCost( 3871 TargetTransformInfo::SK_InsertSubvector, SrcVecTy, /*Mask*/ None, 3872 Offset, 3873 FixedVectorType::get(SrcVecTy->getElementType(), NumScalars)); 3874 else if (!IsIdentity) 3875 Cost += TTI->getShuffleCost(TTI::SK_PermuteSingleSrc, SrcVecTy, 3876 ShuffleMask); 3877 3878 return Cost; 3879 } 3880 case Instruction::ZExt: 3881 case Instruction::SExt: 3882 case Instruction::FPToUI: 3883 case Instruction::FPToSI: 3884 case Instruction::FPExt: 3885 case Instruction::PtrToInt: 3886 case Instruction::IntToPtr: 3887 case Instruction::SIToFP: 3888 case Instruction::UIToFP: 3889 case Instruction::Trunc: 3890 case Instruction::FPTrunc: 3891 case Instruction::BitCast: { 3892 Type *SrcTy = VL0->getOperand(0)->getType(); 3893 InstructionCost ScalarEltCost = 3894 TTI->getCastInstrCost(E->getOpcode(), ScalarTy, SrcTy, 3895 TTI::getCastContextHint(VL0), CostKind, VL0); 3896 if (NeedToShuffleReuses) { 3897 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 3898 } 3899 3900 // Calculate the cost of this instruction. 3901 InstructionCost ScalarCost = VL.size() * ScalarEltCost; 3902 3903 auto *SrcVecTy = FixedVectorType::get(SrcTy, VL.size()); 3904 InstructionCost VecCost = 0; 3905 // Check if the values are candidates to demote. 3906 if (!MinBWs.count(VL0) || VecTy != SrcVecTy) { 3907 VecCost = 3908 ReuseShuffleCost + 3909 TTI->getCastInstrCost(E->getOpcode(), VecTy, SrcVecTy, 3910 TTI::getCastContextHint(VL0), CostKind, VL0); 3911 } 3912 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost ); } } while (false) ; 3913 return VecCost - ScalarCost; 3914 } 3915 case Instruction::FCmp: 3916 case Instruction::ICmp: 3917 case Instruction::Select: { 3918 // Calculate the cost of this instruction. 3919 InstructionCost ScalarEltCost = 3920 TTI->getCmpSelInstrCost(E->getOpcode(), ScalarTy, Builder.getInt1Ty(), 3921 CmpInst::BAD_ICMP_PREDICATE, CostKind, VL0); 3922 if (NeedToShuffleReuses) { 3923 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 3924 } 3925 auto *MaskTy = FixedVectorType::get(Builder.getInt1Ty(), VL.size()); 3926 InstructionCost ScalarCost = VecTy->getNumElements() * ScalarEltCost; 3927 3928 // Check if all entries in VL are either compares or selects with compares 3929 // as condition that have the same predicates. 3930 CmpInst::Predicate VecPred = CmpInst::BAD_ICMP_PREDICATE; 3931 bool First = true; 3932 for (auto *V : VL) { 3933 CmpInst::Predicate CurrentPred; 3934 auto MatchCmp = m_Cmp(CurrentPred, m_Value(), m_Value()); 3935 if ((!match(V, m_Select(MatchCmp, m_Value(), m_Value())) && 3936 !match(V, MatchCmp)) || 3937 (!First && VecPred != CurrentPred)) { 3938 VecPred = CmpInst::BAD_ICMP_PREDICATE; 3939 break; 3940 } 3941 First = false; 3942 VecPred = CurrentPred; 3943 } 3944 3945 InstructionCost VecCost = TTI->getCmpSelInstrCost( 3946 E->getOpcode(), VecTy, MaskTy, VecPred, CostKind, VL0); 3947 // Check if it is possible and profitable to use min/max for selects in 3948 // VL. 3949 // 3950 auto IntrinsicAndUse = canConvertToMinOrMaxIntrinsic(VL); 3951 if (IntrinsicAndUse.first != Intrinsic::not_intrinsic) { 3952 IntrinsicCostAttributes CostAttrs(IntrinsicAndUse.first, VecTy, 3953 {VecTy, VecTy}); 3954 InstructionCost IntrinsicCost = 3955 TTI->getIntrinsicInstrCost(CostAttrs, CostKind); 3956 // If the selects are the only uses of the compares, they will be dead 3957 // and we can adjust the cost by removing their cost. 3958 if (IntrinsicAndUse.second) 3959 IntrinsicCost -= 3960 TTI->getCmpSelInstrCost(Instruction::ICmp, VecTy, MaskTy, 3961 CmpInst::BAD_ICMP_PREDICATE, CostKind); 3962 VecCost = std::min(VecCost, IntrinsicCost); 3963 } 3964 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost ); } } while (false) ; 3965 return ReuseShuffleCost + VecCost - ScalarCost; 3966 } 3967 case Instruction::FNeg: 3968 case Instruction::Add: 3969 case Instruction::FAdd: 3970 case Instruction::Sub: 3971 case Instruction::FSub: 3972 case Instruction::Mul: 3973 case Instruction::FMul: 3974 case Instruction::UDiv: 3975 case Instruction::SDiv: 3976 case Instruction::FDiv: 3977 case Instruction::URem: 3978 case Instruction::SRem: 3979 case Instruction::FRem: 3980 case Instruction::Shl: 3981 case Instruction::LShr: 3982 case Instruction::AShr: 3983 case Instruction::And: 3984 case Instruction::Or: 3985 case Instruction::Xor: { 3986 // Certain instructions can be cheaper to vectorize if they have a 3987 // constant second vector operand. 3988 TargetTransformInfo::OperandValueKind Op1VK = 3989 TargetTransformInfo::OK_AnyValue; 3990 TargetTransformInfo::OperandValueKind Op2VK = 3991 TargetTransformInfo::OK_UniformConstantValue; 3992 TargetTransformInfo::OperandValueProperties Op1VP = 3993 TargetTransformInfo::OP_None; 3994 TargetTransformInfo::OperandValueProperties Op2VP = 3995 TargetTransformInfo::OP_PowerOf2; 3996 3997 // If all operands are exactly the same ConstantInt then set the 3998 // operand kind to OK_UniformConstantValue. 3999 // If instead not all operands are constants, then set the operand kind 4000 // to OK_AnyValue. If all operands are constants but not the same, 4001 // then set the operand kind to OK_NonUniformConstantValue. 4002 ConstantInt *CInt0 = nullptr; 4003 for (unsigned i = 0, e = VL.size(); i < e; ++i) { 4004 const Instruction *I = cast<Instruction>(VL[i]); 4005 unsigned OpIdx = isa<BinaryOperator>(I) ? 1 : 0; 4006 ConstantInt *CInt = dyn_cast<ConstantInt>(I->getOperand(OpIdx)); 4007 if (!CInt) { 4008 Op2VK = TargetTransformInfo::OK_AnyValue; 4009 Op2VP = TargetTransformInfo::OP_None; 4010 break; 4011 } 4012 if (Op2VP == TargetTransformInfo::OP_PowerOf2 && 4013 !CInt->getValue().isPowerOf2()) 4014 Op2VP = TargetTransformInfo::OP_None; 4015 if (i == 0) { 4016 CInt0 = CInt; 4017 continue; 4018 } 4019 if (CInt0 != CInt) 4020 Op2VK = TargetTransformInfo::OK_NonUniformConstantValue; 4021 } 4022 4023 SmallVector<const Value *, 4> Operands(VL0->operand_values()); 4024 InstructionCost ScalarEltCost = 4025 TTI->getArithmeticInstrCost(E->getOpcode(), ScalarTy, CostKind, Op1VK, 4026 Op2VK, Op1VP, Op2VP, Operands, VL0); 4027 if (NeedToShuffleReuses) { 4028 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 4029 } 4030 InstructionCost ScalarCost = VecTy->getNumElements() * ScalarEltCost; 4031 InstructionCost VecCost = 4032 TTI->getArithmeticInstrCost(E->getOpcode(), VecTy, CostKind, Op1VK, 4033 Op2VK, Op1VP, Op2VP, Operands, VL0); 4034 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost ); } } while (false) ; 4035 return ReuseShuffleCost + VecCost - ScalarCost; 4036 } 4037 case Instruction::GetElementPtr: { 4038 TargetTransformInfo::OperandValueKind Op1VK = 4039 TargetTransformInfo::OK_AnyValue; 4040 TargetTransformInfo::OperandValueKind Op2VK = 4041 TargetTransformInfo::OK_UniformConstantValue; 4042 4043 InstructionCost ScalarEltCost = TTI->getArithmeticInstrCost( 4044 Instruction::Add, ScalarTy, CostKind, Op1VK, Op2VK); 4045 if (NeedToShuffleReuses) { 4046 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 4047 } 4048 InstructionCost ScalarCost = VecTy->getNumElements() * ScalarEltCost; 4049 InstructionCost VecCost = TTI->getArithmeticInstrCost( 4050 Instruction::Add, VecTy, CostKind, Op1VK, Op2VK); 4051 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost ); } } while (false) ; 4052 return ReuseShuffleCost + VecCost - ScalarCost; 4053 } 4054 case Instruction::Load: { 4055 // Cost of wide load - cost of scalar loads. 4056 Align alignment = cast<LoadInst>(VL0)->getAlign(); 4057 InstructionCost ScalarEltCost = TTI->getMemoryOpCost( 4058 Instruction::Load, ScalarTy, alignment, 0, CostKind, VL0); 4059 if (NeedToShuffleReuses) { 4060 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 4061 } 4062 InstructionCost ScalarLdCost = VecTy->getNumElements() * ScalarEltCost; 4063 InstructionCost VecLdCost; 4064 if (E->State == TreeEntry::Vectorize) { 4065 VecLdCost = TTI->getMemoryOpCost(Instruction::Load, VecTy, alignment, 0, 4066 CostKind, VL0); 4067 } else { 4068 assert(E->State == TreeEntry::ScatterVectorize && "Unknown EntryState")(static_cast <bool> (E->State == TreeEntry::ScatterVectorize && "Unknown EntryState") ? void (0) : __assert_fail ( "E->State == TreeEntry::ScatterVectorize && \"Unknown EntryState\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4068, __extension__ __PRETTY_FUNCTION__)) ; 4069 VecLdCost = TTI->getGatherScatterOpCost( 4070 Instruction::Load, VecTy, cast<LoadInst>(VL0)->getPointerOperand(), 4071 /*VariableMask=*/false, alignment, CostKind, VL0); 4072 } 4073 if (!NeedToShuffleReuses && !E->ReorderIndices.empty()) { 4074 SmallVector<int> NewMask; 4075 inversePermutation(E->ReorderIndices, NewMask); 4076 VecLdCost += TTI->getShuffleCost( 4077 TargetTransformInfo::SK_PermuteSingleSrc, VecTy, NewMask); 4078 } 4079 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecLdCost, ScalarLdCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecLdCost, ScalarLdCost ); } } while (false) ; 4080 return ReuseShuffleCost + VecLdCost - ScalarLdCost; 4081 } 4082 case Instruction::Store: { 4083 // We know that we can merge the stores. Calculate the cost. 4084 bool IsReorder = !E->ReorderIndices.empty(); 4085 auto *SI = 4086 cast<StoreInst>(IsReorder ? VL[E->ReorderIndices.front()] : VL0); 4087 Align Alignment = SI->getAlign(); 4088 InstructionCost ScalarEltCost = TTI->getMemoryOpCost( 4089 Instruction::Store, ScalarTy, Alignment, 0, CostKind, VL0); 4090 InstructionCost ScalarStCost = VecTy->getNumElements() * ScalarEltCost; 4091 InstructionCost VecStCost = TTI->getMemoryOpCost( 4092 Instruction::Store, VecTy, Alignment, 0, CostKind, VL0); 4093 if (IsReorder) { 4094 SmallVector<int> NewMask; 4095 inversePermutation(E->ReorderIndices, NewMask); 4096 VecStCost += TTI->getShuffleCost( 4097 TargetTransformInfo::SK_PermuteSingleSrc, VecTy, NewMask); 4098 } 4099 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecStCost, ScalarStCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecStCost, ScalarStCost ); } } while (false) ; 4100 return VecStCost - ScalarStCost; 4101 } 4102 case Instruction::Call: { 4103 CallInst *CI = cast<CallInst>(VL0); 4104 Intrinsic::ID ID = getVectorIntrinsicIDForCall(CI, TLI); 4105 4106 // Calculate the cost of the scalar and vector calls. 4107 IntrinsicCostAttributes CostAttrs(ID, *CI, 1); 4108 InstructionCost ScalarEltCost = 4109 TTI->getIntrinsicInstrCost(CostAttrs, CostKind); 4110 if (NeedToShuffleReuses) { 4111 ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost; 4112 } 4113 InstructionCost ScalarCallCost = VecTy->getNumElements() * ScalarEltCost; 4114 4115 auto VecCallCosts = getVectorCallCosts(CI, VecTy, TTI, TLI); 4116 InstructionCost VecCallCost = 4117 std::min(VecCallCosts.first, VecCallCosts.second); 4118 4119 LLVM_DEBUG(dbgs() << "SLP: Call cost " << VecCallCost - ScalarCallCostdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Call cost " << VecCallCost - ScalarCallCost << " (" << VecCallCost << "-" << ScalarCallCost << ")" << " for " << *CI << "\n"; } } while (false) 4120 << " (" << VecCallCost << "-" << ScalarCallCost << ")"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Call cost " << VecCallCost - ScalarCallCost << " (" << VecCallCost << "-" << ScalarCallCost << ")" << " for " << *CI << "\n"; } } while (false) 4121 << " for " << *CI << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Call cost " << VecCallCost - ScalarCallCost << " (" << VecCallCost << "-" << ScalarCallCost << ")" << " for " << *CI << "\n"; } } while (false) ; 4122 4123 return ReuseShuffleCost + VecCallCost - ScalarCallCost; 4124 } 4125 case Instruction::ShuffleVector: { 4126 assert(E->isAltShuffle() &&(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) 4127 ((Instruction::isBinaryOp(E->getOpcode()) &&(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) 4128 Instruction::isBinaryOp(E->getAltOpcode())) ||(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) 4129 (Instruction::isCast(E->getOpcode()) &&(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) 4130 Instruction::isCast(E->getAltOpcode()))) &&(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) 4131 "Invalid Shuffle Vector Operand")(static_cast <bool> (E->isAltShuffle() && (( Instruction::isBinaryOp(E->getOpcode()) && Instruction ::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E ->getOpcode()) && Instruction::isCast(E->getAltOpcode ()))) && "Invalid Shuffle Vector Operand") ? void (0) : __assert_fail ("E->isAltShuffle() && ((Instruction::isBinaryOp(E->getOpcode()) && Instruction::isBinaryOp(E->getAltOpcode())) || (Instruction::isCast(E->getOpcode()) && Instruction::isCast(E->getAltOpcode()))) && \"Invalid Shuffle Vector Operand\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4131, __extension__ __PRETTY_FUNCTION__)) ; 4132 InstructionCost ScalarCost = 0; 4133 if (NeedToShuffleReuses) { 4134 for (unsigned Idx : E->ReuseShuffleIndices) { 4135 Instruction *I = cast<Instruction>(VL[Idx]); 4136 ReuseShuffleCost -= TTI->getInstructionCost(I, CostKind); 4137 } 4138 for (Value *V : VL) { 4139 Instruction *I = cast<Instruction>(V); 4140 ReuseShuffleCost += TTI->getInstructionCost(I, CostKind); 4141 } 4142 } 4143 for (Value *V : VL) { 4144 Instruction *I = cast<Instruction>(V); 4145 assert(E->isOpcodeOrAlt(I) && "Unexpected main/alternate opcode")(static_cast <bool> (E->isOpcodeOrAlt(I) && "Unexpected main/alternate opcode" ) ? void (0) : __assert_fail ("E->isOpcodeOrAlt(I) && \"Unexpected main/alternate opcode\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4145, __extension__ __PRETTY_FUNCTION__)) ; 4146 ScalarCost += TTI->getInstructionCost(I, CostKind); 4147 } 4148 // VecCost is equal to sum of the cost of creating 2 vectors 4149 // and the cost of creating shuffle. 4150 InstructionCost VecCost = 0; 4151 if (Instruction::isBinaryOp(E->getOpcode())) { 4152 VecCost = TTI->getArithmeticInstrCost(E->getOpcode(), VecTy, CostKind); 4153 VecCost += TTI->getArithmeticInstrCost(E->getAltOpcode(), VecTy, 4154 CostKind); 4155 } else { 4156 Type *Src0SclTy = E->getMainOp()->getOperand(0)->getType(); 4157 Type *Src1SclTy = E->getAltOp()->getOperand(0)->getType(); 4158 auto *Src0Ty = FixedVectorType::get(Src0SclTy, VL.size()); 4159 auto *Src1Ty = FixedVectorType::get(Src1SclTy, VL.size()); 4160 VecCost = TTI->getCastInstrCost(E->getOpcode(), VecTy, Src0Ty, 4161 TTI::CastContextHint::None, CostKind); 4162 VecCost += TTI->getCastInstrCost(E->getAltOpcode(), VecTy, Src1Ty, 4163 TTI::CastContextHint::None, CostKind); 4164 } 4165 4166 SmallVector<int> Mask(E->Scalars.size()); 4167 for (unsigned I = 0, End = E->Scalars.size(); I < End; ++I) { 4168 auto *OpInst = cast<Instruction>(E->Scalars[I]); 4169 assert(E->isOpcodeOrAlt(OpInst) && "Unexpected main/alternate opcode")(static_cast <bool> (E->isOpcodeOrAlt(OpInst) && "Unexpected main/alternate opcode") ? void (0) : __assert_fail ("E->isOpcodeOrAlt(OpInst) && \"Unexpected main/alternate opcode\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4169, __extension__ __PRETTY_FUNCTION__)) ; 4170 Mask[I] = I + (OpInst->getOpcode() == E->getAltOpcode() ? End : 0); 4171 } 4172 VecCost += 4173 TTI->getShuffleCost(TargetTransformInfo::SK_Select, VecTy, Mask, 0); 4174 LLVM_DEBUG(dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost))do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dumpTreeCosts(E, ReuseShuffleCost, VecCost, ScalarCost ); } } while (false) ; 4175 return ReuseShuffleCost + VecCost - ScalarCost; 4176 } 4177 default: 4178 llvm_unreachable("Unknown instruction")::llvm::llvm_unreachable_internal("Unknown instruction", "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4178) ; 4179 } 4180} 4181 4182bool BoUpSLP::isFullyVectorizableTinyTree() const { 4183 LLVM_DEBUG(dbgs() << "SLP: Check whether the tree with height "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Check whether the tree with height " << VectorizableTree.size() << " is fully vectorizable .\n" ; } } while (false) 4184 << VectorizableTree.size() << " is fully vectorizable .\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Check whether the tree with height " << VectorizableTree.size() << " is fully vectorizable .\n" ; } } while (false) ; 4185 4186 // We only handle trees of heights 1 and 2. 4187 if (VectorizableTree.size() == 1 && 4188 VectorizableTree[0]->State == TreeEntry::Vectorize) 4189 return true; 4190 4191 if (VectorizableTree.size() != 2) 4192 return false; 4193 4194 // Handle splat and all-constants stores. Also try to vectorize tiny trees 4195 // with the second gather nodes if they have less scalar operands rather than 4196 // the initial tree element (may be profitable to shuffle the second gather) 4197 // or they are extractelements, which form shuffle. 4198 SmallVector<int> Mask; 4199 if (VectorizableTree[0]->State == TreeEntry::Vectorize && 4200 (allConstant(VectorizableTree[1]->Scalars) || 4201 isSplat(VectorizableTree[1]->Scalars) || 4202 (VectorizableTree[1]->State == TreeEntry::NeedToGather && 4203 VectorizableTree[1]->Scalars.size() < 4204 VectorizableTree[0]->Scalars.size()) || 4205 (VectorizableTree[1]->State == TreeEntry::NeedToGather && 4206 VectorizableTree[1]->getOpcode() == Instruction::ExtractElement && 4207 isShuffle(VectorizableTree[1]->Scalars, Mask)))) 4208 return true; 4209 4210 // Gathering cost would be too much for tiny trees. 4211 if (VectorizableTree[0]->State == TreeEntry::NeedToGather || 4212 VectorizableTree[1]->State == TreeEntry::NeedToGather) 4213 return false; 4214 4215 return true; 4216} 4217 4218static bool isLoadCombineCandidateImpl(Value *Root, unsigned NumElts, 4219 TargetTransformInfo *TTI, 4220 bool MustMatchOrInst) { 4221 // Look past the root to find a source value. Arbitrarily follow the 4222 // path through operand 0 of any 'or'. Also, peek through optional 4223 // shift-left-by-multiple-of-8-bits. 4224 Value *ZextLoad = Root; 4225 const APInt *ShAmtC; 4226 bool FoundOr = false; 4227 while (!isa<ConstantExpr>(ZextLoad) && 4228 (match(ZextLoad, m_Or(m_Value(), m_Value())) || 4229 (match(ZextLoad, m_Shl(m_Value(), m_APInt(ShAmtC))) && 4230 ShAmtC->urem(8) == 0))) { 4231 auto *BinOp = cast<BinaryOperator>(ZextLoad); 4232 ZextLoad = BinOp->getOperand(0); 4233 if (BinOp->getOpcode() == Instruction::Or) 4234 FoundOr = true; 4235 } 4236 // Check if the input is an extended load of the required or/shift expression. 4237 Value *LoadPtr; 4238 if ((MustMatchOrInst && !FoundOr) || ZextLoad == Root || 4239 !match(ZextLoad, m_ZExt(m_Load(m_Value(LoadPtr))))) 4240 return false; 4241 4242 // Require that the total load bit width is a legal integer type. 4243 // For example, <8 x i8> --> i64 is a legal integer on a 64-bit target. 4244 // But <16 x i8> --> i128 is not, so the backend probably can't reduce it. 4245 Type *SrcTy = LoadPtr->getType()->getPointerElementType(); 4246 unsigned LoadBitWidth = SrcTy->getIntegerBitWidth() * NumElts; 4247 if (!TTI->isTypeLegal(IntegerType::get(Root->getContext(), LoadBitWidth))) 4248 return false; 4249 4250 // Everything matched - assume that we can fold the whole sequence using 4251 // load combining. 4252 LLVM_DEBUG(dbgs() << "SLP: Assume load combining for tree starting at "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Assume load combining for tree starting at " << *(cast<Instruction>(Root)) << "\n"; } } while (false) 4253 << *(cast<Instruction>(Root)) << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Assume load combining for tree starting at " << *(cast<Instruction>(Root)) << "\n"; } } while (false) ; 4254 4255 return true; 4256} 4257 4258bool BoUpSLP::isLoadCombineReductionCandidate(RecurKind RdxKind) const { 4259 if (RdxKind != RecurKind::Or) 4260 return false; 4261 4262 unsigned NumElts = VectorizableTree[0]->Scalars.size(); 4263 Value *FirstReduced = VectorizableTree[0]->Scalars[0]; 4264 return isLoadCombineCandidateImpl(FirstReduced, NumElts, TTI, 4265 /* MatchOr */ false); 4266} 4267 4268bool BoUpSLP::isLoadCombineCandidate() const { 4269 // Peek through a final sequence of stores and check if all operations are 4270 // likely to be load-combined. 4271 unsigned NumElts = VectorizableTree[0]->Scalars.size(); 4272 for (Value *Scalar : VectorizableTree[0]->Scalars) { 4273 Value *X; 4274 if (!match(Scalar, m_Store(m_Value(X), m_Value())) || 4275 !isLoadCombineCandidateImpl(X, NumElts, TTI, /* MatchOr */ true)) 4276 return false; 4277 } 4278 return true; 4279} 4280 4281bool BoUpSLP::isTreeTinyAndNotFullyVectorizable() const { 4282 // No need to vectorize inserts of gathered values. 4283 if (VectorizableTree.size() == 2 && 4284 isa<InsertElementInst>(VectorizableTree[0]->Scalars[0]) && 4285 VectorizableTree[1]->State == TreeEntry::NeedToGather) 4286 return true; 4287 4288 // We can vectorize the tree if its size is greater than or equal to the 4289 // minimum size specified by the MinTreeSize command line option. 4290 if (VectorizableTree.size() >= MinTreeSize) 4291 return false; 4292 4293 // If we have a tiny tree (a tree whose size is less than MinTreeSize), we 4294 // can vectorize it if we can prove it fully vectorizable. 4295 if (isFullyVectorizableTinyTree()) 4296 return false; 4297 4298 assert(VectorizableTree.empty()(static_cast <bool> (VectorizableTree.empty() ? ExternalUses .empty() : true && "We shouldn't have any external users" ) ? void (0) : __assert_fail ("VectorizableTree.empty() ? ExternalUses.empty() : true && \"We shouldn't have any external users\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4300, __extension__ __PRETTY_FUNCTION__)) 4299 ? ExternalUses.empty()(static_cast <bool> (VectorizableTree.empty() ? ExternalUses .empty() : true && "We shouldn't have any external users" ) ? void (0) : __assert_fail ("VectorizableTree.empty() ? ExternalUses.empty() : true && \"We shouldn't have any external users\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4300, __extension__ __PRETTY_FUNCTION__)) 4300 : true && "We shouldn't have any external users")(static_cast <bool> (VectorizableTree.empty() ? ExternalUses .empty() : true && "We shouldn't have any external users" ) ? void (0) : __assert_fail ("VectorizableTree.empty() ? ExternalUses.empty() : true && \"We shouldn't have any external users\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4300, __extension__ __PRETTY_FUNCTION__)) ; 4301 4302 // Otherwise, we can't vectorize the tree. It is both tiny and not fully 4303 // vectorizable. 4304 return true; 4305} 4306 4307InstructionCost BoUpSLP::getSpillCost() const { 4308 // Walk from the bottom of the tree to the top, tracking which values are 4309 // live. When we see a call instruction that is not part of our tree, 4310 // query TTI to see if there is a cost to keeping values live over it 4311 // (for example, if spills and fills are required). 4312 unsigned BundleWidth = VectorizableTree.front()->Scalars.size(); 4313 InstructionCost Cost = 0; 4314 4315 SmallPtrSet<Instruction*, 4> LiveValues; 4316 Instruction *PrevInst = nullptr; 4317 4318 // The entries in VectorizableTree are not necessarily ordered by their 4319 // position in basic blocks. Collect them and order them by dominance so later 4320 // instructions are guaranteed to be visited first. For instructions in 4321 // different basic blocks, we only scan to the beginning of the block, so 4322 // their order does not matter, as long as all instructions in a basic block 4323 // are grouped together. Using dominance ensures a deterministic order. 4324 SmallVector<Instruction *, 16> OrderedScalars; 4325 for (const auto &TEPtr : VectorizableTree) { 4326 Instruction *Inst = dyn_cast<Instruction>(TEPtr->Scalars[0]); 4327 if (!Inst) 4328 continue; 4329 OrderedScalars.push_back(Inst); 4330 } 4331 llvm::sort(OrderedScalars, [&](Instruction *A, Instruction *B) { 4332 auto *NodeA = DT->getNode(A->getParent()); 4333 auto *NodeB = DT->getNode(B->getParent()); 4334 assert(NodeA && "Should only process reachable instructions")(static_cast <bool> (NodeA && "Should only process reachable instructions" ) ? void (0) : __assert_fail ("NodeA && \"Should only process reachable instructions\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4334, __extension__ __PRETTY_FUNCTION__)) ; 4335 assert(NodeB && "Should only process reachable instructions")(static_cast <bool> (NodeB && "Should only process reachable instructions" ) ? void (0) : __assert_fail ("NodeB && \"Should only process reachable instructions\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4335, __extension__ __PRETTY_FUNCTION__)) ; 4336 assert((NodeA == NodeB) == (NodeA->getDFSNumIn() == NodeB->getDFSNumIn()) &&(static_cast <bool> ((NodeA == NodeB) == (NodeA->getDFSNumIn () == NodeB->getDFSNumIn()) && "Different nodes should have different DFS numbers" ) ? void (0) : __assert_fail ("(NodeA == NodeB) == (NodeA->getDFSNumIn() == NodeB->getDFSNumIn()) && \"Different nodes should have different DFS numbers\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4337, __extension__ __PRETTY_FUNCTION__)) 4337 "Different nodes should have different DFS numbers")(static_cast <bool> ((NodeA == NodeB) == (NodeA->getDFSNumIn () == NodeB->getDFSNumIn()) && "Different nodes should have different DFS numbers" ) ? void (0) : __assert_fail ("(NodeA == NodeB) == (NodeA->getDFSNumIn() == NodeB->getDFSNumIn()) && \"Different nodes should have different DFS numbers\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4337, __extension__ __PRETTY_FUNCTION__)) ; 4338 if (NodeA != NodeB) 4339 return NodeA->getDFSNumIn() < NodeB->getDFSNumIn(); 4340 return B->comesBefore(A); 4341 }); 4342 4343 for (Instruction *Inst : OrderedScalars) { 4344 if (!PrevInst) { 4345 PrevInst = Inst; 4346 continue; 4347 } 4348 4349 // Update LiveValues. 4350 LiveValues.erase(PrevInst); 4351 for (auto &J : PrevInst->operands()) { 4352 if (isa<Instruction>(&*J) && getTreeEntry(&*J)) 4353 LiveValues.insert(cast<Instruction>(&*J)); 4354 } 4355 4356 LLVM_DEBUG({do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4357 dbgs() << "SLP: #LV: " << LiveValues.size();do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4358 for (auto *X : LiveValues)do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4359 dbgs() << " " << X->getName();do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4360 dbgs() << ", Looking at ";do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4361 Inst->dump();do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) 4362 })do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { { dbgs() << "SLP: #LV: " << LiveValues .size(); for (auto *X : LiveValues) dbgs() << " " << X->getName(); dbgs() << ", Looking at "; Inst->dump (); }; } } while (false) ; 4363 4364 // Now find the sequence of instructions between PrevInst and Inst. 4365 unsigned NumCalls = 0; 4366 BasicBlock::reverse_iterator InstIt = ++Inst->getIterator().getReverse(), 4367 PrevInstIt = 4368 PrevInst->getIterator().getReverse(); 4369 while (InstIt != PrevInstIt) { 4370 if (PrevInstIt == PrevInst->getParent()->rend()) { 4371 PrevInstIt = Inst->getParent()->rbegin(); 4372 continue; 4373 } 4374 4375 // Debug information does not impact spill cost. 4376 if ((isa<CallInst>(&*PrevInstIt) && 4377 !isa<DbgInfoIntrinsic>(&*PrevInstIt)) && 4378 &*PrevInstIt != PrevInst) 4379 NumCalls++; 4380 4381 ++PrevInstIt; 4382 } 4383 4384 if (NumCalls) { 4385 SmallVector<Type*, 4> V; 4386 for (auto *II : LiveValues) { 4387 auto *ScalarTy = II->getType(); 4388 if (auto *VectorTy = dyn_cast<FixedVectorType>(ScalarTy)) 4389 ScalarTy = VectorTy->getElementType(); 4390 V.push_back(FixedVectorType::get(ScalarTy, BundleWidth)); 4391 } 4392 Cost += NumCalls * TTI->getCostOfKeepingLiveOverCall(V); 4393 } 4394 4395 PrevInst = Inst; 4396 } 4397 4398 return Cost; 4399} 4400 4401InstructionCost BoUpSLP::getTreeCost(ArrayRef<Value *> VectorizedVals) { 4402 InstructionCost Cost = 0; 4403 LLVM_DEBUG(dbgs() << "SLP: Calculating cost for tree of size "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Calculating cost for tree of size " << VectorizableTree.size() << ".\n"; } } while ( false) 4404 << VectorizableTree.size() << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Calculating cost for tree of size " << VectorizableTree.size() << ".\n"; } } while ( false) ; 4405 4406 unsigned BundleWidth = VectorizableTree[0]->Scalars.size(); 4407 4408 for (unsigned I = 0, E = VectorizableTree.size(); I < E; ++I) { 4409 TreeEntry &TE = *VectorizableTree[I].get(); 4410 4411 InstructionCost C = getEntryCost(&TE, VectorizedVals); 4412 Cost += C; 4413 LLVM_DEBUG(dbgs() << "SLP: Adding cost " << Cdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for bundle that starts with " << *TE.Scalars[0] << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4414 << " for bundle that starts with " << *TE.Scalars[0]do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for bundle that starts with " << *TE.Scalars[0] << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4415 << ".\n"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for bundle that starts with " << *TE.Scalars[0] << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4416 << "SLP: Current total cost = " << Cost << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for bundle that starts with " << *TE.Scalars[0] << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) ; 4417 } 4418 4419 SmallPtrSet<Value *, 16> ExtractCostCalculated; 4420 InstructionCost ExtractCost = 0; 4421 SmallBitVector IsIdentity; 4422 SmallVector<unsigned> VF; 4423 SmallVector<SmallVector<int>> ShuffleMask; 4424 SmallVector<Value *> FirstUsers; 4425 SmallVector<APInt> DemandedElts; 4426 for (ExternalUser &EU : ExternalUses) { 4427 // We only add extract cost once for the same scalar. 4428 if (!ExtractCostCalculated.insert(EU.Scalar).second) 4429 continue; 4430 4431 // Uses by ephemeral values are free (because the ephemeral value will be 4432 // removed prior to code generation, and so the extraction will be 4433 // removed as well). 4434 if (EphValues.count(EU.User)) 4435 continue; 4436 4437 // No extract cost for vector "scalar" 4438 if (isa<FixedVectorType>(EU.Scalar->getType())) 4439 continue; 4440 4441 // Already counted the cost for external uses when tried to adjust the cost 4442 // for extractelements, no need to add it again. 4443 if (isa<ExtractElementInst>(EU.Scalar)) 4444 continue; 4445 4446 // If found user is an insertelement, do not calculate extract cost but try 4447 // to detect it as a final shuffled/identity match. 4448 if (EU.User && isa<InsertElementInst>(EU.User)) { 4449 if (auto *FTy = dyn_cast<FixedVectorType>(EU.User->getType())) { 4450 Optional<int> InsertIdx = getInsertIndex(EU.User, 0); 4451 if (!InsertIdx || *InsertIdx == UndefMaskElem) 4452 continue; 4453 Value *VU = EU.User; 4454 auto *It = find_if(FirstUsers, [VU](Value *V) { 4455 // Checks if 2 insertelements are from the same buildvector. 4456 if (VU->getType() != V->getType()) 4457 return false; 4458 auto *IE1 = cast<InsertElementInst>(VU); 4459 auto *IE2 = cast<InsertElementInst>(V); 4460 // Go though of insertelement instructions trying to find either VU as 4461 // the original vector for IE2 or V as the original vector for IE1. 4462 do { 4463 if (IE1 == VU || IE2 == V) 4464 return true; 4465 if (IE1) 4466 IE1 = dyn_cast<InsertElementInst>(IE1->getOperand(0)); 4467 if (IE2) 4468 IE2 = dyn_cast<InsertElementInst>(IE2->getOperand(0)); 4469 } while (IE1 || IE2); 4470 return false; 4471 }); 4472 int VecId = -1; 4473 if (It == FirstUsers.end()) { 4474 VF.push_back(FTy->getNumElements()); 4475 ShuffleMask.emplace_back(VF.back(), UndefMaskElem); 4476 FirstUsers.push_back(EU.User); 4477 DemandedElts.push_back(APInt::getNullValue(VF.back())); 4478 IsIdentity.push_back(true); 4479 VecId = FirstUsers.size() - 1; 4480 } else { 4481 VecId = std::distance(FirstUsers.begin(), It); 4482 } 4483 int Idx = *InsertIdx; 4484 ShuffleMask[VecId][Idx] = EU.Lane; 4485 IsIdentity.set(IsIdentity.test(VecId) & 4486 (EU.Lane == Idx || EU.Lane == UndefMaskElem)); 4487 DemandedElts[VecId].setBit(Idx); 4488 } 4489 } 4490 4491 // If we plan to rewrite the tree in a smaller type, we will need to sign 4492 // extend the extracted value back to the original type. Here, we account 4493 // for the extract and the added cost of the sign extend if needed. 4494 auto *VecTy = FixedVectorType::get(EU.Scalar->getType(), BundleWidth); 4495 auto *ScalarRoot = VectorizableTree[0]->Scalars[0]; 4496 if (MinBWs.count(ScalarRoot)) { 4497 auto *MinTy = IntegerType::get(F->getContext(), MinBWs[ScalarRoot].first); 4498 auto Extend = 4499 MinBWs[ScalarRoot].second ? Instruction::SExt : Instruction::ZExt; 4500 VecTy = FixedVectorType::get(MinTy, BundleWidth); 4501 ExtractCost += TTI->getExtractWithExtendCost(Extend, EU.Scalar->getType(), 4502 VecTy, EU.Lane); 4503 } else { 4504 ExtractCost += 4505 TTI->getVectorInstrCost(Instruction::ExtractElement, VecTy, EU.Lane); 4506 } 4507 } 4508 4509 InstructionCost SpillCost = getSpillCost(); 4510 Cost += SpillCost + ExtractCost; 4511 for (int I = 0, E = FirstUsers.size(); I < E; ++I) { 4512 if (!IsIdentity.test(I)) { 4513 InstructionCost C = TTI->getShuffleCost( 4514 TTI::SK_PermuteSingleSrc, 4515 cast<FixedVectorType>(FirstUsers[I]->getType()), ShuffleMask[I]); 4516 LLVM_DEBUG(dbgs() << "SLP: Adding cost " << Cdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of insertelement external users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4517 << " for final shuffle of insertelement external users "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of insertelement external users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4518 << *VectorizableTree.front()->Scalars.front() << ".\n"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of insertelement external users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4519 << "SLP: Current total cost = " << Cost << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of insertelement external users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) ; 4520 Cost += C; 4521 } 4522 unsigned VF = ShuffleMask[I].size(); 4523 for (int &Mask : ShuffleMask[I]) 4524 Mask = (Mask == UndefMaskElem ? 0 : VF) + Mask; 4525 InstructionCost C = TTI->getShuffleCost( 4526 TTI::SK_PermuteTwoSrc, cast<FixedVectorType>(FirstUsers[I]->getType()), 4527 ShuffleMask[I]); 4528 LLVM_DEBUG(do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4529 dbgs()do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4530 << "SLP: Adding cost " << Cdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4531 << " for final shuffle of vector node and external insertelement users "do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4532 << *VectorizableTree.front()->Scalars.front() << ".\n"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4533 << "SLP: Current total cost = " << Cost << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Adding cost " << C << " for final shuffle of vector node and external insertelement users " << *VectorizableTree.front()->Scalars.front() << ".\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) ; 4534 Cost += C; 4535 InstructionCost InsertCost = TTI->getScalarizationOverhead( 4536 cast<FixedVectorType>(FirstUsers[I]->getType()), DemandedElts[I], 4537 /*Insert*/ true, 4538 /*Extract*/ false); 4539 Cost -= InsertCost; 4540 LLVM_DEBUG(dbgs() << "SLP: subtracting the cost " << InsertCostdo { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: subtracting the cost " << InsertCost << " for insertelements gather.\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4541 << " for insertelements gather.\n"do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: subtracting the cost " << InsertCost << " for insertelements gather.\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) 4542 << "SLP: Current total cost = " << Cost << "\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: subtracting the cost " << InsertCost << " for insertelements gather.\n" << "SLP: Current total cost = " << Cost << "\n"; } } while (false) ; 4543 } 4544 4545#ifndef NDEBUG 4546 SmallString<256> Str; 4547 { 4548 raw_svector_ostream OS(Str); 4549 OS << "SLP: Spill Cost = " << SpillCost << ".\n" 4550 << "SLP: Extract Cost = " << ExtractCost << ".\n" 4551 << "SLP: Total Cost = " << Cost << ".\n"; 4552 } 4553 LLVM_DEBUG(dbgs() << Str)do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << Str; } } while (false) ; 4554 if (ViewSLPTree) 4555 ViewGraph(this, "SLP" + F->getName(), false, Str); 4556#endif 4557 4558 return Cost; 4559} 4560 4561Optional<TargetTransformInfo::ShuffleKind> 4562BoUpSLP::isGatherShuffledEntry(const TreeEntry *TE, SmallVectorImpl<int> &Mask, 4563 SmallVectorImpl<const TreeEntry *> &Entries) { 4564 // TODO: currently checking only for Scalars in the tree entry, need to count 4565 // reused elements too for better cost estimation. 4566 Mask.assign(TE->Scalars.size(), UndefMaskElem); 4567 Entries.clear(); 4568 // Build a lists of values to tree entries. 4569 DenseMap<Value *, SmallPtrSet<const TreeEntry *, 4>> ValueToTEs; 4570 for (const std::unique_ptr<TreeEntry> &EntryPtr : VectorizableTree) { 4571 if (EntryPtr.get() == TE) 4572 break; 4573 if (EntryPtr->State != TreeEntry::NeedToGather) 4574 continue; 4575 for (Value *V : EntryPtr->Scalars) 4576 ValueToTEs.try_emplace(V).first->getSecond().insert(EntryPtr.get()); 4577 } 4578 // Find all tree entries used by the gathered values. If no common entries 4579 // found - not a shuffle. 4580 // Here we build a set of tree nodes for each gathered value and trying to 4581 // find the intersection between these sets. If we have at least one common 4582 // tree node for each gathered value - we have just a permutation of the 4583 // single vector. If we have 2 different sets, we're in situation where we 4584 // have a permutation of 2 input vectors. 4585 SmallVector<SmallPtrSet<const TreeEntry *, 4>> UsedTEs; 4586 DenseMap<Value *, int> UsedValuesEntry; 4587 for (Value *V : TE->Scalars) { 4588 if (isa<UndefValue>(V)) 4589 continue; 4590 // Build a list of tree entries where V is used. 4591 SmallPtrSet<const TreeEntry *, 4> VToTEs; 4592 auto It = ValueToTEs.find(V); 4593 if (It != ValueToTEs.end()) 4594 VToTEs = It->second; 4595 if (const TreeEntry *VTE = getTreeEntry(V)) 4596 VToTEs.insert(VTE); 4597 if (VToTEs.empty()) 4598 return None; 4599 if (UsedTEs.empty()) { 4600 // The first iteration, just insert the list of nodes to vector. 4601 UsedTEs.push_back(VToTEs); 4602 } else { 4603 // Need to check if there are any previously used tree nodes which use V. 4604 // If there are no such nodes, consider that we have another one input 4605 // vector. 4606 SmallPtrSet<const TreeEntry *, 4> SavedVToTEs(VToTEs); 4607 unsigned Idx = 0; 4608 for (SmallPtrSet<const TreeEntry *, 4> &Set : UsedTEs) { 4609 // Do we have a non-empty intersection of previously listed tree entries 4610 // and tree entries using current V? 4611 set_intersect(VToTEs, Set); 4612 if (!VToTEs.empty()) { 4613 // Yes, write the new subset and continue analysis for the next 4614 // scalar. 4615 Set.swap(VToTEs); 4616 break; 4617 } 4618 VToTEs = SavedVToTEs; 4619 ++Idx; 4620 } 4621 // No non-empty intersection found - need to add a second set of possible 4622 // source vectors. 4623 if (Idx == UsedTEs.size()) { 4624 // If the number of input vectors is greater than 2 - not a permutation, 4625 // fallback to the regular gather. 4626 if (UsedTEs.size() == 2) 4627 return None; 4628 UsedTEs.push_back(SavedVToTEs); 4629 Idx = UsedTEs.size() - 1; 4630 } 4631 UsedValuesEntry.try_emplace(V, Idx); 4632 } 4633 } 4634 4635 unsigned VF = 0; 4636 if (UsedTEs.size() == 1) { 4637 // Try to find the perfect match in another gather node at first. 4638 auto It = find_if(UsedTEs.front(), [TE](const TreeEntry *EntryPtr) { 4639 return EntryPtr->isSame(TE->Scalars); 4640 }); 4641 if (It != UsedTEs.front().end()) { 4642 Entries.push_back(*It); 4643 std::iota(Mask.begin(), Mask.end(), 0); 4644 return TargetTransformInfo::SK_PermuteSingleSrc; 4645 } 4646 // No perfect match, just shuffle, so choose the first tree node. 4647 Entries.push_back(*UsedTEs.front().begin()); 4648 } else { 4649 // Try to find nodes with the same vector factor. 4650 assert(UsedTEs.size() == 2 && "Expected at max 2 permuted entries.")(static_cast <bool> (UsedTEs.size() == 2 && "Expected at max 2 permuted entries." ) ? void (0) : __assert_fail ("UsedTEs.size() == 2 && \"Expected at max 2 permuted entries.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4650, __extension__ __PRETTY_FUNCTION__)) ; 4651 // FIXME: Shall be replaced by GetVF function once non-power-2 patch is 4652 // landed. 4653 auto &&GetVF = [](const TreeEntry *TE) { 4654 if (!TE->ReuseShuffleIndices.empty()) 4655 return TE->ReuseShuffleIndices.size(); 4656 return TE->Scalars.size(); 4657 }; 4658 DenseMap<int, const TreeEntry *> VFToTE; 4659 for (const TreeEntry *TE : UsedTEs.front()) 4660 VFToTE.try_emplace(GetVF(TE), TE); 4661 for (const TreeEntry *TE : UsedTEs.back()) { 4662 auto It = VFToTE.find(GetVF(TE)); 4663 if (It != VFToTE.end()) { 4664 VF = It->first; 4665 Entries.push_back(It->second); 4666 Entries.push_back(TE); 4667 break; 4668 } 4669 } 4670 // No 2 source vectors with the same vector factor - give up and do regular 4671 // gather. 4672 if (Entries.empty()) 4673 return None; 4674 } 4675 4676 // Build a shuffle mask for better cost estimation and vector emission. 4677 for (int I = 0, E = TE->Scalars.size(); I < E; ++I) { 4678 Value *V = TE->Scalars[I]; 4679 if (isa<UndefValue>(V)) 4680 continue; 4681 unsigned Idx = UsedValuesEntry.lookup(V); 4682 const TreeEntry *VTE = Entries[Idx]; 4683 int FoundLane = findLaneForValue(VTE->Scalars, VTE->ReuseShuffleIndices, V); 4684 Mask[I] = Idx * VF + FoundLane; 4685 // Extra check required by isSingleSourceMaskImpl function (called by 4686 // ShuffleVectorInst::isSingleSourceMask). 4687 if (Mask[I] >= 2 * E) 4688 return None; 4689 } 4690 switch (Entries.size()) { 4691 case 1: 4692 return TargetTransformInfo::SK_PermuteSingleSrc; 4693 case 2: 4694 return TargetTransformInfo::SK_PermuteTwoSrc; 4695 default: 4696 break; 4697 } 4698 return None; 4699} 4700 4701InstructionCost 4702BoUpSLP::getGatherCost(FixedVectorType *Ty, 4703 const DenseSet<unsigned> &ShuffledIndices) const { 4704 unsigned NumElts = Ty->getNumElements(); 4705 APInt DemandedElts = APInt::getNullValue(NumElts); 4706 for (unsigned I = 0; I < NumElts; ++I) 4707 if (!ShuffledIndices.count(I)) 4708 DemandedElts.setBit(I); 4709 InstructionCost Cost = 4710 TTI->getScalarizationOverhead(Ty, DemandedElts, /*Insert*/ true, 4711 /*Extract*/ false); 4712 if (!ShuffledIndices.empty()) 4713 Cost += TTI->getShuffleCost(TargetTransformInfo::SK_PermuteSingleSrc, Ty); 4714 return Cost; 4715} 4716 4717InstructionCost BoUpSLP::getGatherCost(ArrayRef<Value *> VL) const { 4718 // Find the type of the operands in VL. 4719 Type *ScalarTy = VL[0]->getType(); 4720 if (StoreInst *SI = dyn_cast<StoreInst>(VL[0])) 4721 ScalarTy = SI->getValueOperand()->getType(); 4722 auto *VecTy = FixedVectorType::get(ScalarTy, VL.size()); 4723 // Find the cost of inserting/extracting values from the vector. 4724 // Check if the same elements are inserted several times and count them as 4725 // shuffle candidates. 4726 DenseSet<unsigned> ShuffledElements; 4727 DenseSet<Value *> UniqueElements; 4728 // Iterate in reverse order to consider insert elements with the high cost. 4729 for (unsigned I = VL.size(); I > 0; --I) { 4730 unsigned Idx = I - 1; 4731 if (isConstant(VL[Idx])) 4732 continue; 4733 if (!UniqueElements.insert(VL[Idx]).second) 4734 ShuffledElements.insert(Idx); 4735 } 4736 return getGatherCost(VecTy, ShuffledElements); 4737} 4738 4739// Perform operand reordering on the instructions in VL and return the reordered 4740// operands in Left and Right. 4741void BoUpSLP::reorderInputsAccordingToOpcode(ArrayRef<Value *> VL, 4742 SmallVectorImpl<Value *> &Left, 4743 SmallVectorImpl<Value *> &Right, 4744 const DataLayout &DL, 4745 ScalarEvolution &SE, 4746 const BoUpSLP &R) { 4747 if (VL.empty()) 4748 return; 4749 VLOperands Ops(VL, DL, SE, R); 4750 // Reorder the operands in place. 4751 Ops.reorder(); 4752 Left = Ops.getVL(0); 4753 Right = Ops.getVL(1); 4754} 4755 4756void BoUpSLP::setInsertPointAfterBundle(const TreeEntry *E) { 4757 // Get the basic block this bundle is in. All instructions in the bundle 4758 // should be in this block. 4759 auto *Front = E->getMainOp(); 4760 auto *BB = Front->getParent(); 4761 assert(llvm::all_of(E->Scalars, [=](Value *V) -> bool {(static_cast <bool> (llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })) ? void (0) : __assert_fail ("llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4764, __extension__ __PRETTY_FUNCTION__)) 4762 auto *I = cast<Instruction>(V);(static_cast <bool> (llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })) ? void (0) : __assert_fail ("llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4764, __extension__ __PRETTY_FUNCTION__)) 4763 return !E->isOpcodeOrAlt(I) || I->getParent() == BB;(static_cast <bool> (llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })) ? void (0) : __assert_fail ("llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4764, __extension__ __PRETTY_FUNCTION__)) 4764 }))(static_cast <bool> (llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })) ? void (0) : __assert_fail ("llvm::all_of(E->Scalars, [=](Value *V) -> bool { auto *I = cast<Instruction>(V); return !E->isOpcodeOrAlt(I) || I->getParent() == BB; })" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4764, __extension__ __PRETTY_FUNCTION__)) ; 4765 4766 // The last instruction in the bundle in program order. 4767 Instruction *LastInst = nullptr; 4768 4769 // Find the last instruction. The common case should be that BB has been 4770 // scheduled, and the last instruction is VL.back(). So we start with 4771 // VL.back() and iterate over schedule data until we reach the end of the 4772 // bundle. The end of the bundle is marked by null ScheduleData. 4773 if (BlocksSchedules.count(BB)) { 4774 auto *Bundle = 4775 BlocksSchedules[BB]->getScheduleData(E->isOneOf(E->Scalars.back())); 4776 if (Bundle && Bundle->isPartOfBundle()) 4777 for (; Bundle; Bundle = Bundle->NextInBundle) 4778 if (Bundle->OpValue == Bundle->Inst) 4779 LastInst = Bundle->Inst; 4780 } 4781 4782 // LastInst can still be null at this point if there's either not an entry 4783 // for BB in BlocksSchedules or there's no ScheduleData available for 4784 // VL.back(). This can be the case if buildTree_rec aborts for various 4785 // reasons (e.g., the maximum recursion depth is reached, the maximum region 4786 // size is reached, etc.). ScheduleData is initialized in the scheduling 4787 // "dry-run". 4788 // 4789 // If this happens, we can still find the last instruction by brute force. We 4790 // iterate forwards from Front (inclusive) until we either see all 4791 // instructions in the bundle or reach the end of the block. If Front is the 4792 // last instruction in program order, LastInst will be set to Front, and we 4793 // will visit all the remaining instructions in the block. 4794 // 4795 // One of the reasons we exit early from buildTree_rec is to place an upper 4796 // bound on compile-time. Thus, taking an additional compile-time hit here is 4797 // not ideal. However, this should be exceedingly rare since it requires that 4798 // we both exit early from buildTree_rec and that the bundle be out-of-order 4799 // (causing us to iterate all the way to the end of the block). 4800 if (!LastInst) { 4801 SmallPtrSet<Value *, 16> Bundle(E->Scalars.begin(), E->Scalars.end()); 4802 for (auto &I : make_range(BasicBlock::iterator(Front), BB->end())) { 4803 if (Bundle.erase(&I) && E->isOpcodeOrAlt(&I)) 4804 LastInst = &I; 4805 if (Bundle.empty()) 4806 break; 4807 } 4808 } 4809 assert(LastInst && "Failed to find last instruction in bundle")(static_cast <bool> (LastInst && "Failed to find last instruction in bundle" ) ? void (0) : __assert_fail ("LastInst && \"Failed to find last instruction in bundle\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4809, __extension__ __PRETTY_FUNCTION__)) ; 4810 4811 // Set the insertion point after the last instruction in the bundle. Set the 4812 // debug location to Front. 4813 Builder.SetInsertPoint(BB, ++LastInst->getIterator()); 4814 Builder.SetCurrentDebugLocation(Front->getDebugLoc()); 4815} 4816 4817Value *BoUpSLP::gather(ArrayRef<Value *> VL) { 4818 // List of instructions/lanes from current block and/or the blocks which are 4819 // part of the current loop. These instructions will be inserted at the end to 4820 // make it possible to optimize loops and hoist invariant instructions out of 4821 // the loops body with better chances for success. 4822 SmallVector<std::pair<Value *, unsigned>, 4> PostponedInsts; 4823 SmallSet<int, 4> PostponedIndices; 4824 Loop *L = LI->getLoopFor(Builder.GetInsertBlock()); 4825 auto &&CheckPredecessor = [](BasicBlock *InstBB, BasicBlock *InsertBB) { 4826 SmallPtrSet<BasicBlock *, 4> Visited; 4827 while (InsertBB && InsertBB != InstBB && Visited.insert(InsertBB).second) 4828 InsertBB = InsertBB->getSinglePredecessor(); 4829 return InsertBB && InsertBB == InstBB; 4830 }; 4831 for (int I = 0, E = VL.size(); I < E; ++I) { 4832 if (auto *Inst = dyn_cast<Instruction>(VL[I])) 4833 if ((CheckPredecessor(Inst->getParent(), Builder.GetInsertBlock()) || 4834 getTreeEntry(Inst) || (L && (L->contains(Inst)))) && 4835 PostponedIndices.insert(I).second) 4836 PostponedInsts.emplace_back(Inst, I); 4837 } 4838 4839 auto &&CreateInsertElement = [this](Value *Vec, Value *V, unsigned Pos) { 4840 // No need to insert undefs elements - exit. 4841 if (isa<UndefValue>(V)) 4842 return Vec; 4843 Vec = Builder.CreateInsertElement(Vec, V, Builder.getInt32(Pos)); 4844 auto *InsElt = dyn_cast<InsertElementInst>(Vec); 4845 if (!InsElt) 4846 return Vec; 4847 GatherSeq.insert(InsElt); 4848 CSEBlocks.insert(InsElt->getParent()); 4849 // Add to our 'need-to-extract' list. 4850 if (TreeEntry *Entry = getTreeEntry(V)) { 4851 // Find which lane we need to extract. 4852 unsigned FoundLane = 4853 std::distance(Entry->Scalars.begin(), find(Entry->Scalars, V)); 4854 assert(FoundLane < Entry->Scalars.size() && "Couldn't find extract lane")(static_cast <bool> (FoundLane < Entry->Scalars.size () && "Couldn't find extract lane") ? void (0) : __assert_fail ("FoundLane < Entry->Scalars.size() && \"Couldn't find extract lane\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4854, __extension__ __PRETTY_FUNCTION__)) ; 4855 if (!Entry->ReuseShuffleIndices.empty()) { 4856 FoundLane = std::distance(Entry->ReuseShuffleIndices.begin(), 4857 find(Entry->ReuseShuffleIndices, FoundLane)); 4858 } 4859 ExternalUses.emplace_back(V, InsElt, FoundLane); 4860 } 4861 return Vec; 4862 }; 4863 Value *Val0 = 4864 isa<StoreInst>(VL[0]) ? cast<StoreInst>(VL[0])->getValueOperand() : VL[0]; 4865 FixedVectorType *VecTy = FixedVectorType::get(Val0->getType(), VL.size()); 4866 Value *Vec = PoisonValue::get(VecTy); 4867 for (int I = 0, E = VL.size(); I < E; ++I) { 4868 if (PostponedIndices.contains(I)) 4869 continue; 4870 Vec = CreateInsertElement(Vec, VL[I], I); 4871 } 4872 // Append instructions, which are/may be part of the loop, in the end to make 4873 // it possible to hoist non-loop-based instructions. 4874 for (const std::pair<Value *, unsigned> &Pair : PostponedInsts) 4875 Vec = CreateInsertElement(Vec, Pair.first, Pair.second); 4876 4877 return Vec; 4878} 4879 4880namespace { 4881/// Merges shuffle masks and emits final shuffle instruction, if required. 4882class ShuffleInstructionBuilder { 4883 IRBuilderBase &Builder; 4884 const unsigned VF = 0; 4885 bool IsFinalized = false; 4886 SmallVector<int, 4> Mask; 4887 4888public: 4889 ShuffleInstructionBuilder(IRBuilderBase &Builder, unsigned VF) 4890 : Builder(Builder), VF(VF) {} 4891 4892 /// Adds a mask, inverting it before applying. 4893 void addInversedMask(ArrayRef<unsigned> SubMask) { 4894 if (SubMask.empty()) 4895 return; 4896 SmallVector<int, 4> NewMask; 4897 inversePermutation(SubMask, NewMask); 4898 addMask(NewMask); 4899 } 4900 4901 /// Functions adds masks, merging them into single one. 4902 void addMask(ArrayRef<unsigned> SubMask) { 4903 SmallVector<int, 4> NewMask(SubMask.begin(), SubMask.end()); 4904 addMask(NewMask); 4905 } 4906 4907 void addMask(ArrayRef<int> SubMask) { 4908 if (SubMask.empty()) 4909 return; 4910 if (Mask.empty()) { 4911 Mask.append(SubMask.begin(), SubMask.end()); 4912 return; 4913 } 4914 SmallVector<int, 4> NewMask(SubMask.size(), SubMask.size()); 4915 int TermValue = std::min(Mask.size(), SubMask.size()); 4916 for (int I = 0, E = SubMask.size(); I < E; ++I) { 4917 if (SubMask[I] >= TermValue || SubMask[I] == UndefMaskElem || 4918 Mask[SubMask[I]] >= TermValue) { 4919 NewMask[I] = UndefMaskElem; 4920 continue; 4921 } 4922 NewMask[I] = Mask[SubMask[I]]; 4923 } 4924 Mask.swap(NewMask); 4925 } 4926 4927 Value *finalize(Value *V) { 4928 IsFinalized = true; 4929 unsigned ValueVF = cast<FixedVectorType>(V->getType())->getNumElements(); 4930 if (VF == ValueVF && Mask.empty()) 4931 return V; 4932 SmallVector<int, 4> NormalizedMask(VF, UndefMaskElem); 4933 std::iota(NormalizedMask.begin(), NormalizedMask.end(), 0); 4934 addMask(NormalizedMask); 4935 4936 if (VF == ValueVF && ShuffleVectorInst::isIdentityMask(Mask)) 4937 return V; 4938 return Builder.CreateShuffleVector(V, Mask, "shuffle"); 4939 } 4940 4941 ~ShuffleInstructionBuilder() { 4942 assert((IsFinalized || Mask.empty()) &&(static_cast <bool> ((IsFinalized || Mask.empty()) && "Shuffle construction must be finalized.") ? void (0) : __assert_fail ("(IsFinalized || Mask.empty()) && \"Shuffle construction must be finalized.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4943, __extension__ __PRETTY_FUNCTION__)) 4943 "Shuffle construction must be finalized.")(static_cast <bool> ((IsFinalized || Mask.empty()) && "Shuffle construction must be finalized.") ? void (0) : __assert_fail ("(IsFinalized || Mask.empty()) && \"Shuffle construction must be finalized.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4943, __extension__ __PRETTY_FUNCTION__)) ; 4944 } 4945}; 4946} // namespace 4947 4948Value *BoUpSLP::vectorizeTree(ArrayRef<Value *> VL) { 4949 unsigned VF = VL.size(); 4950 InstructionsState S = getSameOpcode(VL); 4951 if (S.getOpcode()) { 4952 if (TreeEntry *E = getTreeEntry(S.OpValue)) 4953 if (E->isSame(VL)) { 4954 Value *V = vectorizeTree(E); 4955 if (VF != cast<FixedVectorType>(V->getType())->getNumElements()) { 4956 if (!E->ReuseShuffleIndices.empty()) { 4957 // Reshuffle to get only unique values. 4958 // If some of the scalars are duplicated in the vectorization tree 4959 // entry, we do not vectorize them but instead generate a mask for 4960 // the reuses. But if there are several users of the same entry, 4961 // they may have different vectorization factors. This is especially 4962 // important for PHI nodes. In this case, we need to adapt the 4963 // resulting instruction for the user vectorization factor and have 4964 // to reshuffle it again to take only unique elements of the vector. 4965 // Without this code the function incorrectly returns reduced vector 4966 // instruction with the same elements, not with the unique ones. 4967 4968 // block: 4969 // %phi = phi <2 x > { .., %entry} {%shuffle, %block} 4970 // %2 = shuffle <2 x > %phi, %poison, <4 x > <0, 0, 1, 1> 4971 // ... (use %2) 4972 // %shuffle = shuffle <2 x> %2, poison, <2 x> {0, 2} 4973 // br %block 4974 SmallVector<int> UniqueIdxs; 4975 SmallSet<int, 4> UsedIdxs; 4976 int Pos = 0; 4977 int Sz = VL.size(); 4978 for (int Idx : E->ReuseShuffleIndices) { 4979 if (Idx != Sz && UsedIdxs.insert(Idx).second) 4980 UniqueIdxs.emplace_back(Pos); 4981 ++Pos; 4982 } 4983 assert(VF >= UsedIdxs.size() && "Expected vectorization factor "(static_cast <bool> (VF >= UsedIdxs.size() && "Expected vectorization factor " "less than original vector size." ) ? void (0) : __assert_fail ("VF >= UsedIdxs.size() && \"Expected vectorization factor \" \"less than original vector size.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4984, __extension__ __PRETTY_FUNCTION__)) 4984 "less than original vector size.")(static_cast <bool> (VF >= UsedIdxs.size() && "Expected vectorization factor " "less than original vector size." ) ? void (0) : __assert_fail ("VF >= UsedIdxs.size() && \"Expected vectorization factor \" \"less than original vector size.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4984, __extension__ __PRETTY_FUNCTION__)) ; 4985 UniqueIdxs.append(VF - UsedIdxs.size(), UndefMaskElem); 4986 V = Builder.CreateShuffleVector(V, UniqueIdxs, "shrink.shuffle"); 4987 } else { 4988 assert(VF < cast<FixedVectorType>(V->getType())->getNumElements() &&(static_cast <bool> (VF < cast<FixedVectorType> (V->getType())->getNumElements() && "Expected vectorization factor less " "than original vector size.") ? void (0) : __assert_fail ("VF < cast<FixedVectorType>(V->getType())->getNumElements() && \"Expected vectorization factor less \" \"than original vector size.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4990, __extension__ __PRETTY_FUNCTION__)) 4989 "Expected vectorization factor less "(static_cast <bool> (VF < cast<FixedVectorType> (V->getType())->getNumElements() && "Expected vectorization factor less " "than original vector size.") ? void (0) : __assert_fail ("VF < cast<FixedVectorType>(V->getType())->getNumElements() && \"Expected vectorization factor less \" \"than original vector size.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4990, __extension__ __PRETTY_FUNCTION__)) 4990 "than original vector size.")(static_cast <bool> (VF < cast<FixedVectorType> (V->getType())->getNumElements() && "Expected vectorization factor less " "than original vector size.") ? void (0) : __assert_fail ("VF < cast<FixedVectorType>(V->getType())->getNumElements() && \"Expected vectorization factor less \" \"than original vector size.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 4990, __extension__ __PRETTY_FUNCTION__)) ; 4991 SmallVector<int> UniformMask(VF, 0); 4992 std::iota(UniformMask.begin(), UniformMask.end(), 0); 4993 V = Builder.CreateShuffleVector(V, UniformMask, "shrink.shuffle"); 4994 } 4995 } 4996 return V; 4997 } 4998 } 4999 5000 // Check that every instruction appears once in this bundle. 5001 SmallVector<int> ReuseShuffleIndicies; 5002 SmallVector<Value *> UniqueValues; 5003 if (VL.size() > 2) { 5004 DenseMap<Value *, unsigned> UniquePositions; 5005 unsigned NumValues = 5006 std::distance(VL.begin(), find_if(reverse(VL), [](Value *V) { 5007 return !isa<UndefValue>(V); 5008 }).base()); 5009 VF = std::max<unsigned>(VF, PowerOf2Ceil(NumValues)); 5010 int UniqueVals = 0; 5011 bool HasUndefs = false; 5012 for (Value *V : VL.drop_back(VL.size() - VF)) { 5013 if (isa<UndefValue>(V)) { 5014 ReuseShuffleIndicies.emplace_back(UndefMaskElem); 5015 HasUndefs = true; 5016 continue; 5017 } 5018 if (isConstant(V)) { 5019 ReuseShuffleIndicies.emplace_back(UniqueValues.size()); 5020 UniqueValues.emplace_back(V); 5021 continue; 5022 } 5023 auto Res = UniquePositions.try_emplace(V, UniqueValues.size()); 5024 ReuseShuffleIndicies.emplace_back(Res.first->second); 5025 if (Res.second) { 5026 UniqueValues.emplace_back(V); 5027 ++UniqueVals; 5028 } 5029 } 5030 if (HasUndefs && UniqueVals == 1 && UniqueValues.size() == 1) { 5031 // Emit pure splat vector. 5032 // FIXME: why it is not identified as an identity. 5033 unsigned NumUndefs = count(ReuseShuffleIndicies, UndefMaskElem); 5034 if (NumUndefs == ReuseShuffleIndicies.size() - 1) 5035 ReuseShuffleIndicies.append(VF - ReuseShuffleIndicies.size(), 5036 UndefMaskElem); 5037 else 5038 ReuseShuffleIndicies.assign(VF, 0); 5039 } else if (UniqueValues.size() >= VF - 1 || UniqueValues.size() <= 1) { 5040 ReuseShuffleIndicies.clear(); 5041 UniqueValues.clear(); 5042 UniqueValues.append(VL.begin(), std::next(VL.begin(), NumValues)); 5043 } 5044 UniqueValues.append(VF - UniqueValues.size(), 5045 UndefValue::get(VL[0]->getType())); 5046 VL = UniqueValues; 5047 } 5048 5049 ShuffleInstructionBuilder ShuffleBuilder(Builder, VF); 5050 Value *Vec = gather(VL); 5051 if (!ReuseShuffleIndicies.empty()) { 5052 ShuffleBuilder.addMask(ReuseShuffleIndicies); 5053 Vec = ShuffleBuilder.finalize(Vec); 5054 if (auto *I = dyn_cast<Instruction>(Vec)) { 5055 GatherSeq.insert(I); 5056 CSEBlocks.insert(I->getParent()); 5057 } 5058 } 5059 return Vec; 5060} 5061 5062Value *BoUpSLP::vectorizeTree(TreeEntry *E) { 5063 IRBuilder<>::InsertPointGuard Guard(Builder); 5064 5065 if (E->VectorizedValue) { 5066 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *E->Scalars[0] << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *E->Scalars[0] << ".\n"; } } while (false) ; 5067 return E->VectorizedValue; 5068 } 5069 5070 bool NeedToShuffleReuses = !E->ReuseShuffleIndices.empty(); 5071 unsigned VF = E->Scalars.size(); 5072 if (NeedToShuffleReuses) 5073 VF = E->ReuseShuffleIndices.size(); 5074 ShuffleInstructionBuilder ShuffleBuilder(Builder, VF); 5075 if (E->State == TreeEntry::NeedToGather) { 5076 setInsertPointAfterBundle(E); 5077 Value *Vec; 5078 SmallVector<int> Mask; 5079 SmallVector<const TreeEntry *> Entries; 5080 Optional<TargetTransformInfo::ShuffleKind> Shuffle = 5081 isGatherShuffledEntry(E, Mask, Entries); 5082 if (Shuffle.hasValue()) { 5083 assert((Entries.size() == 1 || Entries.size() == 2) &&(static_cast <bool> ((Entries.size() == 1 || Entries.size () == 2) && "Expected shuffle of 1 or 2 entries.") ? void (0) : __assert_fail ("(Entries.size() == 1 || Entries.size() == 2) && \"Expected shuffle of 1 or 2 entries.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5084, __extension__ __PRETTY_FUNCTION__)) 5084 "Expected shuffle of 1 or 2 entries.")(static_cast <bool> ((Entries.size() == 1 || Entries.size () == 2) && "Expected shuffle of 1 or 2 entries.") ? void (0) : __assert_fail ("(Entries.size() == 1 || Entries.size() == 2) && \"Expected shuffle of 1 or 2 entries.\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5084, __extension__ __PRETTY_FUNCTION__)) ; 5085 Vec = Builder.CreateShuffleVector(Entries.front()->VectorizedValue, 5086 Entries.back()->VectorizedValue, Mask); 5087 } else { 5088 Vec = gather(E->Scalars); 5089 } 5090 if (NeedToShuffleReuses) { 5091 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5092 Vec = ShuffleBuilder.finalize(Vec); 5093 if (auto *I = dyn_cast<Instruction>(Vec)) { 5094 GatherSeq.insert(I); 5095 CSEBlocks.insert(I->getParent()); 5096 } 5097 } 5098 E->VectorizedValue = Vec; 5099 return Vec; 5100 } 5101 5102 assert((E->State == TreeEntry::Vectorize ||(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5104, __extension__ __PRETTY_FUNCTION__)) 5103 E->State == TreeEntry::ScatterVectorize) &&(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5104, __extension__ __PRETTY_FUNCTION__)) 5104 "Unhandled state")(static_cast <bool> ((E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && "Unhandled state" ) ? void (0) : __assert_fail ("(E->State == TreeEntry::Vectorize || E->State == TreeEntry::ScatterVectorize) && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5104, __extension__ __PRETTY_FUNCTION__)) ; 5105 unsigned ShuffleOrOp = 5106 E->isAltShuffle() ? (unsigned)Instruction::ShuffleVector : E->getOpcode(); 5107 Instruction *VL0 = E->getMainOp(); 5108 Type *ScalarTy = VL0->getType(); 5109 if (auto *Store = dyn_cast<StoreInst>(VL0)) 5110 ScalarTy = Store->getValueOperand()->getType(); 5111 else if (auto *IE = dyn_cast<InsertElementInst>(VL0)) 5112 ScalarTy = IE->getOperand(1)->getType(); 5113 auto *VecTy = FixedVectorType::get(ScalarTy, E->Scalars.size()); 5114 switch (ShuffleOrOp) { 5115 case Instruction::PHI: { 5116 auto *PH = cast<PHINode>(VL0); 5117 Builder.SetInsertPoint(PH->getParent()->getFirstNonPHI()); 5118 Builder.SetCurrentDebugLocation(PH->getDebugLoc()); 5119 PHINode *NewPhi = Builder.CreatePHI(VecTy, PH->getNumIncomingValues()); 5120 Value *V = NewPhi; 5121 if (NeedToShuffleReuses) 5122 V = Builder.CreateShuffleVector(V, E->ReuseShuffleIndices, "shuffle"); 5123 5124 E->VectorizedValue = V; 5125 5126 // PHINodes may have multiple entries from the same block. We want to 5127 // visit every block once. 5128 SmallPtrSet<BasicBlock*, 4> VisitedBBs; 5129 5130 for (unsigned i = 0, e = PH->getNumIncomingValues(); i < e; ++i) { 5131 ValueList Operands; 5132 BasicBlock *IBB = PH->getIncomingBlock(i); 5133 5134 if (!VisitedBBs.insert(IBB).second) { 5135 NewPhi->addIncoming(NewPhi->getIncomingValueForBlock(IBB), IBB); 5136 continue; 5137 } 5138 5139 Builder.SetInsertPoint(IBB->getTerminator()); 5140 Builder.SetCurrentDebugLocation(PH->getDebugLoc()); 5141 Value *Vec = vectorizeTree(E->getOperand(i)); 5142 NewPhi->addIncoming(Vec, IBB); 5143 } 5144 5145 assert(NewPhi->getNumIncomingValues() == PH->getNumIncomingValues() &&(static_cast <bool> (NewPhi->getNumIncomingValues() == PH->getNumIncomingValues() && "Invalid number of incoming values" ) ? void (0) : __assert_fail ("NewPhi->getNumIncomingValues() == PH->getNumIncomingValues() && \"Invalid number of incoming values\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5146, __extension__ __PRETTY_FUNCTION__)) 5146 "Invalid number of incoming values")(static_cast <bool> (NewPhi->getNumIncomingValues() == PH->getNumIncomingValues() && "Invalid number of incoming values" ) ? void (0) : __assert_fail ("NewPhi->getNumIncomingValues() == PH->getNumIncomingValues() && \"Invalid number of incoming values\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5146, __extension__ __PRETTY_FUNCTION__)) ; 5147 return V; 5148 } 5149 5150 case Instruction::ExtractElement: { 5151 Value *V = E->getSingleOperand(0); 5152 Builder.SetInsertPoint(VL0); 5153 ShuffleBuilder.addInversedMask(E->ReorderIndices); 5154 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5155 V = ShuffleBuilder.finalize(V); 5156 E->VectorizedValue = V; 5157 return V; 5158 } 5159 case Instruction::ExtractValue: { 5160 auto *LI = cast<LoadInst>(E->getSingleOperand(0)); 5161 Builder.SetInsertPoint(LI); 5162 auto *PtrTy = PointerType::get(VecTy, LI->getPointerAddressSpace()); 5163 Value *Ptr = Builder.CreateBitCast(LI->getOperand(0), PtrTy); 5164 LoadInst *V = Builder.CreateAlignedLoad(VecTy, Ptr, LI->getAlign()); 5165 Value *NewV = propagateMetadata(V, E->Scalars); 5166 ShuffleBuilder.addInversedMask(E->ReorderIndices); 5167 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5168 NewV = ShuffleBuilder.finalize(NewV); 5169 E->VectorizedValue = NewV; 5170 return NewV; 5171 } 5172 case Instruction::InsertElement: { 5173 Builder.SetInsertPoint(VL0); 5174 Value *V = vectorizeTree(E->getOperand(1)); 5175 5176 const unsigned NumElts = 5177 cast<FixedVectorType>(VL0->getType())->getNumElements(); 5178 const unsigned NumScalars = E->Scalars.size(); 5179 5180 // Create InsertVector shuffle if necessary 5181 Instruction *FirstInsert = nullptr; 5182 bool IsIdentity = true; 5183 unsigned Offset = UINT_MAX(2147483647 *2U +1U); 5184 for (unsigned I = 0; I < NumScalars; ++I) { 5185 Value *Scalar = E->Scalars[I]; 5186 if (!FirstInsert && 5187 !is_contained(E->Scalars, cast<Instruction>(Scalar)->getOperand(0))) 5188 FirstInsert = cast<Instruction>(Scalar); 5189 Optional<int> InsertIdx = getInsertIndex(Scalar, 0); 5190 if (!InsertIdx || *InsertIdx == UndefMaskElem) 5191 continue; 5192 unsigned Idx = *InsertIdx; 5193 if (Idx < Offset) { 5194 Offset = Idx; 5195 IsIdentity &= I == 0; 5196 } else { 5197 assert(Idx >= Offset && "Failed to find vector index offset")(static_cast <bool> (Idx >= Offset && "Failed to find vector index offset" ) ? void (0) : __assert_fail ("Idx >= Offset && \"Failed to find vector index offset\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5197, __extension__ __PRETTY_FUNCTION__)) ; 5198 IsIdentity &= Idx - Offset == I; 5199 } 5200 } 5201 assert(Offset < NumElts && "Failed to find vector index offset")(static_cast <bool> (Offset < NumElts && "Failed to find vector index offset" ) ? void (0) : __assert_fail ("Offset < NumElts && \"Failed to find vector index offset\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5201, __extension__ __PRETTY_FUNCTION__)) ; 5202 5203 // Create shuffle to resize vector 5204 SmallVector<int> Mask(NumElts, UndefMaskElem); 5205 if (!IsIdentity) { 5206 for (unsigned I = 0; I < NumScalars; ++I) { 5207 Value *Scalar = E->Scalars[I]; 5208 Optional<int> InsertIdx = getInsertIndex(Scalar, 0); 5209 if (!InsertIdx || *InsertIdx == UndefMaskElem) 5210 continue; 5211 Mask[*InsertIdx - Offset] = I; 5212 } 5213 } else { 5214 std::iota(Mask.begin(), std::next(Mask.begin(), NumScalars), 0); 5215 } 5216 if (!IsIdentity || NumElts != NumScalars) 5217 V = Builder.CreateShuffleVector(V, UndefValue::get(V->getType()), Mask); 5218 5219 if (NumElts != NumScalars) { 5220 SmallVector<int> InsertMask(NumElts); 5221 std::iota(InsertMask.begin(), InsertMask.end(), 0); 5222 for (unsigned I = 0; I < NumElts; I++) { 5223 if (Mask[I] != UndefMaskElem) 5224 InsertMask[Offset + I] = NumElts + I; 5225 } 5226 5227 V = Builder.CreateShuffleVector( 5228 FirstInsert->getOperand(0), V, InsertMask, 5229 cast<Instruction>(E->Scalars.back())->getName()); 5230 } 5231 5232 ++NumVectorInstructions; 5233 E->VectorizedValue = V; 5234 return V; 5235 } 5236 case Instruction::ZExt: 5237 case Instruction::SExt: 5238 case Instruction::FPToUI: 5239 case Instruction::FPToSI: 5240 case Instruction::FPExt: 5241 case Instruction::PtrToInt: 5242 case Instruction::IntToPtr: 5243 case Instruction::SIToFP: 5244 case Instruction::UIToFP: 5245 case Instruction::Trunc: 5246 case Instruction::FPTrunc: 5247 case Instruction::BitCast: { 5248 setInsertPointAfterBundle(E); 5249 5250 Value *InVec = vectorizeTree(E->getOperand(0)); 5251 5252 if (E->VectorizedValue) { 5253 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n"; } } while (false) ; 5254 return E->VectorizedValue; 5255 } 5256 5257 auto *CI = cast<CastInst>(VL0); 5258 Value *V = Builder.CreateCast(CI->getOpcode(), InVec, VecTy); 5259 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5260 V = ShuffleBuilder.finalize(V); 5261 5262 E->VectorizedValue = V; 5263 ++NumVectorInstructions; 5264 return V; 5265 } 5266 case Instruction::FCmp: 5267 case Instruction::ICmp: { 5268 setInsertPointAfterBundle(E); 5269 5270 Value *L = vectorizeTree(E->getOperand(0)); 5271 Value *R = vectorizeTree(E->getOperand(1)); 5272 5273 if (E->VectorizedValue) { 5274 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n"; } } while (false) ; 5275 return E->VectorizedValue; 5276 } 5277 5278 CmpInst::Predicate P0 = cast<CmpInst>(VL0)->getPredicate(); 5279 Value *V = Builder.CreateCmp(P0, L, R); 5280 propagateIRFlags(V, E->Scalars, VL0); 5281 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5282 V = ShuffleBuilder.finalize(V); 5283 5284 E->VectorizedValue = V; 5285 ++NumVectorInstructions; 5286 return V; 5287 } 5288 case Instruction::Select: { 5289 setInsertPointAfterBundle(E); 5290 5291 Value *Cond = vectorizeTree(E->getOperand(0)); 5292 Value *True = vectorizeTree(E->getOperand(1)); 5293 Value *False = vectorizeTree(E->getOperand(2)); 5294 5295 if (E->VectorizedValue) { 5296 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n"; } } while (false) ; 5297 return E->VectorizedValue; 5298 } 5299 5300 Value *V = Builder.CreateSelect(Cond, True, False); 5301 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5302 V = ShuffleBuilder.finalize(V); 5303 5304 E->VectorizedValue = V; 5305 ++NumVectorInstructions; 5306 return V; 5307 } 5308 case Instruction::FNeg: { 5309 setInsertPointAfterBundle(E); 5310 5311 Value *Op = vectorizeTree(E->getOperand(0)); 5312 5313 if (E->VectorizedValue) { 5314 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n"; } } while (false) ; 5315 return E->VectorizedValue; 5316 } 5317 5318 Value *V = Builder.CreateUnOp( 5319 static_cast<Instruction::UnaryOps>(E->getOpcode()), Op); 5320 propagateIRFlags(V, E->Scalars, VL0); 5321 if (auto *I = dyn_cast<Instruction>(V)) 5322 V = propagateMetadata(I, E->Scalars); 5323 5324 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5325 V = ShuffleBuilder.finalize(V); 5326 5327 E->VectorizedValue = V; 5328 ++NumVectorInstructions; 5329 5330 return V; 5331 } 5332 case Instruction::Add: 5333 case Instruction::FAdd: 5334 case Instruction::Sub: 5335 case Instruction::FSub: 5336 case Instruction::Mul: 5337 case Instruction::FMul: 5338 case Instruction::UDiv: 5339 case Instruction::SDiv: 5340 case Instruction::FDiv: 5341 case Instruction::URem: 5342 case Instruction::SRem: 5343 case Instruction::FRem: 5344 case Instruction::Shl: 5345 case Instruction::LShr: 5346 case Instruction::AShr: 5347 case Instruction::And: 5348 case Instruction::Or: 5349 case Instruction::Xor: { 5350 setInsertPointAfterBundle(E); 5351 5352 Value *LHS = vectorizeTree(E->getOperand(0)); 5353 Value *RHS = vectorizeTree(E->getOperand(1)); 5354 5355 if (E->VectorizedValue) { 5356 LLVM_DEBUG(dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n")do { if (::llvm::DebugFlag && ::llvm::isCurrentDebugType ("SLP")) { dbgs() << "SLP: Diamond merged for " << *VL0 << ".\n"; } } while (false) ; 5357 return E->VectorizedValue; 5358 } 5359 5360 Value *V = Builder.CreateBinOp( 5361 static_cast<Instruction::BinaryOps>(E->getOpcode()), LHS, 5362 RHS); 5363 propagateIRFlags(V, E->Scalars, VL0); 5364 if (auto *I = dyn_cast<Instruction>(V)) 5365 V = propagateMetadata(I, E->Scalars); 5366 5367 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5368 V = ShuffleBuilder.finalize(V); 5369 5370 E->VectorizedValue = V; 5371 ++NumVectorInstructions; 5372 5373 return V; 5374 } 5375 case Instruction::Load: { 5376 // Loads are inserted at the head of the tree because we don't want to 5377 // sink them all the way down past store instructions. 5378 bool IsReorder = E->updateStateIfReorder(); 5379 if (IsReorder) 5380 VL0 = E->getMainOp(); 5381 setInsertPointAfterBundle(E); 5382 5383 LoadInst *LI = cast<LoadInst>(VL0); 5384 Instruction *NewLI; 5385 unsigned AS = LI->getPointerAddressSpace(); 5386 Value *PO = LI->getPointerOperand(); 5387 if (E->State == TreeEntry::Vectorize) { 5388 5389 Value *VecPtr = Builder.CreateBitCast(PO, VecTy->getPointerTo(AS)); 5390 5391 // The pointer operand uses an in-tree scalar so we add the new BitCast 5392 // to ExternalUses list to make sure that an extract will be generated 5393 // in the future. 5394 if (getTreeEntry(PO)) 5395 ExternalUses.emplace_back(PO, cast<User>(VecPtr), 0); 5396 5397 NewLI = Builder.CreateAlignedLoad(VecTy, VecPtr, LI->getAlign()); 5398 } else { 5399 assert(E->State == TreeEntry::ScatterVectorize && "Unhandled state")(static_cast <bool> (E->State == TreeEntry::ScatterVectorize && "Unhandled state") ? void (0) : __assert_fail ("E->State == TreeEntry::ScatterVectorize && \"Unhandled state\"" , "/build/llvm-toolchain-snapshot-13~++20210610111127+c5ffc6f8bd6a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp" , 5399, __extension__ __PRETTY_FUNCTION__)) ; 5400 Value *VecPtr = vectorizeTree(E->getOperand(0)); 5401 // Use the minimum alignment of the gathered loads. 5402 Align CommonAlignment = LI->getAlign(); 5403 for (Value *V : E->Scalars) 5404 CommonAlignment = 5405 commonAlignment(CommonAlignment, cast<LoadInst>(V)->getAlign()); 5406 NewLI = Builder.CreateMaskedGather(VecPtr, CommonAlignment); 5407 } 5408 Value *V = propagateMetadata(NewLI, E->Scalars); 5409 5410 ShuffleBuilder.addInversedMask(E->ReorderIndices); 5411 ShuffleBuilder.addMask(E->ReuseShuffleIndices); 5412 V = ShuffleBuilder.finalize(V); 5413 E->VectorizedValue = V; 5414 ++NumVectorInstructions; 5415 return V; 5416 } 5417 case Instruction::Store: { 5418 bool IsReorder = !E->ReorderIndices.empty(); 5419 auto *SI = cast<StoreInst>( 5420 IsReorder ? E->Scalars[E->ReorderIndices.front()] : VL0); 5421 unsigned AS = SI->getPointerAddressSpace(); 5422 5423 setInsertPointAfterBundle(E); 5424 5425 Value *VecValue = vectorizeTree(E->getOperand(0)); 5426 ShuffleBuilder.addMask(E->ReorderIndices); 5427 VecValue = ShuffleBuilder.finalize(VecValue); 5428 5429 Value *ScalarPtr = SI->getPointerOperand(); 5430 Value *VecPtr = Builder.CreateBitCast( 5431 ScalarPtr, VecValue->getType()->getPointerTo(AS)); 5432 StoreInst *ST = Builder.CreateAlignedStore(VecValue, VecPtr, 5433 SI->getAlign()); 5434 5435 // The pointer operand uses an in-tree scalar, so add the new BitCast to 5436 // ExternalUses to make sure that an extract will be generated in the 5437
__label__pos
0.943072
The source code for this tutorial can be found on GitHub, and a live demo can be found here. What You Will Learn In this tutorial we will create a simple real-time chat application. It will feature a chat-panel that stores messages received after you join, a list of currently connected users, and an input field to send messages from. We will be using WebSockets for this, as WebSockets provides us with full-duplex communication channels over a single TCP connection, meaning we won’t have to make additional HTTP requests to send and receive messages. A WebSocket connection stays open, greatly reducing latency (and complexity). Create a Spark Maven Project First, we need to create a Spark Maven project (→ Tutorial) Dependencies Other than Spark, we are going to use a simple Java HTML Builder (j2html) to generate our chat messages. Our POM.xml should have the following dependencies: <dependencies> <dependency> <groupId>com.sparkjava</groupId> <artifactId>spark-core</artifactId> <version>2.3</version> </dependency> <dependency> <groupId>com.j2html</groupId> <artifactId>j2html</artifactId> <version>0.7</version> </dependency> </dependencies> Creating the Java Application We need to keep track of all our users and assign them usernames. We create a map (userUsernameMap) that maps sessions to usernames, an int for the next username (nextUserNumber), and the Spark server code: public class Chat { // this map is shared between sessions and threads, so it needs to be thread-safe (http://stackoverflow.com/a/2688817) static Map<Session, String> userUsernameMap = new ConcurrentHashMap<>(); static int nextUserNumber = 1; //Used for creating the next username public static void main(String[] args) { staticFileLocation("/public"); //index.html is served at localhost:4567 (default port) webSocket("/chat", ChatWebSocketHandler.class); init(); } } We also need to create a few methods for sending messages to all our connected users. We will only send messages to users whose session has the status open (Session::isOpen). We use a stream and a filter to reduce our list (the keySet of our userUsernameMap), then send out a JSON structure containing a HTML message and a list of usernames (the values of our userUsernameMap): //Sends a message from one user to all users, along with a list of current usernames public static void broadcastMessage(String sender, String message) { userUsernameMap.keySet().stream().filter(Session::isOpen).forEach(session -> { try { session.getRemote().sendString(String.valueOf(new JSONObject() .put("userMessage", createHtmlMessageFromSender(sender, message)) .put("userlist", userUsernameMap.values()) )); } catch (Exception e) { e.printStackTrace(); } }); } To create the HTML, we will use j2html. We will create an article-tag containing the name of the sender, the message itself, and a timestamp: //Builds a HTML element with a sender-name, a message, and a timestamp, private static String createHtmlMessageFromSender(String sender, String message) { return article().with( b(sender + " says:"), p(message), span().withClass("timestamp").withText(new SimpleDateFormat("HH:mm:ss").format(new Date())) ).render(); } Setting up Our WebSocketHandler The approach we will use is very straightforward: Add the user to our userUsernameMap when he connects, remove him when he disconnects, and send all his messages to all users. Since we want usernames, but don’t want complexity, we will generate usernames based on when someone connects to the server: @WebSocket public class ChatWebSocketHandler { private String sender, msg; @OnWebSocketConnect public void onConnect(Session user) throws Exception { String username = "User" + Chat.nextUserNumber++; Chat.userUsernameMap.put(user, username); Chat.broadcastMessage(sender = "Server", msg = (username + " joined the chat")); } @OnWebSocketClose public void onClose(Session user, int statusCode, String reason) { String username = Chat.userUsernameMap.get(user); Chat.userUsernameMap.remove(user); Chat.broadcastMessage(sender = "Server", msg = (username + " left the chat")); } @OnWebSocketMessage public void onMessage(Session user, String message) { Chat.broadcastMessage(sender = Chat.userUsernameMap.get(user), msg = message); } } That’s it for the Java part, the rest is HTML and JavaScript. Building a JavaScript Client In order to use demonstrate that our application works, we can build a JavaScript client. First we create our index.html: <!DOCTYPE html> <html> <head> <meta name="viewport" content="width=device-width, initial-scale=1"> <title>WebsSockets</title> <link rel="stylesheet" href="style.css"> </head> <body> <div id="chatControls"> <input id="message" placeholder="Type your message"> <button id="send">Send</button> </div> <ul id="userlist"> <!-- Built by JS --> </ul> <div id="chat"> <!-- Built by JS --> </div> <script src="websocketDemo.js"></script> </body> </html> As you can see, we reference a stylesheet called style.css, which can be found on GitHub. The final step needed for completing our chat application is creating websocketDemo.js: //Establish the WebSocket connection and set up event handlers var webSocket = new WebSocket("ws://" + location.hostname + ":" + location.port + "/chat/"); webSocket.onmessage = function (msg) { updateChat(msg); }; webSocket.onclose = function () { alert("WebSocket connection closed") }; //Send message if "Send" is clicked id("send").addEventListener("click", function () { sendMessage(id("message").value); }); //Send message if enter is pressed in the input field id("message").addEventListener("keypress", function (e) { if (e.keyCode === 13) { sendMessage(e.target.value); } }); //Send a message if it's not empty, then clear the input field function sendMessage(message) { if (message !== "") { webSocket.send(message); id("message").value = ""; } } //Update the chat-panel, and the list of connected users function updateChat(msg) { var data = JSON.parse(msg.data); insert("chat", data.userMessage); id("userlist").innerHTML = ""; data.userlist.forEach(function (user) { insert("userlist", "<li>" + user + "</li>"); }); } //Helper function for inserting HTML as the first child of an element function insert(targetId, message) { id(targetId).insertAdjacentHTML("afterbegin", message); } //Helper function for selecting element by id function id(id) { return document.getElementById(id); } And that’s it! Now try opening localhost:4567 in a couple of different browser windows (that you can see simultaneously) and talk to yourself. Conlusion Well, that was fast! We have a working real-time chat application implemented without polling, written in a total of less than 100 lines of Java and JavaScript. The implementation is very basic though, and we should at least split up the sending of the userlist and the messages (so that we don’t rebuild the user list every time anyone sends a message), but since the focus of this tutorial was supposed to be on WebSockets, I chose to do the implementation as minimal as I could be comfortable with. The source code for this tutorial can be found on GitHub.
__label__pos
0.96901
Java Swing Tutorial - Java JComboBox.hidePopup() Syntax JComboBox.hidePopup() has the following syntax. public void hidePopup() Example In the following code shows how to use JComboBox.hidePopup() method. //from w w w . j a v a2 s . co m import javax.swing.JComboBox; public class Main { public static void main(String[] argv) throws Exception { String[] items = { "item1", "item2" }; JComboBox cb = new JComboBox(items); cb.hidePopup(); } }
__label__pos
0.991101
[BUGFIX] Filelist: Only show paste buttons if clipObj is set [Packages/TYPO3.CMS.git] / typo3 / sysext / filelist / Classes / FileList.php 1 <?php 2 namespace TYPO3\CMS\Filelist; 3 4 /*************************************************************** 5 * Copyright notice 6 * 7 * (c) 1999-2013 Kasper Skårhøj ([email protected]) 8 * All rights reserved 9 * 10 * This script is part of the TYPO3 project. The TYPO3 project is 11 * free software; you can redistribute it and/or modify 12 * it under the terms of the GNU General Public License as published by 13 * the Free Software Foundation; either version 2 of the License, or 14 * (at your option) any later version. 15 * 16 * The GNU General Public License can be found at 17 * http://www.gnu.org/copyleft/gpl.html. 18 * A copy is found in the textfile GPL.txt and important notices to the license 19 * from the author is found in LICENSE.txt distributed with these scripts. 20 * 21 * 22 * This script is distributed in the hope that it will be useful, 23 * but WITHOUT ANY WARRANTY; without even the implied warranty of 24 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 25 * GNU General Public License for more details. 26 * 27 * This copyright notice MUST APPEAR in all copies of the script! 28 ***************************************************************/ 29 30 use TYPO3\CMS\Core\Resource\FolderInterface; 31 32 /** 33 * Class for rendering of File>Filelist 34 * 35 * @author Kasper Skårhøj <[email protected]> 36 */ 37 class FileList extends \TYPO3\CMS\Backend\RecordList\AbstractRecordList { 38 39 /** 40 * Default Max items shown 41 * 42 * @todo Define visibility 43 */ 44 public $iLimit = 40; 45 46 /** 47 * Boolean. Thumbnails on records containing files (pictures) 48 * 49 * @todo Define visibility 50 */ 51 public $thumbs = 0; 52 53 /** 54 * @todo Define visibility 55 */ 56 public $widthGif = '<img src="clear.gif" width="1" height="1" hspace="165" alt="" />'; 57 58 /** 59 * Max length of strings 60 * 61 * @todo Define visibility 62 */ 63 public $fixedL = 30; 64 65 /** 66 * @todo Define visibility 67 */ 68 public $script = ''; 69 70 /** 71 * If TRUE click menus are generated on files and folders 72 * 73 * @todo Define visibility 74 */ 75 public $clickMenus = 1; 76 77 /** 78 * The field to sort by 79 * 80 * @todo Define visibility 81 */ 82 public $sort = ''; 83 84 /** 85 * Reverse sorting flag 86 * 87 * @todo Define visibility 88 */ 89 public $sortRev = 1; 90 91 /** 92 * @todo Define visibility 93 */ 94 public $firstElementNumber = 0; 95 96 /** 97 * @todo Define visibility 98 */ 99 public $clipBoard = 0; 100 101 /** 102 * @todo Define visibility 103 */ 104 public $bigControlPanel = 0; 105 106 /** 107 * @todo Define visibility 108 */ 109 public $JScode = ''; 110 111 /** 112 * @todo Define visibility 113 */ 114 public $HTMLcode = ''; 115 116 /** 117 * @todo Define visibility 118 */ 119 public $totalbytes = 0; 120 121 /** 122 * @todo Define visibility 123 */ 124 public $dirs = array(); 125 126 /** 127 * @todo Define visibility 128 */ 129 public $files = array(); 130 131 /** 132 * @todo Define visibility 133 */ 134 public $path = ''; 135 136 /** 137 * @var \TYPO3\CMS\Core\Resource\Folder 138 */ 139 protected $folderObject; 140 141 /** 142 * Counting the elements no matter what 143 * 144 * @todo Define visibility 145 */ 146 public $eCounter = 0; 147 148 /** 149 * @todo Define visibility 150 */ 151 public $dirCounter = 0; 152 153 /** 154 * @todo Define visibility 155 */ 156 public $totalItems = ''; 157 158 /** 159 * @todo Define visibility 160 */ 161 public $CBnames = array(); 162 163 /** 164 * Initialization of class 165 * 166 * @param \TYPO3\CMS\Core\Resource\Folder $folderObject The folder to work on 167 * @param integer $pointer Pointer 168 * @param boolean $sort Sorting column 169 * @param boolean $sortRev Sorting direction 170 * @param boolean $bigControlPanel Show clipboard flag 171 * @return void 172 * @todo Define visibility 173 */ 174 public function start(\TYPO3\CMS\Core\Resource\Folder $folderObject, $pointer, $sort, $sortRev, $clipBoard = FALSE, $bigControlPanel = FALSE) { 175 $this->script = \TYPO3\CMS\Backend\Utility\BackendUtility::getModuleUrl('file_list'); 176 $this->folderObject = $folderObject; 177 $this->counter = 0; 178 $this->totalbytes = 0; 179 $this->JScode = ''; 180 $this->HTMLcode = ''; 181 $this->path = $folderObject->getIdentifier(); 182 $this->sort = $sort; 183 $this->sortRev = $sortRev; 184 $this->firstElementNumber = $pointer; 185 $this->clipBoard = $clipBoard; 186 $this->bigControlPanel = $bigControlPanel; 187 // Setting the maximum length of the filenames to the user's settings or minimum 30 (= $this->fixedL) 188 $this->fixedL = max($this->fixedL, $GLOBALS['BE_USER']->uc['titleLen']); 189 $GLOBALS['LANG']->includeLLFile('EXT:lang/locallang_common.xlf'); 190 } 191 192 /** 193 * Reading files and directories, counting elements and generating the list in ->HTMLcode 194 * 195 * @return void 196 * @todo Define visibility 197 */ 198 public function generateList() { 199 $this->HTMLcode .= $this->getTable('fileext,tstamp,size,rw,_REF_'); 200 } 201 202 /** 203 * Return the buttons used by the file list to include in the top header 204 * 205 * @param \TYPO3\CMS\Core\Resource\Folder $folderObject 206 * @return array 207 */ 208 public function getButtonsAndOtherMarkers(\TYPO3\CMS\Core\Resource\Folder $folderObject) { 209 $otherMarkers = array( 210 'PAGE_ICON' => '', 211 'TITLE' => '' 212 ); 213 $buttons = array( 214 'level_up' => '', 215 'refresh' => '', 216 'title' => '', 217 'page_icon' => '', 218 'PASTE' => '' 219 ); 220 // Makes the code for the foldericon in the top 221 if ($folderObject) { 222 list($_, $icon, $path) = $this->dirData($folderObject); 223 $title = htmlspecialchars($folderObject->getIdentifier()); 224 // Start compiling the HTML 225 // @todo: how to fix this? $title = $GLOBALS['SOBE']->basicFF->blindPath($title); 226 // If this is some subpage under the mount root.... 227 if ($folderObject->getStorage()->isWithinFileMountBoundaries($folderObject)) { 228 // The icon with link 229 $otherMarkers['PAGE_ICON'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon($icon, array('title' => $title)); 230 $buttons['level_up'] = $this->linkWrapDir(\TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-view-go-up', array('title' => $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:labels.upOneLevel', 1))), $folderObject); 231 // No HTML specialchars here - HTML like <strong> </strong> is allowed 232 $otherMarkers['TITLE'] .= \TYPO3\CMS\Core\Utility\GeneralUtility::removeXSS(\TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($title, -($this->fixedL + 20))); 233 } else { 234 // This is the root page 235 $otherMarkers['PAGE_ICON'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('apps-filetree-root'); 236 $otherMarkers['TITLE'] .= htmlspecialchars(\TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($title, -($this->fixedL + 20))); 237 } 238 if ($this->clickMenus) { 239 $otherMarkers['PAGE_ICON'] = $GLOBALS['SOBE']->doc->wrapClickMenuOnIcon($otherMarkers['PAGE_ICON'], $folderObject->getCombinedIdentifier()); 240 } 241 // Add paste button if clipboard is initialized 242 if ($this->clipObj instanceof \TYPO3\CMS\Backend\Clipboard\Clipboard) { 243 $elFromTable = $this->clipObj->elFromTable('_FILE'); 244 if (count($elFromTable)) { 245 $buttons['PASTE'] = '<a href="' . htmlspecialchars($this->clipObj->pasteUrl('_FILE', $this->folderObject->getCombinedIdentifier())) . '" onclick="return ' . htmlspecialchars($this->clipObj->confirmMsg('_FILE', $this->path, 'into', $elFromTable)) . '" title="' . $GLOBALS['LANG']->getLL('clip_paste', TRUE) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-document-paste-after') . '</a>'; 246 } 247 } 248 249 } 250 $buttons['refresh'] = '<a href="' . htmlspecialchars($this->listURL()) . '" title="' . $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:labels.reload', 1) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-system-refresh') . '</a>'; 251 return array($buttons, $otherMarkers); 252 } 253 254 /** 255 * Wrapping input string in a link with clipboard command. 256 * 257 * @param string $string String to be linked - must be htmlspecialchar'ed / prepared before. 258 * @param string $table table - NOT USED 259 * @param string $cmd "cmd" value 260 * @param string $warning Warning for JS confirm message 261 * @return string Linked string 262 * @todo Define visibility 263 */ 264 public function linkClipboardHeaderIcon($string, $table, $cmd, $warning = '') { 265 $onClickEvent = 'document.dblistForm.cmd.value=\'' . $cmd . '\';document.dblistForm.submit();'; 266 if ($warning) { 267 $onClickEvent = 'if (confirm(' . $GLOBALS['LANG']->JScharCode($warning) . ')){' . $onClickEvent . '}'; 268 } 269 return '<a href="#" onclick="' . htmlspecialchars($onClickEvent) . 'return false;">' . $string . '</a>'; 270 } 271 272 /** 273 * Returns a table with directories and files listed. 274 * 275 * @param array $rowlist Array of files from path 276 * @return string HTML-table 277 * @todo Define visibility 278 */ 279 public function getTable($rowlist) { 280 // TODO use folder methods directly when they support filters 281 $storage = $this->folderObject->getStorage(); 282 $storage->resetFileAndFolderNameFiltersToDefault(); 283 $folders = $storage->getFolderList($this->folderObject->getIdentifier()); 284 $files = $storage->getFileList($this->folderObject->getIdentifier()); 285 // Only render the contents of a browsable storage 286 if ($this->folderObject->getStorage()->isBrowsable()) { 287 $this->sort = trim($this->sort); 288 if ($this->sort !== '') { 289 $filesToSort = array(); 290 foreach ($files as $file) { 291 $fileObject = $storage->getFile($file['identifier']); 292 switch ($this->sort) { 293 case 'size': 294 $sortingKey = $fileObject->getSize(); 295 break; 296 case 'rw': 297 $sortingKey = ($fileObject->checkActionPermission('read') ? 'R' : '' . $fileObject->checkActionPermission('write')) ? 'W' : ''; 298 break; 299 case 'fileext': 300 $sortingKey = $fileObject->getExtension(); 301 break; 302 case 'tstamp': 303 $sortingKey = $fileObject->getModificationTime(); 304 break; 305 default: 306 if ($fileObject->hasProperty($this->sort)) { 307 $sortingKey = $fileObject->getProperty($this->sort); 308 } else { 309 $sortingKey = $fileObject->getName(); 310 } 311 } 312 $i = 0; 313 while (isset($filesToSort[$sortingKey . $i])) { 314 $i++; 315 } 316 $filesToSort[$sortingKey . $i] = $fileObject; 317 } 318 uksort($filesToSort, 'strnatcasecmp'); 319 if (intval($this->sortRev) === 1) { 320 $filesToSort = array_reverse($filesToSort); 321 } 322 $files = $filesToSort; 323 } 324 $this->totalItems = count($folders) + count($files); 325 // Adds the code of files/dirs 326 $out = ''; 327 $titleCol = 'file'; 328 // Cleaning rowlist for duplicates and place the $titleCol as the first column always! 329 $rowlist = \TYPO3\CMS\Core\Utility\GeneralUtility::rmFromList($titleCol, $rowlist); 330 $rowlist = \TYPO3\CMS\Core\Utility\GeneralUtility::uniqueList($rowlist); 331 $rowlist = $rowlist ? $titleCol . ',' . $rowlist : $titleCol; 332 if ($this->bigControlPanel || $this->clipBoard) { 333 $rowlist = str_replace('file,', 'file,_CLIPBOARD_,', $rowlist); 334 } 335 $this->fieldArray = explode(',', $rowlist); 336 $folderObjects = array(); 337 foreach ($folders as $folder) { 338 $folderObjects[] = $storage->getFolder($folder['identifier']); 339 } 340 341 $folderObjects = \TYPO3\CMS\Core\Resource\Utility\ListUtility::resolveSpecialFolderNames($folderObjects); 342 uksort($folderObjects, 'strnatcasecmp'); 343 344 // Directories are added 345 $iOut = $this->formatDirList($folderObjects); 346 if ($iOut) { 347 // Half line is drawn 348 $theData = array( 349 $titleCol => '' 350 ); 351 } 352 // Files are added 353 $iOut .= $this->formatFileList($files, $titleCol); 354 // Header line is drawn 355 $theData = array(); 356 foreach ($this->fieldArray as $v) { 357 if ($v == '_CLIPBOARD_' && $this->clipBoard) { 358 $cells = array(); 359 $table = '_FILE'; 360 $elFromTable = $this->clipObj->elFromTable($table); 361 if (count($elFromTable)) { 362 $cells[] = '<a href="' . htmlspecialchars($this->clipObj->pasteUrl('_FILE', $this->folderObject->getCombinedIdentifier())) . '" onclick="return ' . htmlspecialchars($this->clipObj->confirmMsg('_FILE', $this->path, 'into', $elFromTable)) . '" title="' . $GLOBALS['LANG']->getLL('clip_paste', 1) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-document-paste-after') . '</a>'; 363 } 364 if ($this->clipObj->current != 'normal' && $iOut) { 365 $cells[] = $this->linkClipboardHeaderIcon(\TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-edit-copy', array('title' => $GLOBALS['LANG']->getLL('clip_selectMarked', 1))), $table, 'setCB'); 366 $cells[] = $this->linkClipboardHeaderIcon(\TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-edit-delete', array('title' => $GLOBALS['LANG']->getLL('clip_deleteMarked'))), $table, 'delete', $GLOBALS['LANG']->getLL('clip_deleteMarkedWarning')); 367 $onClick = 'checkOffCB(\'' . implode(',', $this->CBnames) . '\', this); return false;'; 368 $cells[] = '<a class="cbcCheckAll" rel="" href="#" onclick="' . htmlspecialchars($onClick) . '" title="' . $GLOBALS['LANG']->getLL('clip_markRecords', TRUE) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-document-select') . '</a>'; 369 } 370 $theData[$v] = implode('', $cells); 371 } else { 372 // Normal row: 373 $theT = $this->linkWrapSort($GLOBALS['LANG']->getLL('c_' . $v, 1), $this->folderObject->getCombinedIdentifier(), $v); 374 $theData[$v] = $theT; 375 } 376 } 377 378 if (!empty($iOut)) { 379 $out .= '<thead>' . $this->addelement(1, $levelUp, $theData, ' class="t3-row-header"', '') . '</thead>'; 380 $out .= '<tbody>' . $iOut . '</tbody>'; 381 // half line is drawn 382 // finish 383 $out = ' 384 <!-- 385 File list table: 386 --> 387 <table cellpadding="0" cellspacing="0" id="typo3-filelist"> 388 ' . $out . ' 389 </table>'; 390 } 391 } else { 392 /** @var $flashMessage \TYPO3\CMS\Core\Messaging\FlashMessage */ 393 $flashMessage = \TYPO3\CMS\Core\Utility\GeneralUtility::makeInstance('TYPO3\\CMS\\Core\\Messaging\\FlashMessage', $GLOBALS['LANG']->getLL('storageNotBrowsableMessage'), $GLOBALS['LANG']->getLL('storageNotBrowsableTitle'), \TYPO3\CMS\Core\Messaging\FlashMessage::INFO); 394 $out = $flashMessage->render(); 395 } 396 return $out; 397 } 398 399 /** 400 * Gets the number of files and total size of a folder 401 * 402 * @return string 403 * @todo Define visibility 404 */ 405 public function getFolderInfo() { 406 if ($this->counter == 1) { 407 $fileLabel = $GLOBALS['LANG']->getLL('file', TRUE); 408 } else { 409 $fileLabel = $GLOBALS['LANG']->getLL('files', TRUE); 410 } 411 return $this->counter . ' ' . $fileLabel . ', ' . \TYPO3\CMS\Core\Utility\GeneralUtility::formatSize($this->totalbytes, $GLOBALS['LANG']->getLL('byteSizeUnits', TRUE)); 412 } 413 414 /** 415 * This returns tablerows for the directories in the array $items['sorting']. 416 * 417 * @param \TYPO3\CMS\Core\Resource\Folder[] $folders Folders of \TYPO3\CMS\Core\Resource\Folder 418 * @return string HTML table rows. 419 * @todo Define visibility 420 */ 421 public function formatDirList(array $folders) { 422 $out = ''; 423 foreach ($folders as $folderName => $folderObject) { 424 $role = $folderObject->getRole(); 425 if ($role === FolderInterface::ROLE_PROCESSING) { 426 // don't show processing-folder 427 continue; 428 } 429 if ($role !== FolderInterface::ROLE_DEFAULT) { 430 $displayName = '<strong>' . htmlspecialchars($folderName) . '</strong>'; 431 } else { 432 $displayName = htmlspecialchars($folderName); 433 } 434 435 list($flag, $code) = $this->fwd_rwd_nav(); 436 $out .= $code; 437 if ($flag) { 438 // Initialization 439 $this->counter++; 440 list($_, $icon, $path) = $this->dirData($folderObject); 441 // The icon with link 442 $theIcon = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon($icon, array('title' => $folderName)); 443 if ($this->clickMenus) { 444 $theIcon = $GLOBALS['SOBE']->doc->wrapClickMenuOnIcon($theIcon, $folderObject->getCombinedIdentifier()); 445 } 446 // Preparing and getting the data-array 447 $theData = array(); 448 foreach ($this->fieldArray as $field) { 449 switch ($field) { 450 case 'size': 451 $numFiles = $folderObject->getFileCount(); 452 $theData[$field] = $numFiles . ' ' . $GLOBALS['LANG']->getLL(($numFiles === 1 ? 'file' : 'files'), TRUE); 453 break; 454 case 'rw': 455 $theData[$field] = (!$folderObject->checkActionPermission('read') ? ' ' : '<span class="typo3-red"><strong>' . $GLOBALS['LANG']->getLL('read', TRUE) . '</strong></span>') . (!$folderObject->checkActionPermission('write') ? '' : '<span class="typo3-red"><strong>' . $GLOBALS['LANG']->getLL('write', TRUE) . '</strong></span>'); 456 break; 457 case 'fileext': 458 $theData[$field] = $GLOBALS['LANG']->getLL('folder', TRUE); 459 break; 460 case 'tstamp': 461 // @todo: FAL: how to get the mtime info -- $theData[$field] = \TYPO3\CMS\Backend\Utility\BackendUtility::date($theFile['tstamp']); 462 $theData[$field] = '-'; 463 break; 464 case 'file': 465 $theData[$field] = $this->linkWrapDir($displayName, $folderObject); 466 break; 467 case '_CLIPBOARD_': 468 $temp = ''; 469 if ($this->bigControlPanel) { 470 $temp .= $this->makeEdit($folderObject); 471 } 472 $temp .= $this->makeClip($folderObject); 473 $theData[$field] = $temp; 474 break; 475 case '_REF_': 476 $theData[$field] = $this->makeRef($folderObject); 477 break; 478 default: 479 $theData[$field] = \TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($theFile[$field], $this->fixedL); 480 break; 481 } 482 } 483 $out .= $this->addelement(1, $theIcon, $theData, ' class="file_list_normal"'); 484 } 485 $this->eCounter++; 486 $this->dirCounter = $this->eCounter; 487 } 488 return $out; 489 } 490 491 /** 492 * Wraps the directory-titles 493 * 494 * @param string $title String to be wrapped in links 495 * @param \TYPO3\CMS\Core\Resource\Folder $folderObject Folder to work on 496 * @return string HTML 497 * @todo Define visibility 498 */ 499 public function linkWrapDir($title, \TYPO3\CMS\Core\Resource\Folder $folderObject) { 500 $href = $this->backPath . $this->script . '&id=' . rawurlencode($folderObject->getCombinedIdentifier()); 501 $onclick = ' onclick="' . htmlspecialchars(('top.content.nav_frame.hilight_row("file","folder' . \TYPO3\CMS\Core\Utility\GeneralUtility::md5int($folderObject->getCombinedIdentifier()) . '_"+top.fsMod.currentBank)')) . '"'; 502 // Sometimes $code contains plain HTML tags. In such a case the string should not be modified! 503 if (!strcmp($title, strip_tags($title))) { 504 return '<a href="' . htmlspecialchars($href) . '"' . $onclick . ' title="' . htmlspecialchars($title) . '">' . \TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($title, $this->fixedL) . '</a>'; 505 } else { 506 return '<a href="' . htmlspecialchars($href) . '"' . $onclick . '>' . $title . '</a>'; 507 } 508 } 509 510 /** 511 * Wraps filenames in links which opens them in a window IF they are in web-path. 512 * 513 * @param string $code String to be wrapped in links 514 * @param \TYPO3\CMS\Core\Resource\File $fileObject File to be linked 515 * @return string HTML 516 * @todo Define visibility 517 */ 518 public function linkWrapFile($code, \TYPO3\CMS\Core\Resource\File $fileObject) { 519 $fileUrl = $fileObject->getPublicUrl(TRUE); 520 if ($fileUrl) { 521 $aOnClick = 'return top.openUrlInWindow(\'' . $fileUrl . '\', \'WebFile\');'; 522 $code = '<a href="#" title="' . htmlspecialchars($code) . '" onclick="' . htmlspecialchars($aOnClick) . '">' . \TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($code, $this->fixedL) . '</a>'; 523 } 524 return $code; 525 } 526 527 /** 528 * Returns list URL; This is the URL of the current script with id and imagemode parameters, thats all. 529 * The URL however is not relative (with the backpath), otherwise \TYPO3\CMS\Core\Utility\GeneralUtility::sanitizeLocalUrl() would say that 530 * the URL would be invalid 531 * 532 * @return string URL 533 * @todo Define visibility 534 */ 535 public function listURL() { 536 return \TYPO3\CMS\Core\Utility\GeneralUtility::linkThisScript(array( 537 'target' => rawurlencode($this->folderObject->getCombinedIdentifier()), 538 'imagemode' => $this->thumbs 539 )); 540 } 541 542 /** 543 * Returns some data specific for the directories... 544 * 545 * @param \TYPO3\CMS\Core\Resource\Folder $folderObject File information array 546 * @return array (title, icon, path) 547 * @todo Define visibility 548 */ 549 public function dirData(\TYPO3\CMS\Core\Resource\Folder $folderObject) { 550 $title = htmlspecialchars($folderObject->getName()); 551 $icon = 'apps-filetree-folder-default'; 552 $role = $folderObject->getRole(); 553 if ($role === FolderInterface::ROLE_TEMPORARY) { 554 $title = '<strong>' . $GLOBALS['LANG']->getLL('temp', TRUE) . '</strong>'; 555 $icon = 'apps-filetree-folder-temp'; 556 } elseif ($role === FolderInterface::ROLE_RECYCLER) { 557 $icon = 'apps-filetree-folder-recycler'; 558 $title = '<strong>' . $GLOBALS['LANG']->getLL('recycler', TRUE) . '</strong>'; 559 } 560 // Mark the icon as read-only icon if the folder is not writable 561 if ($folderObject->checkActionPermission('write') === FALSE) { 562 $icon = 'apps-filetree-folder-locked'; 563 } 564 return array($title, $icon, $folderObject->getIdentifier()); 565 } 566 567 /** 568 * This returns tablerows for the files in the array $items['sorting']. 569 * 570 * @param \TYPO3\CMS\Core\Resource\File[] $files File items 571 * @return string HTML table rows. 572 * @todo Define visibility 573 */ 574 public function formatFileList(array $files) { 575 $out = ''; 576 foreach ($files as $fileObject) { 577 list($flag, $code) = $this->fwd_rwd_nav(); 578 $out .= $code; 579 if ($flag) { 580 // Initialization 581 $this->counter++; 582 $fileInfo = $fileObject->getStorage()->getFileInfo($fileObject); 583 $this->totalbytes += $fileObject->getSize(); 584 $ext = $fileObject->getExtension(); 585 $fileName = trim($fileObject->getName()); 586 // The icon with link 587 $theIcon = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIconForFile($ext, array('title' => $fileName)); 588 if ($this->clickMenus) { 589 $theIcon = $GLOBALS['SOBE']->doc->wrapClickMenuOnIcon($theIcon, $fileObject->getCombinedIdentifier()); 590 } 591 // Preparing and getting the data-array 592 $theData = array(); 593 foreach ($this->fieldArray as $field) { 594 switch ($field) { 595 case 'size': 596 $theData[$field] = \TYPO3\CMS\Core\Utility\GeneralUtility::formatSize($fileObject->getSize(), $GLOBALS['LANG']->getLL('byteSizeUnits', TRUE)); 597 break; 598 case 'rw': 599 $theData[$field] = '' . (!$fileObject->checkActionPermission('read') ? ' ' : '<span class="typo3-red"><strong>' . $GLOBALS['LANG']->getLL('read', TRUE) . '</strong></span>') . (!$fileObject->checkActionPermission('write') ? '' : '<span class="typo3-red"><strong>' . $GLOBALS['LANG']->getLL('write', TRUE) . '</strong></span>'); 600 break; 601 case 'fileext': 602 $theData[$field] = strtoupper($ext); 603 break; 604 case 'tstamp': 605 $theData[$field] = \TYPO3\CMS\Backend\Utility\BackendUtility::date($fileInfo['mtime']); 606 break; 607 case '_CLIPBOARD_': 608 $temp = ''; 609 if ($this->bigControlPanel) { 610 $temp .= $this->makeEdit($fileObject); 611 } 612 $temp .= $this->makeClip($fileObject); 613 $theData[$field] = $temp; 614 break; 615 case '_REF_': 616 $theData[$field] = $this->makeRef($fileObject); 617 break; 618 case 'file': 619 $theData[$field] = $this->linkWrapFile(htmlspecialchars($fileName), $fileObject); 620 // Thumbnails? 621 if ($this->thumbs && $this->isImage($ext)) { 622 $processedFile = $fileObject->process(\TYPO3\CMS\Core\Resource\ProcessedFile::CONTEXT_IMAGEPREVIEW, array()); 623 if ($processedFile) { 624 $thumbUrl = $processedFile->getPublicUrl(TRUE); 625 $theData[$field] .= '<br /><img src="' . $thumbUrl . '" hspace="2" title="' . htmlspecialchars($fileName) . '" alt="" />'; 626 } 627 } 628 break; 629 default: 630 // @todo: fix the access on the array 631 $theData[$field] = htmlspecialchars(\TYPO3\CMS\Core\Utility\GeneralUtility::fixed_lgd_cs($theFile[$field], $this->fixedL)); 632 break; 633 } 634 } 635 $out .= $this->addelement(1, $theIcon, $theData, ' class="file_list_normal"'); 636 } 637 $this->eCounter++; 638 } 639 return $out; 640 } 641 642 /** 643 * Returns TRUE if $ext is an image-extension according to $GLOBALS['TYPO3_CONF_VARS']['GFX']['imagefile_ext'] 644 * 645 * @param string $ext File extension 646 * @return boolean 647 * @todo Define visibility 648 */ 649 public function isImage($ext) { 650 return \TYPO3\CMS\Core\Utility\GeneralUtility::inList($GLOBALS['TYPO3_CONF_VARS']['GFX']['imagefile_ext'], strtolower($ext)); 651 } 652 653 /** 654 * Wraps the directory-titles ($code) in a link to filelist/mod1/index.php (id=$path) and sorting commands... 655 * 656 * @param string $code String to be wrapped 657 * @param string $folderIdentifier ID (path) 658 * @param string $col Sorting column 659 * @return string HTML 660 * @todo Define visibility 661 */ 662 public function linkWrapSort($code, $folderIdentifier, $col) { 663 if ($this->sort === $col) { 664 // Check reverse sorting 665 $params = '&SET[sort]=' . $col . '&SET[reverse]=' . ($this->sortRev ? '0' : '1'); 666 $sortArrow = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('status-status-sorting-light-' . ($this->sortRev ? 'desc' : 'asc')); 667 } else { 668 $params = '&SET[sort]=' . $col . '&SET[reverse]=0'; 669 $sortArrow = ''; 670 } 671 $href = \TYPO3\CMS\Core\Utility\GeneralUtility::resolveBackPath(($GLOBALS['BACK_PATH'] . $this->script)) . '&id=' . rawurlencode($folderIdentifier) . $params; 672 return '<a href="' . htmlspecialchars($href) . '">' . $code . $sortArrow . '</a>'; 673 } 674 675 /** 676 * Creates the clipboard control pad 677 * 678 * @param \TYPO3\CMS\Core\Resource\File|\TYPO3\CMS\Core\Resource\Folder $fileOrFolderObject Array with information about the file/directory for which to make the clipboard panel for the listing. 679 * @return string HTML-table 680 * @todo Define visibility 681 */ 682 public function makeClip($fileOrFolderObject) { 683 $cells = array(); 684 $fullIdentifier = $fileOrFolderObject->getCombinedIdentifier(); 685 $md5 = \TYPO3\CMS\Core\Utility\GeneralUtility::shortmd5($fullIdentifier); 686 // For normal clipboard, add copy/cut buttons: 687 if ($this->clipObj->current == 'normal') { 688 $isSel = $this->clipObj->isSelected('_FILE', $md5); 689 $cells[] = '<a href="' . htmlspecialchars($this->clipObj->selUrlFile($fullIdentifier, 1, ($isSel == 'copy'))) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon(('actions-edit-copy' . ($isSel == 'copy' ? '-release' : '')), array('title' => $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.copy', 1))) . '</a>'; 690 $cells[] = '<a href="' . htmlspecialchars($this->clipObj->selUrlFile($fullIdentifier, 0, ($isSel == 'cut'))) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon(('actions-edit-cut' . ($isSel == 'cut' ? '-release' : '')), array('title' => $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.cut', 1))) . '</a>'; 691 } else { 692 // For numeric pads, add select checkboxes: 693 $n = '_FILE|' . $md5; 694 $this->CBnames[] = $n; 695 $checked = $this->clipObj->isSelected('_FILE', $md5) ? ' checked="checked"' : ''; 696 $cells[] = '<input type="hidden" name="CBH[' . $n . ']" value="0" />' . '<input type="checkbox" name="CBC[' . $n . ']" value="' . htmlspecialchars($fullIdentifier) . '" class="smallCheckboxes"' . $checked . ' />'; 697 } 698 // Display PASTE button, if directory: 699 $elFromTable = $this->clipObj->elFromTable('_FILE'); 700 if (is_a($fileOrFolderObject, 'TYPO3\\CMS\\Core\\Resource\\Folder') && count($elFromTable)) { 701 $cells[] = '<a href="' . htmlspecialchars($this->clipObj->pasteUrl('_FILE', $fullIdentifier)) . '" onclick="return ' . htmlspecialchars($this->clipObj->confirmMsg('_FILE', $fullIdentifier, 'into', $elFromTable)) . '" title="' . $GLOBALS['LANG']->getLL('clip_pasteInto', TRUE) . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-document-paste-into') . '</a>'; 702 } 703 // Compile items into a DIV-element: 704 return ' <!-- CLIPBOARD PANEL: --> 705 <div class="typo3-clipCtrl"> 706 ' . implode(' 707 ', $cells) . ' 708 </div>'; 709 } 710 711 /** 712 * Creates the edit control section 713 * 714 * @param \TYPO3\CMS\Core\Resource\File|\TYPO3\CMS\Core\Resource\Folder $fileOrFolderObject Array with information about the file/directory for which to make the edit control section for the listing. 715 * @return string HTML-table 716 * @todo Define visibility 717 */ 718 public function makeEdit($fileOrFolderObject) { 719 $cells = array(); 720 $fullIdentifier = $fileOrFolderObject->getCombinedIdentifier(); 721 // Edit metadata of file 722 try { 723 if (is_a($fileOrFolderObject, 'TYPO3\\CMS\\Core\\Resource\\File') && $fileOrFolderObject->isIndexed() && $fileOrFolderObject->checkActionPermission('edit')) { 724 $data = array( 725 'sys_file' => array($fileOrFolderObject->getUid() => 'edit') 726 ); 727 $editOnClick = \TYPO3\CMS\Backend\Utility\BackendUtility::editOnClick(\TYPO3\CMS\Core\Utility\GeneralUtility::implodeArrayForUrl('edit', $data), $GLOBALS['BACK_PATH'], $this->listUrl()); 728 $cells['editmetadata'] = '<a href="#" onclick="' . $editOnClick . '" title="Edit Metadata of this file">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-document-open') . '</a>'; 729 } else { 730 $cells['editmetadata'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('empty-empty'); 731 } 732 } catch (\Exception $e) { 733 $cells['editmetadata'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('empty-empty'); 734 } 735 // Edit file content (if editable) 736 if (is_a($fileOrFolderObject, 'TYPO3\\CMS\\Core\\Resource\\File') && $fileOrFolderObject->checkActionPermission('edit') && \TYPO3\CMS\Core\Utility\GeneralUtility::inList($GLOBALS['TYPO3_CONF_VARS']['SYS']['textfile_ext'], $fileOrFolderObject->getExtension())) { 737 $editOnClick = 'top.content.list_frame.location.href=top.TS.PATH_typo3+\'file_edit.php?target=' . rawurlencode($fullIdentifier) . '&returnUrl=\'+top.rawurlencode(top.content.list_frame.document.location.pathname+top.content.list_frame.document.location.search);return false;'; 738 $cells['edit'] = '<a href="#" onclick="' . $editOnClick . '" title="' . $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.edit') . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-page-open') . '</a>'; 739 } else { 740 $cells['edit'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('empty-empty'); 741 } 742 // rename the file 743 if ($fileOrFolderObject->checkActionPermission('rename')) { 744 $renameOnClick = 'top.content.list_frame.location.href = top.TS.PATH_typo3+\'file_rename.php?target=' . rawurlencode($fullIdentifier) . '&returnUrl=\'+top.rawurlencode(top.content.list_frame.document.location.pathname+top.content.list_frame.document.location.search);return false;'; 745 $cells['rename'] = '<a href="#" onclick="' . $renameOnClick . '" title="' . $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.rename') . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-edit-rename') . '</a>'; 746 } else { 747 $cells['rename'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('empty-empty'); 748 } 749 if (is_a($fileOrFolderObject, 'TYPO3\\CMS\\Core\\Resource\\Folder')) { 750 $infoOnClick = 'top.launchView( \'_FOLDER\', \'' . $fullIdentifier . '\');return false;'; 751 } elseif (is_a($fileOrFolderObject, 'TYPO3\\CMS\\Core\\Resource\\File')) { 752 $infoOnClick = 'top.launchView( \'_FILE\', \'' . $fullIdentifier . '\');return false;'; 753 } 754 $cells['info'] = '<a href="#" onclick="' . $infoOnClick . '" title="' . $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.info') . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('status-dialog-information') . '</a>'; 755 756 // delete the file 757 if ($fileOrFolderObject->checkActionPermission('remove')) { 758 $identifier = $fileOrFolderObject->getIdentifier(); 759 if ($fileOrFolderObject instanceof TYPO3\CMS\Core\Resource\Folder) { 760 $referenceCountText = \TYPO3\CMS\Backend\Utility\BackendUtility::referenceCount('_FILE', $identifier, ' (There are %s reference(s) to this folder!)'); 761 } else { 762 $referenceCountText = \TYPO3\CMS\Backend\Utility\BackendUtility::referenceCount('sys_file', $identifier, ' (There are %s reference(s) to this file!)'); 763 } 764 765 if ($GLOBALS['BE_USER']->jsConfirmation(4)) { 766 $confirmationCheck = 'confirm(' . $GLOBALS['LANG']->JScharCode(sprintf($GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:mess.delete'), basename($identifier)) . $referenceCountText) . ')'; 767 } else { 768 $confirmationCheck = '1 == 1'; 769 } 770 771 $removeOnClick = 'if (' . $confirmationCheck . ') { top.content.list_frame.location.href=top.TS.PATH_typo3+\'tce_file.php?file[delete][0][data]=' . rawurlencode($fileOrFolderObject->getCombinedIdentifier()) . '&vC=' . $GLOBALS['BE_USER']->veriCode() . '&redirect=\'+top.rawurlencode(top.content.list_frame.document.location.pathname+top.content.list_frame.document.location.search);};'; 772 773 $cells['delete'] = '<a href="#" onclick="' . htmlspecialchars($removeOnClick) . '" title="' . $GLOBALS['LANG']->sL('LLL:EXT:lang/locallang_core.xlf:cm.delete') . '">' . \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('actions-edit-delete') . '</a>'; 774 } else { 775 $cells['delete'] = \TYPO3\CMS\Backend\Utility\IconUtility::getSpriteIcon('empty-empty'); 776 } 777 778 // Hook for manipulating edit icons. 779 if (is_array($GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['fileList']['editIconsHook'])) { 780 foreach ($GLOBALS['TYPO3_CONF_VARS']['SC_OPTIONS']['fileList']['editIconsHook'] as $classData) { 781 $hookObject = \TYPO3\CMS\Core\Utility\GeneralUtility::getUserObj($classData); 782 if (!$hookObject instanceof \TYPO3\CMS\Filelist\FileListEditIconHookInterface) { 783 throw new \UnexpectedValueException( 784 '$hookObject must implement interface \\TYPO3\\CMS\\Filelist\\FileListEditIconHookInterface', 785 1235225797 786 ); 787 } 788 $hookObject->manipulateEditIcons($cells, $this); 789 } 790 } 791 // Compile items into a DIV-element: 792 return ' <!-- EDIT CONTROLS: --> 793 <div class="typo3-editCtrl"> 794 ' . implode(' 795 ', $cells) . ' 796 </div>'; 797 } 798 799 /** 800 * Make reference count 801 * 802 * @param \TYPO3\CMS\Core\Resource\File|\TYPO3\CMS\Core\Resource\Folder $fileOrFolderObject Array with information about the file/directory for which to make the clipboard panel for the listing. 803 * @return string HTML 804 * @todo Define visibility 805 */ 806 public function makeRef($fileOrFolderObject) { 807 if ($fileOrFolderObject instanceof \TYPO3\CMS\Core\Resource\FolderInterface) { 808 return '-'; 809 } 810 // Look up the path: 811 $rows = $GLOBALS['TYPO3_DB']->exec_SELECTgetRows('*', 'sys_refindex', 'ref_table=\'sys_file\' AND ref_uid = ' . (integer)$fileOrFolderObject->getUid() . ' AND deleted=0'); 812 return $this->generateReferenceToolTip($rows, '\'_FILE\', ' . \TYPO3\CMS\Core\Utility\GeneralUtility::quoteJSvalue($fileOrFolderObject->getCombinedIdentifier())); 813 } 814 815 } 816 817 ?>
__label__pos
0.997231
Skip to content + styled() Utility for creating styled components. Introduction All Material UI components are styled with the styled() utility. This utility is built on top of the styled() module of @mui/styled-engine and provides additional features. Import path You can use the utility coming from the @mui/system package, or if you are using @mui/material, you can import it from @mui/material/styles. The difference is in the default theme that is used (if no theme is available in the React context). import { styled } from '@mui/system'; // If you are using @mui/material import { styled } from '@mui/material/styles'; What problems does it solve? The utility can be used as a replacement for emotion's or styled-components' styled() utility. It aims to solve the same problem, but also provides the following benefits: 1. It uses a default theme if no theme is available in React context. 2. It supports the theme's styleOverrides and variants to be applied, based on the name applied in the options (can be skipped). 3. It adds support for the the sx prop (can be skipped). 4. It adds by default the shouldForwardProp option (that can be overridden), taking into account: ownerState, theme, sx, and as. API styled(Component, [options])(styles) => Component Arguments 1. Component: The component that will be wrapped. 2. options (object [optional]): • options.shouldForwardProp ((prop: string) => bool [optional]): Indicates whether the prop should be forwarded to the Component. • options.label (string [optional]): The suffix of the style sheet. Useful for debugging. • options.name (string [optional]): The key used under theme.components for specifying styleOverrides and variants. Also used for generating the label. • options.slot (string [optional]): If Root, it automatically applies the theme's variants. • options.overridesResolver ((props: object, styles: Record<string, styles>) => styles [optional]): Function that returns styles based on the props and the theme.components[name].styleOverrides object. • options.skipVariantsResolver (bool): Disables the automatic resolver for the theme.components[name].variants. • options.skipSx (bool [optional]): Disables the sx prop on the component. • The other keys are forwarded to the options argument of emotion's styled([Component], [options]). 3. styles (object | ({ ...props, theme }) => object [optional]): A styles object or a function returning a styles object. The function receives the theme and component's props in an object which is its single argument. Returns Component: The new component created. Basic usage Styled div Press Enter to start editing Styled div with theme Press Enter to start editing Custom components This example demonstrates how you can use the styled API to create custom components, with the same capabilities as the core components: Primary Secondary Press Enter to start editing If you inspect this element with the browser DevTools in development mode, you will notice that the class of the component now ends with the MyThemeComponent-root, which comes from the name and slot options that were provided. browser DevTools showing the rendered component In addition to this, the color, sx, and variant props are not propagated to the generated div element. Removing features If you would like to remove some of the MUI System specific features, you can do it like this: const StyledComponent = styled('div', {}, { name: 'MuiStyled', slot: 'Root', - overridesResolver: (props, styles) => styles.root, // disables theme.components[name].styleOverrides + skipVariantsResolver: true, // disables theme.components[name].variants + skipSx: true, // disables the sx prop }); Create custom styled() utility If you want to have a different default theme for the styled() utility, you can create your own version of it, using the createStyled() utility. import { createStyled, createTheme } from '@mui/system'; const defaultTheme = createTheme({ // your custom theme values }); const styled = createStyled({ defaultTheme }); export default styled; Difference with the sx prop The styled function is an extension of the styled utility provided by the underlying style library used – either Emotion or styled-components. It is guaranteed that it will produce the same output as the styled function coming from the style library for the same input. The sx prop, on the other hand, is a new way of styling your components, focused on fast customization. styled is a function, while sx is a prop available on the components created with styled. Therefore, you will notice the following differences: sx provides more shortcuts than styled With styled: const MyStyledButton = styled('button')({ mx: 1, // ❌ don't use this! This shortcut is only provided by the `sx` prop }); With sx: import Button from '@mui/material/Button'; const MyStyledButton = (props) => ( <Button sx={{ mx: 1, // ✔️ this shortcut is specific to the `sx` prop, }} > {props.children} </Button> ); The style definition varies slightly With styled: const MyStyledButton = styled('button')({ padding: 1, // means "1px", NOT "theme.spacing(1)" }); With sx: import Button from '@mui/material/Button'; const MyStyledButton = (props) => ( <Button sx={{ padding: 1, // means "theme.spacing(1)", NOT "1px" }} > {props.children} </Button> ); Patterns for how to use props differ With styled: const MyStyledButton = styled('button')((props) => ({ backgroundColor: props.myBackgroundColor, })); With sx: import Button from '@mui/material/Button'; const MyStyledButton = (props) => ( <Button sx={{ backgroundColor: props.myCustomColor }}>{props.children}</Button> ); Parameter when using function are different for each field With styled (not recommended): // You may find this syntax in the wild, but for code readability // we recommend using only one top-level function const MyStyledButtonPropsPerField = styled('button')({ backgroundColor: (props) => props.myBackgroundColor, }); With sx: import Button from '@mui/material/Button'; import { lighten } from 'polished'; const MyStyledButton = (props) => ( <Button sx={{ backgroundColor: (theme) => lighten(0.2, theme.palette.primary.main) }} > {props.children} </Button> ); // Note: for direct theme access without modification, you can also use a shortcut by providing the key as a string const MyStyledButton = (props) => ( <Button sx={{ backgroundColor: 'primary.main' }}>{props.children}</Button> ); How can I use the sx syntax with the styled() utility? If you prefer the sx syntax and want to use it in both the sx prop and the styled() utility, you can use the unstable_sx utility from the theme: Styled div with theme Press Enter to start editing The overhead added by using the unstable_sx utility is the same as if you were to use the sx prop on the component. How to use components selector API If you've ever used the styled() API of either emotion or styled-components, you should have been able to use components as selectors. import styled from '@emotion/styled'; const Child = styled.div` color: red; `; const Parent = styled.div` ${Child} { color: green; } `; render( <div> <Parent> <Child>Green because I am inside a Parent</Child> </Parent> <Child>Red because I am not inside a Parent</Child> </div>, ); With MUI System's styled() utility, you can use components as selectors, too. When using @mui/styled-engine-sc (styled-components), nothing needs to be done. When using @mui/styled-engine (emotion), the default engine, there are a few steps you should perform: First, you should install @emotion/babel-plugin. npm install @emotion/babel-plugin Then, configure the plugin to know about the Material UI version of the styled() utility: babel.config.js module.exports = { ... plugins: [ [ "@emotion", { importMap: { "@mui/system": { styled: { canonicalImport: ["@emotion/styled", "default"], styledBaseImport: ["@mui/system", "styled"] } }, "@mui/material/styles": { styled: { canonicalImport: ["@emotion/styled", "default"], styledBaseImport: ["@mui/material/styles", "styled"] } } } } ] ] }; Now you should be able to use components as your selectors!
__label__pos
0.899757
Easy Steps to Update Google Wi-Fi – Seamless Connectivity Guide Updating your Google Wi-Fi is crucial to maintaining a seamless and uninterrupted internet experience. Follow these easy steps to ensure optimal performance and connectivity. Key Takeaways: • Remove your existing Wi-Fi 5 network if you have a Nest Wifi or Google Wifi mesh system. • Factory reset your old Wi-Fi 5 network using the Google Home app. • Set up your new Wi-Fi 6E network on the Nest Wifi Pro, manually adding your custom settings. • Keep your devices connected seamlessly by reusing the same network name and password. • Troubleshoot individual devices that may not automatically reconnect. • Avoid operating both a Nest Wifi Pro network and a Nest Wifi or Google Wifi network at the same location. • Use the Google Home app troubleshooter to fix common Wi-Fi problems. • Contact your Internet Service Provider for further assistance if issues persist. Removing Your Existing Wi-Fi 5 Network If you have a Nest Wifi or Google Wifi mesh system, it is important to remove your existing Wi-Fi 5 network before setting up a new Wi-Fi 6E network on your Nest Wifi Pro. To do this, you can follow these easy steps: 1. Open the Google Home app on your mobile device. 2. Select the settings menu by tapping on the gear icon. 3. Choose the Wi-Fi option to access your network settings. 4. Locate your existing Wi-Fi 5 network and tap on it. 5. Select the option to “Forget network” or “Remove network” to delete it from your device. 6. Confirm the removal of the network when prompted. By removing your Wi-Fi 5 network, you are making way for the new and improved Wi-Fi 6E network on your Nest Wifi Pro, ensuring faster speeds and better connectivity for your devices. Table 1: Steps to Remove Existing Wi-Fi 5 Network Step Description 1 Open Google Home app 2 Access settings menu 3 Select Wi-Fi option 4 Locate existing Wi-Fi 5 network 5 Tap on network and choose “Forget network” or “Remove network” 6 Confirm network removal Once you have successfully removed your Wi-Fi 5 network, you are ready to proceed with setting up your new Wi-Fi 6E network on the Nest Wifi Pro. It is recommended to follow the instructions provided in the next section to ensure a seamless transition and uninterrupted connectivity for your devices. Setting Up Your New Wi-Fi 6E Network After removing your old network, you can now proceed to set up your new Wi-Fi 6E network using the Nest Wifi Pro. This advanced system offers faster speeds and improved connectivity, ensuring a seamless online experience for all your devices. Step 1: Factory Reset Your Nest Wifi Pro Before setting up your new network, it’s important to factory reset your Nest Wifi Pro to remove any previous settings. To do this, locate the reset button on the back of the Nest Wifi Pro router and press it for at least 10 seconds until the light ring starts flashing. This will erase all existing configurations and prepare the device for setup. Step 2: Connect the Nest Wifi Pro Now, you can connect the Nest Wifi Pro to your modem using an Ethernet cable. Plug one end of the cable into the Nest Wifi Pro router’s WAN port and the other end into your modem’s ethernet port. Ensure that the devices are securely connected. Step 3: Set Up Your Network Next, download the Google Home app on your smartphone or tablet. Open the app and follow the on-screen instructions to set up your Nest Wifi Pro network. Make sure to provide all the necessary details when prompted, including your network name and password. You can also choose to manually add your custom settings or details again to your new Wi-Fi 6E network. This ensures that your network preferences, such as parental controls, guest access, and device prioritization, are transferred to the new network. Step 4: Reconnect Your Devices Once the setup is complete, you can reconnect your devices to the new Wi-Fi 6E network using the same network name and password as your old network. This will allow your devices to seamlessly reconnect to the new network without any disruptions. If some devices don’t automatically reconnect, troubleshoot them individually by forgetting the old network and manually connecting them to the new network. Consult the device’s user manual or manufacturer’s website for specific instructions on how to do this. Important Note It’s not recommended to operate both a Nest Wifi Pro network and Nest Wifi or Google Wifi network at the same location. To ensure optimal performance and avoid interference, it’s best to use only one type of network at a time. Troubleshooting and Further Assistance If you encounter any difficulties during the setup process or experience any issues with your new Wi-Fi 6E network, you can use the Google Home app troubleshooter to quickly fix common Wi-Fi problems. Simply open the app, navigate to the Wi-Fi settings, and follow the troubleshooting steps outlined there. If problems persist or you require further assistance, don’t hesitate to contact your Internet Service Provider. They can provide additional support and guidance, ensuring that your new Wi-Fi 6E network is working smoothly and efficiently. Troubleshooting and Further Assistance If you encounter any difficulties during the update or experience ongoing Wi-Fi problems, this section will provide troubleshooting suggestions and guidance on how to seek further assistance. It’s essential to ensure a seamless connectivity experience with your updated Google Wi-Fi. Here are some steps you can take: 1. First, make sure that all the cables are securely connected and that your devices are powered on. Sometimes, a simple check can resolve the issue. 2. If you’re experiencing slow or intermittent internet connectivity, try restarting your modem and router. This can help refresh the network and resolve any temporary glitches. 3. Use the Google Home app troubleshooter to quickly diagnose and fix common Wi-Fi problems. The troubleshooter will guide you through a series of steps to identify the issue and provide solutions tailored to your specific situation. 4. If the troubleshooter doesn’t resolve the problem, contact your Internet Service Provider (ISP) for further assistance. They can help troubleshoot connectivity issues from their end and provide guidance on any potential network disruptions or outages. It’s important to note that operating both a Nest Wifi Pro network and a Nest Wifi or Google Wifi network at the same location is not recommended. These networks may interfere with each other, leading to connectivity issues. Therefore, it’s best to choose one system and ensure that all your devices are connected to the same network for optimal performance. Problem Potential Solution Slow internet speed Restart your modem and router. Intermittent connectivity Check and secure all cable connections. Devices not reconnecting Manually reconnect devices to the new network. No Wi-Fi signal Ensure that your devices are within range of the router. By following these troubleshooting steps and reaching out to your ISP if needed, you can address any Wi-Fi problems and ensure a smooth connectivity experience with your updated Google Wi-Fi. Conclusion Updating your Google Wi-Fi is a crucial step in ensuring uninterrupted internet connectivity and optimal performance. By following the steps outlined in this guide, you can seamlessly update your Wi-Fi network and embrace the future of connectivity. To update your Google Wi-Fi for seamless connectivity, begin by removing your existing Wi-Fi 5 network if you have a Nest Wifi or Google Wifi mesh system. Use the Google Home app to factory reset your old network and remove the existing Wi-Fi network. Next, set up your Nest Wifi Pro network and manually add your custom settings or details to your new Wi-Fi 6E network. It is recommended to reuse the same network name and password as your old network to keep your existing devices connected seamlessly. If some devices don’t automatically reconnect to the new network, troubleshoot them individually. Use the Google Home app troubleshooter to quickly fix common Wi-Fi problems and follow the troubleshooting steps outlined there. If issues persist, don’t hesitate to contact your Internet Service Provider for further assistance. Remember, it’s not recommended to operate both a Nest Wifi Pro network and Nest Wifi or Google Wifi network at the same location. Ensure a smooth update process by following these easy steps and enjoy a reliable and efficient Wi-Fi experience with your updated Google Wi-Fi. FAQ How do I update my Google Wi-Fi for seamless connectivity? To update your Google Wi-Fi, follow these easy steps: I have a Nest Wifi or Google Wifi mesh system. How do I remove my existing Wi-Fi 5 network? To remove your existing Wi-Fi 5 network, you need to: How do I set up my new Wi-Fi 6E network on the Nest Wifi Pro? To set up your new Wi-Fi 6E network, follow these steps: Can I reuse the same network name and password as my old network? Yes, it is recommended to reuse the same network name and password to keep your existing devices connected seamlessly. What should I do if some devices don’t reconnect after the update? If some devices don’t reconnect, troubleshoot them individually to identify and resolve the connectivity issue. Is it recommended to operate both a Nest Wifi Pro network and Nest Wifi or Google Wifi network at the same location? No, it is not recommended to operate both networks simultaneously at the same location. How can I troubleshoot common Wi-Fi problems during the update process? You can use the Google Home app troubleshooter to quickly fix common Wi-Fi problems. Follow the troubleshooting steps outlined in the app. What should I do if issues persist after troubleshooting? If issues persist, contact your Internet Service Provider for further assistance and support.
__label__pos
0.974187
Click here to Skip to main content Click here to Skip to main content Dependency Injection and Unit Of Work using Castle Windsor and NHibernate , 8 May 2014 Rate this: Please Sign up or sign in to vote. An implementation of dependency injection, repository and unit of work patterns using Castle Windsor and NHibernate. Contents Introduction In this article, I will show an implementation of dependency injection, repository and unit of work patterns using Castle Windsor as DI (dependency injection) container and NHibernate as ORM (Object Relational Mapping) tool. • Dependency Injection: Dependency injection is a software design pattern that allows removing hard-coded dependencies and making it possible to change them, whether at run-time or compile-time [1] • Repository: Mediates between the domain and data mapping layers using a collection-like interface for accessing domain objects. [2] • Unit of work: Used to define and manage transactional jobs in your application. [4] You can find many resources, articles and tutorials on dependency injection and unit of work. So, I'll not define them here. This article will be focused on solution of problems described in the next section. Problems When you are developing a data-driven application, you consider some design principles. In this section, I'll describe them and explain the solution briefly. How to open/close connections First problem is that: How and where to open and close connections. Surely, it's best to manage connections in database layer (probably in repositories). So, we can open connection, run database command and close connection in every repository method call. This may be inefficient and useless if we want to share same connection on different methods of same repository or different methods of different repositories (Think a transaction that uses some methods of different repositories). If we are bulding a web site (ASP.NET MVC or Web Forms), we can open the connection on Application_BeginRequest and close it on Application_EndRequest. But this approach has some disadvantages: • We open/close database for every request even if some request does not use database. So, a connection is taken from connection pool for a while even if it's not used. • We open database at the beginning of the request and close it at the end. But, there maybe cases the request is very long and database operation takes a short time in that request. Again, an inefficient usage of connection pool. Issues described above may not be big problems for you (indeed, for me, they are problems). But if you are not developing a web site, if you are developing a windows service that runs many threads those are using database in some periods of time? So, where to open/close connection. How to manage transactions If you are using transactions (most of applications do that) in your database operations, where to begin, commit or rollback the transaction. You can not do in repository methods since your transaction may include many different repository method calls. So, your business layer (that calls the repository methods) can begin/commit/rollback transactions. There are some problems in this approach: • Your business layer includes database specific codes. This breaks single responsibility and layering. • This approach dublicates transaction logic in every business method. As I described in the previous section, you can use Application_BeginRequest and Application_EndRequest to manage transactions. Same problems are here again: You can begin/commit unnecessary transactions. Also, you must handle errors and rollback transactions when needed. Also, if you are not developing a web site, it's hard to find a good place to begin/commit/rollback transactions. So, best approach is to begin a transaction when you really need to begin, commit the transaction when and only if all of your operations succeed and the rollback transaction if any of your operations fail. In this article, I implemented right this approach. Implementation I implemented a simple phone book application using ASP.NET MVC (as web framework), Sql Server (as DBMS), NHibernate (as ORM) and Castle Windsor (as Dependency Injection Container). Entities In our implementation, Entity is mapped to a record in a database table. In domain driven design (DDD), Entity is a persistent object that has a unique identifier [3]. All Entities in our solution is derived from Entity class defined below: public interface IEntity<TPrimaryKey> { TPrimaryKey Id { get; set; } } public class Entity<TPrimaryKey> : IEntity<TPrimaryKey> { public virtual TPrimaryKey Id { get; set; } } An entity has a unique identifier (primary key) that's type can be different (int, long, guid...), so it's a generic class. Our entities (People, Phone, City...) are derived from this class. For example, People class is defined as below: public class Person : Entity<int> { public virtual int CityId { get; set; } public virtual string Name { get; set; } public virtual DateTime BirthDay { get; set; } public virtual string Notes { get; set; } public virtual DateTime RecordDate { get; set; } public Person() { Notes = ""; RecordDate = DateTime.Now; } } As you see, primary key of Person is defined as int. Entity mappings ORM tools (like Entity framework and NHibernate) requires definition of mapping of Entities to database tables. There are many ways to implement it. I used NHibernate Fluent API to accomplish that. So, we define a mapping class for all entitites as shown below for People entity: public class PersonMap : ClassMap<Person> { public PersonMap() { Table("People"); Id(x => x.Id).Column("PersonId"); Map(x => x.CityId); Map(x => x.Name); Map(x => x.BirthDay); Map(x => x.Notes); Map(x => x.RecordDate); } } Repositories (Database layer) Repositories [2] are used to create database layer to abstract data access logic from upper layers. Generally a repository class is created for each entity (or aggregation: group of entities). I created a repository for each entity. First, I defined an interface that must be implemented by all repository classes: /// <summary> /// This interface must be implemented by all repositories to ensure UnitOfWork to work. /// </summary> public interface IRepository { } /// <summary> /// This interface is implemented by all repositories to ensure implementation of fixed methods. /// </summary> /// <typeparam name="TEntity">Main Entity type this repository works on</typeparam> /// <typeparam name="TPrimaryKey">Primary key type of the entity</typeparam> public interface IRepository<TEntity, TPrimaryKey> : IRepository where TEntity : Entity<TPrimaryKey> { /// <summary> /// Used to get a IQueryable that is used to retrive entities from entire table. /// </summary> /// <returns>IQueryable to be used to select entities from database</returns> IQueryable<TEntity> GetAll(); /// <summary> /// Gets an entity. /// </summary> /// <param name="key">Primary key of the entity to get</param> /// <returns>Entity</returns> TEntity Get(TPrimaryKey key); /// <summary> /// Inserts a new entity. /// </summary> /// <param name="entity">Entity</param> void Insert(TEntity entity); /// <summary> /// Updates an existing entity. /// </summary> /// <param name="entity">Entity</param> void Update(TEntity entity); /// <summary> /// Deletes an entity. /// </summary> /// <param name="id">Id of the entity</param> void Delete(TPrimaryKey id); } Thus, all repository classes must implement methods listed above. But, implementation of these methods are almost same in NHibernate. So, we can define a base class for all repositories. Thus, we don't implement same logic for all repositories. See NhRepositoryBase defined as below: /// <summary> /// Base class for all repositories those uses NHibernate. /// </summary> /// <typeparam name="TEntity">Entity type</typeparam> /// <typeparam name="TPrimaryKey">Primary key type of the entity</typeparam> public abstract class NhRepositoryBase<TEntity, TPrimaryKey> : IRepository<TEntity, TPrimaryKey> where TEntity : Entity<TPrimaryKey> { /// <summary> /// Gets the NHibernate session object to perform database operations. /// </summary> protected ISession Session { get { return NhUnitOfWork.Current.Session; } } /// <summary> /// Used to get a IQueryable that is used to retrive object from entire table. /// </summary> /// <returns>IQueryable to be used to select entities from database</returns> public IQueryable<TEntity> GetAll() { return Session.Query<TEntity>(); } /// <summary> /// Gets an entity. /// </summary> /// <param name="key">Primary key of the entity to get</param> /// <returns>Entity</returns> public TEntity Get(TPrimaryKey key) { return Session.Get<TEntity>(key); } /// <summary> /// Inserts a new entity. /// </summary> /// <param name="entity">Entity</param> public void Insert(TEntity entity) { Session.Save(entity); } /// <summary> /// Updates an existing entity. /// </summary> /// <param name="entity">Entity</param> public void Update(TEntity entity) { Session.Update(entity); } /// <summary> /// Deletes an entity. /// </summary> /// <param name="id">Id of the entity</param> public void Delete(TPrimaryKey id) { Session.Delete(Session.Load<TEntity>(id)); } } Session property is used to get session object (database connection object in NHibernate) from NhUnitOfWork.Current.Session. This gets the right Session object for current running transaction, so repositories does not think how to open/close connection and transaction. This mechanism will be explained in later sections. As you see, all CRUD operations are implemented for all repositories as default. So, now we can create a PersonRepository that can select, delete, update and create records just declaring types shown below: public interface IPersonRepository : IRepository<Person, int> { } public class NhPersonRepository : NhRepositoryBase<Person, int>, IPersonRepository { } We can do same for Phone and City entities. If we want to add a custom repository method, we can just add it to related entitiy's repository. For instance, we can add a new method to PhoneRepository to delete phones of given person: public interface IPhoneRepository : IRepository<Phone, int> { /// <summary> /// Deletes all phone numbers for given person id. /// </summary> /// <param name="personId">Id of the person</param> void DeletePhonesOfPerson(int personId); } public class NhPhoneRepository : NhRepositoryBase<Phone, int>, IPhoneRepository { public void DeletePhonesOfPerson(int personId) { var phones = GetAll().Where(phone => phone.PersonId == personId).ToList(); foreach (var phone in phones) { Session.Delete(phone); } } } Unit of work Unit of work is used to define and manage transactional jobs in your application. We define IUnitOfWork interface to define these jobs: /// <summary> /// Represents a transactional job. /// </summary> public interface IUnitOfWork { /// <summary> /// Opens database connection and begins transaction. /// </summary> void BeginTransaction(); /// <summary> /// Commits transaction and closes database connection. /// </summary> void Commit(); /// <summary> /// Rollbacks transaction and closes database connection. /// </summary> void Rollback(); } We implement IUnitOfWork for NHibernate as shown below: /// <summary> /// Implements Unit of work for NHibernate. /// </summary> public class NhUnitOfWork : IUnitOfWork { /// <summary> /// Gets current instance of the NhUnitOfWork. /// It gets the right instance that is related to current thread. /// </summary> public static NhUnitOfWork Current { get { return _current; } set { _current = value; } } [ThreadStatic] private static NhUnitOfWork _current; /// <summary> /// Gets Nhibernate session object to perform queries. /// </summary> public ISession Session { get; private set; } /// <summary> /// Reference to the session factory. /// </summary> private readonly ISessionFactory _sessionFactory; /// <summary> /// Reference to the currently running transcation. /// </summary> private ITransaction _transaction; /// <summary> /// Creates a new instance of NhUnitOfWork. /// </summary> /// <param name="sessionFactory"></param> public NhUnitOfWork(ISessionFactory sessionFactory) { _sessionFactory = sessionFactory; } /// <summary> /// Opens database connection and begins transaction. /// </summary> public void BeginTransaction() { Session = _sessionFactory.OpenSession(); _transaction = Session.BeginTransaction(); } /// <summary> /// Commits transaction and closes database connection. /// </summary> public void Commit() { try { _transaction.Commit(); } finally { Session.Close(); } } /// <summary> /// Rollbacks transaction and closes database connection. /// </summary> public void Rollback() { try { _transaction.Rollback(); } finally { Session.Close(); } } } Static Current property is the key point of the class. As you see, it gets/sets _current field that is marked as ThreadStatic [5]. So, we can use same Unit of work object in same thread. Thus, we can share same connection/transaction between objects. Lastly, we define an attribute that is used to mark a method which must be transactional: /// <summary> /// This attribute is used to indicate that declaring method is transactional (atomic). /// A method that has this attribute is intercepted, a transaction starts before call the method. /// At the end of method call, transaction is commited if there is no exception, othervise it's rolled back. /// </summary> [AttributeUsage(AttributeTargets.Method)] public class UnitOfWorkAttribute : Attribute { } If a method must be transactional, we just mark it with UnitOfWork attribute. Then we will intercept these methods using dependency injection as described later. Service layer In Domain Driven Design, Domain services is used to implement business logic. It can use repositories to perform database tasks. For example, PersonService is defined as below: public class PersonService : IPersonService { private readonly IPersonRepository _personRepository; private readonly IPhoneRepository _phoneRepository; public PersonService(IPersonRepository personRepository, IPhoneRepository phoneRepository) { _personRepository = personRepository; _phoneRepository = phoneRepository; } public void CreatePerson(Person person) { _personRepository.Insert(person); } [UnitOfWork] public void DeletePerson(int personId) { _personRepository.Delete(personId); _phoneRepository.DeletePhonesOfPerson(personId); } //... some other methods are not shown here since it's not needed. See source codes. } Focus on usage of UnitOfWork attrbiute defined in the previous section. DeletePerson method is marked as UnitOfWork. Because it calls two different repository methods and these method calls must be transactional. On the other hand, CreatePerson is not UnitOfWork since it only calls one repository method (Insert of person repository) and this repository method can open/close (manage) it's own transaction. We will see how to implement it in next sections. Dependency injection Dependency Injection (DI) Containers (like Castle Windsor we use) is used to manage dependencies and lifecycles of object is your application. So, it helps you to build loosely coupled components/modules in your application. It's generally initialized on startup of your application. In an ASP.NET application, global.asax is the mostly used place to initialize it: public class MvcApplication : System.Web.HttpApplication { private WindsorContainer _windsorContainer; protected void Application_Start() { InitializeWindsor(); // Other startup logic... } protected void Application_End() { if (_windsorContainer != null) { _windsorContainer.Dispose(); } } private void InitializeWindsor() { _windsorContainer = new WindsorContainer(); _windsorContainer.Install(FromAssembly.This()); ControllerBuilder.Current.SetControllerFactory(new WindsorControllerFactory(_windsorContainer.Kernel)); } } We create a WindsowContainer object (this is the main object to use dependency injection) on application startup and dispose it at the end. In InitializeWindsor method we also change default controller factory of MVC to use dependency injection. So, whenever ASP.NET MVC needs a Controller (in every web request), it creates it using DI [6]. See http://docs.castleproject.org/Windsor.Windsor-tutorial-ASP-NET-MVC-3-application-To-be-Seen.ashx to learn Castle windsor. Here, our controller factory: public class WindsorControllerFactory : DefaultControllerFactory { private readonly IKernel _kernel; public WindsorControllerFactory(IKernel kernel) { _kernel = kernel; } public override void ReleaseController(IController controller) { _kernel.ReleaseComponent(controller); } protected override IController GetControllerInstance(RequestContext requestContext, Type controllerType) { if (controllerType == null) { throw new HttpException(404, string.Format("The controller for path '{0}' could not be found.", requestContext.HttpContext.Request.Path)); } return (IController)_kernel.Resolve(controllerType); } } It's pretty simple and describes itself. The important part is to inject our own object dependencies. It's accomplished by PhoneBookDependencyInstaller class. This class is automatically investigated by Castle windsor since it implements IWindsorInstaller. Remember the _windsorContainer.Install(FromAssembly.This()); line in global.asax file. public class PhoneBookDependencyInstaller : IWindsorInstaller { public void Install(IWindsorContainer container, IConfigurationStore store) { container.Kernel.ComponentRegistered += Kernel_ComponentRegistered; //Register all controllers container.Register( //Nhibernate session factory Component.For<ISessionFactory>().UsingFactoryMethod(CreateNhSessionFactory).LifeStyle.Singleton, //Unitofwork interceptor Component.For<NhUnitOfWorkInterceptor>().LifeStyle.Transient, //All repoistories Classes.FromAssembly(Assembly.GetAssembly(typeof(NhPersonRepository))).InSameNamespaceAs<NhPersonRepository>().WithService.DefaultInterfaces().LifestyleTransient(), //All services Classes.FromAssembly(Assembly.GetAssembly(typeof(PersonService))).InSameNamespaceAs<PersonService>().WithService.DefaultInterfaces().LifestyleTransient(), //All MVC controllers Classes.FromThisAssembly().BasedOn<IController>().LifestyleTransient() ); } /// <summary> /// Creates NHibernate Session Factory. /// </summary> /// <returns>NHibernate Session Factory</returns> private static ISessionFactory CreateNhSessionFactory() { var connStr = ConfigurationManager.ConnectionStrings["PhoneBook"].ConnectionString; return Fluently.Configure() .Database(MsSqlConfiguration.MsSql2008.ConnectionString(connStr)) .Mappings(m => m.FluentMappings.AddFromAssembly(Assembly.GetAssembly(typeof(PersonMap)))) .BuildSessionFactory(); } void Kernel_ComponentRegistered(string key, Castle.MicroKernel.IHandler handler) { //Intercept all methods of all repositories. if (UnitOfWorkHelper.IsRepositoryClass(handler.ComponentModel.Implementation)) { handler.ComponentModel.Interceptors.Add(new InterceptorReference(typeof(NhUnitOfWorkInterceptor))); } //Intercept all methods of classes those have at least one method that has UnitOfWork attribute. foreach (var method in handler.ComponentModel.Implementation.GetMethods()) { if (UnitOfWorkHelper.HasUnitOfWorkAttribute(method)) { handler.ComponentModel.Interceptors.Add(new InterceptorReference(typeof(NhUnitOfWorkInterceptor))); return; } } } } As you see we are registering all components using Windsor's Register method. Notice that we are registering all repository classes with single line. We do same for services and controlles. We're using factory method to create ISessionFactory that is used to create ISession (database connection) objects to use with NHibernate. Finally, at the begining of Install method, we register to Kernel's ComponentRegistered event to inject our interception logic. See Kernel_ComponentRegistered method. If a method is a repository method, we are always using interception for it. Also, if a method has marked with UnitOfWork attribute, it's also intercepted with our NhUnitOfWorkInterceptor class. Interception Interception is a technique to run some codes at the begining of a method call and at the end of the method call. It's generally used for logging, profiling, caching... etc. Using this technique you can dynamically inject your codes into desired methods without changing the method. We use interception to implement Unit of work. If a method is repository method or marked as UnitOfWork attribute (explained in previous section), we open a database connection (Session on NHibernate) and begin a transaction at the begining of the method. If no exception is thrown by the intercepted method, the transaction is commited at the end of the method. If the method throws any exception, whole transaction is rolled back. So, let's see NhUnitOfWorkInterceptor class to know how I implemented it: /// <summary> /// This interceptor is used to manage transactions. /// </summary> public class NhUnitOfWorkInterceptor : IInterceptor { private readonly ISessionFactory _sessionFactory; /// <summary> /// Creates a new NhUnitOfWorkInterceptor object. /// </summary> /// <param name="sessionFactory">Nhibernate session factory.</param> public NhUnitOfWorkInterceptor(ISessionFactory sessionFactory) { _sessionFactory = sessionFactory; } /// <summary> /// Intercepts a method. /// </summary> /// <param name="invocation">Method invocation arguments</param> public void Intercept(IInvocation invocation) { //If there is a running transaction, just run the method if (NhUnitOfWork.Current != null || !RequiresDbConnection(invocation.MethodInvocationTarget)) { invocation.Proceed(); return; } try { NhUnitOfWork.Current = new NhUnitOfWork(_sessionFactory); NhUnitOfWork.Current.BeginTransaction(); try { invocation.Proceed(); NhUnitOfWork.Current.Commit(); } catch { try { NhUnitOfWork.Current.Rollback(); } catch { } throw; } } finally { NhUnitOfWork.Current = null; } } private static bool RequiresDbConnection(MethodInfo methodInfo) { if (UnitOfWorkHelper.HasUnitOfWorkAttribute(methodInfo)) { return true; } if (UnitOfWorkHelper.IsRepositoryMethod(methodInfo)) { return true; } return false; } } The main method is Intercept. It first checks if there is a transaction started before for this thread. If so, it does not starts a new transaction, uses the current transaction (see NhUnitOfWork.Current). Thus, nested calls to methods those have UnitOfWork attribute can use/share same transaction. A transaction is created/commited in only first entrance of a unit of work method. Also, if a method is not transactional, it just calls the method and returns. invocation.Proceed() command perform a call to the intercepted method. If there is no transaction currently, we must start a new transaction and commit the transaction if no error, rollback the transaction on an error case. Finally we set the NhUnitOfWork.Current = null to allow to start other transactions for current thread later if needed. You can also see UnitOfWorkHelper class. Thus, the only open/close connection and begin/commit/rollback transaction codes in entire application is defined in only one point. Web Application I developed a simple phone book application using the patterns described above. It's an ASP.NET MVC application. So, server side is implemented by ASP.NET MVC Controllers, client side is implemented using HTML and Javascript. To run application, first create the database using PhoneBookDb.sql file in the solution (it requires SQL Server 2008 R2 or a newer version). Then check the connection string in web.config if you change the name of the database. Also, it may require to Enable Nuget Package Restore to restore nuget packages. Server side In our application (for demonstration purposes), we define just one controller: HomeController. It gets needed services in it's constructor. Castle Windsor injects these services into controller since we registered service classes and MVC controller factory to Windsor. public class HomeController : Controller { private readonly IPersonService _personService; private readonly IPhoneService _phoneService; private readonly ICityService _cityService; public HomeController(IPersonService personService, ICityService cityService, IPhoneService phoneService) { _personService = personService; _cityService = cityService; _phoneService = phoneService; } public ActionResult Index() { return View(); } public JsonResult PeopleList() { try { var personList = _personService.GetPeopleList(); return Json(new { Result = "OK", Records = personList }); } catch (Exception ex) { return Json(new {Result = "ERROR", Message = ex.Message}); } } public JsonResult DeletePerson(int id) { try { _personService.DeletePerson(id); return Json(new { Result = "OK" }); } catch (Exception ex) { return Json(new { Result = "ERROR", Message = ex.Message }); } } //Other actions... } As you see, we use JSON objects to communicate with the client. Web page makes AJAX request to this controller to perform CRUD (Create, Retrive (select), Update and Delete) operations. See source codes for other actions. Client side Client side is HTML and Javascript. I implemented client side using jTable [7]. It's a jQuery plugin that is developed by me. It automates listing of database records and update/create/delete forms. See jtable.org for more information. See source codes for implementation. Here is the result: Client side jTable implementation Console Application In this section, I'll show how to implement all techniques (mentioned above) for a console application. I created a simple console application that periodically runs, checks and deletes all people who are older than 120 age (Think that I don't want to store old people in my phone book). This is just an example for services those run on background and performs some tasks. It may be a console application or a background Windows Service. Does not matter. Here, our main class that prepares and runs dependency injection and starts our periodically running service: public class PhoneBookRunner : IDisposable { private WindsorContainer _windsorContainer; public void Start() { _windsorContainer = new WindsorContainer(); _windsorContainer.Install(FromAssembly.Containing<PhoneBookDependencyInstaller>()); _windsorContainer.Install(FromAssembly.This()); _windsorContainer.Resolve<IPeriodicServiceTrigger>().Start(); } public void Dispose() { if (_windsorContainer != null) { _windsorContainer.Resolve<IPeriodicServiceTrigger>().Stop(); _windsorContainer.Dispose(); } } } In the Start method, we initialize the Castle Windsor container. Then we are registering dependencies using Installers. See source codes for Program.cs that uses PhoneBookRunner class. PhoneBookDependencyInstaller registers main components of our application. I also added a new installer for this console application: public class BackgroundServiceInstaller : IWindsorInstaller { public void Install(IWindsorContainer container, IConfigurationStore store) { container.Register( Component.For<IDeleteOldRecordsService>().ImplementedBy<DeleteOldRecordsService>().LifestyleTransient(), Component.For<IPeriodicServiceTrigger>().ImplementedBy<PeriodicServiceTrigger>().LifestyleSingleton() ); } } PeriodicServiceTrigger uses a Timer to perform periodic checks: public class PeriodicServiceTrigger : IPeriodicServiceTrigger { private readonly IDeleteOldRecordsService _deleteOldRecordsService; private readonly Timer _timer; public PeriodicServiceTrigger(IDeleteOldRecordsService deleteOldRecordsService) { _deleteOldRecordsService = deleteOldRecordsService; _timer = new Timer(10000); _timer.Elapsed += Timer_Elapsed; } public void Start() { _timer.Start(); } public void Stop() { _timer.Stop(); } private void Timer_Elapsed(object sender, ElapsedEventArgs e) { _deleteOldRecordsService.DeletePeopleOlderThan120Age(); } } And finally we have a IDeleteOldRecordsService that is used to implement business rule: public class DeleteOldRecordsService : IDeleteOldRecordsService { private readonly IPersonRepository _personRepository; private readonly IPersonService _personService; public DeleteOldRecordsService(IPersonRepository personRepository, IPersonService personService) { _personRepository = personRepository; _personService = personService; } [UnitOfWork] public void DeletePeopleOlderThan120Age() { var yearAge120 = DateTime.Now.AddYears(-120); var oldRecords = _personRepository.GetAll().Where(person => person.BirthDay < yearAge120).ToList(); foreach (var oldRecord in oldRecords) { _personService.DeletePerson(oldRecord.Id); } } } As you see, DeletePeopleOlderThan120Age is a UnitOfWork method. So, it runs as transactional. More If you did like the patterns and techniques in this article, you will also like ASP.NET Boilerplate application framework. It implementes all techniques described in this article. It open source and distributed on nuget.. See Introduction to ASP.NET Boilerplate article: http://www.codeproject.com/Articles/768664/Introduction-to-ASP-NET-Boilerplate Article History • 13.02.2013 - Added console application • 10.02.2013 - First version References License This article, along with any associated source code and files, is licensed under The Code Project Open License (CPOL) About the Author Halil ibrahim Kalkan Software Developer Sestek Turkey Turkey I have started programming at 14 years old using Pascal as hobby. Then I interested in web development (HTML, JavaScript, ASP...) before university.   I graduated from Sakarya University Computer Engineering. At university, I learned C++, Visual Basic.NET, C#, ASP.NET and Java. I partly implemented ARP, IP and TCP protocols in Java as my final term project.   Now, I am working in a private company in Istanbul as a senior software architect & developer. Working on Windows and web based software development using C# and ASP.NET MVC.   http://www.halilibrahimkalkan.com Follow on   Twitter   LinkedIn Comments and Discussions   GeneralMy vote of 5 PinmemberSancarBEY24-Apr-13 4:07  GeneralRe: My vote of 5 PinmemberHalil ibrahim Kalkan1-May-13 21:13  General General    News News    Suggestion Suggestion    Question Question    Bug Bug    Answer Answer    Joke Joke    Rant Rant    Admin Admin    Use Ctrl+Left/Right to switch messages, Ctrl+Up/Down to switch threads, Ctrl+Shift+Left/Right to switch pages. | Advertise | Privacy | Mobile Web02 | 2.8.140721.1 | Last Updated 9 May 2014 Article Copyright 2013 by Halil ibrahim Kalkan Everything else Copyright © CodeProject, 1999-2014 Terms of Service Layout: fixed | fluid
__label__pos
0.991653
Online JudgeProblem SetAuthorsOnline ContestsUser Web Board Home Page F.A.Qs Statistical Charts Problems Submit Problem Online Status Prob.ID: Register Update your info Authors ranklist Current Contest Past Contests Scheduled Contests Award Contest User ID: Password:   Register Language: Area of Simple Polygons Time Limit: 1000MSMemory Limit: 65536K Total Submissions: 3897Accepted: 2013 Description There are N, 1 <= N <= 1,000 rectangles in the 2-D xy-plane. The four sides of a rectangle are horizontal or vertical line segments. Rectangles are defined by their lower-left and upper-right corner points. Each corner point is a pair of two nonnegative integers in the range of 0 through 50,000 indicating its x and y coordinates. Assume that the contour of their union is defi ned by a set S of segments. We can use a subset of S to construct simple polygon(s). Please report the total area of the polygon(s) constructed by the subset of S. The area should be as large as possible. In a 2-D xy-plane, a polygon is defined by a finite set of segments such that every segment extreme (or endpoint) is shared by exactly two edges and no subsets of edges has the same property. The segments are edges and their extremes are the vertices of the polygon. A polygon is simple if there is no pair of nonconsecutive edges sharing a point. Example: Consider the following three rectangles: rectangle 1: < (0, 0) (4, 4) >, rectangle 2: < (1, 1) (5, 2) >, rectangle 3: < (1, 1) (2, 5) >. The total area of all simple polygons constructed by these rectangles is 18. Input The input consists of multiple test cases. A line of 4 -1's separates each test case. An extra line of 4 -1's marks the end of the input. In each test case, the rectangles are given one by one in a line. In each line for a rectangle, 4 non-negative integers are given. The first two are the x and y coordinates of the lower-left corner. The next two are the x and y coordinates of the upper-right corner. Output For each test case, output the total area of all simple polygons in a line. Sample Input 0 0 4 4 1 1 5 2 1 1 2 5 -1 -1 -1 -1 0 0 2 2 1 1 3 3 2 2 4 4 -1 -1 -1 -1 -1 -1 -1 -1  Sample Output 18 10 Source [Submit]   [Go Back]   [Status]   [Discuss] Home Page   Go Back  To top All Rights Reserved 2003-2013 Ying Fuchen,Xu Pengcheng,Xie Di Any problem, Please Contact Administrator
__label__pos
0.651657
博客> 第四章字符串和集合 第四章字符串和集合 2018-03-16 16:01 评论:0 阅读:140 易_君 ios swift 字符串和集合: //: Playground - noun: a place where people can play /** 字符概述: let/var 名称:Character = "值" let/var也是区分关键字是否可变。 转义字符:当使用双引号表示某些字符时,有时候输出的结果并不是我们所想要,如"\"字符比较特殊不能够使用双引号表示,这时,需要在字符"\"的前面插入一个转转义符"\",即对字符这样转义"\" 常见的转义字符——说明: \t —— 水平制表符 \n —— 换行 \r —— 回车 \" —— 双引号 \' —— 单引号 \ —— 反斜线 注意:Swift语言采用Unicode编码,它的字符几乎涵盖了大家所知的一切字符,包括图片表情 */ import UIKit //字符 let ch:Character = "c" //与C语言中的字符不同,Swift的字符是使用双引号包含的,并且它的值只能是一个字母、数字、汉字或者符号 print(ch) //转义字符 let cha:Character = "\" print(cha) //字符串 //通过字面量赋值的方式初始化字符串 let tempString = "this is temp string" print(tempString) //通过创建String实例的方式初始化字符串 let anotherString = String() print("--(anotherString)") //仅仅输出了--,anotherString输出为空字符 if anotherString.isEmpty { //isEmpty属性,判断字符串是否为空 print("啥都没有") } //字符串的基本操作 /* 获取字符串的长度: 1.调用lengthOfBytes方法,该方法是获取字符串中字节的个数 2.直接计算字符的个数 / let string = "Hello World" let len1 = string.lengthOfBytes(using: String.Encoding.utf8) print(len1) let len2 = string.characters.count print(len2) /* 字符串的遍历: String表示特定序列的字符值的集合,每一个字符值代表一个Unicode字符。可利用for-in循环来遍历字符串的每一个字符 / for c in string.characters { //通过characters获取了string的字符集合 print(c) } /** 连接字符串和字符: let定义的字符串是不可修改的,无法进行连接操作,只有var定义的字符串可以修改,想要进行修改可以通过如下三种方式: 1.使用加法赋值运算符连接字符串 2.使用加法运算符拼接字符串 3.调用append方法追加字符串 比较String和NSString 在截取字符串的子串的过程中,可能会经常见到str as NSString 或者 str as String来回转换字符串,有些操作用NSString比较方便,但是有些操作恰恰相反,String保留了大部分NSString的API,如hasPrefix,subStringWithRange等。区别: (1)String是一个结构体类型,而NSString是一个继承自NSObject基类对象 (2)两者都可以使用自己的类名来直接进行初始化,从表面上看写法相同,但是NSString的意思是初始化了一个指针指向了这个字符串,但是String是把字符串的字面量赋值给常量或变量 (3)String的字符串拼接比NSString更加方便,NSString需要使用append或者stringWithFormat将两个字符串拼接,而Swift仅仅使用'+'号就能够实现 (4)string可以实现遍历输出一个字符串内部所有字符,而NSString是做不到的,这是因为String继承了CollectionType协议 (5)NSString直接使用length就可以获得字符串的长度,但是Swift是真正的类似于length的方法就是取出character属性获得字符集合,然后使用count属性计算即可 */ var varString = "itcast" //定义了一个可变字符串 varString += ".com" //使用 += 运算符连接.com print(varString) let string1 = "Hello" let string2 = " World" var string3 = string1 + string2 //使用 + 连接字符串组成一个新字符串 print(string3) let character:Character = "!" string3.append(character) print(string3) /* 字符串插值: 是一种构建新字符串的方式,可以在其中包含常量、变量、字面量和表达式,插入的字符串字面量的每一项都在以反斜线为前缀的小括号中 / let multiplier = 3 //将multiplier插入到字符串中 let message = "(multiplier)乘以2.5等于(Double(multiplier) * 2.5)" print(message) /* 格式字符串: y用于表示时间的字符串,一般都要求有着固定的格式,为此,String提供了两个常用的构造函数,将字符串按照一定的格式显示 / let hour = 8 let minute = 5 let seconds = 6 let dateString = String(format: "d:d:d", arguments:[hour,minute,seconds]) print(dateString) let dateString2 = String(format: "d:d:d", hour,minute,seconds) print(dateString2) /* 字符串的截取: as关键字 调用substring(with: NSMakeRange)方法 / let Nstring = "Hello World" //将Nstring转换为OC的字符串使用,并根据范围来截取字串 let subString = (Nstring as NSString).substring(with: NSMakeRange(1, 5)) print(subString) /* 字符串的比较: 1.比较字符或字符串是否相等 2.比较字符串的前缀或后缀 / let emptyStr = "" let emptyStr2 = String() if emptyStr == emptyStr2 { //使用==判断两个字符串是否相等 print("相等") }else{ print("不相等") } let ch1 = "" let ch2 = "c" if ch1 == ch2 { //使用==判断两个字符串是否相等 print("相等") }else{ print("不相等") } //创建一个数组 let docFolder = ["java.docx","JavaBean.docx","Object-C.xlsx","Swift.docx"] //定义一个变量,用于统计docx文档的个数 var worldDocCount = 0 //遍历数组 for doc in docFolder { //如果后缀为.docx,worldDocCount加1 if doc.hasSuffix(".docx") { worldDocCount += 1 } } print("World文档中的个数为:(worldDocCount)") //定义一个变量,用于统计关于Java文档的个数 var javaDocCount = 0 //遍历数组 for doc in docFolder { //转换为小写 let lowercaseStr = doc.lowercased() //如果前缀为"Java","javaDocCount加1 if lowercaseStr.hasPrefix("java") { javaDocCount += 1 } } print("java相关文档个数为:(javaDocCount)") /* 集合Collection Swift语言提供了数组,set和字典三种集合类型,用于存储集合数据。在Swift中通常用两种形式来定义数组类型,一种是Array的形式,Element表示数组中唯一存在的数据类型,另一种是[Element]这样简单的语法。 注意:数组对于存储数据有着具体的要求,数据元素在被存储进入某个数组之前,必须明确数据类型,方法是通过显示的类型标注或者类型推断 / var array1:Array //只能存储String类型的数据 var array2:[String] //等价于array1 //创建一个空数组 var someInts = [Int]() //用构造方式创建了一个Int类型的空数组 /* • 使用字面量构造数组 let 常量名:[类型] = [value1,value2,value3...] let 常量名 = [value1,value2,value3...] */ let array:[String] = ["Hello","Hi"] let arr = ["zhangsan",18] as [Any] //两组数组相加来创建一个新数组 let arr1 = ["张三","李四"] let arr2 = ["王五"] print(arr1 + arr2) //数组常见的操作 //获取数组的长度 let season:[String] = ["春","夏","秋","冬"] let length = season.count //获取数组元素的总个数 print(length) //数组的遍历 for element in season{ //遍历数组 print(element) } /* 数组元素的增加,删除,方法如下: 1.append(newElement:Element)方法,向数组的某位追加元素。 2.insert(_newElement:Element,at i:Int)方法,向某个索引位置插入元素。 3.removeFirst()方法,删除数组的第一个元素。 4.remove(at index:Int)方法,删除数组中指定位置的元素 5.removeLast()方法,删除数组的末尾元素。 6.removeAll(keepingCapacity keepCapacity:Bool = Default)方法,删除所有的元素,并且保留存储空间 / //定义一个可变数组 var var_array = ["zhangsan","lisi"] //在末尾追加一个元素 var_array.append("wangwu") print(var_array) //向指定位置插入一个元素 var_array.insert("zhaoliu", at: 2) print(var_array) //删除第一个元素 var_array.removeFirst() print(var_array) //删除最后一个元素 var_array.removeLast() print(var_array) //删除指定位置的元素 var_array.remove(at: 1) print(var_array) //删除所有元素 //var_array.removeAll() var_array.removeAll(keepingCapacity: true) print(var_array) /* 使用下标语法修改数组 array[索引值],需要注意的是,数组的索引值必须存在的,否则会出现数组越界的运行时错误,由于数组越界问题,不可以使用下标访问的形式在数组的尾部添加新的元素 / //定义一个可变数组 var indexArray = ["One","Two","Three","Four","Five"] //使用下标语法修改1个元素 indexArray[0] = "Six" print(indexArray) //使用下标语法修改多个元素 indexArray[1...2] = ["Seven"] print(indexArray) //使用下标语法修改多个元素 indexArray[2...3] = ["English","Nine"] print(indexArray) //数组的内存分配 //定义一个存放字符串的数组 var charray:[String] //分配空间 charray = [String]() for element in 0..<128 Swift数组可以使用加法赋值运算符(+=)或者加法运算符直接添加拥有相同类型的数组> Element表示Set中允许存储的类型,和数组不同的是,Set没有等价的简化式 */ //创建Set //创建空Set var letters = Set() //letters变量被推断为Set() Set([]) //用数组字面量创建Set var favoriteColor:Set = ["red","blud","greend"] //只允许存储String类型的值 var favoriteColorS:Set = ["红色","蓝色","绿色"] //自动推断类型为Set /* 访问和修改Set 通过使用count属性,可以获取Set中元素的总个数;通过使用Set类型的方法,可以对Set实现增加、查找、删除的操作 / //获取Set中元素的数量 let count = favoriteColor.count //获取元素的个数 print("我喜欢的颜色有(count)种") /* Set的增删查操作 insert(:)方法,添加一个新元素 remove(:)方法,删除一个元素,如果该值是Set的一个元素则删除该元素并返回被删除的元素值,如果Set不包含该值,则返回nil removeAll()方法,删除Set的所有元素 contains(_:)方法,检查Set中是否包含有一个特定的值,如果包含则返回true,否则返回false。 / //1.添加 favoriteColorS.insert("黄色") //2.删除 if let removeGreen = favoriteColorS.remove("绿色") { print("我已经不喜欢(removeGreen)了") }else{ print("我从来没有喜欢过这个颜色") } //3.检查是否包含某元素 if favoriteColorS.contains("黑色"){ print("Ok,我还喜欢这个颜色") }else{ print("黑色太沉重,不是我的菜") } //4.删除所有的元素 favoriteColorS.removeAll() print(favoriteColorS) /*遍历Set 可以使用for-in,由于Set中元素是无序的可以先用sorted()方法返回Set的有序集合/ var colors:Set = ["white","red","black"] print("--无序Set") for color in colors { print(color) } print("--有序Set") for sortedcolor in colors.sorted() { print(sortedcolor) } var numbers:Set = ["1","2","3"] for number in numbers { print(number) } for num in numbers.sorted() { print(num) } /* Set之间的操作 如果把两个Set组合在一起,判断两个Set的共有元素、或者判断两个Set是否全包含、部分包含或者相交包含或者不相交 intersection(:)方法,根据两个Set中共同包含的中创建一个新的Set symmetricDifference(:)方法,根据包含于单个Set且不同时包含于两个Set中的值创建一个新的Set union(:)方法,根据两个Set包含的所有值创建一个新的Set subtracting(:)方法,根据a包含但b不包含的值创建一个新的Set / var a:Set = [1,2,3,4,5] let b:Set = [4,5,6,7,8] //1.intersection(:)方法 print(a.intersection(b).sorted()) //2.symmetricDifference(:)方法 print(a.symmetricDifference(b).sorted()) //3.union(:)方法 print(a.union(b).sorted()) //4.subtracting(:)方法 print(a.subtracting(b).sorted()) /* Set之间的操作 如果把两个Set组合在一起,判断两个Set的共有元素、或者判断两个Set是否全包含、部分包含或者相交包含或者不相交 intersection(:)方法,根据两个Set中共同包含的中创建一个新的Set symmetricDifference(:)方法,根据包含于单个Set且不同时包含于两个Set中的值创建一个新的Set union(:)方法,根据两个Set包含的所有值创建一个新的Set subtracting(:)方法,根据a包含但b不包含的值创建一个新的Set / var a:Set = [1,2,3,4,5] let b:Set = [4,5,6,7,8] //1.intersection(:)方法 print(a.intersection(b).sorted()) //2.symmetricDifference(:)方法 print(a.symmetricDifference(b).sorted()) //3.union(:)方法 print(a.union(b).sorted()) //4.subtracting(:)方法 print(a.subtracting(b).sorted()) /* Set之间的关系 指的是Set元素之间的关系,如果一个Set包含了另一个Set的所有元素,通常称前一个Set为后一个Set的父Set,后一个为前一个Set的子Set,两个Set的所有元素相同,通常说两个Set相等,两个Set没有共同元素时,彼此不关联。 1.是否相等运算符(==):判断两个Set的值是否全部相同 2.isSubset(of:)方法,判断一个Set中的值是否都被包含在另一个Set中 3.isSuperset(of:)方法,判断一个Set中是否包含另一个Set中的所有值 4.isStrictSubset(of:)方法,判断一个Set是否是另一个Set的子集合,并且两个Set不相等 5.isStrictSuperset(of:)方法,判断一个Set是否是另一个Set的父集合,并且两个Set不相等 6.isDisjoint(with_:)方法,判断两个Set是否不含有相同的值(是否有没有交集) / let x:Set = [1,2,3,4,5] let y:Set = [1,2] let z:Set = [4,5,6,7,8] let m:Set = [1,2] print("y是x的子集",y.isSubset(of: x)) print("x是y的父集",x.isSuperset(of: y)) print("y和Z没有交集",y.isDisjoint(with: z)) print("y和m相等",y == m) /** 字典 基本格式: let/var 名称:Dictionary = [key1:value,key2:value,key3:value]一个键值对是一个key值和一个value的结合体,键和值之间以冒号分割,这些键值对构成了一个列表,由方括号并以逗号分割 */ var airports:Dictionary = ["TYO":"Tokyo","DUB":"Doubin"] //只能是型字典 var dict = ["name":"susan","age":18,"heigth":170] as [String : Any] dict["name"] = "xiaohua" //将name对应的值进行修改 dict["gender"] = "Girl" //增加一个键为gender的键值对 print(dict) //字典的遍历 for (k,v) in dict { //遍历字典 print("(k)---(v)") } //字典的合并 var dict1 = ["name":"小花","age":18] as [String : Any] var dict2 = ["title":"老大","name":"小草"] as [String : Any] //将dict2的内容合并到dict1中 for (k,v) in dict2{ dict1[k] = v } print(dict1) //如果合并过程中出现相同的键则,合并项中旧值会被覆盖 /* 比较Swift和Object-C的字典 1.Object-C中使用{}来定义一个字典,而Swift中仍使用[]来定义 2.在Object-C中有NSDictionary和NSMutableDictionary之分,而Swift中也是通过let/var来区分字典是否可变 3.与Object-C字典不同,Swift的字典不仅可以存储对象类型的值,还可以存储基本数据类型,结构体,枚举值。 4.Object-C中的NSDictionary是一个继承自NSObject基类的对象,而Swift中的字典是一个结构体 / 收藏 0 sina weixin mail 回到顶部
__label__pos
0.967018
CFA Level 1 Quantitative Methods - Hypothesis Testing Hypothesis testing provides a basis for taking ideas or theories that someone initially develops about the economy or investing or markets, and then deciding whether these ideas are true or false. More precisely, hypothesis testing helps decide whether the tested ideas are probably true or probably false as the conclusions made with the hypothesis-testing process are never made with 100% confidence - which we found in the sampling and estimating process: we have degrees of confidence - e.g. 95% or 99% - but not absolute certainty. Hypothesis testing is often associated with the procedure for acquiring and developing knowledge known as the scientific method. As such, it relates the fields of investment and economic research (i.e., business topics) to other traditional branches of science (mathematics, physics, medicine, etc.) Hypothesis testing is similar in some respects to the estimation processes presented in the previous section. Indeed, the field of statistical inference, where conclusions on a population are drawn from observing subsets of the larger group, is generally divided into two groups: estimation and hypothesis testing. With estimation, the focus was on answering (with a degree of confidence) the value of a parameter, or else a range within which the parameter most likely falls. Think of estimating as working from general to specific. With hypothesis testing, the focus is shifted: we start my making a statement about the parameter's value, and then the question becomes whether the statement is true or not true. In other words, it starts with a specific value and works the other way to make a general statement. What is a Hypothesis? A hypothesis is a statement made about a population parameter. These are typical hypotheses: "the mean annual return of this mutual fund is greater than 12%", and "the mean return is greater than the average return for the category". Stating the hypothesis is the initial step in a defined seven-step process for hypothesis testing - a process developed based on the scientific method. We indicate each step below. In the remainder of this section of the study guide, we develop a detailed explanation for how to answer each step's question. Hypothesis testing seeks to answer seven questions: 1. What are the null hypothesis and the alternative hypothesis? 2. Which test statistic is appropriate, and what is the probability distribution? 3. What is the required level of significance? 4. What is the decision rule? 5. Based on the sample data, what is the value of the test statistic? 6. Do we reject or fail to reject the null hypothesis? 7. Based on our rejection or inability to reject, what is our investment or economic decision? Null Hypothesis Step #1 in our process involves stating the null and alternate hypothesis. The null hypothesis is the statement that will be tested. The null hypothesis is usually denoted with "H0". For investment and economic research applications, and as it relates to the CFA exam, the null hypothesis will be a statement on the value of a population parameter, usually the mean value if a question relates to return, or the standard deviation if it relates to risk. It can also refer to the value of any random variable (e.g. sales at company XYZ are at least $10 million this quarter). In hypothesis testing, the null hypothesis is initially regarded to be true, until (based on our process) we gather enough proof to either reject the null hypothesis, or fail to reject the null hypothesis. Alternative Hypothesis The alternative hypothesis is a statement that will be accepted as a result of the null hypothesis being rejected. The alternative hypothesis is usually denoted "Ha". In hypothesis testing, we do not directly test the worthiness of the alternate hypothesis, as our testing focus is on the null. Think of the alternative hypothesis as the residual of the null - for example, if the null hypothesis states that sales at company XYZ are at least $10 million this quarter, the alternative hypothesis to this null is that sales will fail to reach the $10 million mark. Between the null and the alternative, it is necessary to account for all possible values of a parameter. In other words, if we gather evidence to reject this null hypothesis, then we must necessarily accept the alternative. If we fail to reject the null, then we are rejecting the alternative. One-Tailed Test The labels "one-tailed" and "two-tailed" refer to the standard normal distribution (as well as all of the t-distributions). The key words for identifying a one-tailed test are "greater than or less than". For example, if our hypothesis is that the annual return on this mutual fund will be greater than 8%, it's a one-tailed test that will be rejected based only on finding observations in the left tail. Figure 2.13 below illustrates a one-tailed test for "greater than" (rejection in left tail). (A one-tailed test for "less than" would look similar to the graph below, with the rejection region for less than in the right tail rather than the left.) Two-Tailed test Characterized by the words "equal to or not equal to". For example, if our hypothesis were that the return on a mutual fund is equal to 8%, we could reject it based on observations in either tail (sufficiently higher than 8% or sufficiently lower than 8%). Choosing the null and the alternate hypothesis: If θ (theta) is the actual value of a population parameter (e.g. mean or standard deviation), and θ0 (theta subzero) is the value of theta according to our hypothesis, the null and alternative hypothesis can be formed in three different ways: Choosing what will be the null and what will be the alternative depends on the case and what it is we wish to prove. We usually have two different approaches to what we could make the null and alternative, but in most cases, it's preferable to make the null what we believe we can reject, and then attempt to reject it. For example, in our case of a one-tailed test with the return hypothesized to be greater than 8%, we could make the greater-than case the null (alternative being less than), or we could make the greater-than case the alternative (with less than the null). Which should we choose? A hypothesis test is typically designed to look for evidence that may possibly reject the null. So in this case, we would make the null hypothesis "the return is less than or equal to 8%", which means we are looking for observations in the left tail. If we reject the null, then the alternative is true, and we conclude the fund is likely to return at least 8%. Test Statistic Step #2 in our seven-step process involves identifying an appropriate test statistic. In hypothesis testing, a test statistic is defined as a quantity taken from a sample that is used as the basis for testing the null hypothesis (rejecting or failing to reject the null). Calculating a test statistic will vary based upon the case and our choice of probability distribution (for example, t-test, z-value). The general format of the calculation is: Formula 2.36 Test statistic = (sample statistic) - (value of parameter according to null) (Standard error of sample statistic) Type I and Type II Errors Step #3 in hypothesis testing involves specifying the significance level of our hypothesis test. The significance level is similar in concept to the confidence level associated with estimating a parameter - both involve choosing the probability of making an error (denoted by α, or alpha), with lower alphas reducing the percentage probability of error. In the case of estimators, the tradeoff of reducing this error was to accept a wider (less precise) confidence interval. In the case of hypothesis testing, choosing lower alphas also involves a tradeoff - in this case, increasing a second type of error. Errors in hypothesis testing come in two forms: Type I and Type II. A type I error is defined as rejecting the null hypothesis when it is true. A type II error is defined as not rejecting the null hypothesis when it is false. As the table below indicates, these errors represent two of the four possible outcomes of a hypothesis test: The reason for separating type I and type II errors is that, depending on the case, there can be serious consequences for a type I error, and there are other cases when type II errors need to be avoided, and it is important to understand which type is more important to avoid. Significance Level Denoted by α, or alpha, the significance level is the probability of making a type I error, or the probability that we will reject the null hypothesis when it is true. So if we choose a significance level of 0.05, it means there is a 5% chance of making a type I error. A 0.01 significance level means there is just a 1% chance of making a type I error. As a rule, a significance level is specified prior to calculating the test statistic, as the analyst conducting the research may use the result of the test statistic calculation to impact the choice of significance level (may prompt a change to higher or lower significance). Such a change would take away from the objectivity of the test. While any level of alpha is permissible, in practice there is likely to be one of three possibilities for significance level: 0.10 (semi-strong evidence for rejecting the null hypothesis), 0.05 (strong evidence), and 0.01 (very strong evidence). Why wouldn't't we always opt for 0.01 or even lower probabilities of type I errors - isn't the idea to reduce and eliminate errors? In hypothesis testing, we have to control two types of errors, with a tradeoff that when one type is reduced, the other type is increased. In other words, by lowering the chances of a type I error, we must reject the null less frequently - including when it is false (a type II error). Actually quantifying this tradeoff is impossible because the probability of a type II error (denoted by β, or beta) is not easy to define (i.e. it changes for each value of θ). Only by increasing sample size can we reduce the probability of both types of errors. Decision Rule Step #4 in the hypothesis-testing process requires stating a decision rule. This rule is crafted by comparing two values: (1) the result of the calculated value of the test statistic, which we will complete in step #5 and (2) a rejection point, or critical value (or values) that is (are) the function of our significance level and the probability distribution being used in the test. If the calculated value of the test statistic is as extreme (or more extreme) than the rejection point, then we reject the null hypothesis, and state that the result is statistically significant. Otherwise, if the test statistic does not reach the rejection point, then we cannot reject the null hypothesis and we state that the result is not statistically significant. A rejection point depends on the probability distribution, on the chosen alpha, and on whether the test in one-tailed or two-tailed. For example, if in our case we are able to use the standard normal distribution (the z-value), if we choose an alpha of 0.05, and we have a two-tailed test (i.e. reject the null hypothesis when the test statistic is either above or below), the two rejection points are taken from the z-values for standard normal distributions: below -1.96 and above +1.96. Thus if the calculated test statistic is in these two rejection ranges, the decision would be to reject the null hypothesis. Otherwise, we fail to reject the null hypothesis. Look Out! Traditionally, it was said that we accepted the null hypothesis; however, the authors have discouraged use of the word "accept", in terms of accepting the null hypothesis, as those terms imply a greater degree of conviction about the null than is warranted. Having made the effort to make this distinction, do not be surprised if this subtle change (which seems inconsequential on the surface) somehow finds its way onto the CFA exam (if you answer "accept the null hypothesis", you get the question wrong, and if you answer "fail to reject the null hypothesis" you score points. Power of a Test The power of a hypothesis test refers to the probability of correctly rejecting the null hypothesis. There are two possible outcomes when the null hypothesis is false: either we (1) reject it (as we correctly should) or (2) we accept it - and make a type II error. Thus the power of a test is also equivalent to 1 minus the beta (β), the probability of a type II error. Since beta isn't quantified, neither is the power of a test. For hypothesis tests, it is sufficient to specify significance level, or alpha. However, given a choice between more than one test statistic (for example, z-test, t-test), we will always choose the test that increases a test's power, all other factors equal. Confidence Intervals vs. Hypothesis Tests Confidence intervals , as a basis for estimating population parameters, were constructed as a function of "number of standard deviations away from the mean". For example, for 95% confidence that our interval will include the population mean (μ), when we use the standard normal distribution (z-statistic), the interval is: (sample mean) ± 1.96 * (standard error), or, equivalently,-1.96*(standard error) < (sample mean) < +1.96*(standard error). Hypothesis tests, as a basis for testing the value of population parameters, are also set up to reject or not reject based on "number of standard deviations away from the mean". The basic structure for testing the null hypothesis at the 5% significance level, again using the standard normal, is -1.96 < [(sample mean - hypothesized population mean) / standard error] < +1.96, or, equivalently,-1.96 * (std. error) < (sample mean) - (hypo. pop. mean) < +1.96 * (std. error). In hypothesis testing, we essentially create an interval within which the null will not be rejected, and we are 95% confident in this interval (i.e. there's a 5% chance of a type I error). By slightly rearranging terms, the structure for a confidence interval and the structure for rejecting/not rejecting a null hypothesis appear very similar - an indication of the relationship between the concepts. Making a Statistical Decision Step #6 in hypothesis testing involves making the statistical decision, which actually compares the test statistic to the value computed as the rejection point; that is, it carries out the decision rule created in step #4. For example, with a significance level of 0.05, using the standard normal distribution, on a two-tailed test (i.e. null is "equal to"; alternative is not equal to), we have rejection points below -1.96 and above +1.96. If our calculated test statistic [(sample mean - hypothesized mean) / standard error] = 0.6, then we cannot reject the null hypothesis. If the calculated value is 3.6, we reject the null hypothesis and accept the alternative. The final step, or step #7, involves making the investment or economic decision (i.e. the real-world decision). In this context, the statistical decision is but one of many considerations. For example, take a case where we created a hypothesis test to determine whether a mutual fund outperformed its peers in a statistically significant manner. For this test, the null hypothesis was that the fund's mean annual return was less than or equal to a category average; the alternative was that it was greater than the average. Assume that at a significance level of 0.05, we were able to establish statistical significance and reject the null hypothesis, thus accepting the alternative. In other words, our statistical decision was that this fund would outperform peers, but what is the investment decision? The investment decision would likely take into account (for example) the risk tolerance of the client and the volatility (risk) measures of the fund, and it would assess whether transaction costs and tax implications make the investment decision worth making. In other words, rejecting/not rejecting a null hypothesis does not automatically require that a decision be carried out; thus there is the need to assess the statistical decision and the economic or investment decision in two separate steps. comments powered by Disqus Trading Center
__label__pos
0.97291
Intereting Posts Использование Xtend для разработки Android (снова) AsyncTask.executeOnExecutor () до уровня API 11 Countdowntimer в минутах и ​​секундах Init: sys_prop: разрешение отказано uid: 1003 имя: service.bootanim.exit Android Notification.Builder: показать уведомление без значка Android – как добавить метод щелчка элемента в ArrayAdapter Стиль отображения контекстного меню Android edittext Android Webview (Phonegap): отключить автокоррект, автокапитализировать и автозаполнение не работает Android Studio 2.0 IllegalArgumentException: кнопка не существует 0 Android Studio: Код ошибки 1: Gradle: выполнение не выполнено для задачи ': app: processDebugResources' Android – расширяемый ListView – с помощью ViewHolder для оптимизации Проект не имеет файла default.properties! Измените свойства проекта, чтобы установить его Проблемы распространения бета-теста Android Android set scrollLeft при обновлении контента в событии прокрутки не обновляет позицию полосы прокрутки Как обрабатывать событие onClick на imageSpan в editText? Как передать изображение по умолчанию предыдущей активности, если изображение не выбрано В моем приложении у меня есть функция камеры. Пользователь может снимать изображение или выбирать из галереи. Активность B @Override protected void onActivityResult(int requestCode, int resultCode, Intent data) { super.onActivityResult(requestCode, resultCode, data); switch (requestCode) { case RESULT_LOAD_IMAGE: if (requestCode == RESULT_LOAD_IMAGE && resultCode == RESULT_OK & null != data) { selectedImage = data.getData(); imageView.setImageURI(selectedImage); } break; case REQUEST_IMAGE_CAPTURE: if (requestCode == REQUEST_IMAGE_CAPTURE && resultCode == RESULT_OK) { try { selectedImage = imageUri; getContentResolver().notifyChange(selectedImage, null); imageView.setImageURI(null); imageView.setImageURI(imageUri); } catch (Exception e) { Toast.makeText(this, "Failed to load", Toast.LENGTH_SHORT) .show(); } } } } submit.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { Intent returnIntent = new Intent(); if(selectedImage!=null) { returnIntent.putExtra("img_uri", selectedImage.toString()); } setResult(Activity.RESULT_OK, returnIntent); finish(); } }); Когда кнопка отправки нажата, она вернется к предыдущей операции. Активность А @Override public void onActivityResult(int requestCode, int resultCode, Intent data) { if (resultCode == RESULT_OK) { if (requestCode == PROJECT_REQUEST_CODE) { imgURI = Uri.parse(data.getStringExtra("img_uri")); if (mClickedPosition == -1) { if (obj != null) obj.addNewItem(imgURI); } else { if (obj != null) obj.changeItem(mClickedPosition, imgURI); } } } } Теперь мой вопрос заключается в том, как передать изображение по умолчанию в Activity A. Если пользователь не выбрал какое-либо изображение? Это xml imageView в imageView B <ImageView android:paddingTop="30dp" android:layout_gravity="center" android:layout_width="330dp" android:layout_height="wrap_content" android:adjustViewBounds="true" android:src="@mipmap/no_image" android:scaleType="fitXY" android:id="@+id/imageView" android:layout_weight="0.50" /> Если изображение не выбрано, я хочу, чтобы @mipmap/no_image возвращался к Activity B. Возможно ли это? Благодарю. Вы можете передать URI вашего изображения, хранящегося в ресурсах, следующим образом: submit.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { Intent returnIntent = new Intent(); if(selectedImage!=null) { returnIntent.putExtra("img_uri", selectedImage.toString()); } else { returnIntent.putExtra("img_uri", Uri.parse("android.resource://your.package.name/mipmap/no_image").toString()); } setResult(Activity.RESULT_OK, returnIntent); finish(); } }); Затем вы можете использовать его, как обычно, Uri, чтобы показать изображение в ImageView . ByDefault установил значение selectedImage в значение «default». Теперь в вашем addNewItem() и changeItem() переданном в imageUri if(imageUri.equalsIgnoreCase("default")) // set your default image that is @mipmap/no_image else // set image you are getting from imageUri по if(imageUri.equalsIgnoreCase("default")) // set your default image that is @mipmap/no_image else // set image you are getting from imageUri Вам не нужно возвращать образ по умолчанию в Activity A. Как и другие авторы, вам просто нужно подготовиться к тому условию, когда пользователь не выбирает какое-либо изображение. Измените свой onActivityResult для Activity A следующим образом: @Override protected void onActivityResult(int requestCode, int resultCode, Intent data) { super.onActivityResult(requestCode, resultCode, data); if(data!=null){ switch (requestCode) { case RESULT_LOAD_IMAGE: if (requestCode == RESULT_LOAD_IMAGE && resultCode == RESULT_OK) { selectedImage = data.getData(); imageView.setImageURI(selectedImage); } break; case REQUEST_IMAGE_CAPTURE: if (requestCode == REQUEST_IMAGE_CAPTURE && resultCode == RESULT_OK) { try { selectedImage = imageUri; getContentResolver().notifyChange(selectedImage, null); imageView.setImageURI(imageUri); } catch (Exception e) { Toast.makeText(this, "Failed to load", Toast.LENGTH_SHORT).show(); } } break; default: //Your default image goes here, when the user doesn't select either of the two options imageView.setImageResource(R.drawable.<default_image>); break; } } Код: @Override protected void onActivityResult(int requestCode, int resultCode, Intent data) { super.onActivityResult(requestCode, resultCode, data); switch (requestCode) { case RESULT_LOAD_IMAGE: if (requestCode == RESULT_LOAD_IMAGE && resultCode == RESULT_OK & null != data) { selectedImage = data.getData(); imageView.setImageURI(selectedImage); }else{ //set default } break; case REQUEST_IMAGE_CAPTURE: if (requestCode == REQUEST_IMAGE_CAPTURE && resultCode == RESULT_OK) { try { selectedImage = imageUri; getContentResolver().notifyChange(selectedImage, null); imageView.setImageURI(null); imageView.setImageURI(imageUri); } catch (Exception e) { Toast.makeText(this, "Failed to load", Toast.LENGTH_SHORT) .show(); } }else{ //set default } default{ //set default } }
__label__pos
0.983022
   HyperLink (PQL) Returns markup text to make the supplied text a hyperlink to the specified URL • Returned Output: Text • Library: PQL \ Dynamic \ Markup • Version: 2020.10.000 • Compatibility: Any content (regardless of data source) in the relevant parts of the application Syntax HyperLink( <Text> , <URL> ) * Click on the function's arguments above for more details on the input values. Comments • Use this function to mark-up dynamically created text programmatically. • The text is the visible text seen in the link. • The URL must be a web address for the site or page. • The text can be any static or formula driven text element. Examples This example returns the following resulting text with a live link to wiki.com: "This is a test" "This is a " + Hyperlink("test", "https://www.wiki.com")
__label__pos
0.675566
Python - still new, can't figure this out #1 I've been going through the Python exercises on here and for the most part I actually 'get' it but while trying to do a very simple Magic 8-ball script I've run into an issue that I saw in exercises and never figured out. Code is below. import random print ("Please enter a yes or no question and hit Return: ") question = input(" ") answer = random.randint(0,2) def reply(a): if answer == 0: print ("Yep") elif answer == 1: print ("Mebbe") elif answer == 2: print ("Nope") #return (a) print ("You asked:",question) print ("The answer is: ",reply) The issue is probably obvious but I am oblivious to it. What is wrong with my function? Why does it not display an answer? I realise I haven't called the function in this bit as I've tried a few ways like return and print (response) and clearly I don't know how to. I can get it to work without a function but part of my reason for doing this was to figure out how to use a function for this. Be gentle. #2 calling a function requires parentheses: reply(answer) between the parentheses supply arguments to satisfy function parameters #3 All that does is return the number, not the answer. Instead of, for example, "The answer is: Yep" it gives me "The answer is: 0". #4 Did you fix your function call? can i see an updated version of your code #5 This is returning the Yep, Nope, Mebbe but also a secondary 0, 1, 2 that is not the one from the original reply. I get, for example, "Yep" then underneath "The answer is: 1" which isn't correct so it seems to be calling "answer" two separate times instead of once. print ("Please enter a yes or no question and hit Return: ") question = input(" ") answer = random.randint(0,2) def reply(a): answer = random.randint(0,2) if answer == 0: print ("Yep") elif answer == 1: print ("Mebbe") elif answer == 2: print ("Nope") return (a) print ("You asked:",question) print ("The answer is: ",reply(answer)) #6 nope, nothing is returned? You only print i ran your code on repl, and it works fine? Where are you running the code? #7 Sorry I should clarify that the "Yep", "Nope", and "Mebbe" are my code's replies. I am running it through IDLE Python 3.6 #8 Difference between printing something and returning something. Make sure you know that difference. #9 Can i see a screenshot of you run this code in your IDLE? #10 Returning gives it back to the function, printing puts it in the console. I think? @stetim94 #11 you should return the answer (yep, mebbe, nope) return means the function hands something back, so using print() on the function call will print what the function returns #12 Ah I see it. Needed to return (answer) instead of return(a) OK this code works. import random print ("Please enter a yes or no question and hit Return: ") question = input(" ") answer = random.randint(0,2) def reply(a): if answer == 0: print ("Yep") elif answer == 1: print ("Mebbe") elif answer == 2: print ("Nope") return (answer) print ("You asked:",question) print ("The answer is: ",reply(answer)) The function prints as expected and I get the number printed too (Yep, answer is 0). How do I get it to print "the answer is: " and then add the actual response instead of the random number? I realise I could just change each to "The answer is Mebbe", "The answer is Yep" but I want to know how to do it so it prints the string, not the integer. #13 got this covered already? your function should return what you want printed. because your function call is printed, what the function returns will be printed #14 print ("The answer is: "),reply(answer) This gives me The answer is: Yep, Nope, Mebbe print ("The answer is: ",reply(answer)) This gives me Yep, Nope, Mebbe The answer is: 0, 1, 2 I know you're trying to let me figure it out but I do not see how to get what I want here which is The answer is: Yep, Nope, Mebbe #15 simply do: print ("The answer is: ",reply(answer)) then return Yep, Nope and Mebbe. Because your function call is now part of print, what the reply function returns will be printed on the same line #16 import random def reply(a): if a == 0: print ("Yep") elif a == 1: print ("Mebbe") elif a == 2: print ("Nope") answer = random.randint(0, 2) print(reply(answer)) #17 Try This :slight_smile: import random print ("Please enter a yes or no question and hit Return: ") question = input(" ") answer = random.randint(0,2) def reply(a): if answer == "0": print ("Yep") elif answer == "1": print ("Mebbe") elif answer == "2": print ("Nope") #return (a) print ("You asked:",question) print ("The answer is: ",reply) #18 i think the solution should be: import random print ("Please enter a yes or no question and hit Return: ") question = input(" ") answer = random.randint(0,2) def reply(a): if answer == 0: return "Yep" elif answer == 1: return "Mebbe" elif answer == 2: return "Nope" print ("You asked:",question) print ("The answer is: ",reply(answer)) this way, the function returns the possible answers, which are then printed (because reply function call is inside a print function call) the function parameter of reply will become reduntant, so we can remove it. #19 Ahhhhhh! I see what you were trying to hint at now. Instead of returning the integer "0, 1, 2" and printing "Yep, Nope, Mebbe" I had to simply return "Yep, Nope, Mebbe" to the function. Thank you @stetim94 that was driving me crazy. #20 i was trying to hint you that 3 answers ago :wink:
__label__pos
0.998672
ANSWERS: 1 • The chance of picking a red marble first is 6/10. After doing so, there are 5 red marbles remaining among 9 total. So the chance of picking a red marble second is 5/9. The probability of both events is therefore (6/10)(5/9) = 30/90 = 1/3 Copyright 2020, Wired Ivy, LLC Answerbag | Terms of Service | Privacy Policy
__label__pos
0.99271
Can you change the timestamp on a picture? Can you change the timestamp on a picture? Can you change the timestamp on a picture? Go up to the menu bar and select "Tools," then select the option "Batch Adjust Time Stamp." Select "EXIF (Exchangeable Image File Format) Date/Time Original" in the pop-up window and click "Next." This will open a window marked "New Time Stamp." Enter the new date and new time, and click the "Apply New Time Stamp" ... How do I change the date a picture was taken? To do either of those things, open Photo Gallery and select one or more photos. Then right-click and choose Change Time Taken. You'll see the Change Time Taken dialog box, which you can use to modify the date or adjust for a different time zone. Can you edit timestamps? Android Police First is the ability to manually edit the timestamp on a photo. A new edit button shows up when you hover over the timestamp on a photo's info and lets you change the date and time the pic was taken at, in case your phone or camera wasn't appropriately set and the photo was stamped inaccurately. Can you see metadata on iPhone Photos? Tap on the Photo Library icon from the lower left of the screen. Note: If prompted, allow the app to access your Photos app. Select the album and the photo you would like to view the metadata for. To view the complete data, tap on Metadata and then View all. How do I turn off timestamp on iPhone Photos? Select the iPhone photo that you want to modify the date, gps or other metadata and click the Quick Action button, then choose the "Copy exif date to the create date", or "Edit date info", "Remove GPS info" or other menu item. How do I change the date on photos in Windows? Right-click the photo you want to change the date for, then click [Properties]. Click the date or time of [Date taken] and enter a number, then press the [Enter] key. Date will be changed. Why do my photos have the wrong date? If the camera had incorrect time settings at the time of shooting, the timestamp in the metadata (EXIF / IPTC) created by the camera will not be correct. ... Under the tab “Set File Date” you can shift the date “Created correction” by hours, minutes and seconds and compensate for the wrong time set on the camera. How do I change the date on a JPEG? Changing the date of a specific photo Right-click the photo you want to change the date for, then click [Properties]. Click the date or time of [Date taken] and enter a number, then press the [Enter] key. Date will be changed. Can you change the metadata of a photo? At the bottom of the photo screen, you'll see four options: share, edit, info, and delete. Go ahead and give the “Info” button a tap—it's the little “i” in a circle. You'll see the photo's EXIF data displayed in a nice, readable format that includes the following data: Date and time taken. How do you change the time on a photo? Next, go to Images and click Adjust Date and Time. This will allow you to change the dates for all the selected photos at once. You can also use this method to adjust the time stamp on each individual photo. How can I add a time stamp to a photo? To add date and time stamp, first, add a photo of BMP, PNG, or JPEG format by using its Choose Image option. Immediately, after loading the photo, you can preview the photo with current date and time on the top left of the photo. To set any other date and time, you can use its date and time changing options present at the bottom of the interface. Can you edit the timestamp on Google Photos? If you’ve already uploaded the pictures on Google Photos, you can still edit the timestamp but you can only do that one image at a time. Also, the date editing option is only available on the Google Photos website and not inside their iPhone or Android apps (yet). Can you change the date on a time stamp? Other metadata—like the original time stamp, for example—is “read-only,” which means you can’t modify it with standard software. Even if you try to fiddle with the binary code using a hex editor, it may not be obvious how to change the date and time without corrupting the file. Related Posts:
__label__pos
0.994105
Education How to Calculate List and Tuple Difference There is an investigation of the list and tuple difference in terms of their ability to modify one another. Tuples cannot be modified, but lists can be. Tuples have a remarkable aptitude for recall since they do not forget what they have learned. Unlike lists, which have access to 46 methods, tuples have access to only 33. [] and [] are used to list and tuple difference, respectively (). This section allows you to compare the syntax list to the tuple. A benefit of lists is their ability to reduce tuples. Lists require more time to build and retrieve than tuples. When juxtaposed, the list and tuple difference becomes less apparent. No similarities exist between the two. Unlike tuples, though, lists can extend or contract dynamically. Despite their distinctions, tuples and key lists have several characteristics. Many objects are categorized and stored using structures. Their system stores a substantial quantity of data objects. With the index, you may be able to easily discover specific information. This table allows you to compare the difference between lists and tuples side-by-side. Examine our accelerated Python course built just for working professionals. Before list and tuple difference in Python, we first define those words. Lists The most often used data structure in Python may monitor duplicate data. Python’s lists and tuples offer an alternative to arrays for storing and organizing data by permitting the creation of collections of things with similar features. It enhances operations on multiple values. Make a folder for each musical genre you own on your PC. The ability of Pythagoras to translate list and tuple difference is useful for database organization. Tuples Tuples have the advantage over flat lists of being able to store and retrieve data in a tree structure. Splitting them with commas simplifies their management. Once made, they cannot be altered. Tuples cannot grow in size, although lists may. Restrict the number of things that can be added and removed. Immutability is primarily advantageous since it assures and accelerates procedures. While the list and tuple difference in Python is related, there are significant differences that make them useful in a variety of situations. The following is a summary of the arguments I made in my most recent blog article comparing tuples vs lists. A database of structures By utilizing square brackets, a list can be introduced (]]). Below is an example of a Python list declaration. The list includes four unique items. [1,2,3,4,5] (list of numbers) (list of numbers) output (list of numbers) “A,” “by,” “ca,” and “d” were added to the alphabet list variable following “I.” Here is a printed version of the alphabet from A to E (alphabets list) Lists can include practically any data. Start with these steps. When you press [a, 1]b, [2, 3], and [4, 2], the resulting lists are merged. ‘c. print(mixed list)’ (mixed list) Lists are useful for organizing data. Essentially, a nested list is a list within a list. placement of homes = [1, 2, 3, [4,5], 6, 7, 8] overlapping (nested list) (nested list) Multiple-Participant Construction in Grammar As shown by the symbol, this is the initial element in a tuple (). Below is Python code displaying a tuple declaration. Consider the number of tuples to be (1,2,3,4,5) The print alphabet tuple consists of the letters A through E. (num tuple). To print the alphabet, type print(alphabets) (alphabets tuple) A single list may include various types of data. Consider the tuple (1, 2, 3, 4, ‘a,’ ‘b,’ ‘c,’ ‘4) as an example for introducing the concept. This sequence is not fully composed of ones. results(non-sequential tuple) (mixed tuple) Lists are useful for organizing data. Essentially, a nested list is a list within a list. The sequence (1, 2, 3, (4,5), 6, 7, 8) is an illustration of a nested tuple (nested tuple) Slight Syntactic Difference The comparison reveals the list and tuple difference. The syntax of Python must be adjusted. Both the syntaxes for lists and tuples, a pair of brackets, and a set of parenthesis, make it clear that these two structures are separate. In the first part, we examine the differences between list syntax and tuple syntax. For instance: (list size = [10,20,30,40]). This can be written as tup num =. (10, 20, 30, 40) Mutability The degree to which a list or tuple is mutable or immutable is simply one of many differences between the two. Python tuples have a fixed size, unlike lists, which can be expanded to meet the needs of a program. Hence, lists have greater functionality than tuples when it comes to accomplishing specific tasks. People who work with lists, such as data scientists, routinely rearrange their lists. Drop everything and leave immediately. The list changes as items or entire groupings are eliminated. Tuples are more adaptable than their components, which are immovable once separated. Immutable tuples cannot be replicated. Changing a single list item modifies only that item. Indexing enables iterative list node updates. Each of the offered values is independently customizable, providing precise control. Operations While both tuples and lists are valuable data structures, lists provide some advantages. Rearranging the order of a list, deleting duplicates, and adding new items are examples. Functions Operations such as sort, len, max, min, any, sum, all, and sorted are accessible for both types of data in Python. Following the examples are brief arguments for various uses. The Max(tuple) method returns the tuple’s greatest element when given a tuple as an input. Min(tuple) returns the tuple element whose value is the smallest. Use the tuple(seq) method to create a tuple from a list of items. CMP is used to compare the contents of two tuples (tuple1, tuple2). Size Tuples in Python consume more memory than lists since they cannot be modified. Compared to arrays, tuples have less storage capacity. This method of constructing tuples is far more efficient than using a list when dealing with numerous distinct components. The “size” of a tuple indicates how much information it can store. Use the usual method of measuring distances along lines to get a length. Length The difference in length between the two data formats is substantial. Unlike lists, tuples have a constant width and height dimensions. Hence, unlike tuples, the size of a list can be adjusted after it has been formed. Methods Python list functions Append(), Insert(), Clear(), Sort(), Pop(), Reverse(), and Remove() are handy (). Certain manipulations operate only on lists, whereas others also operate on tuples. Such resources consist of the count() and index() routines. Debugging For system-wide debugging, tuples are superior to lists due to their immutability. When dealing with simple activities or small amounts of data, a list is more effective than a spreadsheet. As they are editable, lists have an advantage over tuples. Uses Before making a final decision, the programmer must consider factors such as the frequency and nature of data updates. Tuples provide a convenient method for storing and organizing data, similar to dictionaries but without the need for keys. Tuple data is simple to comprehend. Lists may be classified. Tuples save both time and space when compared to infrequently used lists. As lists are immutable, it will be simple to coordinate any upcoming modifications. Conclusion What list and tuple difference? If you’re having problems distinguishing tuples from lists, this post can help you. Although sharing the name of Python data structure, there are significant differences between the two. Lists can expand or contract as necessary, but tuples have fixed dimensions. Tuples are therefore effective for decreasing wait times. Also read  Related Articles Leave a Reply Your email address will not be published. Required fields are marked * Back to top button
__label__pos
0.826386
1. This website uses cookies to improve service and provide a tailored user experience. By using this site, you agree to this use. See our Cookie Policy. Dismiss Notice How much spam can you create? Discussion in 'BlackHat Lounge' started by TheWisdomGuy, Jul 22, 2015. 1. TheWisdomGuy TheWisdomGuy Registered Member Joined: Sep 24, 2014 Messages: 58 Likes Received: 14 How much can you spam on sites like youtube, facebook, other social networks etc before they take any legal action???? I've seen some news but all of them were like spam kings Just google 'online spam jailed'   2. asap1 asap1 BANNED BANNED Joined: Mar 25, 2013 Messages: 4,959 Likes Received: 3,197 As long as you dont mess with their money they can care less and will let the spam filters handle it. If you really start to cause trouble then I guess that's when they come after you.   3. classybabe classybabe Regular Member Joined: Mar 20, 2014 Messages: 261 Likes Received: 67 I don't think they care. I see spam on there all the time and has never seen anyone jailed for spam  
__label__pos
0.560925
Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++ Selamat datang di Blog Belajar Pemrograman, Kali ini saya akan memberi Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++. Berikut adalah source coding program yang bisa anda copy / download : #include <stdio.h> #include <conio.h> #include <string.h> #include <iostream.h> void main() { char nama[20], kelas[10], password[10]; atas: cout<<"SELAMAT DATANG"<<endl; cout<<"SILAHAKN LOGIN TERLEBIH DAHULU"<<endl; cout<<endl; cout<<"USERNAME = ";cin>>nama; cout<<"KELAS    = ";cin>>kelas; cout<<"PASSWORD = ";cin>>password; if(strcmp(password,"1234")==0) { cout<<"TERIMA KASIH ATAS PASTISIPASINYA"<<endl; cout<<endl; cout<<"TEKAN SEMBARANG"<<endl; goto bawah;} else { cout<<"PASSWORD ATAU USERNEME SALAH"<<endl; goto atas;} bawah: getch(); } Untuk lebih jelas nya agar program tidak error pada saat menjaankan, anda bisa melihat video Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++, secara detail dan jelas dibawah ini : Sekian tutorial pemrograman kali ini semoga bermanfaat dan bisa diterapkan dengan baik dan lebih sempurna. Tag : Turbo C++, barlond turbo C++, membuat program di C++, source coding lengkap program C++, coding C++, pemrograman C++, turorial C++, programming C, bahasa pemrograman C++, pemrogramana C++, Tutorials programming di C++, Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++. Cara Menghasilkan Bitcoin Silahkan komentar pada kolom dibawah ini jika ada yang ingin ditanyakan tentang Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++. 0 Response to "Tutorial Membuat Program Penggunaan IF ELSE Pada Bahasa Pemrograman C++" Catatan: Hanya anggota dari blog ini yang dapat mengirim komentar.
__label__pos
0.997676
I am running windoze 7 64. When running the win 7 nandpro or if running any version of nandpro in a virtual xp machine I get an error when trying to use a virtual device. I'm pretty certain my syntax is right, as the same files on an ancient machine running xp works fine. I have to use virtual as my spi programmer isn't nandpro compatible so I have to write changes to the nand dump as a virtual and then flash the whole file back with my usbjtagnt programmer. I'm thinking it's some sort of permissions issue with windoze but haven't found it yet and wonder if anyone had already invented this particular wheel ? Thanks for ANY input.
__label__pos
0.962754
That's where the SUMIF function comes in handy, along with the more capable SUMIFS function. Under the Sort On heading, leave the setting as Cell Values. 2 - Please click on the How-To link at the end of this post and read the instructions on how to format example data so that it is easier for us to read. Excel Advanced Filter Criteria Examples Extract Items in a Range. In Excel, how do i return entire row, but skip rows that don't meet criteria? I have a master sheet with potential expenses listed. Thread starter Bipin619; Start date Feb 25, 2019; B. If you are working for a financial analyst company, you might need to get or import daily stock. However, really ugly xls file, that are more or less just the pdf in xls, but not a nice table structure. Bottom Line: Learn how to use VBA macros to copy & paste data from one Excel workbook to another, including adding data to the bottom of an existing range or replacing data. Become a Web Query Wiz Take a few minutes to think about what mundane or repetitive tasks YOU do on the internet, particularly cases where you go to the same sites over and over to get data. I'm trying to extract a list of unique values, based on a number of criteria, including some "OR's". Extract unique items for dynamic data validation drop down list How can you extract unique items from a list that contains multiple instances of each item, and use this list in a dynamic dropdown so that any time a new item is added to the list, or, if the list expands, your data validation list is updated?. Find and merge data from 2 Excel spreadsheets in 5 steps. All the clients with ID numbers should display on the ID sheet. Please see below table: So what I want is, when I select 'Not required' from one drop-down, and 'Doc 2' from another, it should give me a list of all the Systems that match this criteria. Thanks for your help in advance. Hi all, I am having difficulty putting two logic together, that is have a formula that works to return the data from 2 criteria using simple Match logic, BUT having difficulty using multiple return values in row logic with it. Now, I want to extract data based on some fixed criteria and I want to fill the data on a different worksheet. This function works in any version of Excel in Windows and Mac, and also in Google Sheets. How to pull data from multiple worksheets based on criteria in two lists I have a workbook with 25 worksheets pertaining to different suppliers. For once, I am going to stick with Excel 2003 and earlier because there is a SUMIFS function in Excel 2007 that resolves our query in seconds:. That’s two criteria and SUMIF will not work with multiple conditions. Here is an example that shows how to do it using one criteria (and extract data based on that criteria). All rows with value F in S column and G or P in T column 3. split the data based on name column in sheet1. How do you extract data from a list, that is dependent on a criteria. Note that each of the SUMPRODUCT formulas shown here references cells B3 and C3 to filter on Account and Product Line. When the filter conditions change often, Excel’s filters fall short. In different cells there are, name, age, address, occupation, date reporting etal. I am trying to build something like a dashboard with couple of drop-downs, that would search a table and provide me a list, based on what I have selected. In this example, we will use "=Tablets" in Product column. Page 23-General discussions of Microsoft Excel. Select all the data elements in the Table. The IF portion of the function determines what data matches the specified criteria and the SUM part does the addition. Using xlrd module, one can retrieve information from a spreadsheet. One option will utilize an array formula using an array constant. This guide will help you build formulas that work the first time with over 50 examples. So, the columns that I want to return data from, columns 2, 3, and 4 on the "transactions" are the "date", "quantity", and "description". The list has some duplicate values. Now select the first cell of the column and click on the Data tab. I solve this using formulas. Extract all data for a list heading by entering no criteria under that heading. I can extract data from Table 1 to excel, which has the DATE field. Extract data based on pattern with respect to specific column Excel Programming Extract data based on pattern with respect to specific column Microsoft Office Forums > Microsoft Excel > Excel Programming. Sample data. This post is broken down as follows. 1- Vlookup function 2- Index & Match function “=INDEX (column to return a value from, (MATCH (lookup value, column to lookup against, 0))” I prefer IndexMatch function for many reasons: 1- IndexMatch works in two directions (right to left and left. Fortunately, there are several alternatives. Good afternoon All, I am looking for help with a formula to extract a list from a database based on 2 criteria. But suppose that you want access to only one cell value in the LINEST results, perhaps to accommodate a worksheet layout in a routine report. Assume that You are going to extract Cell A1 of the sheet1 to cell A1 of sheet2. GetElementByID(YourTableID). Use data from a cell as a parameter value. This guide will help you build formulas that work the first time with over 50 examples. Re: Extract data from one worksheet into another if criteria is met I tried this but it didn't work for me. Excel - Return Max/Min Value Since. A second worksheet "VOC_ASST" has column A "Name". But suppose that you want access to only one cell value in the LINEST results, perhaps to accommodate a worksheet layout in a routine report. Select the Data Tab, then Advanced under the Sort & Filter; Select copy to another location, then next to List range, highlight the range with duplicates (A1:A22). I want to filter the data in columns M,N,O and copy the data. The SUM function in Excel allows you to add up the values in a range of cells. See my article Excel Stock Quotes in Excel to see how you can publish a Google Spreadsheet and then use a web query to get the data into Excel. In this article, I shall show you how to pull or extract data from a website into Excel automatically. Re: Extracting entire row to new sheet based on criteria Please read the Forum Rules at the top of the screen - you should not post a question in someone else's thread, but start your own instead. And you want to split the data into multiple worksheets based on the product name column. Access MID Function. Excel 2013 - Extract Data based on date criteria, and not allow duplicates Hi, I am trying to extract "Supplier Code" Column A from Sheet 1, based on Date Criteria in Column D to another sheet and not have duplicates. I want this information for all the 6000+ rows. Can you help me with a macro for copying data based on criteria to a sheet in another workbook? I have an excel sheet with columns from A- O. You can use the advanced filter for more complex criteria filtering. You just select the master table - the one that you want to update. This way, you can dynamically change your filter criteria, and your formula keeps up. This will give us our dynamic column name within the Excel Table. From the drop down list select (Custom…) The Custom AutoFilter dialog opens. How to extract distinct / unique rows in Excel; How to get unique values in Excel. For example, get price for Large Jacket. The above data contains more than 250 rows. This example teaches you how to apply an advanced filter in Excel to only display records that meet complex criteria. Select the range included headers you want to extract rows from, click Data > Filter to add the Filter icons beside headers. How may I retrieve data from an Excel table based on a variable number of criteria? The criteria for database functions need to be laid out in pairs of cells. Extract PDF Pages Based on Content Posted on April 25, 2014 by Karl Heinz Kremer How would we identify pages in a PDF document that contain a certain word and extract those pages into a new document?. Assume that You are going to extract Cell A1 of the sheet1 to cell A1 of sheet2. Based on above mentioned VB I have prepared the model and everything is working fine except a scenerio where I want "all" (OPEN, CLOSED & DISTPUTED) PO to. My data is already in Excel and no need to run Text-To-Column. This function works in any version of Excel in Windows and Mac, and also in Google Sheets. Re: Extract Data from one sheet into another one based on criteria I think you get better help (and so I advice you), if you post a new question on the forum (Excel programming). When we enter our two criteria in the next step, the 1 in the MATCH function simply means: "Look through the rows in the data and return the row number where all of our criteria are TRUE". In the result tab, I have extracted a list of unique customers into column B. Hi all, I am having difficulty putting two logic together, that is have a formula that works to return the data from 2 criteria using simple Match logic, BUT having difficulty using multiple return values in row logic with it. I am trying to build something like a dashboard with couple of drop-downs, that would search a table and provide me a list, based on what I have selected. and transfer your saved. He comes across a formula that looks pretty complicated, but others had claimed it worked for them, so Nathan decides to give it a whirl. To extract multiple matches to separate cells, in separate rows, you can use an array formula based on INDEX and SMALL. In the example shown, the formula in E5 is: Excel formula: Extract multiple matches into separate rows | Exceljet. But this isn't a major problem for me with this data. The first example uses an array formula, thereby avoiding the use of a helper column. Applies To: Microsoft ® Excel ® 2010, 2013 and 2016. If you work with Excel, I am sure you have used (or at least heard about the regular excel filter). Until it meets the minimal functions that the Excel export does, I'd expect Atlassian to keep getting support calls to re-enable the working export. Today, I encountered a situation where I had to separate the full names of our customers into the first and last names. [sum_range] F3:F18. In Ms Access I have one big table that collect all data. Extract data based on pattern with respect to specific column Excel Programming Extract data based on pattern with respect to specific column Microsoft Office Forums > Microsoft Excel > Excel Programming. In this example, we will use "=Tablets" in Product column. My data is on the tab "Database", and list needs to be on the "report sheet" sheet, where I have placed the 2 criteria. See how to join cell content to get a dynamic data set label using the Ampersand Join symbol. One of the common tasks for people working with text data is to extract a substring in Excel. AVERAGEIFS allows you to get an average of a range of data base on specific criteria(s) (hence the IFS in AVERAGETFS). csv database into Excel for further editing. The data form in Excel allows you to add, edit and delete records (rows) and display only those records that meet certain criteria. Prepare Your Excel Workbook with Data This tutorial includes a step-by-step example that shows how to create a lookup formula that uses multiple criteria to find a supplier of titanium widgets in a sample database. Here is an example that shows how to do it using one criteria (and extract data based on that criteria). I want to export data from 4 Access tables to Excel using VBA. It currently pulls back 30 - 40 rows from the 600K rows. Understanding with an example will be easy. FALSE equals a 0. Now, we want user to select Name, Manager& Process from dropdown lists. The data includes four blank rows above the list range that will be used as a criteria range (A1:C4) and a list range (A6:C10). is hard because it requires a good understanding of how Excel handles data. And you want to split the data into multiple worksheets based on the product name column. One of the common tasks for people working with text data is to extract a substring in Excel. Change values and formulas, or add your own values and formulas and watch the results change, live. At the very least, these functions have been used for the. The data form in Excel allows you to add, edit and delete records (rows) and display only those records that meet certain criteria. > Excel Macros and VBA Split Excel Sheet Into Multiple Files Based On Column Using VBA Do you have a big data on excel sheet and you need to distribute that sheet in multiple sheets, based on some data in a column?. csv database into Excel for further editing. Open the Advanced Filter dialogue box in the Data tab (located in the Sort & Filter tools group). See how to create an IF function formula that will pull data from column C based on criteria in column B. Access has a number of powerful tools to enable specific dates and date ranges to be specified in criteria. Step 2: Select data series using Data Series Navigator: click the category or hold your mouse over the selected category to display a direct link to Navigator (Criteria Search), or click the Search button next to the Enter Series box in the Search window. In this Excel tutorial from ExcelIsFun, the 186th installment in their series of Excel magic tricks, you'll see how to extract records that meet two conditions (criteria) and list them in order with no extra rows between records as well as how to create a formula to extract part of a table from a larger table based on two criteria. Understanding with an example will be easy. Tells Excel that your VLOOKUP return (13659) is part of a formula to look for, and creates a universal formula you can drag into other cells. The circumstances, requirements define which approach is the best to apply. Excel - Return Max/Min Value Since. e for ex Feb-15) and to report the relevant cells below. Answer: The date range is specified in cell G1 and G2, see image above. The data can be placed into either a DataGrid control, or an Excel spreadsheet. How do you extract data from a list, that is dependent on a criteria. GetElementByID(YourTableID). Excel has a set of Text Functions that can do wonders. We’ll briefly explore one simple example of using VBA in Excel to extract values from one worksheet , but this is just the tip of the iceberg for what Excel and VBA can accomplish together. This is one of the most used Excel features for those who use Excel for their data analysis job. Now my user want to filter selected data from query and export only filtered data to Microsoft Excel. Don’t Put Data Below the Extract Range. This article shows a simple way of how to create a drop-down list. Instead, there is a set of functions that can extract a data table from a larger data set based on specific criteria that you set. Also see: Excel Magic Trick 758. Copy all the cells in the table below, and paste into cell A1 in a new worksheet in Excel. org Forums is " How Do I extract data from a list that matches a criteria ". All rows with value F in S column and G or P in T column 3. This can be easily done to include more filters - Extract Data based on criteria. This filter criteria will be reapplied after new data is added and the pivot table is refreshed. How To: Extract records with 2 criteria with Excel's AGGREGATE How To: Extract records via partial-text NOT criteria in Excel How To: Use the Advanced Filter tool to extract data in Excel How To: Filter and extract records with multiple criteria in Microsoft Excel 2010. As shown in the pic below, I have a created a drop-down list with country names. You can also use a free tool called Tabula to extract table data from PDF files. Please refer (link) to this question, so other forummember can see (read) which solutions already been given. We can use Advanced Filter feature to filter multiple columns with different criteria. It's one of the most common data crunching task in Excel. How to import data from the web in Excel: If you manually copy and paste data from a web page into Excel, you might find that all the data gets pasted into one cell. it is a duplicate, it contains unnecessary d You want to extract the first/ last time you see the data In the spreadsheet below we need to identify and remove the red block as it is a duplicate of the first line. CSV and Text – you can upload CSV or tabular Text files. And you want to split the data into multiple worksheets based on the product name column. Sample data. Under the Sort On heading, leave the setting as Cell Values. When selecting the list range, ensure that the cell with the list heading is also selected to avoid duplicate values. Are you looking to get unique values (or remove duplicate values) in Excel, and as quick as possible? Have no fear - doing so is easy! Excel formula, Advanced Filter, Remove Duplicates command, PivotTable, and VBA as ways of extract unique values quick. CSV and Text – you can upload CSV or tabular Text files. (It does not matter if you select a table at this point. In Excel, you can filter the rows meeting the criteria first, and then copy them to another location. See my article Excel Stock Quotes in Excel to see how you can publish a Google Spreadsheet and then use a web query to get the data into Excel. In this short article, I will show using Excel how you can get data from another worksheet based on a cell value. Specifies the member property field on which the label filter is based. Delete Rows Based on Column Criteria This example searches a user entered column for a user-entered string and then deletes any rows where the string is found. criteria "="&VLOOKUP return. My data is already in Excel and no need to run Text-To-Column. If you have a list of records contained in a range of cells it is possible to extract all records that meet a specified set of criteria and place them elsewhere in your workbook. Based on above mentioned VB I have prepared the model and everything is working fine except a scenerio where I want "all" (OPEN, CLOSED & DISTPUTED) PO to. If you enter two criteria on the same row in the criteria range, you create an AND statement. More Tutorials. Excel 2010 Solution. He comes across a formula that looks pretty complicated, but others had claimed it worked for them, so Nathan decides to give it a whirl. Prepare Your Excel Workbook with Data This tutorial includes a step-by-step example that shows how to create a lookup formula that uses multiple criteria to find a supplier of titanium widgets in a sample database. I'm trying to extract a list of unique values, based on a number of criteria, including some "OR's". When you need to find and extract a column of data from one table and place it in another, use the VLOOKUP function. How to combine Excel INDEX and MATCH functions to do a lookup based on multiple criteria. Hello, Need assistance with macro to extract data from multiple sheets based on matched criteria and sum up the values. It allows you to find data in one table using some identifier it has in common with another. Copying filtered data to another worksheet should be simple but it's not. To extract only the records for Ice, use the following format: ="=Ice" Download the Sample Workbook. If you've ever used Excel, then you've probably experienced the agony of choosing an incorrect formula to analyze a data set. I need to to so show the different clients/taxes on separate sheets so that I can add details to it on the sheets. To avoid any confusion, first off, let's agree on what we call unique values in Excel. If you are working for a financial analyst company, you might need to get or import daily stock. This filter criteria will be reapplied after new data is added and the pivot table is refreshed. Excel how to insert data based upon criteria. Excel Data Analysis - Filtering - Filtering allows you to extract data that meets the defined criteria from a given Range or table. Hence, the next time you need to copy cells based on certain criteria, you can choose the most suitable method. How to use VBA macro to split data into different files based on selected column in Excel. Tabula will return a spreadsheet file which you probably need to post-process manually. i have searched it in google. Filtering allows you to extract data that meets the defined criteria from a given Range or table. SUMIFS is a great formula, not to just to sum by different criteria but also to grab data out of data tables by criteria. The first example uses an array formula, thereby avoiding the use of a helper column. In order for my macro to function correctly, I need to be able to format the pivot table to the expected format, and I’ve been able to do this for rows, autoformat, etc. 15 Excel Data Analysis Functions You Need to Know. But, if you need to vlookup the relative value based on multiple criteria, how could you deal with it?. Extract data group filtering filters data extracted in the data group. The actual data that is summed, based on what is defined by the. Now my question is how to use excel to get results quickly. Using multiple criteria in Excel Lookup formulas By Emily Warn (This post is written by JP Pinto, the winner of the Great White Shark Award given for the best article written about VLOOLUP during VLOOKUP Week. In the example shown, the formula in E5 is: Excel formula: Extract multiple matches into separate rows | Exceljet. Thinking about it, it may help if I tell you exactly what I'm trying to do. To avoid any confusion, first off, let's agree on what we call unique values in Excel. Function Syntax. I can normally find everything on here but this has me stumped. > Excel Macros and VBA Split Excel Sheet Into Multiple Files Based On Column Using VBA Do you have a big data on excel sheet and you need to distribute that sheet in multiple sheets, based on some data in a column?. I have a file with 600,000 rows of data. Function Syntax. Drop-down list/data validation list: If you have an advanced knowledge of Excel and understand VLOOKUP and formulas, you can use a drop-down list (also known as a data validation list) to create interactive charts. Use multiple criteria to retrieve data from tables that do not have a unique identifier or index for each row. Excel VBA to Consolidate Data which Meets Criteria. Extract records from a data set based on criteria 2. SUMIFS is a great formula, not to just to sum by different criteria but also to grab data out of data tables by criteria. this is my output. What I need to do is extract the Value from the "ending" column based on the matching BE, Major, and Sub. Then, filter your data based on that list, so you don't have to check all the items manually each time. See screenshot: 2. I want to extract from each sheet the name, age, occupation and address based on date reported with the criteria within the 7 days. This tutorial will show you how to list values from a table based on filter criteria using sub-arrays and the SMALL function. In this tutorial we are going to look at two ways of using VLOOKUP to pull data from a range of multiple columns at one time. Filtering allows you to extract data that meets the defined criteria from a given Range or table. The rationale behind creating a database is typically to store data that you need to access for business purposes. Copying filtered data to another worksheet should be simple but it's not. Use a constant parameter value for a query. Sometimes we get the data set from another person and the need to do a lot of work to clean this data set. Following is the picture of criteria that need to be looked for data that occurs between 1/1/2014 & 12/31/2014 Press ALT + A + Q for launching advance filter dialog box & follow the below ranges in the fields This example will extract the data that contains between the two criteria dates. You can use the advanced filter for more complex criteria filtering. Criteria are a key concept in Excel, but building useful criteria for text, numbers, dates, times, etc. Extract a Date from a Date-Time Number in Excel I recently saw a spreadsheet with the following function =LEFT(A2,8) where cell A2 was equal to 08/06/12 12:23 PM. Additionally, we'll describe how to subset a random number or fraction of rows. And of course this workbook will be open, meaning there will be a memory leak. Page 6 of 20 Example 1: Extract Records based on AND Logical Test (AND Criteria) 1) Our Goal is to extract records and fields with this criteria: Records that match this Criteria: Date >=10/1/2016 AND Date <=3/31/2017 AND Customer =Amazon. It also enables you to animate charts. How to Highlight Cells That Meet Certain Criteria in Excel One of the more basic Conditional Formatting rules that you can create is the highlighting of cells that meet some business criteria. Gender like Gender_equals. In this MS Excel tutorial from ExcelIsFun, the 671st installment in their series of digital spreadsheet magic tricks, you'll learn how to use the AGGREGATE function to create a formula for extracting records with two criteria. In this post, I’m going to show you a better way to extract values from a string than using then VBA InStr function with Left, Right or Mid. I attached an example of both table (which are actually in 2 different workbooks) In the worksheet "report", I would like to search in "quotes table" the value input in report/A2 (i. Filtering and Deleting data with multiple criteria using VBA Posted on December 3, 2016 by Vijay | 0 Comments In this post I’m discussing how to delete filtered data using VBA with the criteria you specify. Click here to visit our frequently asked questions about HTML5 video. In this article, I shall show you how to pull or extract data from a website into Excel automatically. Extract unique items for dynamic data validation drop down list How can you extract unique items from a list that contains multiple instances of each item, and use this list in a dynamic dropdown so that any time a new item is added to the list, or, if the list expands, your data validation list is updated?. Data frames can be indexed in several modes. This allows you to do things like find cells that meet certain conditions, like checking if values in a cell are. For examples, assuming that you have a file of sales data for all products. Until it meets the minimal functions that the Excel export does, I'd expect Atlassian to keep getting support calls to re-enable the working export. This example demonstrates the formatting of cells that fall under a hard-coded value of 4000. You can work with sample data and formulas right here, in this Excel for the web workbook. The above data contains more than 250 rows. range: The range of cells that you want to apply the criteria against. I used a date range start and end in two cells in excel to extract the data through MS query. The items on the left of screen with the Red box around it are the criteria for the database. When selecting the list range, ensure that the cell with the list heading is also selected to avoid duplicate values. Change values and formulas, or add your own values and formulas and watch the results change, live. An Excel advanced filter and a macro to extract records from a list. This tutorial will show you how to list values from a table based on filter criteria using sub-arrays and the SMALL function. Filter Excel Data to a Different. There is no SUMPRODUCTIF function in excel, but you can use the SUMPRODUCT function to solve one criteria or multiple criteria questions, in the same column or different columns. Now my question is how to use excel to get results quickly. In this tutorial we'll take a look at how to use an array formula utilizing INDEX, IF, SMALL and ROW functions to extract unique values from a large list into a table based on a specific criteria. Excel Safety is Cause for Concern. (I will explain in detail with an example. Extract row of data based on 2 criteria I have a table which I want to extract a row which meets 2 criteria. This post also covers a method to remove duplicates from a range. Ask Question Asked 3 years, 11 months ago. Excel VBA : Concatenate Text Based on Criteria Deepanshu Bhalla 6 Comments Advanced Excel , Excel Macro , VBA This article explains how to concatenate text based on unique ID values. The rationale behind creating a database is typically to store data that you need to access for business purposes. Extract unique values based on multiple criteria hello community, I'm looking to extract values from a spreadsheet based on multiple data. In this Excel tutorial from ExcelIsFun, the 186th installment in their series of Excel magic tricks, you'll see how to extract records that meet two conditions (criteria) and list them in order with no extra rows between records as well as how to create a formula to extract part of a table from a larger table based on two criteria. The actual data that is summed, based on what is defined by the. Learn all about rich data types in Excel and how to use them to get more from your data such as current financial data, currency rates, cryptocurrency values and geographical data with just a couple clicks. In the example shown, the formula in F5 is: Excel formula: Extract multiple matches into separate columns | Exceljet. Section 1: How to extract from fixed sized strings. Re: Extract data based on multiple criteria Thank you very much for the feedback and rep. One of the common tasks for people working with text data is to extract a substring in Excel. One of Microsoft Excel's many capabilities is the ability to compare two lists of data, identifying matches between the lists and identifying which items are found in only one list. What is the total amount of money earned from the sales transaction above? The Criteria. Re: Extract Data from one sheet into another one based on criteria I think you get better help (and so I advice you), if you post a new question on the forum (Excel programming). - There will be a field to input the File Path on where to extract the Data from (FP_1). Excel VBA : Filter data based on drop down selection Deepanshu Bhalla 7 Comments Excel Macro , VBA Suppose you are asked to add a combo box (drop down) in your sheet and pull values from a range based on selection in the combo box (drop down). In the example shown, the formula in E5 is: Excel formula: Extract multiple matches into separate rows | Exceljet. This way, you can dynamically change your filter criteria, and your formula keeps up. The advanced filter in Excel uses three ranges, the list or Excel Table, the Criteria range containing logical formulas that define the filter, and an optional Extract range. Click here to visit our frequently asked questions about HTML5 video. Copying filtered data to another worksheet should be simple but it's not. In Excel, you can filter the rows meeting the criteria first, and then copy them to another location. Count Unique values based on multiple columns Deepanshu Bhalla Add Comment Advanced Excel , Excel Unique Values This tutorial explains how to count unique values based on multiple columns in Excel. Filter Excel Table (List) - Multiple Criteria, Nested Conditions Filter example: filter horses that are black , or roan in 'color', or have a star in 'body_mark' Practice file - filter-demo. I have a list of data in 'Table1' on an Excel spreadsheet that looks like this: Column A Column B Pizza Sauce 3 Pepperoni 0 Cheese 1 Crust 2 Garlic 0 Sausage 0 From this list I want to be able to create a second list that, based on the value in B, shows the value in A. The criteria range has column labels and includes at least one blank row between the criteria values and the list range. This post also covers a method to remove duplicates from a range. Then click Close. Excel how to insert data based upon criteria. This allows you to do things like find cells that meet certain conditions, like checking if values in a cell are. It is the range of cells to sum together. This tutorial will show you how to list values from a table based on filter criteria using sub-arrays and the SMALL function. You define one metric that you want to add (the first parameter) and then define subsequent criteria ranges and values that will define filter ranges and values that you want to. The DataGrid population is pretty straight forward, but the Excel extraction required some tinkering. In this blog, I will discuss how to delete rows in excel based on the condition. Microsoft Office Excel uses the separate criteria range in the Advanced Filter dialog box as the source for the advanced criteria. How to extract distinct / unique rows in Excel; How to get unique values in Excel. For example, there are data such as the image below. Based on the way this code works, there isn't a way to modify it to do that. For example, get price for Large Jacket. In "Excel-language" the 1 means TRUE. If you are working for a financial analyst company, you might need to get or import daily stock. How may I retrieve data from an Excel table based on a variable number of criteria? The criteria for database functions need to be laid out in pairs of cells. For example, you may have a cell that contains a combination of text and numbers, or a cell that contains two numbers separated by a delimiter such as a comma. from this point I want to pull back only the top 4 or 5 options. they told you can do it in Enterprise Tab in excel, but i have not there in enterprise tab in my excel. Hello EE, Can someone please help with a formula to extract distinct values based on criteria in another column? I want to extract unique values where the data in the name column is not blank. and transfer your saved. Extract unique values based on criteria Kutools for Excel. To set a simple custom autofilter, Enable autofilter for your spreadsheet using the steps in the section above. However, after several failed attempts he decides to search the web to see how people create a unique list in Excel based on a specific condition. Now, we want user to select Name, Manager& Process from dropdown lists. There is no SUMPRODUCTIF function in excel, but you can use the SUMPRODUCT function to solve one criteria or multiple criteria questions, in the same column or different columns. For once, I am going to stick with Excel 2003 and earlier because there is a SUMIFS function in Excel 2007 that resolves our query in seconds:. This function works in any version of Excel in Windows and Mac, and also in Google Sheets. For basic information, see the introduction to Excel Advanced Filter page. The SUMIF, SUMIFS and SUMPRODUCT functions can also be used at the top of a worksheet, or on a dashboard, to summarize data in a table, based on selected criteria. Attached is a spreadsheet describing my problem in greater detail, but essentially i want to extract rows of data depending on quoter name. This way, you can dynamically change your filter criteria, and your formula keeps up. The following examples illustrate how to perform a lookup based on multiple criteria. Advanced users have seen the use of adding more than one ‘criteria’ to what. When selecting the list range, ensure that the cell with the list heading is also selected to avoid duplicate values. Answer: The date range is specified in cell G1 and G2, see image above. The tutorial shows how to use Advanced Filter in Excel and provides a number of non-trivial criteria range examples to create a case-sensitive filter, find matches and differences between two columns, extract records that match a smaller list, and more. How to import data from the web in Excel: If you manually copy and paste data from a web page into Excel, you might find that all the data gets pasted into one cell. Select the range that you will copy rows based on column criteria, and click the Enterprise > Split Data. Computerized Business Associates is independent from Sage and is not authorized to make any statement, representation, or warranties or grant any license or permission on behalf of Sage regarding any product, service, or website content. Add the Form command to the Quick Access Toolbar. An extract range, where Excel displays the data that satisfies the criteria. I have attached my dummy data. Select all the data elements in the Table. Purpose & background. As soon as. Watch Video - Extract Data Using a Drop Down List in Excel In this tutorial, I will show you how to create a drop-down filter in Excel so that you can extract data based on the selection from the drop-down. This can be easily done to include more filters - Extract Data based on criteria. Excel - A Macro to populate data based on criteria - How-To - Excel Excel macro to color cells based on criteria - Forum - Programming Copy a row from one worksheet to another, based on criteria - Forum - Excel. So, use the web query tool to save time, and make your worksheets more dynamic! You can use web query to get refreshable (or static. (I will explain in detail with an example. To avoid any confusion, first off, let's agree on what we call unique values in Excel. The range that Excel is searching to find the criteria, which is the Sales ID column. In this MS Excel tutorial from ExcelIsFun, the 671st installment in their series of digital spreadsheet magic tricks, you'll learn how to use the AGGREGATE function to create a formula for extracting records with two criteria.
__label__pos
0.744142
Beefy Boxes and Bandwidth Generously Provided by pair Networks "be consistent"   PerlMonks   Answer: Size of a hash ( #63039=categorized answer: print w/ replies, xml ) Need Help?? Q&A > hashes > How can I find the number of elements in a hash? contributed by japhy You did that wrong, by the way. You wanted $size = @array, not $#array. And you want to call keys() in scalar context, like it's documented. Comment on Answer: Size of a hash Re: Answer: Size of a hash by Malkavian (Friar) on Mar 08, 2001 at 22:12 UTC My stupid mistake. RTFM strikes again. Thanks. Log In? Username: Password: What's my password? Create A New User Chatterbox? and the web crawler heard nothing... How do I use this? | Other CB clients Other Users? Others meditating upon the Monastery: (9) As of 2014-09-19 10:29 GMT Sections? Information? Find Nodes? Leftovers? Voting Booth? How do you remember the number of days in each month? Results (135 votes), past polls
__label__pos
0.517773
Skip to content HTTPS clone URL Subversion checkout URL You can clone with or . Download ZIP Browse files extracted dependency-tree code to separate package new ender-dependency-graph package • Loading branch information... 1 parent 35920e2 commit e2778948d1c851a27a4e78adb4ec177a44e79e2b @rvagg rvagg committed View 3  .jshintrc @@ -24,5 +24,6 @@ , "proto": true , "expr": true , "es5": true + , "esnext": true , "strict": false -} +} View 275 lib/dependency-tree.js @@ -1,275 +0,0 @@ -/*! - * ENDER - The open module JavaScript framework - * - * Copyright (c) 2011-2012 @ded, @fat, @rvagg and other contributors - * - * Permission is hereby granted, free of charge, to any person obtaining a copy - * of this software and associated documentation files (the "Software"), to deal - * in the Software without restriction, including without limitation the rights - * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell - * copies of the Software, and to permit persons to whom the Software is furnished - * to do so, subject to the following conditions: - * - * The above copyright notice and this permission notice shall be included in all - * copies or substantial portions of the Software. - * - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR - * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE - * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER - * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, - * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE - * SOFTWARE. - */ - - -/****************************************************************************** - * Build a DependencyTree object that represents a complete tree of - * dependencies that exist in node_modules and in any paths that have been - * specified as packages. - * See the tests for what the data structure looks like. - */ - -var path = require('path') - , async = require('async') - , packageUtil = require('ender-repository').util - , util = require('./util') - - , DependencyTree = { - init: function (options, treeData) { - this.options = options - this.treeData = treeData - return this - } - - // recursive walk over the dependency tree, invoke the callback on each - // leaf node, starting depth-first in key-order, which should give us the - // correctly ordered dependencies - // this may trigger the callback on duplicate packages (although the last - // argument of the callback is a uniqueness indicator) - , forEachOrderedDependency: function (packages, callback, _unique, _treeData, _pkglist) { - var ejs - , findPackage = function (name) { - var k - if (_treeData[name]) return _treeData[name] - for (k in _treeData) { - if (_treeData[k].packageJSON.name == name) return _treeData[k] - } - } - - // _treeData, _unique & _pkglist are for internal use - - if (!_treeData) _treeData = this.treeData - if (!_pkglist) { // first call, do top-level stuff - ejs = packages.indexOf(util.getRootPackageName(this.options)) - // take root package from where it is and put it at the front - if (ejs > 0) packages.splice(0, 0, packages.splice(ejs, 1)[0]) - _pkglist = [] - } - packages.forEach(function (p, i) { - var isUnique = _pkglist.indexOf(p) == -1 - , pkg - - if (isUnique || !_unique) { - pkg = findPackage(p) - if (pkg.dependencies && pkg.packageJSON.dependencies) { - this.forEachOrderedDependency( - Object.keys(pkg.packageJSON.dependencies) - , callback - , _unique - , pkg.dependencies - , _pkglist - ) - } - callback(p, pkg.parents, pkg, _pkglist.length, isUnique) // _pkglist.length tells us the call index - _pkglist.push(p) - } - }.bind(this)) - } - - // does the same as the above but doesn't trigger the callback for packages that have - // already been passed. - // this is for SourceBuild assembling - , forEachUniqueOrderedDependency: function (packages, callback) { - return this.forEachOrderedDependency(packages, callback, true) - } - - // gives a list of packages by proper name from package.json, turns a path into a package name - , localizePackageList: function (packages) { - return packages.map(function (p) { - return packageUtil.isPath(p) - && typeof this.treeData[p] == 'object' - && this.treeData[p].packageJSON - && this.treeData[p].packageJSON.name - && this.treeData[this.treeData[p].packageJSON.name] - ? this.treeData[p].packageJSON.name - : packageUtil.cleanName(p) - }.bind(this)) - } - - , allRootPackages: function () { - return Object.keys(this.treeData) - } - } - - // Called for each root package and each sub-package within those packages node_modules - // directories, and so on down all the way. We get dependencies from the package.json - // files and also the directories inside node_modules - , processPackage = function (parents, treeData, processSubtree, pkg, callback) { - var name = packageUtil.cleanName(pkg) - - , getJSONData = function (callback) { - // get dependency list from package.json - packageUtil.readPackageJSON(parents, name, function (err, json) { - if (err) return callback(err) // wrapped in package-util.js - callback(null, { - data : json - , dependencies : packageUtil.getDependenciesFromJSON(json) - }) - }) - } - - , getDirectoryDependencies = function (callback) { - // dependency list from the package's node_modules directory - - if (packageUtil.isPath(name)) { - // not installed under ./node_modules/, 'name' is a dir, so don't fetch deps from dir/node_modules - return callback(null, { dependencies: [] }) - } - - packageUtil.getDependenciesFromDirectory(parents, name, function (err, dependencies) { - if (err) return callback(err) // wrapped in package-util.js - callback(null, { dependencies: dependencies }) - }) - } - - , finish = function (err, data) { - var childPackages - - if (err) { - if (err.code == 'ENOENT') { - treeData[name] = 'missing' - return callback() - } - return callback(err) // wrapped in package-util.js (see getJSONData & getDirectoryDependencies) - } - - // we have the data, now do something with it - treeData[name] = { - packageJSON : data.json.data - , dependencies : {} // init as empty - , parents : parents.slice(0) // make a copy of parents array - } - - // concat dependencies in node_modules with those in package.json but don't duplicate - childPackages = data.dir.dependencies.concat(data.json.dependencies.filter(function (p) { - return data.dir.dependencies.indexOf(p) == -1 - })) - - // processSubtree() is actually just constructDependencyTreePart() - processSubtree( - parents.concat([ pkg ]) - , treeData[name].dependencies - , childPackages - , callback - ) - } - - if (treeData[name]) return callback() // already done this part of the tree - - async.parallel( - { - json : getJSONData - , dir : getDirectoryDependencies - } - , finish - ) - } - - // recursive function, called for each node - , constructDependencyTreePart = function (memoizedProcessor, parents, treeData, childPackages, callback) { - async.forEach( - childPackages - , memoizedProcessor.bind(null, parents, treeData, constructDependencyTreePart.bind(null, memoizedProcessor)) - , callback - ) - } - - // will return a *complete* dependency tree of ./package.json and ./node_modules, - // we may not want everything in the result so we need to walk the tree using the - // forEach*() methods below - , constructDependencyTree = function (options, packages, callback) { - var treeData = {} - - // This bit of unfortunate complexity needs some explaination: we have 2 paths in our - // dep tree construction, we search ./node_modules/* and we individually search through - // the `packages` list--this leads to duplication of directory & package.json reads. - // Even though this doesn't lead to a corrupt dep tree, the duplication is overhead - // we can do without. The cleanest way is to memoize the processPackage() function and - // make sure duplicate calls to it with the same `parents` and `pkg` arguments are - // only handled once. This memoized function is passed around, it's only useful for - // individual calls to `constructDependencyTree()`. - , memoizedProcessor = async.memoize( - processPackage - , function (parents, _t, _p, pkg) { - // just a hash string to match duplicate `parents` and `pkg` arguments - return [''].concat(parents.concat([pkg, ''])).join('$$') - } - ) - - // a special case of the CWD, in case we are in a package to be included, if we - // didn't do this then a plain `ender build` wouldn't work. - // even though this will result in a double-scan of node_modules, processPackage() - // won't allow duplicate scans below that. - , scanRootDirectory = function (callback) { - packageUtil.getDependenciesFromDirectory([], '.', function (err, dependencies) { - if (err) return callback(err) // wrapped in package-utils.js - constructDependencyTreePart(memoizedProcessor, [], treeData, dependencies, callback) - }) - } - , collectTreeParts = function (part, store) { - Object.keys(part).forEach(function (k) { - if (typeof part[k].dependencies == 'object') { - (store[k] || (store[k] = [])).push(part[k]) - collectTreeParts(part[k].dependencies, store) - } - }) - } - , completeTreeParts = function (part, store) { - Object.keys(part).forEach(function (k) { - if (part[k] !== 'missing') return completeTreeParts(part[k].dependencies, store) - if (store[k]) part[k] = store[k][0] - }) - } - // using collectTreeParts() and completeTreeParts() we first assemble a flat collection of - // all packages by name, then we walk the full tree again and fill in any gaps where packages - // may have dependencies that exist elsewhere in ther tree--npm doesn't always give us a - // complete tree where there are duplicates so we have to go looking. - // we end up with a tree that can contain many duplicates but it's a complete tree. - , completeTree = function () { - var flattened = {} - , dependencyTree - collectTreeParts(treeData, flattened) - completeTreeParts(treeData, flattened) - dependencyTree = Object.create(DependencyTree).init(options, treeData) - callback(null, dependencyTree) - } - - async.parallel( - [ scanRootDirectory, constructDependencyTreePart.bind(null, memoizedProcessor, [], treeData, packages) ] - , function (err) { - if (err) return callback(err) // wrapped in package-util.js - completeTree() - } - ) - } - - // take existing tree data and instantiate a DependencyTree object - , create = function (options, treeData) { - return Object.create(DependencyTree).init(options, treeData) - } - -module.exports = { - generate : constructDependencyTree - , create : create -} View 4 lib/install.js @@ -42,7 +42,7 @@ var async = require('async') , repository = require('ender-repository') , packageUtil = repository.util , installUtil = require('./install-util') - , DependencyTree = require('./dependency-tree') + , DependencyGraph = require('ender-dependency-graph') , installPackages = function (options, packages, callback) { var filteredPackages = installUtil.filterPackagesWithoutCwd(packages) @@ -67,7 +67,7 @@ var async = require('async') // keep running this function until everything is installed , installRest = function (callback) { - DependencyTree.generate(options, packages, function (err, _dependencyTree) { + DependencyGraph(options, packages, function (err, _dependencyTree) { if (err) return callback(err) // wrapped in package-utils.js dependencyTree = _dependencyTree View 47 lib/main-info-util.js @@ -80,54 +80,7 @@ var zlib = require('zlib') }) } - // we use `archy` to print the tree (thanks @substack), but we have to turn our - // dependency tree into an archy-compatible tree. - // we only do a 1/2 job here and leave it to main-info-output to complete the - // process and make it perdy. - , buildArchyTree = function (options, packages, dependencyTree) { - var archyTree = { label: 'Active packages:', nodes: [] } - , localPackages = dependencyTree.localizePackageList(packages) - - dependencyTree.forEachOrderedDependency(localPackages, function (packageName, parents, data, index, first) { - var archyTreeNode = archyTree - , found - , newNode - , regexp - , i - , j - - parents = parents.concat([ packageName ]) - - for (i = 0; i < parents.length; i++) { - found = false - regexp = new RegExp('^(.\\[\\d\\dm)?' + parents[i] + '(?:@.*)?$') - - for (j = 0; j < archyTreeNode.nodes.length; j++) { - if (regexp.test(archyTreeNode.nodes[j].label)) { - found = true - archyTreeNode = archyTreeNode.nodes[j] - break - } - } - - if (!found) { - archyTreeNode.nodes.push(newNode = { label: parents[i], nodes: [] }) - archyTreeNode = newNode - } - } - - if (!archyTreeNode.version) { - archyTreeNode.first = first - archyTreeNode.version = data.packageJSON.version - archyTreeNode.description = data.packageJSON.description - } - }) - - return archyTree - } - module.exports = { sizes : sizes , parseContext : parseContext - , buildArchyTree : buildArchyTree } View 32 lib/main-info.js @@ -30,20 +30,19 @@ * by main-info-output, we just provide the data/model here. */ -var async = require('async') - , util = require('./util') - , mainInfoUtil = require('./main-info-util') - , DependencyTree = require('./dependency-tree') +var util = require('./util') + , mainInfoUtil = require('./main-info-util') + , DependencyGraph = require('ender-dependency-graph') - , generateAndPrint = function (args, out, filename, options, packages, dependencyTree, callback) { + , generateAndPrint = function (args, out, filename, options, packages, dependencyGraph, callback) { // the 2 main processes here is generateSpec() to read the build context from the // build file and construct a dependency tree from what we read, and generateSizes() // that gives us the raw, minified and gzipped sizes. var generateSpec = function (callback) { // if generateAndPrint() has been called from a module with existing options, // packages and dependency tree data we can skip the difficult part and return - if (options && packages && dependencyTree) { - return callback(null, { options : options, packages: packages }, dependencyTree) + if (options && packages && dependencyGraph) { + return callback(null, { options : options, packages: packages }, dependencyGraph) } // read 'Build: ...' and 'Packages: ...' from the head of the build file @@ -51,35 +50,30 @@ var async = require('async') if (err) return callback(err) // wrapped in source-build.js // same process to build a dependency tree as in the original build - DependencyTree.generate(context.options, context.options.packages, function (err, dependencyTree) { + DependencyGraph(context.options, context.options.packages, function (err, dependencyGraph) { if (err) return callback(err) // wrapped in package-util.js - callback(null, context, dependencyTree) + callback(null, context, dependencyGraph) }) }) } - , finish = function (context, dependencyTree, sizes) { + , finish = function (context, dependencyGraph, sizes) { // build an `archy` compatible tree representing the packages in the build - var archyTree = mainInfoUtil.buildArchyTree( - context.options - , context.options.packages - , dependencyTree - ) - - out.buildInfo(filename, context.options, context.packages, sizes, archyTree) + var archyGraph = DependencyGraph.archyTree(context.options.packages, dependencyGraph) + out.buildInfo(filename, context.options, context.packages, sizes, archyGraph) callback() } //TODO: should perform an fs.exists() on the file(s) we're going to check, otherwise we get a // random error from one of the fs read operations above - generateSpec(function (err, context, dependencyTree) { + generateSpec(function (err, context, dependencyGraph) { if (err) return callback(err) // wrapped in generateSpec mainInfoUtil.sizes(context.options, filename, function (err, sizes) { if (err) return callback(err) // wrapped in main-info-util.js - finish(context, dependencyTree, sizes) + finish(context, dependencyGraph, sizes) }) }) } View 8 package.json @@ -1,7 +1,7 @@ { "name": "ender" , "description": "Open Module JavaScript Framework" - , "version": "0.9.8-dev" + , "version": "0.9.9-dev" , "authors": [ "Jacob Thornton @fat <[email protected]> (https://github.com/fat)" , "Dustin Diaz @ded <[email protected]> (https://github.com/ded)" @@ -11,8 +11,10 @@ , "main": "./lib/main.js" , "homepage": "https://ender.jit.su" , "dependencies": { - "ender-repository" : "~0.0.2" - , "ender-minify" : "~0.1.0" + "ender-repository" : "0.0.2" + , "ender-minify" : "0.1.0" + , "ender-dependency-graph" + : "0.0.1" , "colors" : "~0.6.0" , "async" : "~0.1.22" , "hogan.js" : "~2.0.0" View 756 test/unit/dependency-tree-test.js @@ -1,756 +0,0 @@ -/*! - * ENDER - The open module JavaScript framework - * - * Copyright (c) 2011-2012 @ded, @fat, @rvagg and other contributors - * - * Permission is hereby granted, free of charge, to any person obtaining a copy - * of this software and associated documentation files (the "Software"), to deal - * in the Software without restriction, including without limitation the rights - * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell - * copies of the Software, and to permit persons to whom the Software is furnished - * to do so, subject to the following conditions: - * - * The above copyright notice and this permission notice shall be included in all - * copies or substantial portions of the Software. - * - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR - * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE - * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER - * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, - * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE - * SOFTWARE. - */ - - -var buster = require('buster') - , path = require('path') - , assert = buster.assert - , packageUtil = require('ender-repository').util - , DependencyTree = require('../../lib/dependency-tree') - - -buster.testCase('Dependency tree', { - 'constructDependencyTree': { - 'setUp': function () { - this.runTest = function (setupTree, jsons, directories, expectedTree, done) { - var packageUtilMock = this.mock(packageUtil) - , setupExpectations = function (parents, setupTree) { - Object.keys(setupTree).forEach(function (p, i) { - if (p == '$id') - return - var id = setupTree[p].$id || p // for doing special indirection - if (jsons[id] != 'missing') { - packageUtilMock.expects('readPackageJSON') - .once() - .withArgs(parents, p) - .callsArgWith(2, null, jsons[id]) - packageUtilMock.expects('getDependenciesFromDirectory') - [/\//.test(id) ? 'never' : 'once']() // we DON'T want to read node_modules dir for path - .withArgs(parents, p) - .callsArgWith(2, null, directories[id]) - setupExpectations(parents.concat([ p ]), setupTree[p]) - } else { - // dir & package.json missing - packageUtilMock.expects('readPackageJSON') - .once() - .withArgs(parents, p) - .callsArgWith(2, { code: 'ENOENT' }) - packageUtilMock.expects('getDependenciesFromDirectory') - [/\//.test(id) ? 'never' : 'once']() // we DON'T want to read node_modules dir for path - .withArgs(parents, p) - .callsArgWith(2, { code: 'ENOENT' }) - } - }) - } - - packageUtilMock.expects('getDependenciesFromDirectory') - .once() - .withArgs([], '.') - .callsArgWith(2, null, Object.keys(setupTree).filter(function (p) { return !/\//.test(p) })) - setupExpectations([], setupTree) - DependencyTree.generate({}, Object.keys(setupTree), function (err, dependencyTree) { - assert.equals(dependencyTree.treeData, expectedTree) - done() - }) - } - } - - , 'test no dependencies': function (done) { - var packages = { - 'pkg1': {} - , 'some/path/to/pkg2': {} - } - , jsons = { - 'pkg1': { name: 'pkg1' } - , 'some/path/to/pkg2': { name: 'pkg2name' } // name is different to dir, dirname shouldn't matter - } - , directories = { - 'pkg1': [] - , 'some/path/to/pkg2': [] - } - , expectedTree = { - 'pkg1': { - packageJSON: jsons['pkg1'] - , parents: [] - , dependencies: {} - } - , 'some/path/to/pkg2': { - packageJSON: jsons['some/path/to/pkg2'] - , parents: [] - , dependencies: {} - } - } - this.runTest(packages, jsons, directories, expectedTree, done) - } - - , 'test complex dependencies': function (done) { - var packages = { - 'pkg1': { - 'foo': { 'bar': {} } - , 'woohoo': {} - } - , 'some/path/to/pkg2': { - 'wee': { - 'hee': { - 'yo': {} - } - } - , 'foo': { 'bar': {} } - } - } - , jsons = { - 'pkg1': { - name: 'pkg1' - , dependencies: [ 'foo', 'woohoo' ] - } - , 'foo': { - name: 'foo' - , dependencies: [ 'bar' ] - } - , 'bar': { name: 'bar' } - , 'woohoo': { name: 'woohoo' } - , 'some/path/to/pkg2': { - name: 'pkg2name' - , dependencies: [ 'wee', 'foo' ] - } - , 'wee': { - name: 'wee' - , dependencies: [ 'hee' ] - } - , 'hee': { - name: 'hee' - , dependencies: [ 'yo' ] - } - , 'yo': { name: 'yo' } - } - , directories = { - 'pkg1': [ 'foo', 'woohoo' ] - , 'pkg': [ 'foo', 'woohoo' ] - , 'some/path/to/pkg2': [ 'wee', 'foo' ] - , 'foo': [ 'bar' ] - , 'bar': [] - , 'woohoo': [] - , 'wee': [ 'hee' ] - , 'hee': [ 'yo' ] - , 'yo': [] - } - , expectedTree = { - 'pkg1': { - packageJSON: jsons['pkg1'] - , parents: [] - , dependencies: { - 'foo': { - packageJSON: jsons['foo'] - , parents: [ 'pkg1' ] - , dependencies: { - 'bar': { - packageJSON: jsons['bar'] - , parents: [ 'pkg1', 'foo' ] - , dependencies: {} - } - } - } - , 'woohoo': { - packageJSON: jsons['woohoo'] - , parents: [ 'pkg1' ] - , dependencies: {} - } - } - } - , 'some/path/to/pkg2': { - packageJSON: jsons['some/path/to/pkg2'] - , parents: [] - , dependencies: { - 'wee': { - packageJSON: jsons['wee'] - , parents: [ 'some/path/to/pkg2' ] - , dependencies: { - 'hee': { - packageJSON: jsons['hee'] - , parents: [ 'some/path/to/pkg2', 'wee' ] - , dependencies: { - 'yo': { - packageJSON: jsons['yo'] - , parents: [ 'some/path/to/pkg2', 'wee', 'hee' ] - , dependencies: {} - } - } - } - } - } - , 'foo': { - packageJSON: jsons['foo'] - , parents: [ 'some/path/to/pkg2' ] - , dependencies: { - 'bar': { - packageJSON: jsons['bar'] - , parents: [ 'some/path/to/pkg2', 'foo' ] - , dependencies: {} - } - } - } - } - } - } - this.runTest(packages, jsons, directories, expectedTree, done) - } - - , 'test dependencies with missing directories': function (done) { - var packages = { - 'pkg1': { - 'foo': { 'bar': {} } - , 'woohoo': {} - } - } - , jsons = { - 'pkg1': { - name: 'pkg1' - , dependencies: [ 'foo', 'woohoo' ] - } - , 'foo': { - name: 'foo' - , dependencies: [ 'bar' ] - } - , 'bar': 'missing' - , 'woohoo': 'missing' - } - , directories = { - 'pkg1': [ 'foo', 'woohoo' ] - , 'pkg': [ 'foo' ] - , 'foo': [] - } - , expectedTree = { - 'pkg1': { - packageJSON: jsons['pkg1'] - , parents: [] - , dependencies: { - 'foo': { - packageJSON: jsons['foo'] - , parents: [ 'pkg1' ] - , dependencies: { - 'bar': 'missing' - } - } - , 'woohoo': 'missing' - } - } - } - this.runTest(packages, jsons, directories, expectedTree, done) - } - - , 'test dependencies in scattered directories': function (done) { - // this test is designed to show that even with missing dependency directories - // the proper dependency tree can be built if the packages are available in the - // tree somewhere. - // - // the $id awkwardness is so that we can specify the simulation of missing - // directories & package.json files. - // - // see the 'directories' object to see what the simulated directory structure is - // yet it should be able to piece together the full expectedTree - var packages = { - 'pkg1': { - 'pkg4': { 'pkg6': { $id: 'pkg6-missing' } } - , 'pkg3': { $id: 'pkg3-missing' } - } - , 'pkg2': {} - , 'pkg3': {} - , 'pkg5': { 'pkg6': {} } - } - , jsons = { - 'pkg1': { name: 'pkg1', dependencies: [ 'pkg4', 'pkg3' ] } - , 'pkg2': { name: 'pkg2', dependencies: [] } - , 'pkg3-missing': 'missing' - , 'pkg3': { name: 'pkg3', dependencies: [] } - , 'pkg4': { name: 'pkg4', dependencies: [ 'pkg6' ] } - , 'pkg5': { name: 'pkg5', dependencies: [ 'pkg6' ] } - , 'pkg6-missing': 'missing' - , 'pkg6': { name: 'pkg6', dependencies: [] } - } - , directories = { - 'pkg1': [ 'pkg4' ] - , 'pkg2': [] - , 'pkg3': [] - , 'pkg4': [] - , 'pkg5': [ 'pkg6' ] - , 'pkg6': [] - } - , expectedTree = { - 'pkg1': { - packageJSON: jsons['pkg1'] - , parents: [] - , dependencies: { - 'pkg4': { - packageJSON: jsons['pkg4'] - , parents: [ 'pkg1' ] - , dependencies: { - 'pkg6': { - packageJSON: jsons['pkg6'] - , parents: [ 'pkg5' ] - , dependencies: {} - } - } - } - , 'pkg3': { - packageJSON: jsons['pkg3'] - , parents: [] - , dependencies: {} - } - } - } - , 'pkg2': { - packageJSON: jsons['pkg2'] - , parents: [] - , dependencies: {} - } - , 'pkg3': { - packageJSON: jsons['pkg3'] - , parents: [] - , dependencies: {} - } - , 'pkg5': { - packageJSON: jsons['pkg5'] - , parents: [] - , dependencies: { - 'pkg6': { - packageJSON: jsons['pkg6'] - , parents: [ 'pkg5' ] - , dependencies: {} - } - } - } - } - this.runTest(packages, jsons, directories, expectedTree, done) - } - } - - , 'forEachOrderedDependency': { - 'no dependencies': { - 'setUp': function () { - this.originalTree = DependencyTree.create({}, { - 'pkg1': { - dependencies: {} - , packageJSON: { name: 'pkg1' } - , parents: [ 'foo' ] - } - , 'some/path/to/pkg2': { - dependencies: {} - , packageJSON: { name: 'pkg2' } - , parents: [ 'foo', 'bar' ] - } - }) - this.callSpy = this.spy() - - this.verifySpy = function () { - assert.equals(this.callSpy.callCount, 2) - assert.equals(this.callSpy.getCall(0).args[0], 'pkg1') - assert.equals(this.callSpy.getCall(0).args[1], [ 'foo' ]) - assert.equals(this.callSpy.getCall(0).args[2], this.originalTree.treeData['pkg1']) - assert.equals(this.callSpy.getCall(1).args[0], 'some/path/to/pkg2') - assert.equals(this.callSpy.getCall(1).args[1], [ 'foo' , 'bar' ]) - assert.equals(this.callSpy.getCall(1).args[2], this.originalTree.treeData['some/path/to/pkg2']) - } - } - - , 'test forEachUniqueOrderedDependency': function () { - this.originalTree.forEachUniqueOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - - , 'test forEachOrderedDependency': function () { - // should do the same thing - this.originalTree.forEachOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - } - - , 'simple dependencies': { - 'setUp': function () { - this.originalTree = DependencyTree.create({}, { - 'apkg-2': { - parents: [] - , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } - , dependencies: { - 'mypkg-1': { - parents: [ 'apkg-2' ] - , packageJSON: { name: 'mypkg-1' } - , dependencies: {} - } - } - } - , 'somepkg-5': { - parents: [] - , packageJSON: { name: 'somepkg-5', dependencies: { 'foo-4': '*' } } - , dependencies: { - 'foo-4': { - parents: [ 'somepkg-5' ] - , packageJSON: { name: 'foo-4', dependencies: { 'bar-3': '*' } } - , dependencies: { - 'bar-3': { - parents: [ 'somepkg-5', 'foo-4' ] - , packageJSON: { name: 'bar-3' } - , dependencies: {} - } - } - } - } - } - , 'apkg-7': { - parents: [] - , packageJSON: { name: 'apkg-7', dependencies: { 'mypkg-6': '*' } } - , dependencies: { - 'mypkg-6': { - parents: [ 'apkg-7' ] - , packageJSON: { name: 'mypkg-6' } - , dependencies: {} - } - } - } - }) - this.callSpy = this.spy() - this.verifySpy = function () { - assert.equals(this.callSpy.args.length, 7) - - this.callSpy.args.forEach(function (c, i) { - assert.equals(c[3], i) - refute.isNull(c[2]) - refute.isNull(c[2].dependencies) // should be the packageJSON, 'dependencies' is a proxy for this - assert.same(c[1], c[2].parents) - assert.match(c[0], new RegExp('-' + (++i) + '$')) - }) - } - } - - , 'test forEachUniqueOrderedDependency': function () { - this.originalTree.forEachUniqueOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - - , 'test forEachOrderedDependency': function () { - // should do the same thing - this.originalTree.forEachOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - } - - , 'ender-js at front': { - 'setUp': function () { - this.originalTree = DependencyTree.create({}, { - 'apkg-3': { - parents: [] - , packageJSON: { name: 'apkg-3', dependencies: { 'mypkg-2': '*' } } - , dependencies: { - 'mypkg-2': { - parents: [ 'apkg-3' ] - , packageJSON: { name: 'mypkg-2' } - , dependencies: {} - } - } - } - , 'somepkg-4': { - parents: [] - , packageJSON: { name: 'somepkg-4' } - , dependencies: {} - } - , 'ender-js': { - parents: [] - , packageJSON: { name: 'ender-js' } - , dependencies: {} - } // it should spit this out first - , 'apkg-6': { - parents: [] - , packageJSON: { name: 'apkg-6', dependencies: { 'mypkg-5': '*' } } - , dependencies: { - 'mypkg-5': { - parents: [ 'apkg-6' ] - , packageJSON: { name: 'mypkg-5' } - , dependencies: {} - } - } - } - }) - this.callSpy = this.spy() - this.verifySpy = function () { - assert.equals(this.callSpy.args.length, 6) - - this.callSpy.args.forEach(function (c, i) { - assert.equals(c[3], i) - refute.isNull(c[2]) - refute.isNull(c[2].dependencies) // should be the packageJSON, 'dependencies' is a proxy for this - assert.same(c[1], c[2].parents) - if (!i) { - assert.equals(c[0], 'ender-js') - assert.same(c[2], this.originalTree.treeData['ender-js']) - } else - assert.match(c[0], new RegExp('-' + (++i) + '$')) - }.bind(this)) - } - } - - , 'test forEachUniqueOrderedDependency': function () { - this.originalTree.forEachUniqueOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - - , 'test forEachOrderedDependency': function () { - // should do the same thing - this.originalTree.forEachOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - this.verifySpy() - } - } - - , 'duplicate dependencies': { - 'setUp': function () { - this.originalTree = DependencyTree.create({}, { - 'apkg-6': { - parents: [] - , packageJSON: { name: 'apkg-6', dependencies: { 'mypkg-5': '*' } } - , dependencies: { - 'mypkg-5': { - parents: [ 'apkg-6' ] - , packageJSON: { name: 'mypkg-5', dependencies: { 'apkg-2': '*', 'apkg-4': '*' } } - , dependencies: { - 'apkg-2': { - parents: [ 'apkg-6', 'mypkg-5' ] - , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } - , dependencies: { - 'mypkg-1': { - parents: [ 'apkg-6', 'mypkg-5', 'apkg-2' ] - , packageJSON: { name: 'mypkg-1' } - , dependencies: {} - } - } - } - , 'apkg-4': { - parents: [ 'apkg-6', 'mypkg-5' ] - , packageJSON: { name: 'apkg-4', dependencies: { 'mypkg-3': '*' } } - , dependencies: { - 'mypkg-3': { - parents: [ 'apkg-6', 'mypkg-5', 'apkg-4' ] - , packageJSON: { name: 'mypkg-3' } - , dependencies: {} - } - } - } - } - } - } - } - , 'somepkg-9': { - parents: [] - , packageJSON: { name: 'somepkg-9', dependencies: { 'foo-8': '*', 'mypkg-3': '*' } } - , dependencies: { - 'foo-8': { - parents: [ 'somepkg-9' ] - , packageJSON: { name: 'foo-8', dependencies: { 'bar-7': '*' } } - , dependencies: { - 'bar-7': { - parents: [ 'somepkg-9', 'foo-8' ] - , packageJSON: { name: 'bar-7' } - , dependencies: {} - } - } - } - , 'mypkg-3': { - parents: [ 'somepkg-9' ] - , packageJSON: { name: 'mypkg-3' } - , dependencies: {} - } - } - } - , 'apkg-2': { - parents: [] - , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } - , dependencies: { - 'mypkg-1': { - parents: [ 'apkg-2' ] - , packageJSON: { name: 'mypkg-1' } - , dependencies: {} - } - } - } - , 'lastpkg-10': { - parents: [] - , packageJSON: { name: 'lastpkg-10' } - , dependencies: {} - } - }) - this.callSpy = this.spy() - } - - // we should only see unique packages here, they have numbers in their names so we can match them - // easily - , 'test forEachUniqueOrderedDependency': function () { - this.originalTree.forEachUniqueOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - - // expect only uniques - assert.equals(this.callSpy.args.length, 10) - - this.callSpy.args.forEach(function (c, i) { - assert.equals(c[3], i) - refute.isNull(c[2]) - refute.isNull(c[2].dependencies) // should be the packageJSON, 'dependencies' is a proxy for this - assert.same(c[1], c[2].parents) - assert.match(c[0], new RegExp('-' + (++i) + '$')) - }) - } - - // in this case we should see all packages in order, not just uniques, but we should get an argument - // for uniqueness - , 'test forEachOrderedDependency': function () { - var expectedPackages = - 'mypkg-1 apkg-2 mypkg-3 apkg-4 mypkg-5 apkg-6 bar-7 foo-8 mypkg-3 somepkg-9 mypkg-1 apkg-2 lastpkg-10' - .split(' ') - , orderedIndex = 1 - - this.originalTree.forEachOrderedDependency(this.originalTree.allRootPackages(), this.callSpy) - - assert.equals(this.callSpy.args.length, expectedPackages.length) - - this.callSpy.args.forEach(function (c, i) { - // use 'orderedIndex' to check if the current package is a dupe or not according to the - // package name - var expectedIsUnique = new RegExp('-' + orderedIndex + '$').test(c[0]) - if (expectedIsUnique) - orderedIndex++ - assert.equals(c[3], i) - refute.isNull(c[2]) - refute.isNull(c[2].dependencies) // should be the packageJSON, 'dependencies' is a proxy for this - assert.same(c[1], c[2].parents) - assert.equals(c[0], expectedPackages[i]) - assert.equals(c[4], expectedIsUnique, 'index ' + i + ' ' + c[0]) - }) - } - } - - , 'additional unnecessary dependencies': { - 'setUp': function () { - this.originalTree = DependencyTree.create({}, { - 'apkg-2': { - parents: [] - , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } - , dependencies: { - 'mypkg-1': { - parents: [ 'apkg-2' ] - , packageJSON: { name: 'mypkg-1' } - , dependencies: {} - } - } - } - , 'somepkg-5': { - parents: [] - , packageJSON: { name: 'somepkg-5', dependencies: { 'foo-4': '*' } } - , dependencies: { - 'foo-4': { - parents: [ 'somepkg-5' ] - , packageJSON: { name: 'foo-4', dependencies: { 'bar-3': '*' } } - , dependencies: { - 'bar-3': { - parents: [ 'somepkg-5', 'foo-4' ] - , packageJSON: { name: 'bar-3' } - , dependencies: {} - } - } - } - } - } - , 'apkg-7': { - parents: [] - , packageJSON: { name: 'apkg-7', dependencies: { 'mypkg-6': '*' } } - , dependencies: { - 'mypkg-6': { - parents: [ 'apkg-7' ] - , packageJSON: { name: 'mypkg-6' } - , dependencies: {} - } - } - } - }) - this.callSpy = this.spy() - this.verifySpy = function () { - assert.equals(this.callSpy.args.length, 5) - - this.callSpy.args.forEach(function (c, i) { - assert.equals(c[3], i) - refute.isNull(c[2]) - refute.isNull(c[2].dependencies) // should be the packageJSON, 'dependencies' is a proxy for this - assert.same(c[1], c[2].parents) - assert.match(c[0], new RegExp('-' + (++i) + '$')) - }) - } - } - - , 'test forEachUniqueOrderedDependency': function () { - this.originalTree.forEachUniqueOrderedDependency([ 'apkg-2', 'somepkg-5' ], this.callSpy) - this.verifySpy() - } - - , 'test forEachOrderedDependency': function () { - // should do the same thing - this.originalTree.forEachOrderedDependency([ 'apkg-2', 'somepkg-5' ], this.callSpy) - this.verifySpy() - } - } - } - - , 'localizePackageList': { - 'test leaves standard package list alone': function () { - assert.equals(DependencyTree.create({}, {}).localizePackageList([ 'one', 'two', 'three' ]), [ 'one', 'two', 'three' ], {}) - } - - , 'test strips out versions from names': function () { - assert.equals(DependencyTree.create({}, {}).localizePackageList([ 'one', '[email protected]', '[email protected]' ]), [ 'one', 'two', 'three' ], {}) - } - - , 'test returns local packages for relative paths': function () { - var originalPackageList = [ 'one', './two', 'three/foo/bar', '/four' ] - , expectedPackageList = [ 'one', 'two', 'three', 'four' ] - , tree = { - 'one': {} - , './two': { packageJSON: { name: 'two' } } - , 'two': {} - , 'three/foo/bar': { packageJSON: { name: 'three' } } - , 'three': {} - , '/four': { packageJSON: { name: 'four' } } - , 'four': {} - } - - assert.equals(DependencyTree.create({}, tree).localizePackageList(originalPackageList, tree), expectedPackageList) - } - - , 'test leaves unlocalizable packages alone': function () { - var originalPackageList = [ 'one', './two', 'three/foo/bar', '/four' ] - , expectedPackageList = [ 'one', './two', 'three', '/four' ] - , tree = { - 'one': {} - , './two': { packageJSON: { name: 'two' } } - , 'three/foo/bar': { packageJSON: { name: 'three' } } - , 'three': {} - , '/four': { packageJSON: { name: 'four' } } - } - - assert.equals(DependencyTree.create({}, tree).localizePackageList(originalPackageList, tree), expectedPackageList) - } - } -}) View 66 test/unit/install-test.js @@ -23,22 +23,27 @@ */ -var testCase = require('buster').testCase - , repository = require('ender-repository') - , install = require('../../lib/install') - , util = require('../../lib/util') - , installUtil = require('../../lib/install-util') - , DependencyTree = require('../../lib/dependency-tree') +var testCase = require('buster').testCase + , repository = require('ender-repository') + , util = require('../../lib/util') + , installUtil = require('../../lib/install-util') + , install + +require('ender-dependency-graph') testCase('Install', { 'setUp': function () { - this.mockUtil = this.mock(util) - this.mockInstallUtil = this.mock(installUtil) - this.mockRepository = this.mock(repository) - this.mockDependencyTree = this.mock(DependencyTree) - - this.optionsArg = { options: 1 } - this.packagesArg = [ 'yee', 'haw' ] // length 2 + this.mockUtil = this.mock(util) + this.mockInstallUtil = this.mock(installUtil) + this.mockRepository = this.mock(repository) + this.dependencyGraphStub = this.stub() + this.originalEDG = require.cache[require.resolve('ender-dependency-graph')].exports + require.cache[require.resolve('ender-dependency-graph')].exports = this.dependencyGraphStub + require.cache[require.resolve('../../lib/install')] = null + install = require('../../lib/install') + + this.optionsArg = { options: 1 } + this.packagesArg = [ 'yee', 'haw' ] // length 2 //this.mockUtil.expects('mkdir').once().withArgs('node_modules').callsArg(1) // done by mkdirp now, probably should be mocked out... @@ -54,13 +59,18 @@ testCase('Install', { }.bind(this) this.expectGenerate = function (dependencyTreeArg) { - this.mockDependencyTree - .expects('generate') - .withArgs(this.optionsArg, this.packagesArg) - .once() - .callsArgWith(2, null, dependencyTreeArg) + this.dependencyGraphStub.callsArgWith(2, null, dependencyTreeArg) }.bind(this) + this.verifyGenerate = function (calls) { + assert.equals(this.dependencyGraphStub.callCount, calls) + for (var i = 0; i < calls; i++) { + assert.equals(this.dependencyGraphStub.getCall(i).args.length, 3) + assert.equals(this.dependencyGraphStub.getCall(i).args[0], this.optionsArg) + assert.equals(this.dependencyGraphStub.getCall(i).args[1], this.packagesArg) + } + } + this.expectFindMissingDependencies = function (dependencyTreeArg, missingDependenciesArg) { this.mockInstallUtil .expects('findMissingDependencies') @@ -86,6 +96,11 @@ testCase('Install', { } } + , tearDown: function () { + require.cache[require.resolve('ender-dependency-graph')].exports = this.originalEDG + require.cache[require.resolve('../../lib/install.js')] = null + } + , 'test basic one package install, already available': function (done) { var filteredPackagesArg = { filteredPackages: 1, length: this.packagesArg.length } , filteredMissingPackagesArg = [] // length 0, nothing to install @@ -103,8 +118,9 @@ testCase('Install', { refute(err) assert.equals(results, []) assert.same(dependencyTree, dependencyTreeArg) + this.verifyGenerate(1) done() - }) + }.bind(this)) } , 'test one package install, not available': function (done) { @@ -135,8 +151,9 @@ testCase('Install', { refute(err) assert.equals(results, [ resultArg ]) assert.same(dependencyTree, dependencyTreeArg2) + this.verifyGenerate(2) done() - }) + }.bind(this)) } , 'test multi package install, multi install loops required': function (done) { @@ -189,8 +206,9 @@ testCase('Install', { refute(err) assert.equals(results, [ resultArg, resultArg2, resultArg3 ]) assert.same(dependencyTree, dependencyTreeArg4) + this.verifyGenerate(4) done() - }) + }.bind(this)) } , 'test multi package install, should only install the same package once': function (done) { @@ -246,8 +264,9 @@ testCase('Install', { refute(err) assert.equals(results, [ resultArg, resultArg2, resultArg3 ]) assert.same(dependencyTree, dependencyTreeArg4) + this.verifyGenerate(4) done() - }) + }.bind(this)) } @@ -285,8 +304,9 @@ testCase('Install', { refute(err) assert.equals(results, [ resultArg, resultArg2 ]) assert.same(dependencyTree, dependencyTreeArg2) + this.verifyGenerate(2) done() - }) + }.bind(this)) } }) View 21 test/unit/install-util-test.js @@ -23,17 +23,16 @@ */ -var buster = require('buster') - , path = require('path') - , assert = buster.assert - , packageUtil = require('ender-repository').util - , DependencyTree = require('../../lib/dependency-tree.js') - , installUtil = require('../../lib/install-util') +var buster = require('buster') + , path = require('path') + , assert = buster.assert + , DependencyGraph = require('ender-dependency-graph') + , installUtil = require('../../lib/install-util') buster.testCase('Install util', { 'findMissingDependencies': { 'test no dependencies': function () { - var originalTree = DependencyTree.create({}, { + var originalTree = DependencyGraph.create({}, { 'pkg1': { dependencies: {}, packageJSON: { name: 'pkg1' }, parents: [ 'foo' ] } , 'some/path/to/pkg2': { dependencies: {}, packageJSON: { name: 'pkg2' },parents: [ 'foo', 'bar' ] } }) @@ -44,7 +43,7 @@ buster.testCase('Install util', { } , 'simple dependencies (no missing)': function () { - var originalTree = DependencyTree.create({}, { + var originalTree = DependencyGraph.create({}, { 'apkg-2': { parents: [] , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } @@ -93,7 +92,7 @@ buster.testCase('Install util', { , 'simple dependencies (with missing)': { 'setUp': function () { - this.originalTree = DependencyTree.create({}, { + this.originalTree = DependencyGraph.create({}, { 'apkg-2': { parents: [] , packageJSON: { name: 'apkg-2', dependencies: { 'mypkg-1': '*' } } @@ -163,7 +162,7 @@ buster.testCase('Install util', { , 'missing deps exist in other branches': { 'setUp': function () { - this.originalTree = DependencyTree.create({}, { + this.originalTree = DependencyGraph.create({}, { 'mypkg-1': 'missing' , 'apkg-2': { parents: [] @@ -237,7 +236,7 @@ buster.testCase('Install util', { } , 'findPathDependencies': function () { - var originalTree = DependencyTree.create({}, { + var originalTree = DependencyGraph.create({}, { 'apkg-2': { parents: [] , packageJSON: { name: 'apkg-2', dependencies: { './mypkg-1': '*' } } View 46 test/unit/main-info-test.js @@ -23,19 +23,18 @@ */ -var testCase = require('buster').testCase - , mainInfo = require('../../lib/main-info') - , mainInfoOut = require('../../lib/output/main-info-output').create() - , mainInfoUtil = require('../../lib/main-info-util') - , DependencyTree = require('../../lib/dependency-tree') - , SourceBuild = require('../../lib/source-build') +var testCase = require('buster').testCase + , mainInfo + , mainInfoOut = require('../../lib/output/main-info-output').create() + , mainInfoUtil = require('../../lib/main-info-util') testCase('Info', { 'setUp': function () { this.runTest = function (options, expectedFilename, done) { var mainInfoOutMock = this.mock(mainInfoOut) , mainInfoUtilMock = this.mock(mainInfoUtil) - , dependencyTreeMock = this.mock(DependencyTree) + , dependencyGraphStub = this.stub() + , archyTreeStub = this.stub() , packagesArg = { packages: 1 } , optionsPackagesArg = { optionsPackages: 1 } , sizesArg = { sizes: 1 } @@ -43,6 +42,13 @@ testCase('Info', { , treeArg = { tree: 1 } , archyTreeArg = { archyTree: 1 } + require('ender-dependency-graph') + this.originalEDG = require.cache[require.resolve('ender-dependency-graph')].exports + require.cache[require.resolve('../../lib/main-info')] = null + require.cache[require.resolve('ender-dependency-graph')].exports = dependencyGraphStub + dependencyGraphStub.archyTree = archyTreeStub + mainInfo = require('../../lib/main-info') + mainInfoUtilMock .expects('sizes') .once() @@ -53,18 +59,10 @@ testCase('Info', { .once() .withArgs(expectedFilename) .callsArgWith(1, null, contextArg) - dependencyTreeMock - .expects('generate') - .once() - .withArgs(contextArg.options, optionsPackagesArg) + dependencyGraphStub.callsArgWith(2, null, treeArg) // important we use packages from context->options->packages which is the command-line packages // and not context->packages which is the full list of packages in the build - .callsArgWith(2, null, treeArg) - mainInfoUtilMock - .expects('buildArchyTree') - .once() - .withExactArgs(contextArg.options, optionsPackagesArg, treeArg) - .returns(archyTreeArg) + archyTreeStub.returns(archyTreeArg) mainInfoOutMock .expects('buildInfo') .once() @@ -72,10 +70,24 @@ testCase('Info', { mainInfo.exec(options, mainInfoOut, function (err) { refute(err) + assert.equals(dependencyGraphStub.callCount, 1) + assert.equals(dependencyGraphStub.getCall(0).args.length, 3) + assert.equals(dependencyGraphStub.getCall(0).args[0], contextArg.options) + assert.equals(dependencyGraphStub.getCall(0).args[1], optionsPackagesArg) + assert.equals(archyTreeStub.callCount, 1) + assert.equals(archyTreeStub.getCall(0).args.length, 2) + assert.equals(archyTreeStub.getCall(0).args[0], optionsPackagesArg) + assert.equals(archyTreeStub.getCall(0).args[1], treeArg) done() }) } } + + , 'tearDown': function () { + require.cache[require.resolve('ender-dependency-graph')].exports = this.originalEDG + require.cache[require.resolve('../../lib/main-info')] = null + } + , 'test no args': function (done) { this.runTest({}, 'ender.js', done) } View 137 test/unit/main-info-util-test.js @@ -113,141 +113,4 @@ testCase('Info util', { done() }) } - - // generates a tree that can be turned into nice output, not fully `archy` - // compatible yet but can be easily transformed by the output routine - , 'test generateArchyTree': function () { - var optionsArg = { options: 1 } - , packagesArg = { packages: 1 } - , treeArg = { tree: 1 } - , localPackagesArg = { localPackages: 1 } - , forEachCallback - , result - , expectedResult = { - label: 'Active packages:' - , nodes: [ - { - label: 'foo' - , version: '1.0.4' - , description: 'barfoo' - , first: true - , nodes: [ - { - label: 'bar' - , version: '1.0.3' - , description: 'barfoo' - , first: true - , nodes: [ - { - label: 'foobar' - , version: '1.0.0' - , description: 'barfoo' - , first: true - , nodes: [] - } - , { - label: 'baz' - , version: '1.0.1' - , description: 'barfoo' - , first: true - , nodes: [] - } - , { - label: 'bing' - , version: '1.0.2' - , description: 'barfoo' - , first: true - , nodes: [] - } - ] - } - ] - } - , { - label: 'fee' - , version: '1.0.5' - , description: 'barfoo' - , first: false - , nodes: [ - { - label: 'fie' - , version: '1.0.6' - , description: 'barfoo' - , first: false - , nodes: [] - } - ] - } - ] - } - - treeArg.localizePackageList = this.stub().returns(localPackagesArg) - treeArg.forEachOrderedDependency = this.spy() - - - result = mainInfoUtil.buildArchyTree(optionsArg, packagesArg, treeArg) - - assert(treeArg.forEachOrderedDependency.calledWith(localPackagesArg)) - assert(treeArg.forEachOrderedDependency.calledOnce) - - forEachCallback = treeArg.forEachOrderedDependency.lastCall.args[1] - - forEachCallback( - 'foobar' - , [ 'foo', 'bar' ] - , { packageJSON: { version: '1.0.0', name: 'fooblah1', description: 'barfoo' }} - , 0 - , true - ) - - forEachCallback( - 'baz' - , [ 'foo', 'bar' ] - , { packageJSON: { version: '1.0.1', name: 'fooblah2', description: 'barfoo' }} - , 0 - , true - ) - - forEachCallback( - 'bing' - , [ 'foo', 'bar' ] - , { packageJSON: { version: '1.0.2', name: 'fooblah3', description: 'barfoo' }} - , 0 - , true - ) - - forEachCallback( - 'bar' - , [ 'foo' ] - , { packageJSON: { version: '1.0.3', name: 'fooblah4', description: 'barfoo' }} - , 0 - , true - ) - - forEachCallback( - 'foo' - , [ ] - , { packageJSON: { version: '1.0.4', name: 'fooblah5', description: 'barfoo' }} - , 0 - , true - ) - - forEachCallback( - 'fee' - , [] - , { packageJSON: { version: '1.0.5', name: 'fooblah6', description: 'barfoo' }} - , 0 - , false - ) - - forEachCallback( - 'fie' - , [ 'fee' ] - , { packageJSON: { version: '1.0.6', name: 'fooblah7', description: 'barfoo' }} - , 0 - , false - ) - - assert.equals(result, expectedResult) - } }) View 11 test/unit/minify-test.js @@ -25,7 +25,6 @@ var buster = require('buster') , assert = buster.assert - , path = require('path') buster.testCase('Minify', { 'setUp': function () { @@ -33,14 +32,14 @@ buster.testCase('Minify', { this.sourceArg = { source: 1 } this.resultArg = { result: 1 } require('ender-minify') - this.originalEM = require.cache[path.join(__dirname, '../../node_modules/ender-minify/lib/main.js')].exports - require.cache[path.join(__dirname, '../../node_modules/ender-minify/lib/main.js')].exports = this.enderMinifyStub + this.originalEM = require.cache[require.resolve('ender-minify')].exports + require.cache[require.resolve('ender-minify')].exports = this.enderMinifyStub this.enderMinifyStub.minifiers = this.originalEM.minifiers this.enderMinifyStub.closureLevels = this.originalEM.closureLevels this.enderMinifyStub.callsArgWith(3, null, this.resultArg) this.runTest = function (minifier, expectedOptions, parsedArgs, done) { - require.cache[path.join(__dirname, '../../lib/minify.js')] = null + require.cache[require.resolve('../../lib/minify.js')] = null require('../../lib/minify').minify(parsedArgs, this.sourceArg, function (err, result) { refute(err) assert.same(result, this.resultArg) @@ -55,8 +54,8 @@ buster.testCase('Minify', { } , tearDown: function () { - require.cache[path.join(__dirname, '../../node_modules/ender-minify/lib/main.js')].exports = this.originalEM - require.cache[path.join(__dirname, '../../lib/minify.js')] = null + require.cache[require.resolve('ender-minify')].exports = this.originalEM + require.cache[require.resolve('../../lib/minify.js')] = null } , 'test basic minify, default to uglify': function (done) { 0 comments on commit e277894 Please sign in to comment. Something went wrong with that request. Please try again.
__label__pos
0.999941
#!/bin/bash # # This script tests a case of cascading replication. # # Creating system,: # node1-(physical replication)->node2-(physical replication)->node3 # # Stop instances pg_ctl stop -D node1 pg_ctl stop -D node2 pg_ctl stop -D node3 # Remove old files rm -rf node1 rm -rf node2 rm -rf node3 rm log_node2 log_node1 log_node3 # Initialize primary initdb -D node1 echo "wal_level = logical" >> node1/postgresql.conf echo "max_replication_slots=3" >> node1/postgresql.conf pg_ctl -D node1 -l log_node1 start psql -d postgres -c "CREATE DATABASE db1"; psql -d db1 -c "CHECKPOINT"; sleep 3 # Initialize standby pg_basebackup -v -R -D node2 echo "Port=9000" >> node2/postgresql.conf pg_ctl -D node2 -l log_node2 start # Initialize another standby pg_basebackup -p 9000 -v -R -D node3 echo "Port=9001" >> node3/postgresql.conf pg_ctl -D node3 -l log_node3 start # Insert a tuple to primary psql -d db1 -c "CREATE TABLE c1(a int)"; psql -d db1 -c "Insert into c1 Values(2)"; sleep 3 # And verify it can be seen on another standby psql -d db1 -p 9001 -c "Select * from c1"; pg_createsubscriber -D node2/ -S "host=localhost port=9000 dbname=postgres" -d postgres -d db1 -r -v pg_ctl -D node2 -l log_node2 start
__label__pos
0.941015
Navigation This version has reached end of life on October 2016. Learn more about upgrading your version of MongoDB. dropUser Definition dropUser Removes the user from the database on which you run the command. The dropUser command has the following syntax: { dropUser: "<user>", writeConcern: { <write concern> } } The dropUser command document has the following fields: Field Type Description dropUser string The name of the user to delete. You must issue the dropUser command while using the database where the user exists. writeConcern document Optional. The level of write concern for the removal operation. The writeConcern document takes the same fields as the getLastError command. Before dropping a user who has the userAdminAnyDatabase role, ensure you have at least another user with user administration privileges. Required Access You must have the dropUser action on a database to drop a user from that database. Example The following sequence of operations in the mongo shell removes reportUser1 from the products database: use products db.runCommand( { dropUser: "reportUser1", writeConcern: { w: "majority", wtimeout: 5000 } } )
__label__pos
0.984325
Tag Archives: Flex A new Parsley Extension: The DynamicServices tag I just pushed a new Parsley extension project to GitHub – support for the <DynamicService /> tag. Typically in a project, I like to structure my services layers pretty cleanly – for each service I would declare the following: • An interface which defines the services contract • A concrete implementation of the interface, which abstracts away the RemoteObject • A stub implementation of the interface, for testing when the services layer is unavailable For example, a simple EchoService might look a little something like this: // IEchoDelegate.as public interface IEchoDelegate { function echoMessage(source:String):AsyncToken; } // EchoDelegate.as public class EchoDelegate { [Inject] public function service:RemoteObject; public function echoMessage(source:String):AsyncToken { return service.echoMessage(source); } } Here, you can see the concrete delegate is really just boiler-plate code.  It’s pretty repetitive, and while I find lots of value in the pattern, the actual implementation can be a bore. So, I put together a Parsley extension to generate these delegates on the fly. Here’s an example: <?xml version="1.0" encoding="utf-8"?> <parsley:Objects> <fx:Declarations> <services:DynamicService type="{IEchoService}" endpoint="http://localhost:8080/testdrive/messagebroker/amf" destination="echoService" /> </fx:Declarations> </parsley:Objects> This definition combines the Interface and the remote object defininition in the context. An implementation of the interface is generated on-the-fly at runtime, and is available to be injected into classes as required. Eg: // EchoCommand.as public class EchoCommand { [Inject] public var service:IEchoService; public function execute(message:EchoMessage):AsyncToken { return service.echo(message.source); } public function result(result:String):void { trace("Received from the server:" + result); } } Source The source for this is available now on github.  There’s also a demo project available here A word of warning Under the covers, this extension makes use of the ASCommons-Bytecode library to build an implementation of the interface. Unfortunately, this dynamic goodness is not free, and incurs a one-off cost at startup when the ByteCode library parses the bytecode of the swf into memory.  As a result, you’ll notice that Parsley takes a little longer to intitialize than you might be used to.  This is a tradeoff that must be considered before using this approach. Tagged , Flapper: An extension library for AOP with Parsley I’ve just thrown a quick project up on Google Code for allowing simple AOP style programming with Parsley: Flapper The project aims to allow simple registration of AOP interceptors by registering a metatag with an interceptor. It uses the awesome as3-commons bytecode library for proxy generation Configuration The aim is to have simple configuration.  You activate Flapper by simply declaring  AspectSupport in your parsley context: <parsley:ContextBuilder> <parsley:FlexConfig type="{ParsleyConfig}" /> <tag:AspectSupport> <tag:MetatagAspectDefinition metatag="Log" aspect="{new LoggingAspect()}" /> </tag:AspectSupport> </parsley:ContextBuilder> In this example, I’m registering the Log metatag with my LoggingAspect. Now, any class which has a method annotated with Log will be proxied, and the LoggingAspect will get invoked before the actual method itself. Eg: public class Calculator { [Log] public function add(a:int,b:int):int { return a + b; } } The aspects themselves are also extremely simple, implementing an Aspect interface. Here’s my LoggingAspect: public class LoggingAspect implements Aspect { public function intercept(pointCut:ProceedingJoinPoint):void { trace(pointCut.targetMember.localName + " called with params: " + pointCut.arguments.join(",")); pointCut.proceed(); var returnValueString:String = pointCut.returnValue ? pointCut.returnValue.toString() : "null"; trace(pointCut.targetMember.localName + " returned " + returnValueString ); } } This simply traces out the method name that was called, and the arguments that were passed in. Usage That’s all the configuration that’s required.  Now, any time I call calculator.add(), the aspect will trace out the call to the console window. Note – classes that will be proxied must be decalred using Parsley’s Object tag, rather than as an instance. Eg: <fx:Declarations> <!-- Don't do this -- it won't work --> <sample:Calculator /> <!-- Instead, decalre using the Object tag --> <parsley:Object type="{Calculator}" /> </fx:Declarations> Usage from here is no different than with any other injected class. Example: [Inject] public var calculator:Calculator; private function addNumbers():void { // Results in a message being traced out "Called add with arguments: 2,3" calculator.add(2,3) } The project and a demo are available to be checked out here. It’s still fairly early days, and the project hasn’t really been tested – but feel free to grab it and have a play. Important Credit This whole project took 2 nights to write.  The main reason it was so quick and simple  is because all the hard work has already been done by two awesome existing projects. The as3-commons team have done an amazing job on the project, but also in providing thorough clear documentation. Equally so, Jens Halm has done an outstanding job on Parsley, making the framework so easy to extend, with a solid API, and killer docs.  Not to mention he’s extremely responsive in the forums.  Both these projects rock. Update: I’ve updated this blog post to reflect the move to the Aspect interface, where previously I was using the ascommons-bytecode IInterceptor.  This has made chaining Aspects easier.  Note – I’m still using ascommons-bytecode underneath, but it’s now abstracted away a little more. Tagged , , dpHibernate and Spring transactions I’ve recently been getting my feet wet in Java, and starting integrating with dpHibernate to facilitate lazy loading across to Flex. Here’s what my stack looks like: • Spring 2.5.6 • Hibernate 3.3 (including annotations, and transaction management) Trying to get dpHibernate wired in, I was getting lots of errors along the lines of the following: No Hibernate Session bound to thread, and configuration does not allow creation of non-transactional one here   The problem here is that using transaction management, things get a lot tighter on how and when you can get access to a hibernate session.  dpHibernate tries to get it’s session by essentially calling sessionManager.getCurrentSession(); However, as far as Spring’s transaction manager (which is wired into the SessionManager) is concerned, the thread which is executing has no session, and has no business trying to create one — specifically, outside of a transaction. After spending several hours trying to work out how to sidestep Springs transaction management, I realized the trick was simply to wire dpHibernate closer to spring.  This is tricky, because dpHibernate is using an extension of the flex’s JavaAdapter, and exists completely outside of the Spring infrastructure. HibernateSerializer makes a call to a dpHibernate object called SessionManager, which returns the current session outside of a valid Spring transaction. So, I ditched the SessionManager completely, and replaced it with a good old fashionsed SessionFactory object, as follows: @Resource private SessionFactory sessionFactory; public void setSessionFactory(SessionFactory sessionFactory) { this.sessionFactory = sessionFactory; } public SessionFactory getSessionFactory() { return sessionFactory; } The @Resource tag there tells Spring to give me a wired up sessionFactory instance. I also decorated the HibernateSerializer with a @Transactional tag, so that Spring knows the context in which this will operate: @Transactional(readOnly=true) public class HibernateSerializer implements ISerializer { Now, if I add the HibernateSerializer as a bean, Spring knows to give it access to a sessionManager, and to do so in the context of an active transaction. Add the bean declaration as follows: <bean id="hibernateSerializerBean" class="net.digitalprimates.persistence.translators.hibernate.HibernateSerializer" /> Nearly there. Finally, I just need to tell dpHibernate to go and fetch this wired up Serializer from Spring, rather than instantiate it’s own. Inside SerializationFactory, I modified the getSerializer() call to read as follows: public static ISerializer getSerializer(String type) { if( HIBERNATESERIALIZER.equals(type) ) { ServletContext ctx = FlexContext.getServletContext(); WebApplicationContext springContext = WebApplicationContextUtils.getRequiredWebApplicationContext(ctx); ISerializer serializer = (ISerializer) springContext.getBean("hibernateSerializerBean"); if (serializer == null) { throw new RuntimeException("bean named hibernateSerializerBean not found"); } return serializer; } throw new RuntimeException("unsupport serialization type: " +type); } Note that I’ve hard coded the name of the bean here for simplicity sakes. Chances are you’ll want to paramaterize that. That’s it! Now, calls that pass through dpHibernate happen with a Spring managed transactional session. Great!   Edit: Espen came up with a simpler alternative over on his blog — check it out : http://espenskogen.wordpress.com/2009/06/04/more-dphibernate/ Tagged , , , Mock Webservice in AS3 An important part of a good test suite is repeatability, and independance from external changes. Therefore, if your method loads data from an external source, that’s a real testing no-no. However, using the features of dynamic classes in AS3 we can whip up a mock webservice which gives us a guaranteed result, every time. Here’s the class: package com.martypitt.testing.mockObjects { import flash.utils.Dictionary; import flash.utils.flash_proxy; import mx.core.Application; import mx.core.mx_internal; import mx.messaging.messages.AsyncMessage; import mx.messaging.messages.IMessage; import mx.rpc.AbstractOperation; import mx.rpc.AsyncToken; import mx.rpc.events.ResultEvent; import mx.rpc.soap.WebService; use namespace flash_proxy; use namespace mx_internal; public dynamic class MockWebService extends WebService { /** The default response is returned if a call is made to a method that has not been explicitly * set using setResponse() */ private var _defaultResult:Object; /** Determines bevahiour if a call is made to a method where the response has not been explicity set. * If true, an error is thrown. * If false (default) the defaultResponse is returned. * */ private var _throwErrorIfResultNotSet:Boolean = false; private var _resultDictionary:Dictionary = new Dictionary(); public function get defaultResult():Object { return _defaultResult; } public function set defaultResult(value:Object):void { _defaultResult = value; } /** responseDictionary holds a list of predefined methods and their associated responses. * */ protected function get resultDictionary():Dictionary { return _resultDictionary; } protected function set resultDictionary(value:Dictionary):void { _resultDictionary = value; } public function get throwErrorIfResultNotSet():Boolean { return _throwErrorIfResultNotSet; } public function set throwErrorIfResultNotSet(value:Boolean):void { _throwErrorIfResultNotSet = value; } public function MockWebService(destination:String=null, rootURL:String=null) { super(destination, rootURL); } /** callProperty is called by the Flex / flash framework when something tries to access * a method or property of a dynamic object which has not been defined in code. * Here we override the default behaviour to return the correct response. * */ override flash_proxy function callProperty(name:*, ... args:Array):* { var response:Object; var methodName:String = getLocalName(name); switch (true) { case (_resultDictionary[methodName] == null && throwErrorIfResultNotSet) : throw new Error(String(methodName) + " does not have an appropriate result set."); case (_resultDictionary[methodName] == null && !throwErrorIfResultNotSet) : response = defaultResult; break; case (_resultDictionary[methodName] != null) : response = _resultDictionary[methodName] break; default : throw new Error("Unhandled switch case scenario in MockWebService"); } var message:IMessage = new AsyncMessage() var token:AsyncToken = new AsyncToken(message) // We want to return the Async token now (just like a normal webservice would) // but we get the app to dispatch the ResultEvent on the next frame // It's important that the result event is dispatched from the operation itself. var operation:AbstractOperation = super.getOperation(methodName); Application.application.callLater(operation.dispatchEvent,[ResultEvent.createEvent(response,token,message)]); return token; } public function setResult(methodName:String,responseValue:Object):void { _resultDictionary[methodName] = responseValue; } } } Nothing overly swanky, but all the magic happens courtesy of this line : override flash_proxy function callProperty(name:*, ... args:Array):* This method gets invoked whenever you call a method on the class which has not been defined in code. We catch the call, and generate an Async token that makes the caller think that we’re busily off chatting to our webserver. The other key line is this: Application.application.callLater(operation.dispatchEvent,[ResultEvent.createEvent(response,token,message)]); In essence, we’re just dispatching a result event here. But, because our method is called syncronously, and webservices behave asyncronously, it’s important that we dispatch the event on the next frame. Hence Application.application.callLater. Also, the result event must come from the operation itself, not our mock service, so that whoever has called this method gets notified of the result event, just like they would if it were a normal webservice. So, with all that done, we simply set up the webservice, and invoke it: var result:XML = new XML(); var ws:MockWebService = new MockWebService(); ws.setResult("callMyRemoteMethod",result) someClass.webservice = ws; It’s important (and good design practice) that you expose the webservice on your class as property that can be modified at runtime. That’s pretty much it. Questions and comments welcomed. Marty Tagged , , Custom Metatags and slicker unit tests in Flex I’m a fan of unit tests. And, although I probably wouldn’t blazen it across my chest like some, I am sold of the benefits. However, coming from .NET, I miss quite a bit of the elegance present in NUnit when working with FlexUnit. Most specifically, the setup of a test suite is kinda clunky. Normally, you have several classes of unit tests, each which have a static suite() method, returning a test suite comprising the unit tests in that class. Check out the following example: public static function suite():TestSuite { var ts:TestSuite = new TestSuite(); ts.addTest( new MyTestClass("testMethodA") ); ts.addTest( new MyTestClass("testMethodB") ); ts.addTest( new MyTestClass("testMethodC") ); ts.addTest( new MyTestClass("testMethodD") ); return ts; } As you can appreciate, it doesn’t take long for this to get long and clunky. Also, maintanence is a pain in the butt — adding tests don’t get run if you forget to add it in the Suite method, changing names causes RTE, etc etc etc. In NUnit (and, presumably JUnit), you just decorate a class with [TestFixture], and a method with [Test], and NUnit takes care of the rest using that groovy cat, Reflection. Luckily for us, Flex gives us the ability to tell the compiler to retain custom metatags in Flex 3. This means we can achieve similar functionailty — at least at the method level. The first step is to set the compiler option. In Flex Builder 3, right click your project, hop into Flex Compiler, and add this to the list of compiler options: -keep-as3-metadata+=Test Then, using flash.utils.describeType() we can find any methods within a class that are decoared with our [Test] method. The following method examines a class, and returns a TestSuite which contains a test for each decorated method. /** * Generates a test suite with all the methods in a class decorated with [Test] metadata * @param clazz * @return * */ public static function fromClass(clazz:Class):TestSuite { var suite:TestSuite = new TestSuite(); var testMethodNames : XMLList = describeType(clazz).method.(metadata.(@name==”Test”)).@name for each (var node:XML in testMethods) { var methodName:String = node.value() var test:Test = new clazz(methodName); suite.addTest(test); } return suite; } Personally, I chose to add this to the TestSuite class in FlexUnit, though you may prefer a seperate util class. Now, getting all the tests for a class is as simple as: var suite:TestSuite = TestSuite.forClass(myClass); Much more readable! Comments welcomed. Marty Update : I modified the disvoery method to use the E4X expression suggested by Theo. I must’ve been having a brain-fart at the time, as I’m a huge fan of the E4X syntax. Nice spotting Theo! Tagged , , Follow Get every new post delivered to your Inbox. Join 189 other followers
__label__pos
0.735107
I have a database which has multiple tables. I want all the table names to be displayed in a drop down html menu. I have a php function which I call on my html document. function namealbums() { $con = mysql_connect("localhost","root",""); if (!$con) { die('Could not connect: ' . mysql_error()); } $query = mysql_query(" USE photos GO SELECT * FROM sys.Tables GO"); echo '<select name="dropdown">'; while ($row = mysql_fetch_array($query)) { echo '<option value="'.$row.'">'.$row.'</option>'; } echo '</select>'; } ?> This is my HTML form which calls the function. <table width="300px" align="center" border="0" cellpadding="3" cellspacing="1" bgcolor="#FFFFFF"> <tr> <td cellpadding="100"> <?php echo namealbums();?> </td> </tr> </table> All i get is a drop down menu with no results in it. Just empty. Member Avatar LastMitch I have a database which has multiple tables. I want all the table names to be displayed in a drop down html menu. I have a php function which I call on my html document. I'm so lost what you are trying to do? Why are you using SQL Server scripts instead of using MYSQL? I assume you are connected to the db and your query works? I also mean this: $query = mysql_query("USE photos GO SELECT * FROM sys.Tables GO"); Where is your id? How can you fetch data without an id? Instead of this: <?php $query = mysql_query("USE photos GO SELECT * FROM sys.Tables GO"); echo '<select name="dropdown">'; while ($row = mysql_fetch_array($query)) { echo '<option value="'.$row.'">'.$row.'</option>'; } echo '</select>'; } ?> Try this: <?php $query = "SELECT id FROM sys.TablesGO"; $result = mysql_query ($query); echo "<select name=dropdown value=''>Dropdown</option>"; while($row = mysql_fetch_array($result)){ echo "<option value=$row[id]>$row[id]</option>"; } echo "</select>"; ?> Sorry I am using mysql not sql. But I didnt have an idea because I didnt know I needed one. I tried your new code and it returned the same result. I have no idea why. Member Avatar LastMitch @garyjohnson Sorry I am using mysql not sql. Read this your example looks exactly like these example: http://odachoo.blogspot.com/2012/04/querying-sql-server-system-catalog-faq.html But I didnt have an idea because I didnt know I needed one. You do need a DB and a query. I tried your new code and it returned the same result. I have no idea why. The reason why it didn't work because the query is wrong and the db is not connected. I think I help you in the past. I am bit surprise that you don't know how to create a table or a db and create a query to fetch the data. The code I provided does that. The only thing you need to do is connect the db and used the query to fetch the data. What is your table? Can you post your table. Im sorry that Im making this confusing lol but I do have a database and a table. For some reason I didnt post all of the function which connects to the databse. This is the whole function. function namealbums() { $con = mysql_connect("localhost","root",""); if (!$con) { die('Could not connect: ' . mysql_error()); } $query = "SELECT id FROM sys.Tables GO"; $result = mysql_query ($query); echo "<select name=dropdown value=''>Dropdown</option>"; while($row = mysql_fetch_array($result)){ echo "<option value=$row[id]>$row[id]</option>"; } echo "</select>"; } ?> I am not sure how im suppose to post my table but it goes like this. The database is called Photos. There are mutliple tables within the database. The columns go as is, pic_id int (11) location char(30) char2 char(30) char1 char(30) name char(30) album char(30) Member Avatar LastMitch @garyjohnson I am not sure how im suppose to post my table but it goes like this. The database is called Photos. There are mutliple tables within the database. The columns go as is, This is fine. What data are you trying to fetch? name or album? This is for name: <?php $conn = mysql_connect("$db_host", "$db_username", "$db_pass") or die ("could not connect to mysql"); mysql_select_db("$db_name") or die ("no database"); $query = "SELECT pic_id, name FROM Yourtable ORDER BY name"; $result = mysql_query ($query); echo "<select name=dropdown value=''>Dropdown</option>"; while($r = mysql_fetch_array($result)){ echo "<option value=$r[pic_id]>$r[name]</option>"; } echo "</select>"; ?> This is for album: <?php $conn = mysql_connect("$db_host", "$db_username", "$db_pass") or die ("could not connect to mysql"); mysql_select_db("$db_name") or die ("no database"); $query = "SELECT pic_id, album FROM Yourtable ORDER BY album"; $result = mysql_query ($query); echo "<select name=dropdown value=''>Dropdown</option>"; while($r = mysql_fetch_array($result)){ echo "<option value=$r[pic_id]>$r[album]</option>"; } echo "</select>"; ?> This is pretty much the outline. It worked! Thank you so much. I dont know why I had so much trouble with this. I have done this before. Thanks Though! 3.php <html> <head> <link href="http://ajax.googleapis.com/ajax/libs/jqueryui/1.8/themes/base/jquery-ui.css" rel="stylesheet" type="text/css"/> <script type="text/javascript" src="http://ajax.googleapis.com/ajax/libs/jquery/1.6.2/jquery.min.js"></script> <script src="http://ajax.googleapis.com/ajax/libs/jqueryui/1.8/jquery-ui.min.js"></script> <script> $(document).ready(function() { $("#datepicker").datepicker({ dateFormat: "dd/mm/yy" }); $("#datepicker2").datepicker({ dateFormat: "dd/mm/yy" }); }); </script> <?php function peralatan() { $conn = mysql_connect("localhost", "root", "") or die(mysql_error()); mysql_select_db("borrow") or die(mysql_error()); $query = "SELECT item FROM inventori ORDER BY item"; $result = mysql_query ($query); echo "<select name=dropdown value=''>Dropdown</option>"; while($r = mysql_fetch_array($result)){ echo "<option value=$r[id]>$r[item]</option>"; } echo "</select>"; } ?> <tr> <td width="871" height="350"> <p align="center" class="style12">&nbsp;</p> <p align="center" class="style12"></p> <table width="428" height="226" border="0" align="center"> <form method="post" action="4.php"> <table width="520" border="0" align="center"> <tr> <td width="137"><span class="style21">Name</span></td> <td width="144"><input type="textbox1" name="Name[]"></td> <td width="82">&nbsp;</td> <td width="144">&nbsp;</td> </tr> <tr> <td><span class="style21">Startdate</span></td> <td><input type="textbox6" name="Startdate[]" id="datepicker"></td> <td align="center"><span class="style21">sehingga</span></td> <td><input type="textbox7" name="Enddate[]" id="datepicker2"></td> </tr> </table> <br /> <table width="518" border="1" align="center"> <tr align="center"> <th width="28" scope="col"><div align="center" class="style21 style24">Bil.</div></th> <th width="153" scope="col"><span class="style25">Item</span></th> <th width="156" scope="col"><span class="style25">Model</span></th> <th width="153" scope="col"><span class="style25">No. Inventori</span></th> </tr> <tr align="center"> <td>1.</td> <td><select name="dropdown" name="item1[]"> <?php echo item();?></td> <td><input type="textbox8" name="item1[]"></td> <td><input type="textbox8" name="item1[]"></td> </tr> <tr align="center"> <td>2.</td> <td><input type="textbox9" name="item2[]"></td> <td><input type="textbox9" name="item2[]"></td> <td><input type="textbox9" name="item2[]"></td> </tr> <tr align="center"> <td>3.</td> <td><input type="textbox10" name="item3[]"></td> <td><input type="textbox10" name="item3[]"></td> <td><input type="textbox10" name="item3[]"></td> </tr> <tr align="center"> <td>4.</td> <td><input type="textbox11" name="item4[]"></td> <td><input type="textbox11" name="item4[]"></td> <td><input type="textbox11" name="item4[]"></td> </tr> <tr align="center"> <td>5.</td> <td><input type="textbox12" name="item5[]"></td> <td><input type="textbox12" name="item5[]"></td> <td><input type="textbox12" name="item5[]"></td> </tr> </table> <br /> <center><input type="submit" name="submit" value="submit"><input type="reset" name="Reset" value="reset"></center> </form> </body> </html> 4.php <?php // Make a MySQL Connection mysql_connect("localhost", "root", "") or die(mysql_error()); mysql_select_db("pinjaman") or die(mysql_error()); $TextBox1 = implode(', ', $_POST['Name']); $TextBox6 = implode(', ', $_POST['Startdate']); $TextBox7 = implode(', ', $_POST['Enddate']); $dropdown = implode(', ', $_POST['item1']); $TextBox8 = implode(', ', $_POST['item1']); $TextBox8 = implode(', ', $_POST['item1']); $TextBox9 = implode(', ', $_POST['item2']); $TextBox9 = implode(', ', $_POST['item2']); $TextBox9 = implode(', ', $_POST['item2']); $TextBox10 = implode(', ', $_POST['item3']); $TextBox10 = implode(', ', $_POST['item3']); $TextBox10 = implode(', ', $_POST['item3']); $TextBox11 = implode(', ', $_POST['item4']); $TextBox11 = implode(', ', $_POST['item4']); $TextBox11 = implode(', ', $_POST['item4']); if(isset($_POST['submit'])) { $query="INSERT INTO epal VALUES ('" . $TextBox1 . "','" . $TextBox6 . "','" . $TextBox7 . "','" . $dropdown . "','" . $TextBox8 . "','" . $TextBox9 . "','" . $TextBox10 . "','" . $TextBox11 . "')"; mysql_query($query) or die (mysql_error() ); echo "Complete"; } ?> i try your drop down list.it work! but i have a problem to insert the data to database.really need a help.
__label__pos
0.954878
lithium\test\Report::render() public method Renders the test output (e.g. layouts and filter templates). Parameters • string $template name of the template (i.e. 'layout'). • string|array $data array from _data() method. Returns string Filter This method can be filtered. Source public function render($template, $data = []) { $config = $this->_config; if ($template === 'stats' && !$data) { $data = $this->stats(); } $template = Libraries::locate('test.templates', $template, [ 'filter' => false, 'type' => 'file', 'suffix' => ".{$config['format']}.php" ]); if ($template === null) { $message = "Templates for format `{$config['format']}` not found in `test/templates`."; throw new TemplateException($message); } $params = compact('template', 'data', 'config'); return Filters::run(__CLASS__, __FUNCTION__, $params, function($params) { extract($params['data']); ob_start(); include $params['template']; return ob_get_clean(); }); }
__label__pos
0.911756
How many pushups in 6 months? Whatsapp Puzzles: If we do 1 push up repetition per day, 3 sets, then increase it by one rep per week for same three sets, how many push ups will we be doing in 6 months?       Answer: 7215 Solution: Let’s assume 30 days in a month, so 6*30 = 180 days in 6 months, i.e. 180/7 = 25weeks + 5 days. number of push ups in first week 3*7 = 21 so total number of push ups in 25 weeks = 21 + 21*2 + 21*3 …. 21*25 its an arithmetic progression with a1 = 21 and d = 21 its Sum = n(a1+an)/2 = 25(21 + 21*25)/2 = 21*25*26/2 = 6825 Push ups done in 5 days = 21*26*5/7 = 26*15 = 390 Total push ups in 6 months = 6825 + 390 = 7215. You may also like...
__label__pos
1
1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More. Linking Directly To A Pic In A Table Discussion in 'Web Design' started by redrubies, Jan 21, 2013. 1. redrubies redrubies Supreme Member Joined: Jan 17, 2011 Messages: 1,424 Likes Received: 2,560 Location: USA I think I screwed something up. If you have a table with rows of thumbnails, how do you link to one product in the table? Say you want to put a link on twitter of a pic in the second row of the table, third box. What is the code for the link to go directly to that pic? Thank you!   2. gabfire gabfire Newbie Joined: Aug 10, 2012 Messages: 26 Likes Received: 9 You can link directly to it, but it won't be very specific if there are 3 items in the row, it'll just put that row at the top of the page on load. The way you do it is as follows: In your row where the pic is (in your example 2nd row, 3rd column) you'll have to anchor the image. So use this HTML: Code: <a name="pic2ndrow3rdcolumn"><img src="blah/blah/blah.jpg" alt="blah"></a> You could wrap just the header of the field or whatever, but that is the anchor the browser will jump to. Now in your link on Twitter, you'll do this: Code: Hey guys, check out url/pagewithtable.html#pic2ndrow3rdcolumn That will cause the browser to load the HTML and then jump to that anchor tag.   • Thanks Thanks x 1 3. srb888 srb888 Elite Member Joined: Jul 30, 2008 Messages: 3,260 Likes Received: 5,067 Gender: Male Occupation: WebzSurfer Location: Sun, Mon, Tue, WTF, Sat!!! :) Code: [URL]http://www.w3schools.com/tags/tag_img.asp[/URL] Is this what you are looking for?   • Thanks Thanks x 1 4. redrubies redrubies Supreme Member Joined: Jan 17, 2011 Messages: 1,424 Likes Received: 2,560 Location: USA Oh! I got it now. Thank you!   5. gabfire gabfire Newbie Joined: Aug 10, 2012 Messages: 26 Likes Received: 9 No, that's not at all what they're asking. Also - w3schools is a horrible HTML reference;) They've gotten better over the past few years, but there are still tons of things that are outright wrong on there:)   6. srb888 srb888 Elite Member Joined: Jul 30, 2008 Messages: 3,260 Likes Received: 5,067 Gender: Male Occupation: WebzSurfer Location: Sun, Mon, Tue, WTF, Sat!!! :) I gave the same reference as you gave. The link has the same reference, <img src="asdfg" alt="asdfg">, and that's the start point from where you build the code (as you mentioned in your post). The site also has a easy "try the code yourself" on it which helps to build upon the basics, and so regardless of what you think of the site, it's still a good starting point for a beginner. If you know any related site that's better for learners and has a facility to test code like the above site, then please give a reference to that on the forum, so that members like OP and even I can benefit from it.  
__label__pos
0.756001
5.1.3. Multiplication of Fractions and Decimal Numbers by 10, 100, and 1000 Multiply . • A 523.5 • B524 • C 5.235 • D 5 235 Example Multiply . Solution To multiply a decimal number by 10, move the decimal point 1 place to the right. 0 correct 0 incorrect 0 skipped
__label__pos
0.910102
If you're developing an application that requires a lot of forms, you'll often find yourself repeating the same code. There are good reasons to avoid using a form package: loading a bunch of classes and running through layers of validation will never be as fast as just working with markup and post values.  But if your project doesn't have to be incredibly fast, and you're already taking the standard precautions against resource drain (smart caching, efficient class loading, minimal bootstrapping), a form package can save you coding time, or help you standardize forms across an application without relying on copy/paste. Once you've decided to use a form package, you'll have to figure out which one is best for the task at hand. As with so many other programming decisions, rolling your own is the way to ruthlessly optimize for the things you care about, but it requires the kind of time investment that might not be feasible. Choosing an existing package may be as simple as noticing one that's part of a library you're already using, or it might take some extensive research on what's out there and what the strengths are for each. For the purposes of this article, we'll be focusing upon the two big tasks of form building: display and validation. The examples will use Zend_Form, which is large and complex enough to include most of the common ways of doing each. And we'll talk about custom elements, which can be incredibly useful and require attention to both. Views One important question to ask yourself when choosing or building a form package is how it's going to fit into the rest of your architecture, and into the rhythms of development for the project. Most packages offer you the option of auto-building the form markup from the element objects you've added. This has a lot of appeal for back-end developers, for whom writing markup is the least interesting part of the endeavor, and you can get a lot of mileage out of them on plenty of projects. An administrative back-end is ideal for this kind of use. It's used only by people intimately familiar with the terminology you'll be using for labels, so you're unlikely to have do much UI engineering that would require fine-grain control over the layout. Admin areas tend to have a lot of forms that should all share a look and feel. Using auto-build views means you can describe the look in your base class--any tweaks to the default view can be applied to every form. You can also make use of auto-build views for a project where you must do demos while you're working on the underlying API: people (particularly people without programming knowledge, but sometimes even programmers do it, too) instinctively associate the design polish of a form with the state of completion of the code underneath it. The raw look of default forms can work in your favor to remind everyone that your API is still under development. There are, however, good reasons not to let it persist into your production code. If you've inherited a project that uses a form package in this way within an MVC framework, you'll find view blocks that look something like this: <div id="some-form"> <?php echo $this->form->render(); ?> </div> This is not precisely helpful when you or your front-end developer have been asked to add an image next to the CVV field. Where is all the markup? How do you know where to add it? Instead, you need to find your way into the form object to have a look at some abstracted markup: $cvv = new Zend_Form_Element_Text('cvv'); $cvv->addValidator(new My_Custom_CVV_Validator()); $cvv->addDescription('A 3 or 4 digit number located on the back of the card.'); $cvv->addLabel('CVV/Security Code'); $cvv->addDecorator('HTMLTag', array('tag' => 'span', 'class' => 'short-entry')); $this->addElement($cvv); $this->addDisplayGroup( array('card_name','card_type','card_num','exp_month','exp_year','cvv'), 'credit_card', array('displayGroupClass' => 'card') ); This sort of complexity has a tendency to accumulate over time, too. Before long, you may find yourself with code that breaks a form into columns, wraps tags around elements or just shoehorns blocks of markup in wherever they'll fit. At that point, I'd recommend sleeping in locked rooms, because your front-end developer owns several large kitchen knives and knows where you live. How, then, do you balance your client's need for you to bring up forms quickly with your front-end developer's need to have a form's markup all in one file? Personally, I like to keep the overall layout in the view with the important parts plugged in. For example: <form id="new_widget" method="<?php echo $this->form->getMethod(); ?>" action="<?php echo $this->form->getAction(); ?>" enctype="application/x-www-form-urlencoded"<?php if ($this->form->isErrors()) { echo ' class="errors"'; } ?>> <fieldset class="column"> <legend>Details</legend> <ul class="elements"> <li id="elem_title"> <label for="title">Title:</label> <p class="note">(will be used to build the url slug)</p> <?php echo $this->form->title->renderViewHelper(); ?> <?php echo $this->form->title->renderErrors(); ?> </li> . . .and so on. You can see how easy it would be to loop through the elements rather than grabbing them individually, if you were so inclined. Validation Let's consider a simple scenario: you must validate a postal code based on country: U.S. against /^\d{5}(-\d{4})?$/, CA against /^[ABCEGHJKLMNPRSTVXY]\d[A-Z] \d[A-Z]\d$/, etc. There are four common places to handle validation: 1. Using validators attached to the elements 2. Using a central method for the form 3. Outside of the form class entirely, in the controller code 4. Using javascript automatically built by the form package 5. Using javascript elsewhere (e.g., at the bottom of the view) Element-level validation is the way to go if your country is coming from somewhere outside the form-if, for example, you're using country code subdomains. You can have your form class accept a country code on build, and set the appropriate regex for your element when you add it. (Most packages come with a regex validator, and it's a simple one to write if you're rolling your own.) But what if you're getting the country from a select box? You'll need a set of regexes and a way to switch between them based upon the value of the select. For this, you'll need custom validation. Some packages have a built-in "callback" option for element validation, while others use custom classes. Zend Form sends the whole form values array to its validation classes as $context, but this is relatively uncommon. Most packages require you to fall back to the central validation method instead. Centralized validation is typically implemented as a method you can extend (the parent version runs all of the element-level validation) to include rules that apply to the form as a whole. You can use it to switch out which fields are required, too, if the package allows you to update that information on the fly. The main disadvantage to making this your primary method of customized validation is that it's not very re-use friendly. The logic to test a postal code based upon country will have to be copied between forms, or made into a function or class that can be included. (This is another advantage of Zend Form: the validators extend Zend_Validate rather than anything in the Form package, so they can be used on non-form data, without requiring any of the Zend Form classes.) For obvious reasons, breaking out of the form to do validation outside of it isn't ideal. If it ends up there, it should be an indicator that your package is too hard to use. Auto-generated javascript is often more trouble than it's worth: you don't usually have much control over how the error messages are displayed, and just like auto-build views, they make front-end developers sad. Worst of all, they don't save you much time. If you're only using the validators that are already part of the package, it's relatively painless, but any custom validation has to be written twice-once for the backend and once for the front end. That gets old pretty fast. You can get most of the benefit of automatic Javascript by using an AJAX call on submit, but of course it only saves you the difference between the cost of a page load and the cost of an AJAX call. You don't get the benefit of not contacting the server at all, but it's usually sufficient for most forms. Using jquery as an example var formvals = {}; // grab these with .val() as necessary $('#myform').submit(function (e) { e.preventDefault(); $.post({ url: '/path/to/script', data: formvals, error: function (ret) { showGenericError(ret); }, success: function (ret) { if (ret.success) { // go to the next page, show an message, etc } else if ('messages' in ret) { // show them on the form } else { showGenericError(ret); } } }); }); You would just need to add something to the backend to pull up your form class, run the validation method on the post data, and return some json to indicate success and or failure, with error messages keyed by element name. Your front-end dev can do whatever he or she wants with them in the callback. Note that using a real submit button also means it will degrade gracefully; if the user has javascript turned off, the form will behave as usual. Custom javascript is usually best written ad-hoc, per form, with reusable parts included in a central file. Combining the AJAX method above with custom non-ajax validation (like checking password strength as you type) means the front-end and back-end remain separate, which can be a significant benefit when the front-end and back-end developers must work independently. Custom elements Most of the time, regular HTML form elements are fine (and if they aren't, there's usually a jquery plugin to fix that), but occasionally you'll have to do something much more complicated and then include it on multiple forms-for example, an image selector that taps into a pool of user-uploaded images. Often the code to support it will be extensive and involve not only markup but also validation, display, and filtering logic. Rather than try to keep a library of functions to do each piece, you can build a custom element that knows about all those parts and slot itself into the regular form flow. In Zend_Form, this usually means creating a very simple custom form class: class My_Element extends Zend_Form_Element { public $helper = 'myElement'; public function init () { parent::init(); $this->addValidator(new My_Validate_MyElement()); } } And a view helper (assuming you've already set the view helper path): class My_Helper_MyElement extends Zend_View_Helper_FormElement { public function myElement($name, $value='', $attribs=array(), $options=null) { $info = $this->_getInfo($name, $value, $attribs, $options); $content = '<div id="my_elem_' . $info['name'] . '">'; // show the value in an interesting way... // load a jquery plugin, whatever $content .= '</div>'; return $content; } } And, if necessary, a validator: class My_Validate_MyElement extends Zend_Validate_Abstract { const MYBAD = 'myBad'; protected $_messageTemplates = array( self::MYBAD => "%value% is wrong because...', ); public function isValid($value, $context = array()) { if (/* some test */) { $this->_error(self::MYBAD); return false; } return true; } } Then you can add this element wherever you need it. . .like so: $this->addElement(new My_Element()); Much less painful all around. In conclusion As you go about bending forms to your will, bear in mind that the choices you make will affect your whole team, and anyone who works on the code after you. Separating layout from your class structure will ensure that your front-end developer doesn't have to learn about the Decorator Pattern, and modular validation will make for a easy transition when you have to offer the form as a web service. These packages aren't as annoying as they can sometimes seem, and they can provide speed now and flexibility later, provided you use them with care.
__label__pos
0.897444
/* * Copyright (c) 2001 Markus Friedl. All rights reserved. * Copyright (c) 2001 Damien Miller. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "includes.h" RCSID("$OpenBSD: sftp-common.c,v 1.3 2001/06/26 17:27:24 markus Exp $"); #include "buffer.h" #include "bufaux.h" #include "getput.h" #include "log.h" #include "xmalloc.h" #include "sftp.h" #include "sftp-common.h" /* Clear contents of attributes structure */ void attrib_clear(Attrib *a) { a->flags = 0; a->size = 0; a->uid = 0; a->gid = 0; a->perm = 0; a->atime = 0; a->mtime = 0; } /* Convert from struct stat to filexfer attribs */ void stat_to_attrib(struct stat *st, Attrib *a) { attrib_clear(a); a->flags = 0; a->flags |= SSH2_FILEXFER_ATTR_SIZE; a->size = st->st_size; a->flags |= SSH2_FILEXFER_ATTR_UIDGID; a->uid = st->st_uid; a->gid = st->st_gid; a->flags |= SSH2_FILEXFER_ATTR_PERMISSIONS; a->perm = st->st_mode; a->flags |= SSH2_FILEXFER_ATTR_ACMODTIME; a->atime = st->st_atime; a->mtime = st->st_mtime; } /* Decode attributes in buffer */ Attrib * decode_attrib(Buffer *b) { static Attrib a; attrib_clear(&a); a.flags = buffer_get_int(b); if (a.flags & SSH2_FILEXFER_ATTR_SIZE) a.size = buffer_get_int64(b); if (a.flags & SSH2_FILEXFER_ATTR_UIDGID) { a.uid = buffer_get_int(b); a.gid = buffer_get_int(b); } if (a.flags & SSH2_FILEXFER_ATTR_PERMISSIONS) a.perm = buffer_get_int(b); if (a.flags & SSH2_FILEXFER_ATTR_ACMODTIME) { a.atime = buffer_get_int(b); a.mtime = buffer_get_int(b); } /* vendor-specific extensions */ if (a.flags & SSH2_FILEXFER_ATTR_EXTENDED) { char *type, *data; int i, count; count = buffer_get_int(b); for (i = 0; i < count; i++) { type = buffer_get_string(b, NULL); data = buffer_get_string(b, NULL); debug3("Got file attribute \"%s\"", type); xfree(type); xfree(data); } } return &a; } /* Encode attributes to buffer */ void encode_attrib(Buffer *b, Attrib *a) { buffer_put_int(b, a->flags); if (a->flags & SSH2_FILEXFER_ATTR_SIZE) buffer_put_int64(b, a->size); if (a->flags & SSH2_FILEXFER_ATTR_UIDGID) { buffer_put_int(b, a->uid); buffer_put_int(b, a->gid); } if (a->flags & SSH2_FILEXFER_ATTR_PERMISSIONS) buffer_put_int(b, a->perm); if (a->flags & SSH2_FILEXFER_ATTR_ACMODTIME) { buffer_put_int(b, a->atime); buffer_put_int(b, a->mtime); } } /* Convert from SSH2_FX_ status to text error message */ const char * fx2txt(int status) { switch (status) { case SSH2_FX_OK: return("No error"); case SSH2_FX_EOF: return("End of file"); case SSH2_FX_NO_SUCH_FILE: return("No such file or directory"); case SSH2_FX_PERMISSION_DENIED: return("Permission denied"); case SSH2_FX_FAILURE: return("Failure"); case SSH2_FX_BAD_MESSAGE: return("Bad message"); case SSH2_FX_NO_CONNECTION: return("No connection"); case SSH2_FX_CONNECTION_LOST: return("Connection lost"); case SSH2_FX_OP_UNSUPPORTED: return("Operation unsupported"); default: return("Unknown status"); }; /* NOTREACHED */ }
__label__pos
0.998854
Pages Friday, 27 November 2020 How to Design Detection Logic - Part 1    In this first part we are going to share with you some common logical and high level steps we tend to follow to design detection logic for a certain attack technique. To make it simple and straightforward we will start with some definitions (to align) and then analyze the following diagram that summarizes big chunks of the process. Definitions: • attack technique: group of small blocks (primitives) chained to bypass a certain security control (e.g. steal secrets, elevate privileges, execute code remote or locally). • datasource: mainly logs (host and network) and OS telemetry such as processes execution, file modifications, network connection. • detection resilience: high level qualitative metric to measure how easy for an attacker to bypass a certain detection logic (e.g. to detect LSASS memory dump creation we monitor file creation with the name "lsass.dmp". this can be easily bypassed if the attacker has control over the file name). • unique changes: if a certain attack primitive  performs a change that happens a lot and in normal conditions (e.g. create a file with extension .tmp or .js in the user temporary directory)  then this change is not unique enough and hence can't be used as an indicator of suspicious activity. • context: if a certain change is unique enough to use it as an indicator of suspicious activity, we still have to assess if it provides enough context or it can be associated to 100 techniques. Step A, consist of identifying all building blocks of certain attack technique, in our example we have 8 primitives for the attack technique X (often involves reading documentation and source code if available and needed). Step B, consist of identifying what's necessary for the technique success and what's optional from an attacker perspective for the success of the technique, in our example out of 8 primitives only 5 are needed (still green) and the rest are optional and if omitted the technique still works.  Step C, consist of identifying what's under the attacker control and what's not (e.g. in PM1 the technique needs drop a dll file in the system32 directory, the default name is abc.dll (still can be used as signature) but the attacker controls the name and can set it to more than 20 different unique names). In our example out of 5 necessary PMs, only 3 are non modifiable (still green) and 2 are modifiable (marked as dark green). Step D, consist of mapping the 5 necessary PM to the relevant datasources we have at our disposition, (e.g. in PM8 Explorer.exe will perform a network connection but we don't collect processes network telemetry). In our example out of 5 PMs we have telemetry for only  3 PMs and the 2 others are opportunities for improvement (marked in purple) and if we encounter a medium to high number of techniques that requires the same type of telemetry then it's worth using it as a justification to enable visibility on those gaps. Step E, mainly consist of identifying what's normal (happens a lot and if enabled as a detection will DoS your mailbox and SIEM), exclusion opportunities and what's unique enough to use it as an indicator of suspicious activity. This usually involves querying the history of existing datastores and if the number of hits is medium to low then its worth moving to the next step. In our example out of 3 remaining PMs we are left with 2 . Step F,  In this step we are are left with 2/8 PMs, that can serve as our initial detection scope, we need to assess the detection opportunities we have in term of performance impact, alert context and enrichment options. for instance if PM4 alone is indeed indicative of something suspicious still it can be also associated to other unrelated malicious techniques (context), and for PM5 we need to create a rule that matches against 100 different file names (query time and high performance impact).   Following those steps in order is not necessary, and we may have missed (unintentionally) some other important steps. It usually comes to having a good understanding of the offensive technique, filtering out normal behavior while in the same time balancing detection resilience, alert context and performance impact. Also not always we have guarantees to come up with a detection for a TTP, but the ultimate goal is to capture gaps and potential opportunities of improvement. In the upcoming parts we will try to cover each step in details with some practical examples. 
__label__pos
0.93447
[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index] Re: samba mac os x y linux Ya lo envie hace un tiempo, pero es todo un placer para mi enviarlo de nuevo, yo creo que esta incompleto El s?, 09-08-2003 a las 23:37, trujo escribió: > Si lo demas no funciona ¿Por que no nos envias el fichero smb.conf? > -- > Por favor, NO utilice formatos de archivo propietarios para el > intercambio de documentos, como DOC y XLS, sino HTML, RTF, TXT, > CSV o cualquier otro que no obligue a utilizar un programa de un > fabricante concreto para tratar la información contenida en él. > SALUD > # # Sample configuration file for the Samba suite for Debian GNU/Linux. # # # This is the main Samba configuration file. You should read the # smb.conf(5) manual page in order to understand the options listed # here. Samba has a huge number of configurable options most of which # are not shown in this example # # Any line which starts with a ; (semi-colon) or a # (hash) # is a comment and is ignored. In this example we will use a # # for commentary and a ; for parts of the config file that you # may wish to enable # # NOTE: Whenever you modify this file you should run the command # "testparm" to check that you have not many any basic syntactic # errors. # #======================= Global Settings ======================= [global] ## Browsing/Identification ### # Change this to the workgroup/NT-domain name your Samba server will part of workgroup = wretched_wkg # server string is the equivalent of the NT Description field server string = %h server (Samba %v) # Windows Internet Name Serving Support Section: # WINS Support - Tells the NMBD component of Samba to enable its WINS Server ; wins support = no # WINS Server - Tells the NMBD components of Samba to be a WINS Client # Note: Samba can be either a WINS Server, or a WINS Client, but NOT both ; wins server = w.x.y.z # This will prevent nmbd to search for NetBIOS names through DNS. dns proxy = no # What naming service and in what order should we use to resolve host names # to IP addresses ; name resolve order = lmhosts host wins bcast #### Debugging/Accounting #### # This tells Samba to use a separate log file for each machine # that connects log file = /var/log/samba/log.%m # Put a capping on the size of the log files (in Kb). max log size = 1000 # If you want Samba to only log through syslog then set the following # parameter to 'yes'. ; syslog only = no # We want Samba to log a minimum amount of information to syslog. Everything # should go to /var/log/samba/log.{smbd,nmbd} instead. If you want to log # through syslog you should set the following parameter to something higher. syslog = 0 # Do something sensible when Samba crashes: mail the admin a backtrace panic action = /usr/share/samba/panic-action %d ####### Authentication ####### # "security = user" is always a good idea. This will require a Unix account # in this server for every user accessing the server. See # /usr/share/doc/samba-doc/htmldocs/ServerType.html in the samba-doc # package for details. ; security = user # You may wish to use password encryption. See the section on # 'encrypt passwords' in the smb.conf(5) manpage before enabling. encrypt passwords = true # If you are using encrypted passwords, Samba will need to know what # password database type you are using. passdb backend = tdbsam guest obey pam restrictions = yes ; guest account = nobody invalid users = root # This boolean parameter controls whether Samba attempts to sync the Unix # password with the SMB password when the encrypted SMB password in the # passdb is changed. ; unix password sync = no # For Unix password sync to work on a Debian GNU/Linux system, the following # parameters must be set (thanks to Augustin Luton <[email protected]> for # sending the correct chat script for the passwd program in Debian Potato). passwd program = /usr/bin/passwd %u passwd chat = *Enter\snew\sUNIX\spassword:* %n\n *Retype\snew\sUNIX\spassword:* %n\n . # This boolean controls whether PAM will be used for password changes # when requested by an SMB client instead of the program listed in # 'passwd program'. The default is 'no'. ; pam password change = no ########## Printing ########## # If you want to automatically load your printer list rather # than setting them up individually then you'll need this ; load printers = yes # lpr(ng) printing. You may wish to override the location of the # printcap file ; printing = bsd ; printcap name = /etc/printcap # CUPS printing. See also the cupsaddsmb(8) manpage in the # cupsys-client package. ; printing = cups ; printcap name = cups # When using [print$], root is implicitly a 'printer admin', but you can # also give this right to other users to add drivers and set printer # properties ; printer admin = @ntadmin ######## File sharing ######## # Name mangling options ; preserve case = yes ; short preserve case = yes ############ Misc ############ # Using the following line enables you to customise your configuration # on a per machine basis. The %m gets replaced with the netbios name # of the machine that is connecting ; include = /home/samba/etc/smb.conf.%m # Most people will find that this option gives better performance. # See smb.conf(5) and /usr/share/doc/samba-doc/htmldocs/speed.html # for details # You may want to add the following on a Linux system: # SO_RCVBUF=8192 SO_SNDBUF=8192 socket options = TCP_NODELAY # The following parameter is useful only if you have the linpopup package # installed. The samba maintainer and the linpopup maintainer are # working to ease installation and configuration of linpopup and samba. ; message command = /bin/sh -c '/usr/bin/linpopup "%f" "%m" %s; rm %s' & # Domain Master specifies Samba to be the Domain Master Browser. If this # machine will be configured as a BDC (a secondary logon server), you # must set this to 'no'; otherwise, the default behavior is recommended. ; domain master = auto # Some defaults for winbind (make sure you're not using the ranges # for something else.) ; idmap uid = 10000-20000 ; idmap gid = 10000-20000 ; template shell = /bin/bash #======================= Share Definitions ======================= [homes] comment = Home Directories browseable = no # By default, the home directories are exported read-only. Change next # parameter to 'yes' if you want to be able to write to them. writable = no # File creation mask is set to 0700 for security reasons. If you want to # create files with group=rw permissions, set next parameter to 0775. create mask = 0700 # Directory creation mask is set to 0700 for security reasons. If you want to # create dirs. with group=rw permissions, set next parameter to 0775. directory mask = 0700 # Un-comment the following and create the netlogon directory for Domain Logons # (you need to configure Samba to act as a domain controller too.) ;[netlogon] ; comment = Network Logon Service ; path = /home/samba/netlogon ; guest ok = yes ; writable = no ; share modes = no [printers] comment = All Printers browseable = no path = /tmp printable = yes public = no writable = no create mode = 0700 # Windows clients look for this share name as a source of downloadable # printer drivers [print$] comment = Printer Drivers path = /var/lib/samba/printers browseable = yes read only = yes guest ok = no # Uncomment to allow remote administration of Windows print drivers. # Replace 'ntadmin' with the name of the group your admin users are # members of. ; write list = root, @ntadmin # A sample share for sharing your CD-ROM with others. ;[cdrom] ; comment = Samba server's CD-ROM ; writable = no ; locking = no ; path = /cdrom ; public = yes # The next two parameters show how to auto-mount a CD-ROM when the # cdrom share is accesed. For this to work /etc/fstab must contain # an entry like this: # # /dev/scd0 /cdrom iso9660 defaults,noauto,ro,user 0 0 # # The CD-ROM gets unmounted automatically after the connection to the # # If you don't want to use auto-mounting/unmounting make sure the CD # is mounted on /cdrom # ; preexec = /bin/mount /cdrom ; postexec = /bin/umount /cdrom Reply to:
__label__pos
0.793434
View Javadoc 1 /* 2 * Copyright 2012 The Netty Project 3 * 4 * The Netty Project licenses this file to you under the Apache License, 5 * version 2.0 (the "License"); you may not use this file except in compliance 6 * with the License. You may obtain a copy of the License at: 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT 12 * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the 13 * License for the specific language governing permissions and limitations 14 * under the License. 15 */ 16 package io.netty.handler.stream; 17 18 import io.netty.buffer.Unpooled; 19 import io.netty.channel.Channel; 20 import io.netty.channel.ChannelFuture; 21 import io.netty.channel.ChannelFutureListener; 22 import io.netty.channel.ChannelHandler; 23 import io.netty.channel.ChannelHandlerAdapter; 24 import io.netty.channel.ChannelHandlerContext; 25 import io.netty.channel.ChannelPipeline; 26 import io.netty.channel.ChannelProgressivePromise; 27 import io.netty.channel.ChannelPromise; 28 import io.netty.util.ReferenceCountUtil; 29 import io.netty.util.internal.logging.InternalLogger; 30 import io.netty.util.internal.logging.InternalLoggerFactory; 31 32 import java.nio.channels.ClosedChannelException; 33 import java.util.ArrayDeque; 34 import java.util.Queue; 35 36 /** 37 * A {@link ChannelHandler} that adds support for writing a large data stream 38 * asynchronously neither spending a lot of memory nor getting 39 * {@link OutOfMemoryError}. Large data streaming such as file 40 * transfer requires complicated state management in a {@link ChannelHandler} 41 * implementation. {@link ChunkedWriteHandler} manages such complicated states 42 * so that you can send a large data stream without difficulties. 43 * <p> 44 * To use {@link ChunkedWriteHandler} in your application, you have to insert 45 * a new {@link ChunkedWriteHandler} instance: 46 * <pre> 47 * {@link ChannelPipeline} p = ...; 48 * p.addLast("streamer", <b>new {@link ChunkedWriteHandler}()</b>); 49 * p.addLast("handler", new MyHandler()); 50 * </pre> 51 * Once inserted, you can write a {@link ChunkedInput} so that the 52 * {@link ChunkedWriteHandler} can pick it up and fetch the content of the 53 * stream chunk by chunk and write the fetched chunk downstream: 54 * <pre> 55 * {@link Channel} ch = ...; 56 * ch.write(new {@link ChunkedFile}(new File("video.mkv")); 57 * </pre> 58 * 59 * <h3>Sending a stream which generates a chunk intermittently</h3> 60 * 61 * Some {@link ChunkedInput} generates a chunk on a certain event or timing. 62 * Such {@link ChunkedInput} implementation often returns {@code null} on 63 * {@link ChunkedInput#readChunk(ChannelHandlerContext)}, resulting in the indefinitely suspended 64 * transfer. To resume the transfer when a new chunk is available, you have to 65 * call {@link #resumeTransfer()}. 66 */ 67 public class ChunkedWriteHandler extends ChannelHandlerAdapter { 68 69 private static final InternalLogger logger = 70 InternalLoggerFactory.getInstance(ChunkedWriteHandler.class); 71 72 private final Queue<PendingWrite> queue = new ArrayDeque<PendingWrite>(); 73 private volatile ChannelHandlerContext ctx; 74 private PendingWrite currentWrite; 75 76 public ChunkedWriteHandler() { 77 } 78 79 /** 80 * @deprecated use {@link #ChunkedWriteHandler()} 81 */ 82 @Deprecated 83 public ChunkedWriteHandler(int maxPendingWrites) { 84 if (maxPendingWrites <= 0) { 85 throw new IllegalArgumentException( 86 "maxPendingWrites: " + maxPendingWrites + " (expected: > 0)"); 87 } 88 } 89 90 @Override 91 public void handlerAdded(ChannelHandlerContext ctx) throws Exception { 92 this.ctx = ctx; 93 } 94 95 /** 96 * Continues to fetch the chunks from the input. 97 */ 98 public void resumeTransfer() { 99 final ChannelHandlerContext ctx = this.ctx; 100 if (ctx == null) { 101 return; 102 } 103 if (ctx.executor().inEventLoop()) { 104 try { 105 doFlush(ctx); 106 } catch (Exception e) { 107 if (logger.isWarnEnabled()) { 108 logger.warn("Unexpected exception while sending chunks.", e); 109 } 110 } 111 } else { 112 // let the transfer resume on the next event loop round 113 ctx.executor().execute(new Runnable() { 114 115 @Override 116 public void run() { 117 try { 118 doFlush(ctx); 119 } catch (Exception e) { 120 if (logger.isWarnEnabled()) { 121 logger.warn("Unexpected exception while sending chunks.", e); 122 } 123 } 124 } 125 }); 126 } 127 } 128 129 @Override 130 public void write(ChannelHandlerContext ctx, Object msg, ChannelPromise promise) throws Exception { 131 queue.add(new PendingWrite(msg, promise)); 132 } 133 134 @Override 135 public void flush(ChannelHandlerContext ctx) throws Exception { 136 if (!doFlush(ctx)) { 137 // Make sure to flush at least once. 138 ctx.flush(); 139 } 140 } 141 142 @Override 143 public void channelInactive(ChannelHandlerContext ctx) throws Exception { 144 doFlush(ctx); 145 ctx.fireChannelInactive(); 146 } 147 148 @Override 149 public void channelWritabilityChanged(ChannelHandlerContext ctx) throws Exception { 150 if (ctx.channel().isWritable()) { 151 // channel is writable again try to continue flushing 152 doFlush(ctx); 153 } 154 ctx.fireChannelWritabilityChanged(); 155 } 156 157 private void discard(Throwable cause) { 158 for (;;) { 159 PendingWrite currentWrite = this.currentWrite; 160 161 if (this.currentWrite == null) { 162 currentWrite = queue.poll(); 163 } else { 164 this.currentWrite = null; 165 } 166 167 if (currentWrite == null) { 168 break; 169 } 170 Object message = currentWrite.msg; 171 if (message instanceof ChunkedInput) { 172 ChunkedInput<?> in = (ChunkedInput<?>) message; 173 try { 174 if (!in.isEndOfInput()) { 175 if (cause == null) { 176 cause = new ClosedChannelException(); 177 } 178 currentWrite.fail(cause); 179 } else { 180 currentWrite.success(in.length()); 181 } 182 closeInput(in); 183 } catch (Exception e) { 184 currentWrite.fail(e); 185 logger.warn(ChunkedInput.class.getSimpleName() + ".isEndOfInput() failed", e); 186 closeInput(in); 187 } 188 } else { 189 if (cause == null) { 190 cause = new ClosedChannelException(); 191 } 192 currentWrite.fail(cause); 193 } 194 } 195 } 196 197 private boolean doFlush(final ChannelHandlerContext ctx) throws Exception { 198 final Channel channel = ctx.channel(); 199 if (!channel.isActive()) { 200 discard(null); 201 return false; 202 } 203 204 boolean flushed = false; 205 while (channel.isWritable()) { 206 if (currentWrite == null) { 207 currentWrite = queue.poll(); 208 } 209 210 if (currentWrite == null) { 211 break; 212 } 213 final PendingWrite currentWrite = this.currentWrite; 214 final Object pendingMessage = currentWrite.msg; 215 216 if (pendingMessage instanceof ChunkedInput) { 217 final ChunkedInput<?> chunks = (ChunkedInput<?>) pendingMessage; 218 boolean endOfInput; 219 boolean suspend; 220 Object message = null; 221 try { 222 message = chunks.readChunk(ctx); 223 endOfInput = chunks.isEndOfInput(); 224 225 if (message == null) { 226 // No need to suspend when reached at the end. 227 suspend = !endOfInput; 228 } else { 229 suspend = false; 230 } 231 } catch (final Throwable t) { 232 this.currentWrite = null; 233 234 if (message != null) { 235 ReferenceCountUtil.release(message); 236 } 237 238 currentWrite.fail(t); 239 closeInput(chunks); 240 break; 241 } 242 243 if (suspend) { 244 // ChunkedInput.nextChunk() returned null and it has 245 // not reached at the end of input. Let's wait until 246 // more chunks arrive. Nothing to write or notify. 247 break; 248 } 249 250 if (message == null) { 251 // If message is null write an empty ByteBuf. 252 // See https://github.com/netty/netty/issues/1671 253 message = Unpooled.EMPTY_BUFFER; 254 } 255 256 ChannelFuture f = ctx.write(message); 257 if (endOfInput) { 258 this.currentWrite = null; 259 260 // Register a listener which will close the input once the write is complete. 261 // This is needed because the Chunk may have some resource bound that can not 262 // be closed before its not written. 263 // 264 // See https://github.com/netty/netty/issues/303 265 f.addListener(new ChannelFutureListener() { 266 @Override 267 public void operationComplete(ChannelFuture future) throws Exception { 268 currentWrite.progress(chunks.progress(), chunks.length()); 269 currentWrite.success(chunks.length()); 270 closeInput(chunks); 271 } 272 }); 273 } else if (channel.isWritable()) { 274 f.addListener(new ChannelFutureListener() { 275 @Override 276 public void operationComplete(ChannelFuture future) throws Exception { 277 if (!future.isSuccess()) { 278 closeInput((ChunkedInput<?>) pendingMessage); 279 currentWrite.fail(future.cause()); 280 } else { 281 currentWrite.progress(chunks.progress(), chunks.length()); 282 } 283 } 284 }); 285 } else { 286 f.addListener(new ChannelFutureListener() { 287 @Override 288 public void operationComplete(ChannelFuture future) throws Exception { 289 if (!future.isSuccess()) { 290 closeInput((ChunkedInput<?>) pendingMessage); 291 currentWrite.fail(future.cause()); 292 } else { 293 currentWrite.progress(chunks.progress(), chunks.length()); 294 if (channel.isWritable()) { 295 resumeTransfer(); 296 } 297 } 298 } 299 }); 300 } 301 } else { 302 ctx.write(pendingMessage, currentWrite.promise); 303 this.currentWrite = null; 304 } 305 306 // Always need to flush 307 ctx.flush(); 308 flushed = true; 309 310 if (!channel.isActive()) { 311 discard(new ClosedChannelException()); 312 break; 313 } 314 } 315 316 return flushed; 317 } 318 319 static void closeInput(ChunkedInput<?> chunks) { 320 try { 321 chunks.close(); 322 } catch (Throwable t) { 323 if (logger.isWarnEnabled()) { 324 logger.warn("Failed to close a chunked input.", t); 325 } 326 } 327 } 328 329 private static final class PendingWrite { 330 final Object msg; 331 final ChannelPromise promise; 332 333 PendingWrite(Object msg, ChannelPromise promise) { 334 this.msg = msg; 335 this.promise = promise; 336 } 337 338 void fail(Throwable cause) { 339 ReferenceCountUtil.release(msg); 340 promise.tryFailure(cause); 341 } 342 343 void success(long total) { 344 if (promise.isDone()) { 345 // No need to notify the progress or fulfill the promise because it's done already. 346 return; 347 } 348 349 if (promise instanceof ChannelProgressivePromise) { 350 // Now we know what the total is. 351 ((ChannelProgressivePromise) promise).tryProgress(total, total); 352 } 353 354 promise.trySuccess(); 355 } 356 357 void progress(long progress, long total) { 358 if (promise instanceof ChannelProgressivePromise) { 359 ((ChannelProgressivePromise) promise).tryProgress(progress, total); 360 } 361 } 362 } 363 }
__label__pos
0.999998
user2575373 user2575373 - 8 months ago 45 Android Question Android Change language of speech to text to Japanese not working Hi I'm making a Japanese learning app for android. One of the features is to speak to the app in Japanese to check if you are saying words correctly. I got it working with promptSpeechInput but I did not like the ui getting in the way so I decided to go another rout and have my Fragment implement RecognitionListener . For some reason now Japanese is not working and it shows English words. I'm not sure what's wrong. Here is my code for my Speech Fragment public class SpeechFragment extends Fragment implements RecognitionListener { private TextView textViewInput; private ToggleButton buttonSpeak; private SpeechRecognizer speech = null; private Intent recognizerIntent; public SpeechFragment() { // Required empty public constructor } @Override public View onCreateView(LayoutInflater inflater, ViewGroup container, Bundle savedInstanceState) { // Inflate the layout for this fragment View view = inflater.inflate(R.layout.fragment_speech, container, false); speech = SpeechRecognizer.createSpeechRecognizer(this.getContext()); speech.setRecognitionListener(this); recognizerIntent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,RecognizerIntent.LANGUAGE_MODEL_FREE_FORM); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE, Locale.JAPANESE); recognizerIntent.putExtra(RecognizerIntent.EXTRA_MAX_RESULTS, 3); textViewInput = (TextView) view.findViewById(R.id.textViewInput); buttonSpeak = (ToggleButton) view.findViewById(R.id.buttonSpeak); buttonSpeak.setOnCheckedChangeListener(new CompoundButton.OnCheckedChangeListener() { @Override public void onCheckedChanged(CompoundButton buttonView, boolean isChecked) { if (isChecked) { speech.startListening(recognizerIntent); } else { speech.stopListening(); } } }); return view; } @Override public void onResume() { super.onResume(); } @Override public void onPause() { super.onPause(); if (speech != null) { speech.destroy(); } } @Override public void onBeginningOfSpeech() { textViewInput.setText("Speak"); } @Override public void onBufferReceived(byte[] buffer) { } @Override public void onEndOfSpeech() { buttonSpeak.setChecked(false); } @Override public void onError(int errorCode) { String errorMessage = getErrorText(errorCode); textViewInput.setText(errorMessage); buttonSpeak.setChecked(false); } @Override public void onEvent(int arg0, Bundle arg1) { } @Override public void onPartialResults(Bundle arg0) { } @Override public void onReadyForSpeech(Bundle arg0) { } @Override public void onResults(Bundle results) { ArrayList<String> matches = results .getStringArrayList(SpeechRecognizer.RESULTS_RECOGNITION); String text = ""; for (String result : matches) text += result + "\n"; textViewInput.setText(text); } @Override public void onRmsChanged(float rmsdB) { } public static String getErrorText(int errorCode) { String message; switch (errorCode) { case SpeechRecognizer.ERROR_AUDIO: message = "Audio recording error"; break; case SpeechRecognizer.ERROR_CLIENT: message = "Client side error"; break; case SpeechRecognizer.ERROR_INSUFFICIENT_PERMISSIONS: message = "Insufficient permissions"; break; case SpeechRecognizer.ERROR_NETWORK: message = "Network error"; break; case SpeechRecognizer.ERROR_NETWORK_TIMEOUT: message = "Network timeout"; break; case SpeechRecognizer.ERROR_NO_MATCH: message = "No match"; break; case SpeechRecognizer.ERROR_RECOGNIZER_BUSY: message = "RecognitionService busy"; break; case SpeechRecognizer.ERROR_SERVER: message = "error from server"; break; case SpeechRecognizer.ERROR_SPEECH_TIMEOUT: message = "No speech input"; break; default: message = "Didn't understand, please try again."; break; } return message; } } Answer Try like this Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,RecognizerIntent.LANGUAGE_MODEL_FREE_FORM); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE, Locale.JAPANESE); intent.putExtra(RecognizerIntent.EXTRA_MAX_RESULTS, 3); getActivity().startActivityForResult(intent,requestCode); After that override the onActivityResult() method in your activity file(where you called the fragment) @Override protected void onActivityResult(int requestCode, int resultCode, Intent data) { super.onActivityResult(requestCode, resultCode, data); ArrayList<String> words=data.getExtras().getStringArrayList(RecognizerIntent.EXTRA_RESULTS); //Here you can get the spoken words }
__label__pos
0.999862
10 thoughts on “StackWalker moved to CodePlex 1. frank28_nfls very nice! I personally discovered your software about 1 year ago, and was using it in my several projects, and found that it really helps! But I always wondering why you didn’t use any VCS to manage the source code and release. Now finally, lol btw, is there any update in Codeplex since RC9? thanks again for your work, anyway~ Reply 2. Alexander Hello, I need you help I use this code to process exception I found some times the result of StackWalker doesn’t like vs2005’s result? nRetCode = StackWalk64( IMAGE_FILE_MACHINE_I386, m_hProcess, m_hThread, &m_Stackframe, &m_Context, ReadProcessMemory, SymFunctionTableAccess64, SymGetModuleBase64, NULL ); how to explain it? Wish you EMail Reply 3. Deepak Hi Jochen, I am using your tool LeakFinder from codeproject on Win32 and Win64. It is working perefectly fine when run on Win32 and I am getting following output (as expected) : ######## CRT-Leaks ######## ————— Key: 63, 32 bytes ——— d:\tempproj\leakfinder-rc5\main.cpp (44): main f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (327): __tmainCRTStartup f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (196): mainCRTStartup 7C817067 (kernel32): (filename not available): RegisterWaitForInputIdle ————— Key: 64, 10 bytes ——— d:\tempproj\leakfinder-rc5\main.cpp (18): Func5 d:\tempproj\leakfinder-rc5\main.cpp (23): Func4 d:\tempproj\leakfinder-rc5\main.cpp (27): Func3 d:\tempproj\leakfinder-rc5\main.cpp (31): Func2 d:\tempproj\leakfinder-rc5\main.cpp (35): Func1 d:\tempproj\leakfinder-rc5\main.cpp (47): main f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (327): __tmainCRTStartup f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (196): mainCRTStartup 7C817067 (kernel32): (filename not available): RegisterWaitForInputIdle ######## COM-Leaks ######## ————— Key: 00178CA8, 999 bytes ——— 775833D5 (ole32): (filename not available): CoRevokeMallocSpy 774FD073 (ole32): (filename not available): CoTaskMemAlloc d:\tempproj\leakfinder-rc5\main.cpp (39): DoCoMalloc d:\tempproj\leakfinder-rc5\main.cpp (46): main f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (327): __tmainCRTStartup f:\sp\vctools\crt_bld\self_x86\crt\src\crt0.c (196): mainCRTStartup 7C817067 (kernel32): (filename not available): RegisterWaitForInputIdle ‘LeakFinder_VC5.exe’: Unloaded ‘C:\WINDOWS\system32\dbghelp.dll’ The program ‘[5256] LeakFinder_VC5.exe: Native’ has exited with code 0 (0x0). However, when run on Win64 output is as under : ######## CRT-Leaks ######## ————— Key: 65, 48 bytes ——— ————— Key: 66, 10 bytes ——— ######## COM-Leaks ######## ————— Key: 000000000024A600, 999 bytes ——— f:\sp\vctools\crt_bld\self_64_amd64\crt\src\mlock.c (349): _lock 0000000140082670 (LeakFinder_VC5): (filename not available): `dynamic atexit destructor for ‘zzz_LeakFinder” f:\sp\vctools\crt_bld\self_64_amd64\crt\src\crt0dat.c (553): doexit f:\sp\vctools\crt_bld\self_64_amd64\crt\src\crt0dat.c (399): exit f:\sp\vctools\crt_bld\self_64_amd64\crt\src\crt0.c (333): __tmainCRTStartup f:\sp\vctools\crt_bld\self_64_amd64\crt\src\crt0.c (196): mainCRTStartup 000000007735F56D (kernel32): (filename not available): BaseThreadInitThunk 0000000077593281 (ntdll): (filename not available): RtlUserThreadStart ‘LeakFinder_VC5.exe’: Unloaded ‘C:\Windows\System32\dbghelp.dll’ The thread ‘Win64 Thread’ (0x964) has exited with code 0 (0x0). The program ‘[3400] LeakFinder_VC5.exe: Native’ has exited with code 0 (0x0). exited with code 0 (0x0). On debugging on Win64 I found out that ShowCallstack() in StackWalker.cpp wouldn’t get enogh frames. I am not able to figure out what could be the reason. Could you please provide some clue and let me know if you any further information from me. Also, if you could tell why Rbp in context on Win64 is coming out as 0. Thanks in advance, Deepak. Reply 4. Pimsock Hi, Thanks for making Leak Finder available. I am building a little DLL wrapper around it so that I can (hopefully) use it in a VB6-based project. I found that I had to make a couple small changes to the source code to get it to compile (under VC++ 6), and I thought I should let you know about them: In StackWalker.h, #if _MSC_VER < 1400, you do this: #define strcpy_s strcpy #define strncpy_s strncpy To get it to compile, I had to change it to this: #define strcpy_s(a, b, c) strcpy(a, c) #define strncpy_s(a, b, c, d) strncpy(a, c, d) I hope this helps. If I’m misunderstanding something, please let me know. In any case, once again, thanks for making this. Reply 5. jkalmbach Post author Just take a look at the main-trunk in the Source!!! It is already solved 😉 I need to make a new release… Reply 6. Richard Just downloaded StackWalker and tried building on Visual Studio 6. I appreciate that you still provide project files to build it on such an old compiler, but there is a fault. For _MSC_VER < 1400 you provide local #defines for some of the _s functions. VC6 (_MSC_VER = 1200) uses these. However, you then call strcpy_s with three parameters and strncpy_s with four. strcpy only takes two and strncpy only takes three, so the #defines actually need to be : #define strcpy_s( a, b, c ) strcpy( a, c ) #define strncpy_s( a, b, c, d ) strncpy( a, c, b ) Regards, Richard. Reply 7. jkalmbach Post author You should just downlod the source-code directly from the VCS… this is fixed since 5 months…. I need to make a new download… Reply Leave a Reply Your email address will not be published. Required fields are marked * Captcha *
__label__pos
0.779535
How to format custom proc macro? Sorry if anyone has asked about this. I mean, is there a way to make a crate like macro crate make custom proc macro to be formatted by rustfmt? If you are wanting to turn a proc_macro::TokenStream into formatted source code (e.g. so you can generate code and save it to a file), you can always use std::process::Command to invoke rustfmt and pass it your token_stream.to_string() as stdin. Otherwise, the prettyplease crate is nice when you want "good enough" formatting to make your generated code readable. 1 Like I'm sorry I misunderstood before. What I want is to format custom syntax inside macro braces, like custom_lambda! { x => { 10 } } prettyplease seems unable to format complex custom syntax. I wonder if there is a way I could "guide" rustfmt to format my custom syntax? Not as far as I know. Thanks
__label__pos
0.999963
Testing CRA with Moralis React Hello Moralis Community I am trying to test react application and react moralis using testing-library.The problem is the react components wrapped inside MoralisProvider of react-moralis interacting with moralis database. Could anyone please help? Can we see your code and how you’re importing and using Moralis / react-moralis? What command did you run to get to that error? These are the instructions on using react-moralis: Connect with React - Moralis 1 Like import { render, screen } from "@testing-library/react"; import CreateAuction from "pages/assets/single/CreateAuction"; import { MoralisProvider } from "react-moralis"; import { createMemoryHistory } from "history"; import { BrowserRouter as Router } from "react-router-dom"; window.matchMedia = window.matchMedia || function () { return { addListener: jest.fn(), removeListener: jest.fn(), }; }; // const Moralis = require('moralis/node') describe("Creatin Auction", () => { test("should show auction form", async () => { const history = createMemoryHistory(); const route = "/collection/alphabet/asset/yCog7AaY2IMQCFVtBRmYSBp6/auction/createEdit"; render( <MoralisProvider appId={process.env.REACT_APP_MORALIS_APP} serverUrl={process.env.REACT_APP_MORALIS_SERVER} > <Router location={history.location} navigator={history}> <CreateAuction /> </Router> </MoralisProvider> ); expect(screen.getByText(/auction/i)).toBeInTheDocument(); }); }); I run npm run test No window.ethereum found warning will show if there is no crypto wallet extension installed in the browser. is that the error you are referring to? The error I do mention is Sorry for unclear question :pray: Did you import the moralis sdk with the cdn scripts or did you install the moralis npm packages? Is that error breaking the app? Yes already installed, it doesn’t work only when I run test. import { MoralisProvider } from "react-moralis"; render( <MoralisProvider appId={process.env.REACT_APP_MORALIS_APP} serverUrl={process.env.REACT_APP_MORALIS_SERVER} > <Router location={history.location} navigator={history}> <CreateAuction /> </Router> </MoralisProvider> ); did you try adding require('parse/node') in your js file, to see if that changes the anything in the test also try adding require('moralis/node') if parse do not work Thanks @johnversus, I will try and let you know once it is done. @johnversus The issue is the still same shown in the attached image . ohh…This solution seems to have worked for some users. Not sure if there is any other fix. Are you able to share a repo of your project so the issue can be replicated? @alex Here is the repository to replicate issue. The test is passing, I feel these warnings can be ignored. Seems to be an issue with Parse (which Moralis is based on) when used with these Jest tests. Yes @alex, When I deal with database, Parse(based by Moralis) cause issue.I could not get any records from database even though there is. import React from "react"; import { useMoralisQuery } from "react-moralis"; const Home = () => { // Even though there is record related to bdJhoxg2dXAAbff6IBtx9JOH, no data is returned. const { data: collections } = useMoralisQuery("collection", (query) => query.equalTo("objectId", "bdJhoxg2dXAAbff6IBtx9JOH") ); console.log("collections: ", collections) return(<h3>Home Page</h3>); } export default Home; Your code should be working. Make sure the Class name of collection is correct. Also it is case sensitive. Also you should use a useEffect to log (in general). useEffect(() => { console.log('collections: ', collections); }, [collections]); @alex still the same :cry:
__label__pos
0.997022
20 Is there an established algorithm for finding redundant edges in a graph? For example, I'd like to find that a->d and a->e are redundant, and then get rid of them, like this: alt text => alt text Edit: Strilanc was nice enough to read my mind for me. "Redundant" was too strong of a word, since in the example above, neither a->b or a->c is considered redundant, but a->d is. • Can we instead consider B--->C to be redundant? – Zach Scrivena Feb 4 '09 at 7:52 • Does redundant mean "an edge X->Y is redundant if there is a non edge path from X to Y" or are you simply looking for a spanning tree ? – David Lehavi Feb 4 '09 at 8:24 • @Zach: No, B->C is not redundant, because if it is removed there is no path in the resulting graph from B to C. – ShreevatsaR Feb 4 '09 at 14:52 • Sorry to have made your comments incorrect, but I've updated with a better example. – Ryan Fox Feb 5 '09 at 0:55 26 You want to compute the smallest graph which maintains vertex reachability. This is called the transitive reduction of a graph. The wikipedia article should get you started down the right road. • Thanks, that's exactly what I'm looking for. The Wikipedia article even mentions 'tred' for Graphviz, which is especially handy, since that's what I'm working with. – Ryan Fox Feb 5 '09 at 0:50 • There it is. I could see the transitive closure was close. – Charlie Martin Feb 5 '09 at 2:09 1 Check this: Minimum Spanning Tree • If all he needs is get rid of redundant edges, he doesn't have to worry about a minimum spanning tree. Any ole spanning tree will do. – Frederick The Fool Feb 4 '09 at 6:56 • 1 Also remember "Given a connected, undirected graph, a spanning tree of that graph is a subgraph which is a tree and connects all the vertices together." Yet his graph isn't undirected. – Robert Gould Feb 4 '09 at 7:10 1 Several ways to attack this, but first you're going to need to define the problem a little more precisely. First, the graph you have here is acyclic and directed: will this always be true? Next, you need to define what you mean by a "redundant edge". In this case, you start with a graph which has two paths a->c: one via b and one direct one. From this I infer that by "redundant" you mean something like this. Let G=< V, E > be a graph, with V the set of vertices and E ⊆ V×V the set of edges. It kinda looks like you're defining all edges from vi to vj shorter than the longest edge as "redundant". So the easiest thing would be to use depth first search, enumerate the paths, and when you find a new one that's longer, save it as the best candidate. I can't imagine what you want it for, though. Can you tell? 1 A sub-graph of a given graph which contains no "redundant edges" is called a 'spanning tree' of that graph. For any given graph, multiple spanning trees are possible. So, in order to get rid of redundant edges, all you need to do is find any one spanning tree of your graph. You can use any depth-first-search or breadth-first-search algorithm and continue searching till you have visited every vertex in the graph. • It is late, but is what he describes really a spanning tree? – Charlie Martin Feb 4 '09 at 6:56 • Yes. He wants to have a sub-graph which contains all the vertices of the original graph with only one way to reach from one vertex to another. That's exactly what a spanning tree is. – Frederick The Fool Feb 4 '09 at 7:04 • 2 No, even in the reduced graph there are 2 ways to go from a to d. – xuanji Feb 20 '14 at 7:01 0 I had a similar problem and ended up solving it this way: My data structure is made of dependends dictionary, from a node id to a list of nodes that depend on it (ie. its followers in the DAG). Note it works only for a DAG - that is directed, acyclic graph. I haven't calculated the exact complexity of it, but it swallowed my graph of several thousands in a split second. _transitive_closure_cache = {} def transitive_closure(self, node_id): """returns a set of all the nodes (ids) reachable from given node(_id)""" global _transitive_closure_cache if node_id in _transitive_closure_cache: return _transitive_closure_cache[node_id] c = set(d.id for d in dependents[node_id]) for d in dependents[node_id]: c.update(transitive_closure(d.id)) # for the non-pythonists - update is update self to Union result _transitive_closure_cache[node_id] = c return c def can_reduce(self, source_id, dest_id): """returns True if the edge (source_id, dest_id) is redundant (can reach from source_id to dest_id without it)""" for d in dependents[source_id]: if d.id == dest_id: continue if dest_id in transitive_closure(d.id): return True # the dest node can be reached in a less direct path, then this link is redundant return False # Reduce redundant edges: for node in nodes: dependents[node.id] = [d for d in dependents[node.id] if not can_reduce(node.id, d.id)] • just wanted to comment on previous answers - Reducing the redundant edges is NOT the same as Spanning Tree, not even the same as Minimum Spanning Tree. And if one path from A to B is longer than another path from A to B it doesn't mean anything about what edges (if any) are redundant. In his example above you can construct a spanning tree without edge a->b but it is not redundant. – Iftah Jun 29 '11 at 7:37 -1 I think the easiest way to do that, actually imagine how it would look in the real work, imagine if you have joints, Like (A->B)(B->C)(A->C), imagine if distance between near graphs is equals 1, so (A->B) = 1, (B->C) = 1, (A->C) = 2. So you can remove joint (A->C). In other words, minimize. This is just my idea how I would think about it at start. There are various articles and sources on the net, you can look at them and go deeper. Resources, that Will help you: Algorithm for Removing Redundant Edges in the Dual Graph of a Non-Binary CSP Graph Data Structure and Basic Graph Algorithms Google Books, On finding minimal two connected Subgraphs Graph Reduction Redundant trees for preplanned recovery in arbitraryvertex-redundant or edge-redundant graphs Your Answer By clicking "Post Your Answer", you acknowledge that you have read our updated terms of service, privacy policy and cookie policy, and that your continued use of the website is subject to these policies. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.850205
Link Search Menu Expand Document ROLite 0.1.0 This first draft of ROLite RO-Crate is more of an example playground to initiate discussion. This example is based on DataCrate and BioSchemas, bringing in Research Object vocabularies where appropriate. Table of contents 1. Explaining by example: a simple dataset 1. Preamble 2. Dataset description 3. File attributions 1. Multiple people 4. Next step: Workflow research object Explaining by example: a simple dataset simple-dataset-0.1.0 shows an example of a fairly minimal dataset, which can live in GitHub, be distributed as a ZIP archive, posted to Zenodo etc. Root folder of RO The files of the dataset are under data/, which mean this folder structure could later be upgraded to be a BagIt archive (RFC8493). data/ folder Here are our files described in this ROLite example: • repository-sizes.tsv and repository-sizes.ods - our main dataset in raw tab-separated format and organized as a spreadsheet • repository-sizes-chart.png - a chart from the spreadsheet • logs/ - various log files To describe these resources and their relations, ROLite has a single file manifest.jsonld The JSON file is split into a couple of logical parts that in total make out the schema.org description in JSON-LD of both the dataset and the Research Object. Preamble { "@context": ["http://schema.org", "https://w3id.org/bundle/context"], "@type": ["ro:ResearchObject", "Dataset"], "@id": ".", "identifier": "976287d8886dbc477272976658598b1764493a3e", } The @context means we are using JSON keys from both schema.org and Research Object. The default vocabulary is schema.org, but the RO vocabularies can override (shown using the ro: etc below) TODO: Alternative bundle-independent context for research objects The @type shows we are dealing with a Research Object, in particular one that is also a http://schema.org/Dataset. The @id shows that the Research Object is the current directory. Note: other RO serializations have manifest.json in subdirectories like .ro or metadata/; here it is in the root folder to avoid relative climbs like ../data/. While the current directory is a good root, it is also a relative identifier, meaning that any copy or publication of this RO will effectively have a different @id. This is on purpose, in case they get edited. However it is always good practice to assign a more durable and unique identifier (see Identifiers for the 21st century), which can be provided using the http://schema.org/identifier property. The value should ideally be a URI (e.g. https://doi.org/10.17632/6wtpgr3kbj.1), but in this example use a local (but unique) git commit ID. Dataset description "name": "Dataset of repository sizes in CWL Viewer", "description": "This is a simple dataset showing the size in bytes of the repositories examined by https://view.commonwl.org/ sinze September 2018", "keywords": "CWL, repository, example", "temporalCoverage": "2018-09/2019-01", "contentLocation": { "@id": "http://sws.geonames.org/9884115/", "name": "Kilburn Building, Manchester, UK" } File attributions A key aspect of Research Objects is attribution. The manifest makes it possible to attribute different people for different files in the research object, but also for making the collection of the reserarch object: { "@type": ["ro:ResearchObject", "Dataset"], "@id": ".", "creator": { "@type": "Person", "@id": "https://orcid.org/0000-0001-9842-9718", "name": "Stian Soiland-Reyes", "email": "[email protected]" } } Any attributes from Person may be provided, minimally name. Consider that names across the world varies significantly, don’t assume anything about the order of parts of their name. Thus properties like familyName or givenName should only be listed if provided as such by the person. As many people have the same name, and a person may have multiple email addresses (even multiple names!), attributions should also provide a ORCID identifier using @id to disambiguate people when possible. Tip: Always providing a @id on a Person avoids duplicating their other attributes on later mentions of the same person. In ROLite, if a file does not list a creator, and is within the Research Object’s folders, it’s creator can reasonably be assumed to be the creator of the containing research object. However, where appropriate, the Research Object manifest allows overriding with more precise attribution per resource. For instance, if this chart was created by Alice: { "@id": "data/repository-sizes-chart.png", "@type": ["ImageObject"], "description": "Line chart of repository sizes. Exported from spreadsheet." "creator": { "@type": "Person", "name": "Alice W Land", } } Multiple people In some cases there are multiple people involved in making the creative work, even if they didn’t physically save the file (e.g. collaborative editing). In this case it can be useful to distinguish their roles using http://schema.org/author - for instance if Thomas only re-typed the file, which data where collected on paper by Baudoin: { "@id": "data/repository-sizes.tsv" "creator": { "@type": "Person", "name": "Thomas" }, "author": { "@type": "Person", "name": "Baudoin" } } Other contributors, (e.g. they fixed something in the spreadsheet) can be indicated with contributor. All creator, author and contributor properties MAY be JSON lists [] to support multiple people - however as classical files are usually (last) saved by a single person it is expected that creator is a single object. Tip: The creator is expected to be present in the author or contributor list as they will have contributed to the creative work (the content of the file) rather than just its bytes. Exception: typists, converters, archivists. If neither author or contributor are provided, then the creator can be reasonably assumed to be the implied author. Sometimes it is useful to indicate the software and software version used to create a file. This can be indicated as a http://schema.org/SoftwareApplication with the key pav:createdWith in conjunction with the creator of the person that used the software. { "@id": "data/repository-sizes.ods" "creator": { "@type": "Person", "name": "Thomas" } "pav:createdWith": { "@type": "SoftwareApplication", "name": "Apache OpenOffice", "softwareVersion": "4.1.6", "url": "https://www.openoffice.org/" } } Tip: If the file was created automatically without a person controlling it, then the SoftwareApplication would instead be the creator. Next step: Workflow research object The example workflow-0.1.0 shows a more complex example of a Research Object containing a workflow. In this example there is both workflow/ tools/ and test/ - but no data/ payload. This example shows how RO Lite can be used with a non-trivial pre-existing directory structure. In particular the workflow/ folder is actually a KNIME workspace folder that contain a KNIME workflow and its contained input and output data in the internal format used by KNIME. The Research Object manifest.jsonld only highlight a couple of these files: "aggregates": [ { "@id": "workflow/workflow.knime", "@type": ["SoftwareSourceCode", "wfdesc:Workflow"], "name": "RetroPath 2.0 Knime workflow", "description": "KNIME implementation of RetroPath2.0 workflow", "creator": { "@type": "Person", "name": "Thomas Duigou", "email": "[email protected]" }, "programmingLanguage": { "name": "KNIME Analytics Platform", "url": "https://www.knime.com/knime-software/knime-analytics-platform", "version": "3.6" }, "url": "https://www.myexperiment.org/workflows/4987" } # ... As before, the http://schema.org/creator indicates who created the KNIME workflow file. The workflow/workflow.knime file is in KNIME’s internal XML format, indicated here as both as a Research Object Workflow as well as a http://schema.org/SoftwareSourceCode. Tip: To enable schema.org parsing, e.g. by Google’s Structured Data Testing tool, a corresponding schema.org type must be provided. http://schema.org/CreativeWork or http://schema.org/Thing are useful fallbacks. Here http://schema.org/programmingLanguage is used informally to indicate the workflow system used. Providing an identifier with @id mau enable further programmatic understanding of the execution environment. In addition the workflow/ folder, the KNIME workspace that captures the internal workflow state, is listed as an ExampleRun. As this is hard to call this programmatic state a creative work it is just indicated as a https://schema.org/Thing as a fallback, here enabling http://schema.org/description { "@id": "workflow/", "@type": ["Thing", "roterms:ExampleRun"], "description": "KNIME workspace after executing RetroPath2.0 workflow" }
__label__pos
0.751429
Dilogarithm From Wikipedia, the free encyclopedia (Redirected from Spence's function) The dilogarithm along the real axis In mathematics, the dilogarithm (or Spence's function), denoted as Li2(z), is a particular case of the polylogarithm. Two related special functions are referred to as Spence's function, the dilogarithm itself: and its reflection. For |z| < 1, an infinite series also applies (the integral definition constitutes its analytical extension to the complex plane): Alternatively, the dilogarithm function is sometimes defined as In hyperbolic geometry the dilogarithm can be used to compute the volume of an ideal simplex. Specifically, a simplex whose vertices have cross ratio z has hyperbolic volume The function D(z) is sometimes called the Bloch-Wigner function.[1] Lobachevsky's function and Clausen's function are closely related functions. William Spence, after whom the function was named by early writers in the field, was a Scottish mathematician working in the early nineteenth century.[2] He was at school with John Galt,[3] who later wrote a biographical essay on Spence. Analytic structure[edit] Using the former definition above, the dilogarithm function is analytic everywhere on the complex plane except at , where it has a logarithmic branch point. The standard choice of branch cut is along the positive real axis . However, the function is continuous at the branch point and takes on the value . Identities[edit] [4] [5] [4] [5] [4] Particular value identities[edit] [5] [5] [5] [5] [5] Special values[edit] where is the Riemann zeta function. In particle physics[edit] Spence's Function is commonly encountered in particle physics while calculating radiative corrections. In this context, the function is often defined with an absolute value inside the logarithm: See also[edit] Notes[edit] 1. ^ Zagier p. 10 2. ^ "William Spence - Biography". 3. ^ "Biography – GALT, JOHN – Volume VII (1836-1850) – Dictionary of Canadian Biography". 4. ^ a b c Zagier 5. ^ a b c d e f g Weisstein, Eric W. "Dilogarithm". MathWorld. References[edit] Further reading[edit] External links[edit]
__label__pos
0.97569
in reply to Re^2: Create Arrays On-the-Fly in thread Create Arrays On-the-Fly Of course, if you're always looking up data by MAC, it could make more sense to structure it as a giant hash, keyed by the MAC, of hash (refs) of the data. That would make looking up particular MAC's faster and a bit cleaner. It may be easier to read too, depending on your own preferences. #!/usr/bin/env perl5 use strict; use warnings; my %clients; my $curmac; while(<DATA>) { chomp; # Assume a format of "Type Data" my ($type, $data) = split / /; # If it's a MAC address, assume later lines refer to this MAC if($type eq "MAC") { $curmac = $data; next; } # RSSI and SNR just get saved $clients{$curmac}{RSSI} = $data if $type eq "RSSI"; $clients{$curmac}{SNR} = $data if $type eq "SNR"; } # Show it use Data::Dumper; print Dumper \%clients; # Find a given entry sub showmacsnr { my $mac = shift; my $ent = $clients{$mac}; if($ent) { print "MAC $mac has SNR $ent->{SNR}\n"; } else { print "Can't find MAC $mac\n"; } } showmacsnr("ba:98:76:54:32:10"); showmacsnr("thi:is:is:nt:re:al"); # Sample data __DATA__ MAC 01:23:45:67:89:ab RSSI 12 SNR 18 MAC ba:98:76:54:32:10 RSSI 7 SNR 3 % ./tst.pl $VAR1 = { 'ba:98:76:54:32:10' => { 'SNR' => '3', 'RSSI' => '7' }, '01:23:45:67:89:ab' => { 'SNR' => '18', 'RSSI' => '12' } }; MAC ba:98:76:54:32:10 has SNR 3 Can't find MAC thi:is:is:nt:re:al
__label__pos
0.812789
Presentation is loading. Please wait. Presentation is loading. Please wait. 1 VR4121 64-BIT MICROPROCESSOR โดย นางสาว พิลาวัณย์ พลับรู้การ นางสาว เพ็ญพรรณ อัศวนพเกียรติ Similar presentations Presentation on theme: "1 VR4121 64-BIT MICROPROCESSOR โดย นางสาว พิลาวัณย์ พลับรู้การ นางสาว เพ็ญพรรณ อัศวนพเกียรติ"— Presentation transcript: 1 1 VR BIT MICROPROCESSOR โดย นางสาว พิลาวัณย์ พลับรู้การ นางสาว เพ็ญพรรณ อัศวนพเกียรติ 2 Agenda Overview of VR4121 VR4121 pipeline 3 Overview of VR4121 4 Introduction to VR bit microprocessor in VR series manufactured by NEC. Designed for high performance handheld portable computing devices. Designed around the MIPS RISC architecture developed by MIPS technologies. 5 Block Diagram HSP RTC DSU ICU PMU CMU DCU DMAU GIU KIU LED AIU PIU SIU FIR D/A A/D PLL VR4120 CPU core 131/168 MHz BUS CONTROL UNIT OSB I/O keyboard RS232 driver LED Touch panel LCD module PC card ROM/ Flash EDO/ SDRAM BUS CONTROL UNIT HSP RTC DSU ICU PMU CMU DCU DMAU GIU KIU LED AIU PIU SIU FIR 6 VR4120 CPU core CPU Coprocessor 0 Instruction cache Data cache CPU bus interface Clock generator 7 CPU registers R0 = 0 R1 R2 R R31 = Link Address 63 0 HI 63 0 LO 63 0 PC 63 0 General purpose registers Multiply / divide registers Program counter 8 Data format and addressing Double word 64 bits Word 32 bits Half word 16 bits Byte 8 bits VR4121 supports the little-endian order only 9 CPU instruction set Load and store instructions Computational instructions Jump and branch instructions Coprocessor 0 instructions Special instructions Load and store instructions Computational instructions Jump and branch instructions Special instructions 32 bit length instruction MIPS III16 bit length instruction MIPS 16 10 Memory Management Unit : MMU Virtual addresses are translated into physical addresses using an on-chip TLB. The on-chip TLB is a fully- associative memory that holds 32 entries. These pages can have five different sizes, 1 K, 4 K, 16 K, 64 K, and 256 K, and can be specified in each entry. 11 Cache memory (1) Cache controller I-cache D-cache Main memory VR4120 CPU core 12 Cache memory (2) The instruction cache is 16 Kbytes and the data cache is 8 Kbytes. The line size for the instruction/data cache is 4 words (16 bytes). The VR4120 CPU core uses write- back policy. 13 Conclusion : Features of VR4121 Employs 64-bit RISC CPU Core (VR 4120 equivalent) Internal 64-bit data processing. Optimized 6-stage pipeline. On-chip cache memory Instruction cache: 16 Kbytes, Data cache: 8 Kbytes. Address space Physical address space: 32 bits, Virtual address space: 40 bits. 14 Conclusion : Features of VR4121 Memory controller (supports ROM, EDO-type DRAM, SDRAM, SROM, and flash memory) Keyboard, Touch panel, audio interface. DMA, interrupt controller. Serial interface. General-purpose ports. IrDA interface. 15 Conclusion : Features of VR4121 Effective power management features, which include the following four operating modes:Full-speed mode, Standby mode, Suspend mode, Hibernate mode. External input clock: kHz, MHz (for internal CPU core and peripheral unit operation), 48 MHz(dedicated for IrDA interface) 16 VR4121 Pipeline 17 Pipeline Stages –The pipeline is controlled by PClock. Pipeline in MIPS III(32-bit length) instruction mode –The execution of each instruction takes at least 5 PCycles. IF RF EX DC WB Stage PClock Pcycle 18 V R 4121 Pipeline Pipeline in MIPS16 (16-bit length) instruction mode –The execution of each instruction requires at least 6 PCycles. Stage PClock Pcycle IF IT RF EX DCWB 19 Delay in Pipeline Branch Delay, a one-cycle branch delay occurs when: –Target address is calculated by a jump instruction. –Branch condition of branch instruction is met and then logical operation starts for branch-destination comparison. Load Delay –A load instruction that does not allow its result to be used by the instruction immediately following is called a delayed load instruction. 20 Delay in Pipeline (Branch delay slot) Target IF RF EX DC WB Branch Branch delay Branch Delay (in MIPS III Instruction Mode) 21 Delay in Pipeline Branch delay IF IT RF EX DCWB IF IT RF EX DCWB IF IT RF EX DCWB (Branch delay slot) Target Branch Delay (in MIPS 16 Instruction Mode) Branch 22 Interlock and Exception Handling Pipeline flow is interrupted when cache misses or exceptions occur, or when data dependencies are detected. Faults ExceptionsInterlocks StallAbortSlip 23 Exception conditions Example: Interrupt exception, ITLB exception. When an exception condition occurs, the relevant instruction and all those that follow it in the pipeline are cancelled. IF RF EX DC WB IF RF IF Exception 1 2 Exception vector Discard stage Interpret 24 Stall conditions Example: Instruction TLB Miss and Instruction Cache Miss. When a stall occurs, the processor will resolve the condition and then the pipeline will continue. 25 Stall Conditions IF RF EX DC DC IF RF EX EX IF RF RF EX EX EX DC WB RF RF RF EX DCWB IF RF EX DC WBWB DC DC DC WB WB WB WB Data Cache Miss Stall Detect data cache miss 2. Starting moving data cache line to write buffer 3. Get last word into cache and restart pipeline 26 Stall conditions IF RF EX DC DC IF RF EX EX IF RF RF EX EX EX DC WB RF RF RF EX DCWB DC DC DC WB WB WB WB Cache Instruction Stall 1 2 IF RF EX DC WBWB 1. Cache instruction start 2. Cache instruction complete 27 Slip conditions If all of the source operands are available (either from the register file or via the internal bypass logic) and all the hardware resources necessary to complete the instruction will be available whenever required, then the instruction "run"; otherwise, the instruction will "slip". Example: Load Data Interlock and MD Busy Interlock. 28 Slip conditions Load Data Interlock IF RF EX DC WB IF RF EX DC DC2WB 12 IF RF EX DC DC2WB IF RF RF RF EXDC WB 1 Bypass Load A Load B Add A,B 1. Detect load interlock 2.Get the target data 29 Slip conditions IF RF EX DC WB 121 Bypass IF RF EX EX EXEXWBDC 11 IF RF EX EX EXEXWBDC MUL/DIV Multiply/Division opration MFHI/MFLO 1. Detect MD busy Interlock 2. Get target data MD Busy Interlock 30 Bypassing Operand bypass allows an instruction in the EX stage to continue without having to wait for data or conditions to be written to the register file at the end of the WB stage. 31 V R Series Supported Instruction 32 End of the presentation... Any Question ???? Download ppt "1 VR4121 64-BIT MICROPROCESSOR โดย นางสาว พิลาวัณย์ พลับรู้การ นางสาว เพ็ญพรรณ อัศวนพเกียรติ" Similar presentations Ads by Google
__label__pos
0.898158
Creating a Blazor WebAssembly Project Creating a Blazor WebAssembly Project In this article, we'll learn about Blazor WebAssembly and we'll see how to create a project using Visual Studio and dotnet CLI. Starting with Visual Studio 16.6.3 and .NET core 3.1.301 you can create Blazor applications for both server and client/WebAssembly. What's Blazor WebAssembly? WebAssembly or WASM is a binary instruction set supported by modern browsers. It can be used to run code generated from high level programming languages such as C# which means it provides the possibility for developers to create client-side web applications with these programming languages without resorting to JavaScript JavaScript or using plugins, Java applets. With Blazor WebAssembly, you can run the code the web browser as a SPA (Single Page Application). Creating a New Blazor WebAssembly Project You can use Visual Studio to create a new Blazor project by following these steps: • First, create a new project and look for the Blazor App template • You’ll be able to choose between creating a Blazor Server App or Blazor WebAssembly App If you prefer the command-line interface, you can use the dotnet CLI to create your project. Head over to a new command-line interface and run the following command: $ dotnet new blazorwasm -o your-project-name Next, navigate to your project's folder and run a local development server using the following commands: $ cd your-project-name $ dotnet run You can then open your web browser and visit your the application from the https://localhost:5001/ address. ✋If you have any questions about this article, ask them in our GitHub Discussions 👈 community. You can also Gitter ❤️ Like our page and subscribe to our feed for updates!
__label__pos
0.551939
Home Excel formula help Highlighted kmmmm_03 New Contributor I'm setting up a spreadsheet with an indirect formula which I can get to work for the first line but any line after this seems to just reference the original cell and not what is populated for that line. How do I fix this? 6 Replies Highlighted Please specify your formula. @Twifoo =INDIRECT($F$2) Highlighted Use relative reference for the row, like this: =INDIRECT($F2) Highlighted @kmmmm_03  I might not be the best person to answer because it is many years since I last used a direct cell reference as part of a spreadsheet solution.    I would recommend using a Template sheet with as much structure as possible built in through the use of defined names and Tables and structured references to the variable sized panels.  Copying a template sheet then automatically creates local names for each resulting worksheet.     Different areas within the Summary sheet can be used to reference different sheets using the same name, by looking up the name associated with as set of rows numbers, e.g. 'project.name', from a table of steering data 'Project', which comprises a set of row indices and the associated name, which acts rather like a contents page in a book. = LOOKUP(@RowNum, Project ) The sheet tabs are your project codes, which allows names such as = INDIRECT(project.name&"!Status") to direct to any sheet.     'Panels' might be defined by Tables which would allow them to accommodate different sized data sets, the size being readable using ROWS(Table1).  Table names must be unique across the workbook but the can use the same name as the sheet tab. This means ''project.data' = INDIRECT(project.name) can represent an entire table (on different sheets according to the setting of 'project.name').  A data column within each table could then be referenced by 'project.task' which refers to = INDIRECT(project.name&"[Task]") or, alternatively, it would be possible to use = INDEX( project.data, 0, 1 ) for the first column.   All this structuring is alien to normal spreadsheet practice because it requires planning and structure whereas the traditional approach is more of a 'bottom-up' process in which user actions can create infinite variation. Highlighted You must have posted your reply to an incorrect conversation. Please verify. Highlighted @Twifoo  I think you are correct.  An Excel 'running out of memory' error crashed my computer a number of times yesterday and I guess I didn't pick up the pieces correctly.  My post was intended as a reply to a question that showed an image of a sheet with a detailed formula involving = INDIRECT( Sheetname , CELL("address", … ) ) Related Conversations I need help with a formula please. Bernz in Excel on 1 Replies Use Data Validation List To Recall Values In A Table DoctorDre225 in Excel on 1 Replies conditional formatting tonylinka in Excel on 2 Replies
__label__pos
0.737805
Search Images Maps Play YouTube News Gmail Drive More » Sign in Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader. Patents 1. Advanced Patent Search Publication numberUS6040843 A Publication typeGrant Application numberUS 09/181,850 Publication dateMar 21, 2000 Filing dateOct 29, 1998 Priority dateOct 19, 1995 Fee statusLapsed Also published asUS5835102 Publication number09181850, 181850, US 6040843 A, US 6040843A, US-A-6040843, US6040843 A, US6040843A InventorsMichael E. Anderson, Stephen C. Kenyon, John W. Monroe, John J. Smith Original AssigneeSparta, Inc. Export CitationBiBTeX, EndNote, RefMan External Links: USPTO, USPTO Assignment, Espacenet System for transmission and recovery of digital data using video graphics display processor and method of operation thereof US 6040843 A Abstract A system (30) for outputting digital data stored in a memory of a computer (12) in accordance with the invention includes a graphics display processor (20), coupled to the memory, for processing the digital data stored in the memory to produce at least one serial data stream including clock information, which is a function of a clock signal representative of a rate at which the at least one display formatted serial data stream is outputted, and display information for use in controlling a video monitor; a video channel (24), coupled to the display processor, for outputting the at least one serial data stream produced by the graphics display processor; a data processing memory (52); and a data processing system (40, 44 and 50) for processing the at least one serial data stream in response to the clock information and for removing at least the display information and controlling storing of the at least one of the at least one serial data stream with the display information removed which contains the digital data read from the memory of the computer system. Images(43) Previous page Next page Claims(27) We claim: 1. A computer system comprising: a computer system memory which stores digital data; a video monitor which displays information; a graphics display processor, coupled to the computer system memory, which processes the digital data stored in the computer system memory to produce at least one data stream including the digital data and which outputs information to be displayed by the video monitor; a video channel, coupling the video monitor to the graphics display processor, which outputs the information to be displayed to the video monitor and which outputs the at least one data stream including the digital data; and a system, coupled to the video channel, which processes the at least one data stream including the digital data outputted from the video channel. 2. A computer system in accordance with claim 1 wherein: one of the at least one data stream contains a clock signal and display information and another of the at least one data stream contains the digital data and the display information; and the system processes the at least one data stream in response to the clock information, removes at least the display information and controls storing of the at least one of the at least one data stream with the display information removed in another memory of the system coupled to the video channel. 3. A computer system in accordance with claim 1 herein: one of the at least one data stream includes the digital data with at least a portion of the one of the at least one data stream being encoded with self-clocking information which permits the clock signal representative of a rate at which the one of the at least one data stream is outputted by the video channel to be derived from processing the portion of the one of the at least one data stream outputted from the video channel by the system coupled to the video channel. 4. A computer system in accordance with claim 1 therein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 5. A computer system in accordance with claim 2 wherein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 6. A computer system in accordance with claim 3 wherein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 7. A computer system in accordance with claim 1 wherein: the at least one data stream is a serial data stream. 8. A computer system in accordance with claim 2 wherein: the at least one data stream is a serial data stream. 9. A computer system in accordance with claim 3 wherein: the at least one data stream is a serial data stream. 10. A computer system in accordance with claim 4 wherein: the at least one data stream is a serial data stream. 11. A computer system in accordance with claim 5 wherein: the at least one data stream is a serial data stream. 12. A computer system in accordance with claim 6 wherein: the at least one data stream is a serial data stream. 13. A system in accordance with claim 1 wherein: the system is removable from the computing system without altering the operation of the computing system. 14. A system in accordance with claim 2 wherein: the system is removable from the computing system without altering the operation of the computing system. 15. A system in accordance with claim 3 wherein: the system is removable from the computing system without altering the operation of the computing system. 16. In a computer system including a computer system memory which stores digital data, a video monitor which displays information, a graphics display processor, coupled to the computer system memory, which processes the digital data stored in the computer system memory to produce at least one data stream including the digital data and which outputs information to be displayed by the video monitor, a video channel, coupling the video monitor to the graphics display processor, which outputs the information to be displayed to the video monitor and which outputs the at least one data stream including the digital data, and a system, coupled to the video channel, which processes the at least one data stream including the digital data, a method comprising: outputting digital data stored in the computer system memory to the graphics display processor; processing the outputted digital data with the graphics display processor to produce at least one data stream containing the digital data; and outputting the at least one data stream from the video channel to the system. 17. A method in accordance with claim 16 wherein: one of the at least one data stream contains a clock signal and display information and another of the at least one data stream contains the digital data and the display information; and the system processes the at least one data stream in response to the clock information, removes at least the display information and controls storing of the at least one of the at least one data stream with the display information removed in another memory of the system coupled to the video channel. 18. A method in accordance with claim 16 wherein: one of the at least one data stream includes the digital data with at least a portion of the one of the at least one data stream being encoded with self-clocking information which permits the clock signal representative of a rate at which the one of the at least one data stream is outputted by the video channel to be derived from processing the portion of the one of the at least one data stream outputted from the video channel by the system coupled to the video channel. 19. A method in accordance with claim 16 wherein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 20. A method in accordance with claim 17 wherein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 21. A method in accordance with claim 18 wherein: the system coupling the video channel to the graphics display processor comprises a removable memory used for backup of the computer system memory. 22. A method in accordance with claim 17 wherein: the at least one data stream is a serial data stream. 23. A method in accordance with claim 17 wherein: the at least one data stream is a serial data stream. 24. A method in accordance with claim 18 wherein: the at least one data stream is a serial data stream. 25. A method in accordance with claim 19 wherein: the at least one data stream is a serial data stream. 26. A method in accordance with claim 20 wherein: the at least one data stream is a serial data stream. 27. A method in accordance with claim 21 wherein: the at least one data stream is a serial data stream. Description Like reference numerals identify like parts throughout the drawings. BEST MODE FOR CARRYING OUT THE INVENTION FIG. 2 illustrates a block diagram of an embodiment 30 of the present invention. The hardware of the CPU 12, disk controller 13, hard drive 14, floppy disk drive 14, dynamic random access memory 16, graphics adaptor card 18, video random access memory 19, graphics display processor 20, video monitor 22 and video output channel 24 having N parallel lines are identical to the prior art described above in conjunction with FIG. 1. The present invention differs from the prior art in that it utilizes the high speed data outputting capacity of the video channel 24 to process and output at least one serial digital data stream comprised of data stored in the memory space associated with the CPU 12 for applications other than driving a video monitor 22, such as, but not limited to, restoration of data stored within the memory space of the CPU whether in the internal hard drive and/or in the floppy disk drive 15 and/or D RAM 16. The graphics display processor 20 in accordance with the invention outputs at least one serial data stream containing digital data originally stored within the address space of the CPU 12, such as data stored in the hard drive 14 or the floppy disk drive 15 and clock information. The clock information is a function of a clock signal representative of a rate at which the at least one serial data stream is outputted by the video channel 24. The clock information is either an alternating series of one and zero bits outputted on a line of the video channel 24 at an identical rate at which data bits are read out on another line of the video channel or alternatively, at least a portion of each line of each video frame encoded with self-clocking information, such as Manchester code, which may be processed to derive a clock signal as explained below. Connected in parallel to the video channel 24 is a video data acquisition subsystem 32 as described below with respect to FIG. 3. The video data acquisition subsystem 32 contains a processor and associated first and second frame buffer memories which are periodically used to store frames of information outputted by the first and second frame buffers of the graphics display processor 20 as is described below. The video data acquisition subsystem 32 is connected via interface 33 to a system control and storage processor and memory 34 which also contains first and second frame buffers which store frames outputted by the first and second frame buffers of the video data acquisition subsystem 32. The system control, storage processor and memory 34 writes information into a removable storage 36 which may be used to write information back into the internal hard drive 14 for restoration purposes as described below. The restore block 60 represents a processor used for writing information stored on the removable storage 36 back into the address space of the hard drive 14 of the CPU 12 as is described in detail below. The present invention uses the programmable capability of the graphics display processor 20 to transmit data via the video channel 24 to the video data acquisition subsystem 32 while maintaining synchronism between the CPU 12, graphics display processor 20, video data acquisition subsystem 32 and system control and storage processor and memory 34. An example of protocols which may be used and the hardware required to receive, process and convert the data which has been formatted into a video display format by the graphics display processor 20 back into its original form as read from the address space of the CPU 12 to provide restoration of the internal hard drive is described below. The graphics display processor 20 is programmed to operate in a graphics mode. For example, the VGA 640 space of 480 rows of lines with each containing 640 pixels of information may be utilized. Each of the N parallel lines which are present in the video channel 24, which define the color resolution achievable for display of graphics information on the video monitor 22, has data outputted in a binary state of a one or a zero. The data outputted from the address space of the frame buffers of the graphics display processor 20 associated with the hard drive 14 may be visualized as a linear series of data bits having a length in this example of 307,200 bits per frame (640 graphics display processor 20 formats the groups of 307,200 bits per frame which were read from the memory space of the internal hard drive 14 and which are transmitted from the graphics display processor on the video channel 24. During each refresh cycle of the graphics display processor 20, information contained in the video random access memory 19 is continually transmitted to the graphics display processor. The frame rate may be within the conventional range, such as 60-72 per second, which defines the refresh cycle of the graphics display processor 20. Assuming that there are 60 frames per second, a refresh rate equal to 60 processor 20 over the video channel 24 on each of the N parallel lines. This rate represents a rate approaching two megabytes per second. While it is not necessary in practicing the invention, the video monitor 22 may remain attached to display the data outputted by the video channel 24 which has been retrieved from the address space of the internal hard drive of the CPU 12 and processed for display purposes by the graphics display processor 20. As has been stated above, the video channel 24 is also connected to the video data acquisition subsystem 32. In the preferred application of the invention, only a single serial data stream of the N serial data streams are connected to the video data acquisition subsystem 32 as is described below with that line containing the digital data with at least a portion of the one of the at least one serial data stream connected to the video data acquisition subsystem 32 being encoded with self-clocking information which permits the clock signal representative of a rate at which the one of the at least one serial data stream is outputted by the video channel 24 to be derived from processing at least the portion of the one of the at least one serial data stream outputted from the video channel as described below in conjunction with FIGS. 3 and 4. Alternatively, two lines of the N lines of the video channel 24 may be connected to the video data acquisition subsystem 32 with the first line containing the data which has been retrieved from the address space of the CPU 12 and the display information added by the graphics display processor 20 and the other line containing an alternating series of ones and zero bits at the same bit rate at which data is outputted on the first line which is a clock signal for synchronizing the processing of the data by the video data acquisition subsystem 32. The video data acquisition subsystem 32 processes the at least one of the at least one serial data stream, including processing of the clock information, to produce the clock signal representative of a bit rate at which the at least one serial data stream is outputted by the video channel to the video data acquisition subsystem in response to the clock information contained in the at least one serial data stream as described below in conjunction with FIG. 3. As is described below, the video data acquisition subsystem 32 further removes the self-clocking information and converts the serially outputted information received on the at least one serial data stream into parallel information having the same number of bits as each addressable location within the address space of the CPU 12. Thus, the video data acquisition subsystem 32 converts the data back into a parallel format have a number of bits per word corresponding to the number of bits of data per word stored within the address space of the CPU 12. The words may have differing number of bits depending upon the architecture of the CPU 12. Synchronization is maintained between the CPU 12 sending the data to the graphics display processor 20, the graphics display processor sending out video formatted information on the video channel 24 and the video data acquisition subsystem 32 hardware receiving the video formatted information and converting it back into data. Several techniques are employed to establish and maintain this synchronism. The basic output unit of the graphics display processor 20 is a video frame. The video frame includes all data transmitted between consecutive vertical sync (VS) pulses. VS pulses are distinguished from data by a variety of techniques and are usually defined by a slightly higher (and longer) voltage level. Sometimes, a single connector is used to carry this information to the video monitor 22 and sometimes the information is mixed with data on a single connector (such as sync on green). The graphics display processor 20 generates VS at the start of each video refresh. In the example described above, VS is generated 60 times each second. With each VS, the graphics display processor 20 sets a bit in a status register in the graphics adaptor card 18. The CPU 12 application program monitors the VS bit and determines the exact time each video refresh cycle begins. When the video data acquisition subsystem 32 receives VS, it resynchronizes to begin receiving the next set of 307,200 bits sent by the graphics display processor, which represents a full frame of video formatted data, which includes the data retrieved from the address space of the CPU 12, as well as the clocking information and display information, as described below, in conjunction with FIG. 4. The VS alone cannot provide synchronization down to the level of individual pixels. The VS alerts the video data acquisition subsystem 32 to begin pixel synchronization and provides frame synchronization between the CPU 12 and the hardware of the video data acquisition subsystem 32. The video data acquisition subsystem 32 transfers the video data outputted from the graphics display processor 20 on the video channel 24. The video data acquisition subsystem may be implemented with hardware and software modules for performing specific functions on the data to be transferred. The first module of the software controls the graphics display processor 20 which may be, for example without limitation, pages 1-65 of the attached Appendix. This software is used to read data from the hard drive 14, convert it into the transmission format as described below in conjunction with FIG. 4, and write it into the video random access memory 19. The data is then transmitted using the video output channel 24 from the source CPU 12 to the video data acquisition subsystem 32. The hardware of the video data acquisition subsystem 32 converts the original data from the formatted video signal and makes it available to the system control, storage processor and memory 34 and signals the processor within the system control, storage processor and memory when a frame of data is ready for transmission thereto. The second module is the software which controls the system control, storage processor and memory 34 which may be, for example, without limitation, pages 66-197 of the attached Appendix, and which functions to transfer data to a removable storage 36 which is a disk or other storage medium. Furthermore, the system control, storage processor and memory may be used to maintain multiple images and to provide a user interface backup and restore operations as described herein. The video data acquisition subsystem 32 converts the input serial data stream outputted from the video channel 24 back into parallel data having the same number of bits which are read out from each address location in the address space of the CPU 12 and makes the converted data available for storage in the storage of the system control, storage processor and memory 34 and removable storage 36. The input serial video signal is conditioned and converted into a TTL level bit stream by processing performed by input conditioner 38, sync detector 40 and clock recovery circuit 42. The serial data is stripped of framing information, which was applied by the graphics display processor 20 in accordance with the conventional function thereof, and is converted to multiple bit words and held for temporary storage by the data recovery circuit 44 and the data stores 46 as described below in detail. The processed data is outputted alternately in a framed format from the data stores 46 which function as frame buffers. The output from the data stores 46 is applied to the interface 33 which is coupled to the system control storage processor and memory 34. The system control, storage processor and memory 34 is connected to the removable storage 36 which is connected to the restore function 60 as described below. The input conditioner 38 converts the high speed video signal outputted on the video channel 24 into a digital data stream for processing by the remainder of the video data acquisition subsystem hardware. This function is accomplished in two steps using high speed operational amplifier circuits. The first high speed operational amplifier circuit is used to provide a ground reference to the input video signal from the video channel 24 which may be "floating" with reference to ground of the video data acquisition subsystem 32. A differential amplifier is used to compare the video input signal to the return signal and outputs the difference thereof. After this is accomplished, the processed video signal may be converted to a TTL level. The second operational high speed operational amplifier circuit is in a configuration of a Schmitt Trigger. The Schmitt Trigger allows four separate compare points for high and low transitions to provide a high degree of noise immunity. The use of a Schmitt Trigger is necessary to provide adequate processing when poor quality video input signals are received which are often generated by old or low quality video cards present in the video source system. After signal processing by the input conditioner, a pair of output signals are produced in the form of a Manchester encoded digital bit stream which is suitable for processing by digital electronics as described below. The clock recovery circuit 42 recovers clocking information from the Manchester encoded bit stream. The clock recovery is accomplished by detecting mid-bit transitions in the data which are present in the Manchester encoded bit stream to provide a clock edge which is slightly delayed from these mid-bit transitions. An edge detector circuit is used to output a short pulse corresponding to each transition in the Manchester encoded data stream. Thereafter, the short output pulses are sent through a pulse blanking circuit which removes all pulses occurring between the aforementioned slightly delayed signal transitions produced by the edge detector. The pulse blanking is important for proper clock recovery and should be stable over time to permit processing of pixel rates produced by the graphics display processor 20 which are commonly in frequencies of between 25 and 32 million pixels per second as measured between blanking intervals. The sync detector circuit 40 does not modify the Manchester data stream applied thereto as an input and functions to detect the sync signal which is the first 64 bits of each line of the video formatted frames outputted by their video graphics display processor 20 as described below in conjunction with FIG. 4. Once the sync word contained in the first 64 bits of each line of bits is detected in the Manchester data stream, the beginning of a valid formatted line of video has been detected. Once the position of the sync information is detected, a START signal is sent to the data recovery circuit 44 to allow the data recovery circuit to start processing the input bits which are received from the input conditioner 38. The data recovery circuit 44 performs four operations on the Manchester data which is outputted from the input conditioner 38. The first operation is to convert the serial Manchester data stream into a word format having a number of bits identical to the number of bits stored at each addressable location of the address space of the CPU 12. This function is accomplished by strobing bits into a serial to parallel converter with the recovered clock pulses as indicated by the CLOCK output from the clock recovery circuit 42. This processing also strips the Manchester data outputted by the input conditioner 38 of the Manchester encoding of the data. The second operation is to deformat the input lines of video of each frame by stripping off the sync, scan line, control channel and trigger bits as described below in conjunction with FIG. 4. This function is accomplished by routing the bits of each of the fields of FIG. 4 to an appropriate address based upon their position within the scan line having the format of FIG. 4. Frame and line number information are routed to registers used for control of data storage. The third operation is to route the data which is the last field within the line format of FIG. 4 to one of the data stores 46 using an address based on the line number received in the header. Addresses are then incremented for each incoming word until the line is complete. The least significant bit of the frame number is used to determine which of the data stores 40 in which the frame of data will be written. The least significant bit is also routed from the data recovery circuit 44 as the DATA READY signal to the interface 33 to function as a switch between the two data stores during read out. The fourth operation is to signal the interface 33 that the frame is ready for read out which is accomplished by the reception of the aforementioned DATA READY signal. The data stores 46 are identical and perform identical functions. Each stores one complete frame of video data which has been stripped of all of the bits as described below in conjunction with FIG. 4 except the 512 data bits therein. The two data stores 46 perform the function of frame buffers for read out through the interface 33. Control over which store 46 is to be dedicated to data recovery and which store 46 is to be available for read out is based upon the value of the least significant bit of the frame number. When a data store 46 is dedicated to the data recovery process, data is written into the store using local bus control signals that come from the data recovery circuit 44. When a data store 46 is dedicated to the interface 33, data is read from the store using control signals that come from the interface. A data store may not simultaneously receive data from the data recovery circuit 44 and output the data to the interface 33. The interface 33 provides access to the DATA READY signal and the data stores 46 to the system control, storage processor and memory 34 as described above. The interface 33 may have different circuit implementations without changing its performance. The interface 33 may be configured in different ways, but it must have sufficient bandwidth to handle the nominal data rate produced by the video channel 24 and to permit the system control, storage processor and memory 34 to respond to the DATA READY signal without any significant time delay. If the entire frame of data is not read out from the video data acquisition subsystem 32 before the next DATA READY signal, unread data will be corrupted by the next incoming frame. FIG. 4 illustrates an example of a scan line protocol for encoding individual lines of the video formatted frames outputted by the video graphics display processor 20. There are 640 pixels in each of the 480 scan lines which are transmitted in five fields as illustrated in FIG. 4. It should be understood that when Manchester coding is used to encode each bit. Each data bit in each of the fields of FIG. 4 from the address space of the CPU 12 will require two bits to be encoded in a Manchester coded format. The Manchester encoding of each data bit retrieved from the address space of the CPU 12, which is outputted on the video channel 24 under control of the graphics display processor 20, is performed by CPU 12 under control of an application program. The fields are a scan sync field containing 64 bits, a scan line field containing 32 bits, a control channel containing 16 bits, a trigger field containing 16 bits and a data field containing 512 bits. Thus, because of the use of Manchester coding which requires two bits to encode each data bit in order to provide self-clocking information, the scan sync field will be comprised of 32 bits, the scan line field will be comprised of 16 bits, the control channel will be comprised of 8 bits, the trigger field will be comprise of 8 bits and the data field will be comprised of 256 bits for a total of 320 bits or 40 bytes of actual data. The scan sync field is optimized to contain a constant pattern of alternating ones and zeros to permit the clock signal to be derived by the clock recovery circuit 42 by permitting a phase lock loop to lock an oscillator to a frequency at which the scan field bits alternate between a one and zero value as described above. The 64 bits in the scan sync field stabilize the phase lock loop of the oscillator within the clock recovery circuit 42 which provides the clock signal on the output thereof to the sync detector 40 with sufficient stability to permit a remainder of each line of a video frame to be processed without further synchronization information. After the initial pixel clock rate is established for each line within the clock recovery circuit 42 by processing the scan sync field, the phase lock loop circuit maintains the video data acquisition subsystem 32 within synchronization for the remainder of the scan line. If synchronization is lost at the end of the line, it will be reacquired on the next line by processing the next 64 bits of the scan sync field. The scan line field of 32 bits is an address of each scan line in the frame. Numbering starts at zero for the first scan line transmitted following VS. Each scan line address is incremented by one. This technique allows the video data acquisition subsystem 32 to determine the total number of scan lines transmitted with each frame and allows processing when the video frame format is unknown to contain a specified or fixed number of scan lines. The first scan line is detected by the presence of a zero in the scan line field. The line address can also be used to generate addresses at which each scan line's data is stored in the data stores 46. The control channel of 16 bits provides a mechanism for the program of the CPU 12 to send an "out-of-band" stream of data or control signals to either the video data acquisition subsystem 32 or system control, storage processor and memory 34. With each frame occurring at 60 frames per second, up to eight bits of control information, or data, can be passed to the video data acquisition subsystem 32 and system control, storage processor and memory 34. The high order four bits may be used to interrupt the operation of the storage control, storage processor and memory 34. All eight bits are available to the system control and storage processor 34 as a byte-wide status register. The control channel can be used in numerous ways but the four high order bits can be used to provide an efficient method of identifying the contents of each frame. The four low-order bits can be used to pass data directly to the system control, storage processor and memory 34. Quantities, such as the amount of data remaining, the frame dimensions, the refresh rate, etc., may be encoded with this field. Frame synchronization is maintained between the CPU 12 and the video data acquisition subsystem 32 for another reason. The CPU 12 controls providing data to the video random access memory 19. The CPU 12 may be able to keep up with the data requirements of the video random access memory 19 and the graphics display processor 20 and therefore send out 307,200 bits per frame as described above. If the CPU 12 cannot keep up with the requirements of the video random access memory 19, the graphics display processor 20 will retransmit whatever is currently in the video random access memory resulting in some frames being sent out more than once. The video data acquisition subsystem 32 must be able to distinguish between new and old frames of data to determine which frames to store and which to ignore. The trigger field of 16 bits provides the bits necessary for the synchronization. The application program of the CPU 12 increments a counter in the trigger field only when the entire 640 bits frame has been coded and is ready to be transmitted to the video data acquisition subsystem 32. By the time the video data acquisition subsystem 32 receives the trigger field as explained above, it will have already stored the previous frame in its internal memory which includes the data stores 46 which function as first and second frame buffers. As explained above, the video data acquisition subsystem 32 monitors the state of the least significant bit of the trigger field counter maintained by the CPU 12 for changes from one frame to the next. Since the field contains an incrementing counter, the low-order bit toggles between zero and one each time the counter is incremented. As stated above, the dual-ported data stores 46 of the video data acquisition subsystem 32 have two frame buffers with each being large enough to hold an entire 640 bits by 480 line decoded data frame. Each of the data stores 46 holds a single frame of decoded data. While the state of the trigger bit (the low-order bit and the trigger field) remains fixed, the video data acquisition subsystem 32 continues to overwrite the data in the current buffer with each successive frame. The video data acquisition subsystem 32 stops writing into the current buffer of the data stores 46 with new data and begins writing into the other frame buffer when the trigger bit toggles. As the trigger bit toggles, the video data acquisition subsystem 32 begins writing data from the next frame into the other frame buffer of the data stores 46. When a complete data frame has been stored, a DATA READY signal is sent to indicate that a new data block is available for transfer to the system control and storage processor and memory 34 via the interface 33. The dual-ported memory in the form of the data stores 46 permits the CPU within the storage control, storage processor and memory 34 to read data from one memory frame buffer of the dual-ported memory while a new data frame is being stored in the other frame buffer and to read and write data into the first and second frame buffers of the system control, storage processor and memory 34 in the same fashion as data is being read and written into the data stores 46. Once a complete frame has been transferred to the system control, storage processor and memory 34 over the interface 33, the two frame buffers of the data stores 46 are interchanged in function and the process is repeated. If the CPU 12 program causes the trigger bit to toggle after a valid frame has been stored in the video random access memory 19 connected to the graphics display processor 20 and prior to writing the first bit of the next frame, the required CPU 12 to system control and storage processor and memory 34 synchronization is achieved. The CPU 12 program can write as much of the next frame as desired following the trigger bit since the frame buffer switch will take effect before the first data bit of the new frame is stored by the video data acquisition subsystem 32. It is not necessary for the CPU 12 to wait until the storage control and storage processor 34 has received the trigger bit before writing the next frame of data. The graphics display processor 20 performs the transmitting first, followed by the trigger bit and then additional data bits. The data field of 512 bits in FIG. 4 transfers actual data from within the memory space of the CPU 12. With the video format given in the above example, 480 scan lines times 256 data bits (32 bytes) per line, can be transferred with each frame. Therefore, as stated above, up to 921,600 bites per second can be transmitted from the CPU 12 to the video data acquisition subsystem 32 continuously if the CPU 12 keeps up with the graphics display processor 20. Faster CPU's 12 are capable of performing this task. One reason a fast CPU 12 may not keep up with the graphics display processor 20 is that, during backup, it has to read data from the hard drive 14 which takes additional time, with several frames being required to perform the complete disk read. This is particularly true when large disk blocks are being read to optimize the disk I/O. While the CPU 12 is accessing the hard drive 14, calibration data can be continuously sent to the storage control and storage processor and memory 34 for validation. The storage control and storage processor and memory 34 will differentiate between calibration data and valid CPU data by the contents of the control field. If calibration data is found to be incorrect, the operator must be notified by the storage control and storage processor 34 so that the backup process can be terminated. By sending calibration data every time the hard drive 14 is accessed, the validity of the backup is periodically checked throughout the entire backup process. If frame by frame calibration is desired, the low-order four bits of the control field can be used. This allows for validation of the control field but, does not validate bits in other regions of the scan line, such as the data field. Furthermore, the first and second frame buffers of the data stores 46 provide temporary storage and synchronization between the video data acquisition subsystem 32 and the system control, storage processor and memory 34. The system control, storage processor and memory 34 must be sufficiently fast to keep up with the average data transfer rate thereto but may at times be unable to keep up with each transmitted individual frame while storing data in its storage device. The use of first and second frame buffers in the system control, storage processor and memory 34 allows the capture of data sent to it. This is sufficient to keep up with the video data acquisition subsystem 32. When the video data acquisition subsystem 32 notifies the system control, storage processor and memory 34 that one of the data stores 46 is ready to transfer data, the system control and storage processor will transfer the contents of one of the data stores 46 of the video data acquisition subsystem 32 to its own frame buffer space which is comprised of first and second frame buffers and block it for optimal output to the removable storage 36. The first and second frame buffers of the system control and storage processor and memory 34 function in the same fashion as the data stores 46 which function as frame buffers of the video data acquisition subsystem 32 which allow one frame buffer to be written into while the other is being filled and outputting information to the removable storage device 36. The system control, storage processor and memory 34 includes a set of input/output registers to allow the CPU therein to control operating parameters within the video data acquisition subsystem 32 and to monitor its status. Parameters such as the approximate video formats expected and the number of bits in each field of the data packet format of FIG. 4 can be passed from the system control, storage processor and memory 34 to the video data acquisition subsystem 32 to allow adapting to a wide variety of PCs. The restore block 60 functions to perform restoration as follows. When the backing up of files of the CPU 12 is required, the format of the backup data stored in the removable storage 36 is important. The format used allows the original CPU 12 internal hard disk drive to be reconstructed on a file-by-file basis or to have the entire disk restored as a bit image. Two methods may be used to perform this task. In the first method, the data written into the system control, storage processor and memory 34 represents an exact image of the original data stored in the disk of the CPU 12. The PC restore program executed by block 60 uses a PC-based device driver to map individual disk read commands into the disk read commands required to read each sector off of the removable storage 36 as though it still resided on the original CPU 12 disk. The second method also uses data written in the removable storage device 36 which is an exact image of the original CPU 12 disk partition. The program executed by the block 60 uses a driver which processes the removable storage 36 as an extended partition of its own containing one or more logical drives. The user will be able to change to the direct logical drive and allow the operating system to read the files directly as though there were files contained by that operating system. The storage control, storage processor and memory 34 is a fast general-purpose single card computer containing associated memory and further functioning to store collected data on removable storage 36 or to transmit the data by a communications channel (not illustrated) to support diverse applications for the data stored in the internal hard drive 12. Pages 66-197 of the Appendix contain a computer code listing that in association with the circuit schematic of FIGS. 5A-5B, 6A-6D, 7A-7E, 8A-8F, 10A-10C, 11A-11F, 12A-12F and 13A-13F are an embodiment of the present invention. Furthermore, pages 1-65 contain a computer code listing which may be used to control the graphics display 20 which functions in conjunction with the remaining parts of FIG. 3 including the system control, storage processor and memory 34 to practice the present invention. The system control, storage processor and memory 34 is responsible for setting up the mode of operation of the video data acquisition subsystem 32 and monitoring the progress of the backup operation when original data stored within the memory space of the CPU 12 is being restored. The system control, storage processor and memory 34 monitors the amount of data transmitted and computes the estimated time to complete the backup process. The system control, storage processor and memory 34 stores all the collected data and validates calibration data. When operator feedback is utilized, the system control, storage processor and memory 34 is responsible for the generation and formatting of user messages. As has been stated above, while a preferred embodiment of the present invention, as described above in conjunction with FIG. 4 utilizes the scan sync field to produce the local clock signal necessary for completing processing of the transmitted at least one serial data stream into parallel data having the same number of bits as the number of bits stored in each addressable location of the CPU 12 memory, it is also possible to transmit on one line of the video channel 24 data which has not been formatted with self-clocking information and to transmit on another line of the video channel an alternating sequence of ones and zeros at the same bit rate as the data on the one channel for the purpose of functioning as a clock recovery 42 to be applied directly to the data recovery 44. While the invention has been described in terms of its preferred embodiments, it should be understood that numerous modifications may be made thereto without departing from the spirit and scope of the invention as defined in the appended claims. It is intended that all such modifications fall within the scope of the appended claims. ##SPC1## BRIEF DESCRIPTION OF THE DRAWINGS FIG. 1 is a block diagram of a prior art personal computing system including a graphics display processor. FIG. 2 is a block diagram of an embodiment of the present invention. FIG. 3 is an expanded block diagram of an embodiment of the video data acquisition subsystem 32 and the interface 33, system control, storage processor and memory 34, removable storage 36 and restore function 60 of FIG. 2. FIG. 4 is a block diagram of the protocol of encoding data for the scan lines of frames outputted by the graphics display processor. FIGS. 5A-5B, 6A-6D, 7, 7A-7E, 8, 8A-8F, 9, 10A-10C, 11A-11F, 12A-12F and 13A-13F are a circuit schematic of an embodiment of a system for practicing the invention as illustrated in FIG. 3. TECHNICAL FIELD The invention relates to a system and method for using a video graphics display processor associated with a processor, such as a PC, to achieve high speed digital data transfer to an external device such as for purposes of backup of memory of the processor. BACKGROUND ART FIG. 1 illustrates a block diagram of a computing system 10 including associated hardware for providing graphics display capability. The computing system 10 may be a personal computer. Communications in the system are transmitted on bus 11 between the hardware elements described below. The system 10 includes a CPU 12. A disk controller 13 is coupled to bus 11 and to hard drive 14 and floppy disk drive 15. The memory space further includes a dynamic random access memory 16 which is also connected to the bus 11 and which provides high speed reading and writing of data to support data processing performed by the system 10. The memory space is used for diverse functions as known in the art. The hard drive 14 has a much larger storage capacity than the floppy disk 15 and because of its capacity, a substantial time is required for its back up because of the absence of a high speed data port which is available for restoration of the memory space therein. The floppy disk memory 15 is the widely used floppy disk memory for storing information which is processed in accordance with the myriad of functions conventionally performed by the CPU 12. Associated with CPU 12 is a graphics adaptor card 18 which is coupled to bus 11 and which is bidirectionally connected to a video random access memory 19. The video random access memory is also connected to a graphics display processor 20 which continually reads data to be displayed from the video random access memory and formats information for display by a video monitor 22. As is indicated on the video channel 24 by the notation "N", the output from the graphics display processor 20, which is connected to the video monitor 22, is N bits wide which is indicative of the number of bits to produce a color display of a selected number of colors in a color palate encoded by N parallel bits on the N lines of the output 24. The video channel 24 is representative of typically 8 or 24 parallel lines each of which transmits a bit in a word which commands the color encoded by the word to be displayed by the video monitor 22 for each pixel of display data stored in the frame buffer of the video random access memory 19. The video random access memory 19 functions as a dual ported memory coupled to the bus and graphics display processor which permits the CPU 12 to control writing of information stored in the memory space of the CPU such as that stored in the hard drive 14 while the graphics display processor 20 is retrieving information from the video random access memory for purposes of formatting with appropriate video synchronization information for display by the video monitor 22. Typically, the graphics display processor is programmed to operate in a graphics mode. For example, the VGA 640 data space of 480 rows (scan lines). Each of the scan lines contains 640 bits (pixels) of information. Each pixel further is displayed with a programmable color specified by the value of the N bits which are outputted by the N parallel lines of the video channel 24. Thus, the video channel 24 can be thought of as transmitting N serial information streams each having a bit value of zero or one which bit values are combined to command the color of display of each pixel displayed by the video monitor 22. The graphics display processor 20 has first and second frame buffers which function to store information which is outputted by the video random access memory 19 to one of the frame buffers while the other of the frame buffers is driving the display of the video monitor 22 through the outputting of the display formatted data on the video channel 24. Standard non-interlaced monitors 22 typically refresh data at rates of 60-72 frames. Thus, each serial data stream of the N serial data streams outputted by the video channel 24 has a data rate of approximately 20 megabits per second or more. The use of the graphics display processor 20 to send display data to the video monitor 22 over a video channel 24 has been well known for many years. The representation of a video image to be displayed on the video monitor 22 is created by the CPU 12 controlling the writing of the data pattern into the video random access memory 19 where it is read by the graphics display processor 20. The CPU 12 creates proper patterns for display from the address space of system memory including data stored in the hard drive 14 and the bootable backup floppy disk memory 15. The graphics display processor 20 repeatedly scans the video random access memory 19 and processes the pattern of information stored and readout from the frame buffer of the video random access memory into the series of data streams having N parallel bits which are outputted on the video channel 24 to produce color pixels of N bit resolution on the video monitor 22. The video monitor 22 displays the graphical or textual data which has been stored in the memory space of the system 10 and processed by the video random access memory 19 and graphics display processor 20 into a format suitable for display. Graphics display processors 20 support a variety of video formats. Well defined protocols are known for programming these known variety of video formats. Currently, the graphics display processor 20 has been developed to perform the single purpose of displaying the data stored in the memory space of the CPU 12 and converting it into a suitable display format for display on the video monitor 22 by the operations performed by the video random access memory 19 and the graphics display processor. The extremely high data rates which are necessary to drive the display of the video monitor 22 at frame rates which are typically, as explained above, between 60 and 72 frames per second have not been applied to other applications which use the video channel 24 as a high speed data output device. The use of backup procedures to replicate and safeguard information stored in the internal hard drive have become more and more important as the storage capacity of hard drives has rapidly expanded in the last few years. The speed at which backup may be accomplished is a critical factor. As memory drives become larger, the time required to backup the internal hard drive increases. The increased time discourages users from performing backup of the hard drive on a regular basis. PCs having large internal hard drives and PCs not supporting high-speed I/O devices present a particular problem. Today's laptops with large internal drives are good examples of where backup of data is a problem since the backing up of the stored data must be done either via the parallel or serial port which is present on the PC. Currently, rapid backup of computer disk information requires the use of internal hardware devices capable of transferring information from the computer's data bus to an external storage device in a compressed or otherwise proprietary format. The most popular techniques available in the order of increasing transfer rates include the following: (1) Serial communication ports Serial communication ports typically can transfer data at speeds up to 11.5 K/Bytes per second. Serial ports are included on all PCs, are bidirectional and can be used for both backup and restore operations. (2) Parallel communication ports Parallel ports can transfer data at up to 30 K-Bytes per second. Occasionally, some input capability exists, but at much slower speeds dependent on the PC manufacturer's design. Generally, these ports are included on all PCs. Newer designs using parallel integrated circuits allow bidirectional data flow and at higher rates than their predecessors. (3) Floppy disk drives Usually, PCs come with at least one floppy disk drive. These devices will support a continuous transfer rate of about 45 K-Bytes per second for large data sets. The practical transfer rate is limited by mechanical track-to-track access times and the fact that the media needs to be manually changed about after a megabyte has been written. Floppy drives are bidirectional and can be used for both backup and restore purposes. (4) Floppy/hard disk controllers Most PCs come with a disk controller capable of supporting both floppy and hard disk drives. The floppy drive controller can support about 300-500 K-Bytes per second in short bursts but not for continuous periods. The controller is limited by a 16-bit byte count register which requires reloading after 64 K-Bytes have been transferred. The disk controllers are bidirectional and can be used for both backup and restoration. (5) External or internal magnetic data cartridges Today, the most popular backup devices use a magnetic data cartridge. These devices either use the PC's floppy disk controller or a separate external or internal interface controller. These devices can maintain about 500 K-Bytes per second without compression or about 1 M-Bytes per second when using compression techniques. These devices are typically optional equipment and cost approximately $200 for 250 K-Byte of backup capability. Both backup and restore are provided with many options for individual and group file selections available. The problem for many PCs, including laptops and palmtops, is that there is no internal space to hold the extra drive and no external connector to allow connection to an external drive. (6) External or internal disk drives Occasionally, users will install a second hard drive for the purpose of backing up or replicating data sets. This is the fastest backup technique available today and whenever it is possible, sustained transfer rates in excess 500 K-Bytes per second are easily accomplished. DISCLOSURE OF INVENTION The present invention is a process for outputting digital data stored in memory space of a computer having a graphics display processor and further, a system for outputting digital data stored in the memory space of a computer which utilizes the graphics display processor to format data originally stored in the computer memory space to produce at least one serial data stream including the digital data stored in the memory space of the computer and clock information which is a function of a clock signal representative of a rate at which at least one serial data stream is outputted by a video channel associated with the graphics display processor for displaying information formatted for display on a video monitor. With the invention, the high data transfer rates on the video channel which are produced by the graphics display processor for displaying display formatted information with a video monitor are used to output information which is not for display purposes, such as, but not limited to, providing information from the computer memory space for processing in a format to provide restoration of the data stored in the hard drive of the computer memory space to perform backup thereof. The present invention has substantial advantages over the prior art as discussed above as follows. The invention utilizes a graphics display processor to transfer digital data originally stored in the computer memory space to an external device where the data may be processed for diverse applications. The present invention uses the architecture of a conventional computer system, such as a PC, and does not require any additional cost for additional hardware as a consequence of graphics display processors being standard equipment on virtually all PCs. No additional internal hardware is required to perform backup of the CPU memory in accordance with the present invention. No disassembly of the PC is required other than the possible disconnection of the monitor cable. PC backup and restore programs can be stored on a bootable, low-density floppy or other device, such as a PCMCIA card. Co-processor support is not required by a backup program for restoring data of the CPU. Less than 256 K-Bytes of internal CPU memory is used. The backup and recovery of files on a hard drive is possible using the present invention even if the system will not boot from the drive. Furthermore, recovery is possible from damaged boot sectors. The backing up of the memory space of the CPU is easy to use in that all that is required is the insertion of a floppy disk and the turning of the power on. No knowledge is required of the particular type of PC disk controller in use or the encoding format employed by the controller. The only system BIOS disk service used is the "read logical disk sectors" which is provided by all PC BIOS integrated circuits and is not dependent on the operating system. No knowledge of the operating system is required. High sustained data transfer rates such as 921.6 KBytes per second using 60 Hz VGA graphics display formats are possible. A single DOS backup program can be utilized for all Intel-based PCs. The present invention is extendible to systems using microprocessors other than those manufactured by Intel with the interface to the graphics display processors of non-Intel manufactured processors being almost identical to that of Intel-based graphics display processors. The present invention is not dependent upon the memory drive or speed with the invention being based on the video data display rates rather than the speed of an individual CPU or disk drive. As a result, the dual increase of video rates and the speed of the graphics display processors which is ongoing will not obsolete the present invention. The present invention is especially applicable to backup procedures for replicating and safeguarding information contained on PCs which have large internal hard drives and PCs which do not have high-speeds I/O devices built in, such as today's laptops, which have large internal drives. For example, a 340 M-Byte hard disk may be backed up in under seven minutes using the aforementioned 921.6 K-Bytes per second rate. In accordance with the invention, more information may be backed up than with currently available techniques such as disk partition, boot, and FAT sectors, as well as the deleted files are preserved in the backup copy in addition to all of the normal files. The present invention does not depend on the CPU's operating system or the make or model of the graphics display processor or hard drive in the CPU. In large network installations where giga-bytes of storage are involved, the backup process is generally automated, but can take many hours to perform. The present invention can reduce the time required for backup even in large computer systems of this type. A processor for outputting digital data stored in a memory of a computer having a graphics display processor in accordance with the invention includes reading digital data from the memory and processing the digital data to produce at least one serial data stream with the at least one serial data stream including the digital data and clock information, the clock information being a function of a clock signal representative of a rate at which the at least one serial data stream is outputted by a video channel; and serially outputting the at least one serial data stream on the video channel under the control of the graphics processor. The at least one serial data stream also includes display information which permits the at least one serial data stream to be displayed by a video monitor connectable to the video channel; processing at least one of the at least one serial data stream after outputting by the video channel to remove the display information from the processed at least one serial data stream; and producing the clock signal representative of a bit rate at which the at least one serial data stream is outputted by the video channel in response to the clock information in the one of the at least one serial data stream. One of the at least one serial data stream contains the clock signal and the display information and another of the at least one serial data stream contains the digital data and the display information or one of the at least one serial data stream includes the digital data with at least a portion of the one of the at least one serial data stream being encoded with self-clocking information which permits the clock signal representative of a rate at which the one of the at least one serial data stream is outputted by the video channel to be derived from processing at least the portion of the one of the at least one serial data stream outputted from the video channel. The invention further includes processing the one of the at least one serial data stream to remove the self-clocking information. The processing the one of the at least one serial data stream to remove the self-clocking information includes converting the one of the at least one serial data stream into parallel digital data having a number of bits corresponding to a number of bits stored at each addressable location of the memory from which the digital data was read. The at least one serial data stream is outputted in frames formatted for video display, each frame having a set number of lines with each line having bits disposed between periodically occurring horizontal synchronization information with at least a group of bits in each line being encoded with the self-clocking information; the frames are outputted under control of the graphics display processor on the video channel and stored in another memory; and the frames stored in the another memory are read out from the another memory in response to detection of storing the set number of lines in the another memory. Each frame is stored in one of a first and a second frame buffer of the another memory while another frame is being read out of another of the first and second frame buffer with sequential frames stored in the another memory being read out alternatively from the first and second frame buffers during storing of sequential frames outputted on the video channel. The frames read out from the another memory in response to detection of the storing of the set number of lines in the another memory are stored in a memory of a processing system in response to an interrupt signal produced in response to the detection of the storing of the set number of lines in the another memory. The at least one serial data stream comprises a sequence of frames with each frame being serially read out as a series of lines under control of the graphics display processor, each line being formatted into a packet, including the clock information comprising a sync field for producing the clock signal, a scan line field for encoding an address of each line within each frame, a trigger field for encoding a number of a frame within the sequence of frames being outputted on the video channel, and a data field containing data from the block of digital data; and wherein each frame is transmitted with a vertical synchronization pulse and a horizontal synchronization pulse is transmitted with each line. The sync field is processed to produce the clock signal; and the sequence of frames are processed with a clock signal to remove the clock information and to convert each packet into parallel information formatted into groups of bits with each group of bits being equal in number to a number of bits stored at each addressable location in the memory from which the block of digital data was read. The sequence of frames is stored in another memory having first and second frame buffers; and when the trigger field changes in magnitude by one indicating storing of a complete frame from one of the sequence of frames in one of the frame buffers of the another memory, the complete frame is read out from the one of the first and second frame buffers and storing of a subsequent one of the frames is begun in another of the first and second frame buffers while the complete frame is being read out. An interrupt is produced in response to the change in magnitude of the trigger field by one; and the interrupt is received by a processing system which initiates storing of the frame read out from the one of the first and second frame buffers in response to the interrupt in a memory of the processing system. The memory of the processing system has first and second processing system frame buffers and the first and second processing system frame buffers store a sequence of frames in response to the interrupt to cause each of the first and second processing frame buffers to alternatively store a frame. The one serial data stream comprises a sequence of frames with each frame being serially read out as a series of lines under control of the graphics display processor; each frame is processed to remove the self-clocking information while retaining the digital data; the processed frames are stored in a backup memory; and the frames stored in the backup memory are read back into the memory of the computer to restore the digital data originally stored in the memory of the computer. A system for outputting digital data stored in a memory of a computer in accordance with the invention includes a graphics display processor, coupled to the memory, for processing the digital data stored in the memory to produce at least one serial data stream including clock information, which is a function of a clock signal representative of a rate at which the at least one display formatted serial data stream is outputted, and display information for use in controlling a video monitor; a video channel, coupled to the display processor, for outputting the at least one serial data stream produced by the graphics display processor; a data processor memory; and a data processing system, coupled to the video channel and to the data processor memory, for processing the at least one serial data stream in response to the clock information and for removing at least the display information and controlling storing of the at least one of the at least one serial data stream with display information removed in the data processing memory which contains the digital data read from the memory of the computer system. One of the at least one serial data streams contain the clock signal and display information and another of the at least one serial data stream contains the digital data and the display information or one of the at least one serial data stream includes the digital data with at least a portion of the one of the at least one serial data stream being encoded with self-clocking information which permits the clock signal representative of a rate at which the one of the at least one serial data stream is outputted by the video channel to be derived from processing the portion of the one of the at least one serial data stream outputted by the video channel. The invention further includes a storage processor memory; and a storage processor, coupled to the data processing system and to the storage processor memory, for controlling storing of data read from the memory of the data processing system in the storage processor memory in response to the data processing system memory storing a data block of set size. The set size is a full frame of information formatted by the graphics display processor for display by the video monitor. The at least one serial data stream encoded with self-clocking information comprises frames having a set number of lines and bits disposed between periodically occurring horizontal synchronization information with at least a group of bits in each line being encoded with the self-clocking information. Each line is formatted into a packet including a sync field for use in detecting the clock signal and a data field containing the digital data stored in the memory of the computer; and the data processing system comprises a clock, responsive to the sync field, for producing the clock signal and a data separator, responsive to the clock signal and the lines, for removing the self-clocking information and converting the lines into parallel digital data having a number of bits equal to a number of bits stored at each addressable location of the memory of the computer. Each packet further includes a scan line field for encoding an address of each line within each frame and a trigger bit field for encoding a number of a frame within a sequence of frames outputted by the video channel; and each frame is outputted with a vertical synchronization pulse transmitted with each frame at a horizontal synchronization pulse transmitted with each line. The data processing system memory comprises first and second frame buffers; and wherein when the trigger field changes in magnitude by one indicating storing of a complete frame from one of the sequence of frames in one of the frame buffers of the data processing system memory, the data processing system causes the complete frame to be read out from one of the first and second frame buffers and controls storing of another one of the frames and another of the first and second frame buffers of the data processing system memory. The data processing system produces an interrupt in response to the change in magnitude of the trigger field by one; and in response to reception of the interrupt from the data processor, the storage processor initiates storing of a frame read out from the one of the first and second frame buffers of the data processing system memory in the storage processor memory The storage processor memory comprises first and second frame buffers, each storage processor frame buffer storing a frame in response to the reception of the interrupt from the data processing system with the storage processor first and second frame buffers alternatively storing and outputting a frame. The storage processor memory is a backup memory and the storage processor writes frames stored in the storage processor back into the memory of the computer to restore the original digital data. A system for outputting digital data stored in a memory of the computer in accordance with the invention includes a graphics display processor, coupled to the memory, for processing the digital data stored in the memory to produce at least one serial data stream including the digital data and the clock information which is a function of a clock signal representative of a rate at which at least one serial data stream is outputted; and a video channel, coupled to the display processor, for outputting the at least one serial data stream produced by the graphics display processor. The clock information comprises at least a portion of one of the at least one serial data stream encoded with self-clocking information which permits the clock signal to be derived from processing at least the portion of one of the at least one serial data stream. The at least one serial data stream further comprises display information for use in controlling a video monitor. This is a continuation of application Ser. No. 08/545,508, filed Oct. 19, 1995, now U.S. Pat. No. 5,835,102. Patent Citations Cited PatentFiling datePublication dateApplicantTitle US5608424 *Oct 25, 1993Mar 4, 1997Nintendo Co., Ltd.Moving picture display apparatus and external memory used therefor US5673204 *Jul 6, 1995Sep 30, 1997Sun Microsystems, Inc.Loopback video preview for a computer display US5684968 *Jun 16, 1995Nov 4, 1997International Business Machines CorporationMethod and apparatus for processing multiple images concurrently Referenced by Citing PatentFiling datePublication dateApplicantTitle US6574424 *Jun 25, 1999Jun 3, 2003International Business Machines CorporationMethod and apparatus for a randomizer for DVD video US7433901 *Sep 17, 2004Oct 7, 2008International Business Machines CorporationApparatus, system, and method for using multiple criteria to determine collocation granularity for a data source US7827112 *Feb 28, 2002Nov 2, 2010Sony CorporationDigital signal processing apparatus, digital signal processing method, information center, and data delivery system US7853530 *Nov 15, 2006Dec 14, 2010Sony CorporationDigital signal processing apparatus, digital signal processing method, information center, and data delivery system US7925202 *Mar 7, 2007Apr 12, 2011Thomson LicensingPortable communication device for an advanced display US20110221757 *Dec 28, 2010Sep 15, 2011Via Technologies, Inc.Graphics Display Systems and Methods Classifications U.S. Classification345/530, 714/E11.12 International ClassificationG06F11/14 Cooperative ClassificationG06F11/1456 European ClassificationG06F11/14A10H Legal Events DateCodeEventDescription May 8, 2012FPExpired due to failure to pay maintenance fee Effective date: 20120321 Mar 21, 2012LAPSLapse for failure to pay maintenance fees Oct 31, 2011REMIMaintenance fee reminder mailed Jul 18, 2007FPAYFee payment Year of fee payment: 8 Sep 4, 2003FPAYFee payment Year of fee payment: 4
__label__pos
0.751779
cosmetic: simplify some constructs [libav.git] / ffserver.c CommitLineData 85f07f22 FB 1/* 2 * Multiple format streaming server 773a21b8 3 * Copyright (c) 2000, 2001, 2002 Fabrice Bellard 85f07f22 4 * b78e7197 DB 5 * This file is part of FFmpeg. 6 * 7 * FFmpeg is free software; you can redistribute it and/or 773a21b8 FB 8 * modify it under the terms of the GNU Lesser General Public 9 * License as published by the Free Software Foundation; either b78e7197 10 * version 2.1 of the License, or (at your option) any later version. 85f07f22 11 * b78e7197 12 * FFmpeg is distributed in the hope that it will be useful, 85f07f22 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 773a21b8 FB 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 15 * Lesser General Public License for more details. 85f07f22 16 * 773a21b8 17 * You should have received a copy of the GNU Lesser General Public b78e7197 18 * License along with FFmpeg; if not, write to the Free Software 5509bffa 19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 85f07f22 20 */ 364a9607 21 0f4e8165 RB 22#include "config.h" 23#if HAVE_CLOSESOCKET != 1 24#define closesocket close 25#endif 26#include <string.h> 27#include <stdlib.h> 773a21b8 FB 28#include "avformat.h" 29 85f07f22 30#include <stdarg.h> 85f07f22 FB 31#include <unistd.h> 32#include <fcntl.h> 33#include <sys/ioctl.h> b0c858d8 34#ifdef HAVE_SYS_POLL_H 85f07f22 35#include <sys/poll.h> b0c858d8 36#endif 85f07f22 FB 37#include <errno.h> 38#include <sys/time.h> 4568325a 39#undef time //needed because HAVE_AV_CONFIG_H is defined on top 85f07f22 40#include <time.h> 5eb765ef 41#include <sys/wait.h> 85f07f22 42#include <signal.h> 18b67ae5 43#ifdef HAVE_DLFCN_H 2effd274 44#include <dlfcn.h> 6638d424 45#endif 2effd274 46 0bdacf29 47#include "network.h" f1cc88a5 48#include "version.h" 2effd274 49#include "ffserver.h" 1df93ae9 50#include "random.h" f7d78f36 51#include "avstring.h" 85f07f22 52 c367d067 MN 53#undef exit 54 85f07f22 FB 55/* maximum number of simultaneous HTTP connections */ 56#define HTTP_MAX_CONNECTIONS 2000 57 58enum HTTPState { 59 HTTPSTATE_WAIT_REQUEST, 60 HTTPSTATE_SEND_HEADER, 61 HTTPSTATE_SEND_DATA_HEADER, 2effd274 62 HTTPSTATE_SEND_DATA, /* sending TCP or UDP data */ 85f07f22 63 HTTPSTATE_SEND_DATA_TRAILER, 115329f1 64 HTTPSTATE_RECEIVE_DATA, 2effd274 65 HTTPSTATE_WAIT_FEED, /* wait for data from the feed */ 2effd274 FB 66 HTTPSTATE_READY, 67 68 RTSPSTATE_WAIT_REQUEST, 69 RTSPSTATE_SEND_REPLY, bc351386 70 RTSPSTATE_SEND_PACKET, 85f07f22 FB 71}; 72 73const char *http_state[] = { 2effd274 FB 74 "HTTP_WAIT_REQUEST", 75 "HTTP_SEND_HEADER", 76 85f07f22 FB 77 "SEND_DATA_HEADER", 78 "SEND_DATA", 79 "SEND_DATA_TRAILER", 80 "RECEIVE_DATA", 81 "WAIT_FEED", 2effd274 FB 82 "READY", 83 84 "RTSP_WAIT_REQUEST", 85 "RTSP_SEND_REPLY", bc351386 86 "RTSP_SEND_PACKET", 85f07f22 FB 87}; 88 cde25790 89#define IOBUFFER_INIT_SIZE 8192 85f07f22 90 85f07f22 91/* timeouts are in ms */ 2effd274 FB 92#define HTTP_REQUEST_TIMEOUT (15 * 1000) 93#define RTSP_REQUEST_TIMEOUT (3600 * 24 * 1000) 94 85f07f22 FB 95#define SYNC_TIMEOUT (10 * 1000) 96 5eb765ef 97typedef struct { 0c1a9eda 98 int64_t count1, count2; c3f58185 99 int64_t time1, time2; 5eb765ef PG 100} DataRateData; 101 85f07f22 FB 102/* context associated with one connection */ 103typedef struct HTTPContext { 104 enum HTTPState state; 105 int fd; /* socket file descriptor */ 106 struct sockaddr_in from_addr; /* origin */ 107 struct pollfd *poll_entry; /* used when polling */ c3f58185 108 int64_t timeout; 0c1a9eda 109 uint8_t *buffer_ptr, *buffer_end; 85f07f22 110 int http_error; edfdd798 111 int post; 85f07f22 112 struct HTTPContext *next; 42a63c6a 113 int got_key_frame; /* stream 0 => 1, stream 1 => 2, stream 2=> 4 */ 0c1a9eda 114 int64_t data_count; 85f07f22 FB 115 /* feed input */ 116 int feed_fd; 117 /* input format handling */ 118 AVFormatContext *fmt_in; c3f58185 119 int64_t start_time; /* In milliseconds - this wraps fairly often */ 0c1a9eda 120 int64_t first_pts; /* initial pts value */ e240a0bb FB 121 int64_t cur_pts; /* current pts value from the stream in us */ 122 int64_t cur_frame_duration; /* duration of the current frame in us */ 123 int cur_frame_bytes; /* output frame size, needed to compute 124 the time at which we send each 125 packet */ 126 int pts_stream_index; /* stream we choose as clock reference */ 127 int64_t cur_clock; /* current clock reference value in us */ 85f07f22 FB 128 /* output format handling */ 129 struct FFStream *stream; cde25790 PG 130 /* -1 is invalid stream */ 131 int feed_streams[MAX_STREAMS]; /* index of streams in the feed */ 132 int switch_feed_streams[MAX_STREAMS]; /* index of streams in the feed */ 133 int switch_pending; 2effd274 134 AVFormatContext fmt_ctx; /* instance of FFStream for one user */ 85f07f22 135 int last_packet_sent; /* true if last data packet was sent */ 7434ba6d 136 int suppress_log; 5eb765ef 137 DataRateData datarate; 3120d2a2 138 int wmp_client_id; 7434ba6d PG 139 char protocol[16]; 140 char method[16]; 141 char url[128]; cde25790 142 int buffer_size; 0c1a9eda 143 uint8_t *buffer; 2effd274 FB 144 int is_packetized; /* if true, the stream is packetized */ 145 int packet_stream_index; /* current stream for output in state machine */ 115329f1 146 2effd274 147 /* RTSP state specific */ 0c1a9eda 148 uint8_t *pb_buffer; /* XXX: use that in all the code */ 2effd274 FB 149 ByteIOContext *pb; 150 int seq; /* RTSP sequence number */ 115329f1 151 2effd274 FB 152 /* RTP state specific */ 153 enum RTSPProtocol rtp_protocol; 154 char session_id[32]; /* session id */ 155 AVFormatContext *rtp_ctx[MAX_STREAMS]; e240a0bb 156 bc351386 FB 157 /* RTP/UDP specific */ 158 URLContext *rtp_handles[MAX_STREAMS]; 159 160 /* RTP/TCP specific */ 161 struct HTTPContext *rtsp_c; 162 uint8_t *packet_buffer, *packet_buffer_ptr, *packet_buffer_end; 85f07f22 FB 163} HTTPContext; 164 a4d70941 PG 165static AVFrame dummy_frame; 166 85f07f22 FB 167/* each generated stream is described here */ 168enum StreamType { 169 STREAM_TYPE_LIVE, 170 STREAM_TYPE_STATUS, cde25790 171 STREAM_TYPE_REDIRECT, 85f07f22 FB 172}; 173 8256c0a3 PG 174enum IPAddressAction { 175 IP_ALLOW = 1, 176 IP_DENY, 177}; 178 179typedef struct IPAddressACL { 180 struct IPAddressACL *next; 181 enum IPAddressAction action; efa04ce2 182 /* These are in host order */ 8256c0a3 PG 183 struct in_addr first; 184 struct in_addr last; 185} IPAddressACL; 186 85f07f22 FB 187/* description of each stream of the ffserver.conf file */ 188typedef struct FFStream { 189 enum StreamType stream_type; 190 char filename[1024]; /* stream filename */ 2effd274 FB 191 struct FFStream *feed; /* feed we are using (can be null if 192 coming from file) */ e240a0bb FB 193 AVFormatParameters *ap_in; /* input parameters */ 194 AVInputFormat *ifmt; /* if non NULL, force input format */ bd7cf6ad 195 AVOutputFormat *fmt; 8256c0a3 196 IPAddressACL *acl; 85f07f22 197 int nb_streams; 42a63c6a 198 int prebuffer; /* Number of millseconds early to start */ c3f58185 199 int64_t max_time; /* Number of milliseconds to run */ 79c4ea3c 200 int send_on_key; 85f07f22 FB 201 AVStream *streams[MAX_STREAMS]; 202 int feed_streams[MAX_STREAMS]; /* index of streams in the feed */ 203 char feed_filename[1024]; /* file name of the feed storage, or 204 input file name for a stream */ 2ac887ba PG 205 char author[512]; 206 char title[512]; 207 char copyright[512]; 208 char comment[512]; cde25790 209 pid_t pid; /* Of ffmpeg process */ 5eb765ef 210 time_t pid_start; /* Of ffmpeg process */ cde25790 211 char **child_argv; 85f07f22 212 struct FFStream *next; 6edd6884 213 int bandwidth; /* bandwidth, in kbits/s */ 2effd274 FB 214 /* RTSP options */ 215 char *rtsp_option; 829ac53d FB 216 /* multicast specific */ 217 int is_multicast; 218 struct in_addr multicast_ip; 219 int multicast_port; /* first port used for multicast */ 6edd6884 FB 220 int multicast_ttl; 221 int loop; /* if true, send the stream in loops (only meaningful if file) */ 829ac53d 222 85f07f22 223 /* feed specific */ 2effd274 224 int feed_opened; /* true if someone is writing to the feed */ 85f07f22 225 int is_feed; /* true if it is a feed */ e322ea48 226 int readonly; /* True if writing is prohibited to the file */ a6e14edd 227 int conns_served; 0c1a9eda 228 int64_t bytes_served; 6b0bdc75 229 int64_t feed_max_size; /* maximum storage size, zero means unlimited */ 0c1a9eda ZK 230 int64_t feed_write_index; /* current write position in feed (it wraps round) */ 231 int64_t feed_size; /* current size of feed */ 85f07f22 FB 232 struct FFStream *next_feed; 233} FFStream; 234 235typedef struct FeedData { 236 long long data_count; 237 float avg_frame_size; /* frame size averraged over last frames with exponential mean */ 238} FeedData; 239 18405874 AB 240static struct sockaddr_in my_http_addr; 241static struct sockaddr_in my_rtsp_addr; 2effd274 242 33f5e2ec AB 243static char logfilename[1024]; 244static HTTPContext *first_http_ctx; 245static FFStream *first_feed; /* contains only feeds */ 246static FFStream *first_stream; /* contains all streams, including feeds */ 85f07f22 247 2effd274 FB 248static void new_connection(int server_fd, int is_rtsp); 249static void close_connection(HTTPContext *c); 250 251/* HTTP handling */ 252static int handle_connection(HTTPContext *c); 85f07f22 253static int http_parse_request(HTTPContext *c); 5eb765ef 254static int http_send_data(HTTPContext *c); 85f07f22 FB 255static void compute_stats(HTTPContext *c); 256static int open_input_stream(HTTPContext *c, const char *info); 257static int http_start_receive_data(HTTPContext *c); 258static int http_receive_data(HTTPContext *c); 2effd274 FB 259 260/* RTSP handling */ 261static int rtsp_parse_request(HTTPContext *c); 262static void rtsp_cmd_describe(HTTPContext *c, const char *url); 0df65975 263static void rtsp_cmd_options(HTTPContext *c, const char *url); 2effd274 FB 264static void rtsp_cmd_setup(HTTPContext *c, const char *url, RTSPHeader *h); 265static void rtsp_cmd_play(HTTPContext *c, const char *url, RTSPHeader *h); 266static void rtsp_cmd_pause(HTTPContext *c, const char *url, RTSPHeader *h); 267static void rtsp_cmd_teardown(HTTPContext *c, const char *url, RTSPHeader *h); 268 829ac53d 269/* SDP handling */ 115329f1 270static int prepare_sdp_description(FFStream *stream, uint8_t **pbuffer, 829ac53d FB 271 struct in_addr my_ip); 272 2effd274 273/* RTP handling */ 115329f1 274static HTTPContext *rtp_new_connection(struct sockaddr_in *from_addr, bc351386 FB 275 FFStream *stream, const char *session_id, 276 enum RTSPProtocol rtp_protocol); 115329f1 277static int rtp_new_av_stream(HTTPContext *c, bc351386 FB 278 int stream_index, struct sockaddr_in *dest_addr, 279 HTTPContext *rtsp_c); 85f07f22 280 cde25790 281static const char *my_program_name; d6562d2c 282static const char *my_program_dir; cde25790 283 2ac887ba 284static int ffserver_debug; 2effd274 285static int ffserver_daemon; 2ac887ba 286static int no_launch; 5eb765ef 287static int need_to_start_children; 2ac887ba 288 33f5e2ec AB 289static int nb_max_connections; 290static int nb_connections; 85f07f22 291 33f5e2ec AB 292static int max_bandwidth; 293static int current_bandwidth; 42a63c6a 294 c3f58185 295static int64_t cur_time; // Making this global saves on passing it around everywhere 5eb765ef 296 1df93ae9 AB 297static AVRandomState random_state; 298 85f07f22 FB 299static FILE *logfile = NULL; 300 115329f1 301static void __attribute__ ((format (printf, 1, 2))) http_log(const char *fmt, ...) 85f07f22 FB 302{ 303 va_list ap; 304 va_start(ap, fmt); 115329f1 305 7434ba6d 306 if (logfile) { 85f07f22 307 vfprintf(logfile, fmt, ap); 7434ba6d PG 308 fflush(logfile); 309 } 85f07f22 FB 310 va_end(ap); 311} 312 6edd6884 313static char *ctime1(char *buf2) 7434ba6d 314{ 7434ba6d 315 time_t ti; 6edd6884 316 char *p; 7434ba6d 317 7434ba6d PG 318 ti = time(NULL); 319 p = ctime(&ti); 320 strcpy(buf2, p); 321 p = buf2 + strlen(p) - 1; 322 if (*p == '\n') 323 *p = '\0'; 6edd6884 FB 324 return buf2; 325} 326 327static void log_connection(HTTPContext *c) 328{ 329 char buf2[32]; 330 115329f1 331 if (c->suppress_log) 6edd6884 FB 332 return; 333 115329f1 DB 334 http_log("%s - - [%s] \"%s %s %s\" %d %"PRId64"\n", 335 inet_ntoa(c->from_addr.sin_addr), 336 ctime1(buf2), c->method, c->url, 6edd6884 337 c->protocol, (c->http_error ? c->http_error : 200), c->data_count); cde25790 PG 338} 339 0c1a9eda 340static void update_datarate(DataRateData *drd, int64_t count) 5eb765ef PG 341{ 342 if (!drd->time1 && !drd->count1) { 343 drd->time1 = drd->time2 = cur_time; 344 drd->count1 = drd->count2 = count; eeffbdea 345 } else if (cur_time - drd->time2 > 5000) { 33a4ecbe AB 346 drd->time1 = drd->time2; 347 drd->count1 = drd->count2; 348 drd->time2 = cur_time; 349 drd->count2 = count; 5eb765ef PG 350 } 351} 352 353/* In bytes per second */ 0c1a9eda 354static int compute_datarate(DataRateData *drd, int64_t count) 5eb765ef PG 355{ 356 if (cur_time == drd->time1) 357 return 0; 115329f1 358 5eb765ef PG 359 return ((count - drd->count1) * 1000) / (cur_time - drd->time1); 360} 361 a782f209 362 cde25790 PG 363static void start_children(FFStream *feed) 364{ 2ac887ba PG 365 if (no_launch) 366 return; 367 cde25790 368 for (; feed; feed = feed->next) { 5eb765ef PG 369 if (feed->child_argv && !feed->pid) { 370 feed->pid_start = time(0); 371 cde25790 PG 372 feed->pid = fork(); 373 374 if (feed->pid < 0) { 375 fprintf(stderr, "Unable to create children\n"); 376 exit(1); 377 } 378 if (!feed->pid) { 379 /* In child */ 380 char pathname[1024]; 381 char *slash; 382 int i; 383 5eb765ef PG 384 for (i = 3; i < 256; i++) { 385 close(i); 386 } cde25790 387 5eb765ef 388 if (!ffserver_debug) { 2ac887ba PG 389 i = open("/dev/null", O_RDWR); 390 if (i) 391 dup2(i, 0); 392 dup2(i, 1); 393 dup2(i, 2); 5eb765ef PG 394 if (i) 395 close(i); 2ac887ba 396 } cde25790 397 f7d78f36 398 av_strlcpy(pathname, my_program_name, sizeof(pathname)); cde25790 PG 399 400 slash = strrchr(pathname, '/'); 401 if (!slash) { 402 slash = pathname; 403 } else { 404 slash++; 405 } 406 strcpy(slash, "ffmpeg"); 407 d6562d2c PG 408 /* This is needed to make relative pathnames work */ 409 chdir(my_program_dir); 410 a4d70941 PG 411 signal(SIGPIPE, SIG_DFL); 412 cde25790 PG 413 execvp(pathname, feed->child_argv); 414 415 _exit(1); 416 } 417 } 418 } 7434ba6d PG 419} 420 2effd274 FB 421/* open a listening socket */ 422static int socket_open_listen(struct sockaddr_in *my_addr) 85f07f22 423{ 2effd274 424 int server_fd, tmp; 85f07f22 FB 425 426 server_fd = socket(AF_INET,SOCK_STREAM,0); 427 if (server_fd < 0) { 428 perror ("socket"); 429 return -1; 430 } 115329f1 431 85f07f22 FB 432 tmp = 1; 433 setsockopt(server_fd, SOL_SOCKET, SO_REUSEADDR, &tmp, sizeof(tmp)); 434 2effd274 435 if (bind (server_fd, (struct sockaddr *) my_addr, sizeof (*my_addr)) < 0) { b17d099d PG 436 char bindmsg[32]; 437 snprintf(bindmsg, sizeof(bindmsg), "bind(port %d)", ntohs(my_addr->sin_port)); 438 perror (bindmsg); d96633bb 439 closesocket(server_fd); 85f07f22 FB 440 return -1; 441 } 115329f1 442 85f07f22 FB 443 if (listen (server_fd, 5) < 0) { 444 perror ("listen"); d96633bb 445 closesocket(server_fd); 85f07f22 FB 446 return -1; 447 } ba472aaf 448 ff_socket_nonblock(server_fd, 1); 2effd274 FB 449 450 return server_fd; 451} 452 6edd6884 FB 453/* start all multicast streams */ 454static void start_multicast(void) 455{ 456 FFStream *stream; 457 char session_id[32]; 458 HTTPContext *rtp_c; 459 struct sockaddr_in dest_addr; 460 int default_port, stream_index; 461 462 default_port = 6000; 463 for(stream = first_stream; stream != NULL; stream = stream->next) { 464 if (stream->is_multicast) { 465 /* open the RTP connection */ 1df93ae9 AB 466 snprintf(session_id, sizeof(session_id), "%08x%08x", 467 av_random(&random_state), av_random(&random_state)); 6edd6884 FB 468 469 /* choose a port if none given */ 470 if (stream->multicast_port == 0) { 471 stream->multicast_port = default_port; 472 default_port += 100; 473 } 474 475 dest_addr.sin_family = AF_INET; 476 dest_addr.sin_addr = stream->multicast_ip; 477 dest_addr.sin_port = htons(stream->multicast_port); 478 115329f1 479 rtp_c = rtp_new_connection(&dest_addr, stream, session_id, bc351386 480 RTSP_PROTOCOL_RTP_UDP_MULTICAST); 6edd6884 FB 481 if (!rtp_c) { 482 continue; 483 } 484 if (open_input_stream(rtp_c, "") < 0) { 115329f1 485 fprintf(stderr, "Could not open input stream for stream '%s'\n", 6edd6884 FB 486 stream->filename); 487 continue; 488 } 489 6edd6884 490 /* open each RTP stream */ 115329f1 491 for(stream_index = 0; stream_index < stream->nb_streams; 6edd6884 492 stream_index++) { 115329f1 493 dest_addr.sin_port = htons(stream->multicast_port + 6edd6884 494 2 * stream_index); bc351386 495 if (rtp_new_av_stream(rtp_c, stream_index, &dest_addr, NULL) < 0) { 115329f1 496 fprintf(stderr, "Could not open output stream '%s/streamid=%d'\n", 0fa45e19 FB 497 stream->filename, stream_index); 498 exit(1); 6edd6884 FB 499 } 500 } 501 502 /* change state to send data */ 503 rtp_c->state = HTTPSTATE_SEND_DATA; 504 } 505 } 506} 2effd274 FB 507 508/* main loop of the http server */ 509static int http_server(void) 510{ 511 int server_fd, ret, rtsp_server_fd, delay, delay1; 512 struct pollfd poll_table[HTTP_MAX_CONNECTIONS + 2], *poll_entry; 513 HTTPContext *c, *c_next; 514 515 server_fd = socket_open_listen(&my_http_addr); 516 if (server_fd < 0) 517 return -1; 85f07f22 518 2effd274 FB 519 rtsp_server_fd = socket_open_listen(&my_rtsp_addr); 520 if (rtsp_server_fd < 0) 521 return -1; 115329f1 522 85f07f22 FB 523 http_log("ffserver started.\n"); 524 cde25790 PG 525 start_children(first_feed); 526 85f07f22 FB 527 first_http_ctx = NULL; 528 nb_connections = 0; 6edd6884 FB 529 530 start_multicast(); 531 85f07f22 FB 532 for(;;) { 533 poll_entry = poll_table; 534 poll_entry->fd = server_fd; 535 poll_entry->events = POLLIN; 536 poll_entry++; 537 2effd274 FB 538 poll_entry->fd = rtsp_server_fd; 539 poll_entry->events = POLLIN; 540 poll_entry++; 541 85f07f22 FB 542 /* wait for events on each HTTP handle */ 543 c = first_http_ctx; 2effd274 544 delay = 1000; 85f07f22 FB 545 while (c != NULL) { 546 int fd; 547 fd = c->fd; 548 switch(c->state) { 2effd274 FB 549 case HTTPSTATE_SEND_HEADER: 550 case RTSPSTATE_SEND_REPLY: bc351386 551 case RTSPSTATE_SEND_PACKET: 85f07f22 FB 552 c->poll_entry = poll_entry; 553 poll_entry->fd = fd; 2effd274 554 poll_entry->events = POLLOUT; 85f07f22 FB 555 poll_entry++; 556 break; 85f07f22 FB 557 case HTTPSTATE_SEND_DATA_HEADER: 558 case HTTPSTATE_SEND_DATA: 559 case HTTPSTATE_SEND_DATA_TRAILER: 2effd274 FB 560 if (!c->is_packetized) { 561 /* for TCP, we output as much as we can (may need to put a limit) */ 562 c->poll_entry = poll_entry; 563 poll_entry->fd = fd; 564 poll_entry->events = POLLOUT; 565 poll_entry++; 566 } else { e240a0bb FB 567 /* when ffserver is doing the timing, we work by 568 looking at which packet need to be sent every 569 10 ms */ 570 delay1 = 10; /* one tick wait XXX: 10 ms assumed */ 571 if (delay1 < delay) 572 delay = delay1; 2effd274 573 } 85f07f22 574 break; 2effd274 575 case HTTPSTATE_WAIT_REQUEST: 85f07f22 576 case HTTPSTATE_RECEIVE_DATA: 85f07f22 577 case HTTPSTATE_WAIT_FEED: 2effd274 578 case RTSPSTATE_WAIT_REQUEST: 85f07f22 FB 579 /* need to catch errors */ 580 c->poll_entry = poll_entry; 581 poll_entry->fd = fd; a6e14edd 582 poll_entry->events = POLLIN;/* Maybe this will work */ 85f07f22 FB 583 poll_entry++; 584 break; 585 default: 586 c->poll_entry = NULL; 587 break; 588 } 589 c = c->next; 590 } 591 592 /* wait for an event on one connection. We poll at least every 593 second to handle timeouts */ 594 do { 2effd274 595 ret = poll(poll_table, poll_entry - poll_table, delay); 8da4034f AB 596 if (ret < 0 && ff_neterrno() != FF_NETERROR(EAGAIN) && 597 ff_neterrno() != FF_NETERROR(EINTR)) 53e2f9ca 598 return -1; e8d658df 599 } while (ret < 0); 115329f1 600 c3f58185 601 cur_time = av_gettime() / 1000; 85f07f22 602 5eb765ef PG 603 if (need_to_start_children) { 604 need_to_start_children = 0; 605 start_children(first_feed); 606 } 607 85f07f22 608 /* now handle the events */ 2effd274 FB 609 for(c = first_http_ctx; c != NULL; c = c_next) { 610 c_next = c->next; 611 if (handle_connection(c) < 0) { 85f07f22 612 /* close and free the connection */ 7434ba6d 613 log_connection(c); 2effd274 614 close_connection(c); 85f07f22 FB 615 } 616 } 617 85f07f22 618 poll_entry = poll_table; 2effd274 619 /* new HTTP connection request ? */ 85f07f22 620 if (poll_entry->revents & POLLIN) { 2effd274 621 new_connection(server_fd, 0); 85f07f22 FB 622 } 623 poll_entry++; 2effd274 FB 624 /* new RTSP connection request ? */ 625 if (poll_entry->revents & POLLIN) { 626 new_connection(rtsp_server_fd, 1); 627 } 85f07f22 FB 628 } 629} 630 2effd274 FB 631/* start waiting for a new HTTP/RTSP request */ 632static void start_wait_request(HTTPContext *c, int is_rtsp) 85f07f22 633{ 2effd274 FB 634 c->buffer_ptr = c->buffer; 635 c->buffer_end = c->buffer + c->buffer_size - 1; /* leave room for '\0' */ 636 637 if (is_rtsp) { 638 c->timeout = cur_time + RTSP_REQUEST_TIMEOUT; 639 c->state = RTSPSTATE_WAIT_REQUEST; 640 } else { 641 c->timeout = cur_time + HTTP_REQUEST_TIMEOUT; 642 c->state = HTTPSTATE_WAIT_REQUEST; 643 } 644} 645 646static void new_connection(int server_fd, int is_rtsp) 647{ 648 struct sockaddr_in from_addr; 649 int fd, len; 650 HTTPContext *c = NULL; 651 652 len = sizeof(from_addr); 115329f1 653 fd = accept(server_fd, (struct sockaddr *)&from_addr, 2effd274 FB 654 &len); 655 if (fd < 0) 656 return; ba472aaf 657 ff_socket_nonblock(fd, 1); 2effd274 FB 658 659 /* XXX: should output a warning page when coming 660 close to the connection limit */ 661 if (nb_connections >= nb_max_connections) 662 goto fail; 115329f1 663 2effd274 FB 664 /* add a new connection */ 665 c = av_mallocz(sizeof(HTTPContext)); 666 if (!c) 667 goto fail; 115329f1 668 2effd274 FB 669 c->fd = fd; 670 c->poll_entry = NULL; 671 c->from_addr = from_addr; 672 c->buffer_size = IOBUFFER_INIT_SIZE; 673 c->buffer = av_malloc(c->buffer_size); 674 if (!c->buffer) 675 goto fail; 8bc80f8b PG 676 677 c->next = first_http_ctx; 678 first_http_ctx = c; 2effd274 679 nb_connections++; 115329f1 680 2effd274 FB 681 start_wait_request(c, is_rtsp); 682 683 return; 684 685 fail: 686 if (c) { 687 av_free(c->buffer); 688 av_free(c); 689 } d96633bb 690 closesocket(fd); 2effd274 FB 691} 692 693static void close_connection(HTTPContext *c) 694{ 695 HTTPContext **cp, *c1; 696 int i, nb_streams; 697 AVFormatContext *ctx; 698 URLContext *h; 699 AVStream *st; 700 701 /* remove connection from list */ 702 cp = &first_http_ctx; 703 while ((*cp) != NULL) { 704 c1 = *cp; 705 if (c1 == c) { 706 *cp = c->next; 707 } else { 708 cp = &c1->next; 709 } 710 } 711 bc351386 FB 712 /* remove references, if any (XXX: do it faster) */ 713 for(c1 = first_http_ctx; c1 != NULL; c1 = c1->next) { 714 if (c1->rtsp_c == c) 715 c1->rtsp_c = NULL; 716 } 717 2effd274 FB 718 /* remove connection associated resources */ 719 if (c->fd >= 0) d96633bb 720 closesocket(c->fd); 2effd274 FB 721 if (c->fmt_in) { 722 /* close each frame parser */ 723 for(i=0;i<c->fmt_in->nb_streams;i++) { 724 st = c->fmt_in->streams[i]; 01f4895c MN 725 if (st->codec->codec) { 726 avcodec_close(st->codec); 2effd274 FB 727 } 728 } 729 av_close_input_file(c->fmt_in); 730 } 731 732 /* free RTP output streams if any */ 733 nb_streams = 0; 115329f1 734 if (c->stream) 2effd274 735 nb_streams = c->stream->nb_streams; 115329f1 736 2effd274 FB 737 for(i=0;i<nb_streams;i++) { 738 ctx = c->rtp_ctx[i]; 739 if (ctx) { 740 av_write_trailer(ctx); 741 av_free(ctx); 742 } 743 h = c->rtp_handles[i]; 744 if (h) { 745 url_close(h); 746 } 747 } 115329f1 748 b88ba823 MH 749 ctx = &c->fmt_ctx; 750 87638494 751 if (!c->last_packet_sent) { 87638494 PG 752 if (ctx->oformat) { 753 /* prepare header */ 754 if (url_open_dyn_buf(&ctx->pb) >= 0) { 755 av_write_trailer(ctx); bc351386 756 url_close_dyn_buf(&ctx->pb, &c->pb_buffer); 87638494 PG 757 } 758 } 759 } 760 115329f1 761 for(i=0; i<ctx->nb_streams; i++) 0bd53967 762 av_free(ctx->streams[i]); f0ef6240 763 edfdd798 764 if (c->stream && !c->post && c->stream->stream_type == STREAM_TYPE_LIVE) 6edd6884 765 current_bandwidth -= c->stream->bandwidth; 5400e092 AB 766 767 /* signal that there is no feed if we are the feeder socket */ 768 if (c->state == HTTPSTATE_RECEIVE_DATA && c->stream) { 769 c->stream->feed_opened = 0; 770 close(c->feed_fd); 771 } 772 2effd274 773 av_freep(&c->pb_buffer); bc351386 774 av_freep(&c->packet_buffer); 2effd274 FB 775 av_free(c->buffer); 776 av_free(c); 777 nb_connections--; 778} 779 780static int handle_connection(HTTPContext *c) 781{ 782 int len, ret; 115329f1 783 85f07f22 FB 784 switch(c->state) { 785 case HTTPSTATE_WAIT_REQUEST: 2effd274 786 case RTSPSTATE_WAIT_REQUEST: 85f07f22 FB 787 /* timeout ? */ 788 if ((c->timeout - cur_time) < 0) 789 return -1; 790 if (c->poll_entry->revents & (POLLERR | POLLHUP)) 791 return -1; 792 793 /* no need to read if no events */ 794 if (!(c->poll_entry->revents & POLLIN)) 795 return 0; 796 /* read the data */ 1bc1cfdd 797 read_loop: c60202df 798 len = recv(c->fd, c->buffer_ptr, 1, 0); 85f07f22 799 if (len < 0) { 8da4034f AB 800 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 801 ff_neterrno() != FF_NETERROR(EINTR)) 85f07f22 FB 802 return -1; 803 } else if (len == 0) { 804 return -1; 805 } else { 94d9ad5f 806 /* search for end of request. */ 0c1a9eda 807 uint8_t *ptr; 85f07f22 FB 808 c->buffer_ptr += len; 809 ptr = c->buffer_ptr; 810 if ((ptr >= c->buffer + 2 && !memcmp(ptr-2, "\n\n", 2)) || 811 (ptr >= c->buffer + 4 && !memcmp(ptr-4, "\r\n\r\n", 4))) { 812 /* request found : parse it and reply */ 2effd274 FB 813 if (c->state == HTTPSTATE_WAIT_REQUEST) { 814 ret = http_parse_request(c); 815 } else { 816 ret = rtsp_parse_request(c); 817 } 818 if (ret < 0) 85f07f22 FB 819 return -1; 820 } else if (ptr >= c->buffer_end) { 821 /* request too long: cannot do anything */ 822 return -1; 1bc1cfdd 823 } else goto read_loop; 85f07f22 FB 824 } 825 break; 826 827 case HTTPSTATE_SEND_HEADER: 828 if (c->poll_entry->revents & (POLLERR | POLLHUP)) 829 return -1; 830 2effd274 831 /* no need to write if no events */ 85f07f22 FB 832 if (!(c->poll_entry->revents & POLLOUT)) 833 return 0; c60202df 834 len = send(c->fd, c->buffer_ptr, c->buffer_end - c->buffer_ptr, 0); 85f07f22 835 if (len < 0) { 8da4034f AB 836 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 837 ff_neterrno() != FF_NETERROR(EINTR)) { 85f07f22 838 /* error : close connection */ 2effd274 839 av_freep(&c->pb_buffer); 85f07f22 FB 840 return -1; 841 } 842 } else { 843 c->buffer_ptr += len; 2e04edb3 PG 844 if (c->stream) 845 c->stream->bytes_served += len; a6e14edd 846 c->data_count += len; 85f07f22 847 if (c->buffer_ptr >= c->buffer_end) { 2effd274 848 av_freep(&c->pb_buffer); 85f07f22 849 /* if error, exit */ 2effd274 850 if (c->http_error) { 85f07f22 851 return -1; 2effd274 FB 852 } 853 /* all the buffer was sent : synchronize to the incoming stream */ 85f07f22 FB 854 c->state = HTTPSTATE_SEND_DATA_HEADER; 855 c->buffer_ptr = c->buffer_end = c->buffer; 856 } 857 } 858 break; 859 860 case HTTPSTATE_SEND_DATA: 861 case HTTPSTATE_SEND_DATA_HEADER: 862 case HTTPSTATE_SEND_DATA_TRAILER: 2effd274 FB 863 /* for packetized output, we consider we can always write (the 864 input streams sets the speed). It may be better to verify 865 that we do not rely too much on the kernel queues */ 866 if (!c->is_packetized) { 867 if (c->poll_entry->revents & (POLLERR | POLLHUP)) 868 return -1; 115329f1 869 2effd274 FB 870 /* no need to read if no events */ 871 if (!(c->poll_entry->revents & POLLOUT)) 872 return 0; 873 } 5eb765ef 874 if (http_send_data(c) < 0) 85f07f22 875 return -1; 638831aa AB 876 /* close connection if trailer sent */ 877 if (c->state == HTTPSTATE_SEND_DATA_TRAILER) 878 return -1; 85f07f22 FB 879 break; 880 case HTTPSTATE_RECEIVE_DATA: 881 /* no need to read if no events */ 882 if (c->poll_entry->revents & (POLLERR | POLLHUP)) 883 return -1; 884 if (!(c->poll_entry->revents & POLLIN)) 885 return 0; 886 if (http_receive_data(c) < 0) 887 return -1; 888 break; 889 case HTTPSTATE_WAIT_FEED: 890 /* no need to read if no events */ a6e14edd 891 if (c->poll_entry->revents & (POLLIN | POLLERR | POLLHUP)) 85f07f22 FB 892 return -1; 893 894 /* nothing to do, we'll be waken up by incoming feed packets */ 895 break; 2effd274 896 2effd274 FB 897 case RTSPSTATE_SEND_REPLY: 898 if (c->poll_entry->revents & (POLLERR | POLLHUP)) { 899 av_freep(&c->pb_buffer); 900 return -1; 901 } 902 /* no need to write if no events */ 903 if (!(c->poll_entry->revents & POLLOUT)) 904 return 0; c60202df 905 len = send(c->fd, c->buffer_ptr, c->buffer_end - c->buffer_ptr, 0); 2effd274 906 if (len < 0) { 8da4034f AB 907 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 908 ff_neterrno() != FF_NETERROR(EINTR)) { 2effd274 FB 909 /* error : close connection */ 910 av_freep(&c->pb_buffer); 911 return -1; 912 } 913 } else { 914 c->buffer_ptr += len; 915 c->data_count += len; 916 if (c->buffer_ptr >= c->buffer_end) { 917 /* all the buffer was sent : wait for a new request */ 918 av_freep(&c->pb_buffer); 919 start_wait_request(c, 1); 920 } 921 } 922 break; bc351386 FB 923 case RTSPSTATE_SEND_PACKET: 924 if (c->poll_entry->revents & (POLLERR | POLLHUP)) { 925 av_freep(&c->packet_buffer); 926 return -1; 927 } 928 /* no need to write if no events */ 929 if (!(c->poll_entry->revents & POLLOUT)) 930 return 0; c60202df AB 931 len = send(c->fd, c->packet_buffer_ptr, 932 c->packet_buffer_end - c->packet_buffer_ptr, 0); bc351386 933 if (len < 0) { 8da4034f AB 934 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 935 ff_neterrno() != FF_NETERROR(EINTR)) { bc351386 FB 936 /* error : close connection */ 937 av_freep(&c->packet_buffer); 938 return -1; 939 } 940 } else { 941 c->packet_buffer_ptr += len; 942 if (c->packet_buffer_ptr >= c->packet_buffer_end) { 943 /* all the buffer was sent : wait for a new request */ 944 av_freep(&c->packet_buffer); 945 c->state = RTSPSTATE_WAIT_REQUEST; 946 } 947 } 948 break; 2effd274 FB 949 case HTTPSTATE_READY: 950 /* nothing to do */ 951 break; 85f07f22 FB 952 default: 953 return -1; 954 } 955 return 0; 956} 957 3120d2a2 PG 958static int extract_rates(char *rates, int ratelen, const char *request) 959{ 960 const char *p; 961 962 for (p = request; *p && *p != '\r' && *p != '\n'; ) { 963 if (strncasecmp(p, "Pragma:", 7) == 0) { 964 const char *q = p + 7; 965 966 while (*q && *q != '\n' && isspace(*q)) 967 q++; 968 969 if (strncasecmp(q, "stream-switch-entry=", 20) == 0) { 970 int stream_no; 971 int rate_no; 972 973 q += 20; 974 cde25790 975 memset(rates, 0xff, ratelen); 3120d2a2 PG 976 977 while (1) { 978 while (*q && *q != '\n' && *q != ':') 979 q++; 980 981 if (sscanf(q, ":%d:%d", &stream_no, &rate_no) != 2) { 982 break; 983 } 984 stream_no--; 985 if (stream_no < ratelen && stream_no >= 0) { 986 rates[stream_no] = rate_no; 987 } 988 989 while (*q && *q != '\n' && !isspace(*q)) 990 q++; 991 } 992 993 return 1; 994 } 995 } 996 p = strchr(p, '\n'); 997 if (!p) 998 break; 999 1000 p++; 1001 } 1002 1003 return 0; 1004} 1005 cde25790 1006static int find_stream_in_feed(FFStream *feed, AVCodecContext *codec, int bit_rate) 3120d2a2 PG 1007{ 1008 int i; cde25790 PG 1009 int best_bitrate = 100000000; 1010 int best = -1; 1011 1012 for (i = 0; i < feed->nb_streams; i++) { 01f4895c 1013 AVCodecContext *feed_codec = feed->streams[i]->codec; cde25790 PG 1014 1015 if (feed_codec->codec_id != codec->codec_id || 1016 feed_codec->sample_rate != codec->sample_rate || 1017 feed_codec->width != codec->width || 1018 feed_codec->height != codec->height) { 1019 continue; 1020 } 1021 1022 /* Potential stream */ 1023 115329f1 1024 /* We want the fastest stream less than bit_rate, or the slowest cde25790 PG 1025 * faster than bit_rate 1026 */ 1027 1028 if (feed_codec->bit_rate <= bit_rate) { 1029 if (best_bitrate > bit_rate || feed_codec->bit_rate > best_bitrate) { 1030 best_bitrate = feed_codec->bit_rate; 1031 best = i; 1032 } 1033 } else { 1034 if (feed_codec->bit_rate < best_bitrate) { 1035 best_bitrate = feed_codec->bit_rate; 1036 best = i; 1037 } 1038 } 1039 } 1040 1041 return best; 1042} 1043 1044static int modify_current_stream(HTTPContext *c, char *rates) 1045{ 1046 int i; 1047 FFStream *req = c->stream; 1048 int action_required = 0; 3120d2a2 1049 001bcd29 PG 1050 /* Not much we can do for a feed */ 1051 if (!req->feed) 1052 return 0; 1053 3120d2a2 1054 for (i = 0; i < req->nb_streams; i++) { 01f4895c 1055 AVCodecContext *codec = req->streams[i]->codec; 3120d2a2 1056 3120d2a2 PG 1057 switch(rates[i]) { 1058 case 0: cde25790 1059 c->switch_feed_streams[i] = req->feed_streams[i]; 3120d2a2 PG 1060 break; 1061 case 1: cde25790 1062 c->switch_feed_streams[i] = find_stream_in_feed(req->feed, codec, codec->bit_rate / 2); 3120d2a2 PG 1063 break; 1064 case 2: cde25790 PG 1065 /* Wants off or slow */ 1066 c->switch_feed_streams[i] = find_stream_in_feed(req->feed, codec, codec->bit_rate / 4); 1067#ifdef WANTS_OFF 1068 /* This doesn't work well when it turns off the only stream! */ 1069 c->switch_feed_streams[i] = -2; 1070 c->feed_streams[i] = -2; 1071#endif 3120d2a2 PG 1072 break; 1073 } 3120d2a2 1074 cde25790 PG 1075 if (c->switch_feed_streams[i] >= 0 && c->switch_feed_streams[i] != c->feed_streams[i]) 1076 action_required = 1; 1077 } 3120d2a2 1078 cde25790 PG 1079 return action_required; 1080} 3120d2a2 1081 3120d2a2 1082 cde25790 PG 1083static void do_switch_stream(HTTPContext *c, int i) 1084{ 1085 if (c->switch_feed_streams[i] >= 0) { 115329f1 1086#ifdef PHILIP cde25790 PG 1087 c->feed_streams[i] = c->switch_feed_streams[i]; 1088#endif 3120d2a2 1089 cde25790 1090 /* Now update the stream */ 3120d2a2 1091 } cde25790 1092 c->switch_feed_streams[i] = -1; 3120d2a2 1093} 7434ba6d 1094 2effd274 FB 1095/* XXX: factorize in utils.c ? */ 1096/* XXX: take care with different space meaning */ 1097static void skip_spaces(const char **pp) 1098{ 1099 const char *p; 1100 p = *pp; 1101 while (*p == ' ' || *p == '\t') 1102 p++; 1103 *pp = p; 1104} 1105 1106static void get_word(char *buf, int buf_size, const char **pp) 1107{ 1108 const char *p; 1109 char *q; 1110 1111 p = *pp; 1112 skip_spaces(&p); 1113 q = buf; 1114 while (!isspace(*p) && *p != '\0') { 1115 if ((q - buf) < buf_size - 1) 1116 *q++ = *p; 1117 p++; 1118 } 1119 if (buf_size > 0) 1120 *q = '\0'; 1121 *pp = p; 1122} 1123 8256c0a3 PG 1124static int validate_acl(FFStream *stream, HTTPContext *c) 1125{ 1126 enum IPAddressAction last_action = IP_DENY; 1127 IPAddressACL *acl; 1128 struct in_addr *src = &c->from_addr.sin_addr; 2bd8416e 1129 unsigned long src_addr = src->s_addr; 8256c0a3 PG 1130 1131 for (acl = stream->acl; acl; acl = acl->next) { efa04ce2 1132 if (src_addr >= acl->first.s_addr && src_addr <= acl->last.s_addr) { 8256c0a3 PG 1133 return (acl->action == IP_ALLOW) ? 1 : 0; 1134 } 1135 last_action = acl->action; 1136 } 1137 1138 /* Nothing matched, so return not the last action */ 1139 return (last_action == IP_DENY) ? 1 : 0; 1140} 1141 829ac53d FB 1142/* compute the real filename of a file by matching it without its 1143 extensions to all the stream filenames */ 1144static void compute_real_filename(char *filename, int max_size) 1145{ 1146 char file1[1024]; 1147 char file2[1024]; 1148 char *p; 1149 FFStream *stream; 1150 1151 /* compute filename by matching without the file extensions */ f7d78f36 1152 av_strlcpy(file1, filename, sizeof(file1)); 829ac53d FB 1153 p = strrchr(file1, '.'); 1154 if (p) 1155 *p = '\0'; 1156 for(stream = first_stream; stream != NULL; stream = stream->next) { f7d78f36 1157 av_strlcpy(file2, stream->filename, sizeof(file2)); 829ac53d FB 1158 p = strrchr(file2, '.'); 1159 if (p) 1160 *p = '\0'; 1161 if (!strcmp(file1, file2)) { f7d78f36 1162 av_strlcpy(filename, stream->filename, max_size); 829ac53d FB 1163 break; 1164 } 1165 } 1166} 1167 1168enum RedirType { 1169 REDIR_NONE, 1170 REDIR_ASX, 1171 REDIR_RAM, 1172 REDIR_ASF, 1173 REDIR_RTSP, 1174 REDIR_SDP, 1175}; 1176 85f07f22 FB 1177/* parse http request and prepare header */ 1178static int http_parse_request(HTTPContext *c) 1179{ 1180 char *p; 829ac53d 1181 enum RedirType redir_type; 85f07f22 1182 char cmd[32]; bae79c04 1183 char info[1024], filename[1024]; 85f07f22 FB 1184 char url[1024], *q; 1185 char protocol[32]; 1186 char msg[1024]; 1187 const char *mime_type; 1188 FFStream *stream; 42a63c6a 1189 int i; 3120d2a2 1190 char ratebuf[32]; cde25790 1191 char *useragent = 0; 85f07f22 FB 1192 1193 p = c->buffer; 2effd274 1194 get_word(cmd, sizeof(cmd), (const char **)&p); f7d78f36 1195 av_strlcpy(c->method, cmd, sizeof(c->method)); 7434ba6d 1196 85f07f22 1197 if (!strcmp(cmd, "GET")) edfdd798 1198 c->post = 0; 85f07f22 1199 else if (!strcmp(cmd, "POST")) edfdd798 1200 c->post = 1; 85f07f22 FB 1201 else 1202 return -1; 1203 2effd274 1204 get_word(url, sizeof(url), (const char **)&p); f7d78f36 1205 av_strlcpy(c->url, url, sizeof(c->url)); 7434ba6d 1206 2effd274 1207 get_word(protocol, sizeof(protocol), (const char **)&p); 85f07f22 FB 1208 if (strcmp(protocol, "HTTP/1.0") && strcmp(protocol, "HTTP/1.1")) 1209 return -1; 7434ba6d 1210 f7d78f36 1211 av_strlcpy(c->protocol, protocol, sizeof(c->protocol)); 90f9c440 AB 1212 1213 if (ffserver_debug) bb270c08 1214 http_log("New connection: %s %s\n", cmd, url); 115329f1 1215 85f07f22 1216 /* find the filename and the optional info string in the request */ bae79c04 1217 p = strchr(url, '?'); 85f07f22 1218 if (p) { f7d78f36 1219 av_strlcpy(info, p, sizeof(info)); 85f07f22 FB 1220 *p = '\0'; 1221 } else { 1222 info[0] = '\0'; 1223 } 1224 f7d78f36 1225 av_strlcpy(filename, url + ((*url == '/') ? 1 : 0), sizeof(filename)-1); bae79c04 1226 cde25790 PG 1227 for (p = c->buffer; *p && *p != '\r' && *p != '\n'; ) { 1228 if (strncasecmp(p, "User-Agent:", 11) == 0) { 1229 useragent = p + 11; 1230 if (*useragent && *useragent != '\n' && isspace(*useragent)) 1231 useragent++; 1232 break; 1233 } 1234 p = strchr(p, '\n'); 1235 if (!p) 1236 break; 1237 1238 p++; 1239 } 1240 829ac53d FB 1241 redir_type = REDIR_NONE; 1242 if (match_ext(filename, "asx")) { 1243 redir_type = REDIR_ASX; 7434ba6d 1244 filename[strlen(filename)-1] = 'f'; c2ce254c 1245 } else if (match_ext(filename, "asf") && cde25790 PG 1246 (!useragent || strncasecmp(useragent, "NSPlayer", 8) != 0)) { 1247 /* if this isn't WMP or lookalike, return the redirector file */ 829ac53d FB 1248 redir_type = REDIR_ASF; 1249 } else if (match_ext(filename, "rpm,ram")) { 1250 redir_type = REDIR_RAM; 42a63c6a 1251 strcpy(filename + strlen(filename)-2, "m"); 829ac53d FB 1252 } else if (match_ext(filename, "rtsp")) { 1253 redir_type = REDIR_RTSP; bae79c04 1254 compute_real_filename(filename, sizeof(filename) - 1); 829ac53d FB 1255 } else if (match_ext(filename, "sdp")) { 1256 redir_type = REDIR_SDP; bae79c04 1257 compute_real_filename(filename, sizeof(filename) - 1); 42a63c6a 1258 } 115329f1 1259 bae79c04 AB 1260 // "redirect" / request to index.html 1261 if (!strlen(filename)) f7d78f36 1262 av_strlcpy(filename, "index.html", sizeof(filename) - 1); bae79c04 1263 85f07f22 FB 1264 stream = first_stream; 1265 while (stream != NULL) { 8256c0a3 1266 if (!strcmp(stream->filename, filename) && validate_acl(stream, c)) 85f07f22 FB 1267 break; 1268 stream = stream->next; 1269 } 1270 if (stream == NULL) { d445a7e9 1271 snprintf(msg, sizeof(msg), "File '%s' not found", url); 85f07f22 FB 1272 goto send_error; 1273 } 42a63c6a 1274 cde25790 PG 1275 c->stream = stream; 1276 memcpy(c->feed_streams, stream->feed_streams, sizeof(c->feed_streams)); 1277 memset(c->switch_feed_streams, -1, sizeof(c->switch_feed_streams)); 1278 1279 if (stream->stream_type == STREAM_TYPE_REDIRECT) { 1280 c->http_error = 301; 1281 q = c->buffer; d445a7e9 PG 1282 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 301 Moved\r\n"); 1283 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Location: %s\r\n", stream->feed_filename); 1284 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: text/html\r\n"); 1285 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1286 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<html><head><title>Moved</title></head><body>\r\n"); 1287 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "You should be <a href=\"%s\">redirected</a>.\r\n", stream->feed_filename); 1288 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "</body></html>\r\n"); cde25790 PG 1289 1290 /* prepare output buffer */ 1291 c->buffer_ptr = c->buffer; 1292 c->buffer_end = q; 1293 c->state = HTTPSTATE_SEND_HEADER; 1294 return 0; 1295 } 1296 3120d2a2 PG 1297 /* If this is WMP, get the rate information */ 1298 if (extract_rates(ratebuf, sizeof(ratebuf), c->buffer)) { cde25790 PG 1299 if (modify_current_stream(c, ratebuf)) { 1300 for (i = 0; i < sizeof(c->feed_streams) / sizeof(c->feed_streams[0]); i++) { 1301 if (c->switch_feed_streams[i] >= 0) 1302 do_switch_stream(c, i); 1303 } 1304 } 3120d2a2 PG 1305 } 1306 755bfeab 1307 /* If already streaming this feed, do not let start another feeder. */ d0a5513b AB 1308 if (stream->feed_opened) { 1309 snprintf(msg, sizeof(msg), "This feed is already being received."); 1310 goto send_error; 1311 } 1312 edfdd798 1313 if (c->post == 0 && stream->stream_type == STREAM_TYPE_LIVE) { 6edd6884 1314 current_bandwidth += stream->bandwidth; 42a63c6a 1315 } 115329f1 1316 edfdd798 1317 if (c->post == 0 && max_bandwidth < current_bandwidth) { 42a63c6a PG 1318 c->http_error = 200; 1319 q = c->buffer; d445a7e9 PG 1320 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 Server too busy\r\n"); 1321 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: text/html\r\n"); 1322 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1323 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<html><head><title>Too busy</title></head><body>\r\n"); 33f5e2ec AB 1324 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<p>The server is too busy to serve your request at this time.</p>\r\n"); 1325 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<p>The bandwidth being served (including your stream) is %dkbit/sec, and this exceeds the limit of %dkbit/sec.</p>\r\n", 6edd6884 1326 current_bandwidth, max_bandwidth); d445a7e9 1327 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "</body></html>\r\n"); 42a63c6a PG 1328 1329 /* prepare output buffer */ 1330 c->buffer_ptr = c->buffer; 1331 c->buffer_end = q; 1332 c->state = HTTPSTATE_SEND_HEADER; 1333 return 0; 1334 } 115329f1 1335 829ac53d 1336 if (redir_type != REDIR_NONE) { 7434ba6d 1337 char *hostinfo = 0; 115329f1 1338 7434ba6d PG 1339 for (p = c->buffer; *p && *p != '\r' && *p != '\n'; ) { 1340 if (strncasecmp(p, "Host:", 5) == 0) { 1341 hostinfo = p + 5; 1342 break; 1343 } 1344 p = strchr(p, '\n'); 1345 if (!p) 1346 break; 1347 1348 p++; 1349 } 1350 1351 if (hostinfo) { 1352 char *eoh; 1353 char hostbuf[260]; 1354 1355 while (isspace(*hostinfo)) 1356 hostinfo++; 1357 1358 eoh = strchr(hostinfo, '\n'); 1359 if (eoh) { 1360 if (eoh[-1] == '\r') 1361 eoh--; 1362 1363 if (eoh - hostinfo < sizeof(hostbuf) - 1) { 1364 memcpy(hostbuf, hostinfo, eoh - hostinfo); 1365 hostbuf[eoh - hostinfo] = 0; 1366 1367 c->http_error = 200; 1368 q = c->buffer; 829ac53d FB 1369 switch(redir_type) { 1370 case REDIR_ASX: d445a7e9 PG 1371 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 ASX Follows\r\n"); 1372 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: video/x-ms-asf\r\n"); 1373 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1374 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<ASX Version=\"3\">\r\n"); 7c054ea7 1375 //q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<!-- Autogenerated by ffserver -->\r\n"); 115329f1 1376 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<ENTRY><REF HREF=\"http://%s/%s%s\"/></ENTRY>\r\n", 42a63c6a 1377 hostbuf, filename, info); d445a7e9 1378 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "</ASX>\r\n"); 829ac53d FB 1379 break; 1380 case REDIR_RAM: d445a7e9 PG 1381 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 RAM Follows\r\n"); 1382 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: audio/x-pn-realaudio\r\n"); 1383 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1384 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "# Autogenerated by ffserver\r\n"); 115329f1 1385 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "http://%s/%s%s\r\n", 42a63c6a 1386 hostbuf, filename, info); 829ac53d FB 1387 break; 1388 case REDIR_ASF: d445a7e9 PG 1389 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 ASF Redirect follows\r\n"); 1390 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: video/x-ms-asf\r\n"); 1391 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1392 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "[Reference]\r\n"); 115329f1 1393 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Ref1=http://%s/%s%s\r\n", cde25790 1394 hostbuf, filename, info); 829ac53d FB 1395 break; 1396 case REDIR_RTSP: 1397 { 1398 char hostname[256], *p; 1399 /* extract only hostname */ f7d78f36 1400 av_strlcpy(hostname, hostbuf, sizeof(hostname)); 829ac53d FB 1401 p = strrchr(hostname, ':'); 1402 if (p) 1403 *p = '\0'; d445a7e9 1404 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 RTSP Redirect follows\r\n"); 829ac53d 1405 /* XXX: incorrect mime type ? */ d445a7e9 PG 1406 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: application/x-rtsp\r\n"); 1407 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 115329f1 DB 1408 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "rtsp://%s:%d/%s\r\n", 1409 hostname, ntohs(my_rtsp_addr.sin_port), 829ac53d FB 1410 filename); 1411 } 1412 break; 1413 case REDIR_SDP: 1414 { 0c1a9eda 1415 uint8_t *sdp_data; 829ac53d FB 1416 int sdp_data_size, len; 1417 struct sockaddr_in my_addr; 1418 d445a7e9 PG 1419 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 OK\r\n"); 1420 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: application/sdp\r\n"); 1421 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 829ac53d FB 1422 1423 len = sizeof(my_addr); 1424 getsockname(c->fd, (struct sockaddr *)&my_addr, &len); 115329f1 1425 829ac53d 1426 /* XXX: should use a dynamic buffer */ 115329f1 DB 1427 sdp_data_size = prepare_sdp_description(stream, 1428 &sdp_data, 829ac53d FB 1429 my_addr.sin_addr); 1430 if (sdp_data_size > 0) { 1431 memcpy(q, sdp_data, sdp_data_size); 1432 q += sdp_data_size; 1433 *q = '\0'; 1434 av_free(sdp_data); 1435 } 1436 } 1437 break; 1438 default: 0f4e8165 1439 abort(); 829ac53d 1440 break; 2effd274 1441 } 7434ba6d PG 1442 1443 /* prepare output buffer */ 1444 c->buffer_ptr = c->buffer; 1445 c->buffer_end = q; 1446 c->state = HTTPSTATE_SEND_HEADER; 1447 return 0; 1448 } 1449 } 1450 } 1451 d445a7e9 1452 snprintf(msg, sizeof(msg), "ASX/RAM file not handled"); 7434ba6d 1453 goto send_error; 85f07f22 FB 1454 } 1455 a6e14edd 1456 stream->conns_served++; 7434ba6d 1457 85f07f22 FB 1458 /* XXX: add there authenticate and IP match */ 1459 edfdd798 1460 if (c->post) { 85f07f22 FB 1461 /* if post, it means a feed is being sent */ 1462 if (!stream->is_feed) { 7434ba6d PG 1463 /* However it might be a status report from WMP! Lets log the data 1464 * as it might come in handy one day 1465 */ 1466 char *logline = 0; 3120d2a2 1467 int client_id = 0; 115329f1 1468 7434ba6d PG 1469 for (p = c->buffer; *p && *p != '\r' && *p != '\n'; ) { 1470 if (strncasecmp(p, "Pragma: log-line=", 17) == 0) { 1471 logline = p; 1472 break; 1473 } 3120d2a2 PG 1474 if (strncasecmp(p, "Pragma: client-id=", 18) == 0) { 1475 client_id = strtol(p + 18, 0, 10); 1476 } 7434ba6d PG 1477 p = strchr(p, '\n'); 1478 if (!p) 1479 break; 1480 1481 p++; 1482 } 1483 1484 if (logline) { 1485 char *eol = strchr(logline, '\n'); 1486 1487 logline += 17; 1488 1489 if (eol) { 1490 if (eol[-1] == '\r') 1491 eol--; 7906085f 1492 http_log("%.*s\n", (int) (eol - logline), logline); 7434ba6d PG 1493 c->suppress_log = 1; 1494 } 1495 } 3120d2a2 1496 cde25790 PG 1497#ifdef DEBUG_WMP 1498 http_log("\nGot request:\n%s\n", c->buffer); 3120d2a2 PG 1499#endif 1500 1501 if (client_id && extract_rates(ratebuf, sizeof(ratebuf), c->buffer)) { 1502 HTTPContext *wmpc; 1503 1504 /* Now we have to find the client_id */ 1505 for (wmpc = first_http_ctx; wmpc; wmpc = wmpc->next) { 1506 if (wmpc->wmp_client_id == client_id) 1507 break; 1508 } 1509 2d563d2f AB 1510 if (wmpc && modify_current_stream(wmpc, ratebuf)) 1511 wmpc->switch_pending = 1; 3120d2a2 1512 } 115329f1 1513 d445a7e9 1514 snprintf(msg, sizeof(msg), "POST command not handled"); cb275dd9 1515 c->stream = 0; 85f07f22 FB 1516 goto send_error; 1517 } 1518 if (http_start_receive_data(c) < 0) { d445a7e9 1519 snprintf(msg, sizeof(msg), "could not open feed"); 85f07f22 FB 1520 goto send_error; 1521 } 1522 c->http_error = 0; 1523 c->state = HTTPSTATE_RECEIVE_DATA; 1524 return 0; 1525 } 1526 cde25790 1527#ifdef DEBUG_WMP 3120d2a2 1528 if (strcmp(stream->filename + strlen(stream->filename) - 4, ".asf") == 0) { cde25790 1529 http_log("\nGot request:\n%s\n", c->buffer); 3120d2a2 PG 1530 } 1531#endif 1532 85f07f22 FB 1533 if (c->stream->stream_type == STREAM_TYPE_STATUS) 1534 goto send_stats; 1535 1536 /* open input stream */ 1537 if (open_input_stream(c, info) < 0) { d445a7e9 1538 snprintf(msg, sizeof(msg), "Input stream corresponding to '%s' not found", url); 85f07f22 FB 1539 goto send_error; 1540 } 1541 1542 /* prepare http header */ 1543 q = c->buffer; d445a7e9 1544 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 200 OK\r\n"); 85f07f22 FB 1545 mime_type = c->stream->fmt->mime_type; 1546 if (!mime_type) 087fa475 1547 mime_type = "application/x-octet-stream"; d445a7e9 1548 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Pragma: no-cache\r\n"); 85f07f22 FB 1549 1550 /* for asf, we need extra headers */ 8256c0a3 1551 if (!strcmp(c->stream->fmt->name,"asf_stream")) { 3120d2a2 1552 /* Need to allocate a client id */ 3120d2a2 1553 1df93ae9 1554 c->wmp_client_id = av_random(&random_state) & 0x7fffffff; 3120d2a2 1555 d445a7e9 1556 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Server: Cougar 4.1.0.3923\r\nCache-Control: no-cache\r\nPragma: client-id=%d\r\nPragma: features=\"broadcast\"\r\n", c->wmp_client_id); 85f07f22 1557 } d445a7e9 PG 1558 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-Type: %s\r\n", mime_type); 1559 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 115329f1 1560 85f07f22 FB 1561 /* prepare output buffer */ 1562 c->http_error = 0; 1563 c->buffer_ptr = c->buffer; 1564 c->buffer_end = q; 1565 c->state = HTTPSTATE_SEND_HEADER; 1566 return 0; 1567 send_error: 1568 c->http_error = 404; 1569 q = c->buffer; d445a7e9 PG 1570 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "HTTP/1.0 404 Not Found\r\n"); 1571 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "Content-type: %s\r\n", "text/html"); 1572 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "\r\n"); 1573 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<HTML>\n"); 1574 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<HEAD><TITLE>404 Not Found</TITLE></HEAD>\n"); 1575 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "<BODY>%s</BODY>\n", msg); 1576 q += snprintf(q, q - (char *) c->buffer + c->buffer_size, "</HTML>\n"); 85f07f22 FB 1577 1578 /* prepare output buffer */ 1579 c->buffer_ptr = c->buffer; 1580 c->buffer_end = q; 1581 c->state = HTTPSTATE_SEND_HEADER; 1582 return 0; 1583 send_stats: 1584 compute_stats(c); 1585 c->http_error = 200; /* horrible : we use this value to avoid 1586 going to the send data state */ 1587 c->state = HTTPSTATE_SEND_HEADER; 1588 return 0; 1589} 1590 0c1a9eda 1591static void fmt_bytecount(ByteIOContext *pb, int64_t count) 2ac887ba PG 1592{ 1593 static const char *suffix = " kMGTP"; 1594 const char *s; 1595 1596 for (s = suffix; count >= 100000 && s[1]; count /= 1000, s++) { 1597 } 1598 4733abcb 1599 url_fprintf(pb, "%"PRId64"%c", count, *s); 2ac887ba PG 1600} 1601 85f07f22 FB 1602static void compute_stats(HTTPContext *c) 1603{ 1604 HTTPContext *c1; 1605 FFStream *stream; 2effd274 1606 char *p; 85f07f22 1607 time_t ti; 2effd274 FB 1608 int i, len; 1609 ByteIOContext pb1, *pb = &pb1; cde25790 1610 2effd274 FB 1611 if (url_open_dyn_buf(pb) < 0) { 1612 /* XXX: return an error ? */ cde25790 1613 c->buffer_ptr = c->buffer; 2effd274 FB 1614 c->buffer_end = c->buffer; 1615 return; cde25790 1616 } 85f07f22 1617 2effd274 FB 1618 url_fprintf(pb, "HTTP/1.0 200 OK\r\n"); 1619 url_fprintf(pb, "Content-type: %s\r\n", "text/html"); 1620 url_fprintf(pb, "Pragma: no-cache\r\n"); 1621 url_fprintf(pb, "\r\n"); 115329f1 1622 2effd274 1623 url_fprintf(pb, "<HEAD><TITLE>FFServer Status</TITLE>\n"); cde25790 1624 if (c->stream->feed_filename) { 2effd274 1625 url_fprintf(pb, "<link rel=\"shortcut icon\" href=\"%s\">\n", c->stream->feed_filename); cde25790 1626 } 2effd274 FB 1627 url_fprintf(pb, "</HEAD>\n<BODY>"); 1628 url_fprintf(pb, "<H1>FFServer Status</H1>\n"); 85f07f22 1629 /* format status */ 2effd274 FB 1630 url_fprintf(pb, "<H2>Available Streams</H2>\n"); 1631 url_fprintf(pb, "<TABLE cellspacing=0 cellpadding=4>\n"); 1632 url_fprintf(pb, "<TR><Th valign=top>Path<th align=left>Served<br>Conns<Th><br>bytes<Th valign=top>Format<Th>Bit rate<br>kbits/s<Th align=left>Video<br>kbits/s<th><br>Codec<Th align=left>Audio<br>kbits/s<th><br>Codec<Th align=left valign=top>Feed\n"); 85f07f22 FB 1633 stream = first_stream; 1634 while (stream != NULL) { 42a63c6a PG 1635 char sfilename[1024]; 1636 char *eosf; 1637 a6e14edd 1638 if (stream->feed != stream) { f7d78f36 1639 av_strlcpy(sfilename, stream->filename, sizeof(sfilename) - 10); a6e14edd PG 1640 eosf = sfilename + strlen(sfilename); 1641 if (eosf - sfilename >= 4) { 1642 if (strcmp(eosf - 4, ".asf") == 0) { 1643 strcpy(eosf - 4, ".asx"); 1644 } else if (strcmp(eosf - 3, ".rm") == 0) { 1645 strcpy(eosf - 3, ".ram"); d2a067d1 1646 } else if (stream->fmt == &rtp_muxer) { 829ac53d FB 1647 /* generate a sample RTSP director if 1648 unicast. Generate an SDP redirector if 1649 multicast */ 2effd274 FB 1650 eosf = strrchr(sfilename, '.'); 1651 if (!eosf) 1652 eosf = sfilename + strlen(sfilename); 829ac53d FB 1653 if (stream->is_multicast) 1654 strcpy(eosf, ".sdp"); 1655 else 1656 strcpy(eosf, ".rtsp"); a6e14edd 1657 } 42a63c6a 1658 } 115329f1 DB 1659 1660 url_fprintf(pb, "<TR><TD><A HREF=\"/%s\">%s</A> ", a6e14edd 1661 sfilename, stream->filename); 2effd274 1662 url_fprintf(pb, "<td align=right> %d <td align=right> ", 2ac887ba 1663 stream->conns_served); 2effd274 1664 fmt_bytecount(pb, stream->bytes_served); a6e14edd PG 1665 switch(stream->stream_type) { 1666 case STREAM_TYPE_LIVE: 1667 { 1668 int audio_bit_rate = 0; 1669 int video_bit_rate = 0; 58445440 ZK 1670 const char *audio_codec_name = ""; 1671 const char *video_codec_name = ""; 1672 const char *audio_codec_name_extra = ""; 1673 const char *video_codec_name_extra = ""; a6e14edd PG 1674 1675 for(i=0;i<stream->nb_streams;i++) { 1676 AVStream *st = stream->streams[i]; 01f4895c MN 1677 AVCodec *codec = avcodec_find_encoder(st->codec->codec_id); 1678 switch(st->codec->codec_type) { a6e14edd 1679 case CODEC_TYPE_AUDIO: 01f4895c 1680 audio_bit_rate += st->codec->bit_rate; a6e14edd PG 1681 if (codec) { 1682 if (*audio_codec_name) 1683 audio_codec_name_extra = "..."; 1684 audio_codec_name = codec->name; 1685 } 1686 break; 1687 case CODEC_TYPE_VIDEO: 01f4895c 1688 video_bit_rate += st->codec->bit_rate; a6e14edd PG 1689 if (codec) { 1690 if (*video_codec_name) 1691 video_codec_name_extra = "..."; 1692 video_codec_name = codec->name; 1693 } 1694 break; e240a0bb 1695 case CODEC_TYPE_DATA: 01f4895c 1696 video_bit_rate += st->codec->bit_rate; e240a0bb 1697 break; a6e14edd 1698 default: 0f4e8165 1699 abort(); 79c4ea3c 1700 } 85f07f22 1701 } 115329f1 1702 url_fprintf(pb, "<TD align=center> %s <TD align=right> %d <TD align=right> %d <TD> %s %s <TD align=right> %d <TD> %s %s", a6e14edd 1703 stream->fmt->name, 6edd6884 1704 stream->bandwidth, a6e14edd PG 1705 video_bit_rate / 1000, video_codec_name, video_codec_name_extra, 1706 audio_bit_rate / 1000, audio_codec_name, audio_codec_name_extra); 1707 if (stream->feed) { 2effd274 1708 url_fprintf(pb, "<TD>%s", stream->feed->filename); a6e14edd 1709 } else { 2effd274 1710 url_fprintf(pb, "<TD>%s", stream->feed_filename); a6e14edd 1711 } 2effd274 1712 url_fprintf(pb, "\n"); 85f07f22 1713 } a6e14edd PG 1714 break; 1715 default: 2effd274 1716 url_fprintf(pb, "<TD align=center> - <TD align=right> - <TD align=right> - <td><td align=right> - <TD>\n"); a6e14edd 1717 break; 85f07f22 1718 } 85f07f22 FB 1719 } 1720 stream = stream->next; 1721 } 2effd274 1722 url_fprintf(pb, "</TABLE>\n"); a6e14edd PG 1723 1724 stream = first_stream; 1725 while (stream != NULL) { 1726 if (stream->feed == stream) { 2effd274 1727 url_fprintf(pb, "<h2>Feed %s</h2>", stream->filename); cde25790 1728 if (stream->pid) { 2effd274 1729 url_fprintf(pb, "Running as pid %d.\n", stream->pid); cde25790 1730 2effd274 FB 1731#if defined(linux) && !defined(CONFIG_NOCUTILS) 1732 { 1733 FILE *pid_stat; 1734 char ps_cmd[64]; 1735 1736 /* This is somewhat linux specific I guess */ 115329f1 DB 1737 snprintf(ps_cmd, sizeof(ps_cmd), 1738 "ps -o \"%%cpu,cputime\" --no-headers %d", 2effd274 1739 stream->pid); 115329f1 1740 2effd274 FB 1741 pid_stat = popen(ps_cmd, "r"); 1742 if (pid_stat) { 1743 char cpuperc[10]; 1744 char cpuused[64]; 115329f1 DB 1745 1746 if (fscanf(pid_stat, "%10s %64s", cpuperc, 2effd274 FB 1747 cpuused) == 2) { 1748 url_fprintf(pb, "Currently using %s%% of the cpu. Total time used %s.\n", 1749 cpuperc, cpuused); 1750 } 1751 fclose(pid_stat); cde25790 1752 } cde25790 PG 1753 } 1754#endif 1755 2effd274 1756 url_fprintf(pb, "<p>"); cde25790 1757 } 2effd274 1758 url_fprintf(pb, "<table cellspacing=0 cellpadding=4><tr><th>Stream<th>type<th>kbits/s<th align=left>codec<th align=left>Parameters\n"); a6e14edd PG 1759 1760 for (i = 0; i < stream->nb_streams; i++) { 1761 AVStream *st = stream->streams[i]; 01f4895c 1762 AVCodec *codec = avcodec_find_encoder(st->codec->codec_id); b29f97d1 1763 const char *type = "unknown"; b582f314 PG 1764 char parameters[64]; 1765 1766 parameters[0] = 0; a6e14edd 1767 01f4895c 1768 switch(st->codec->codec_type) { a6e14edd PG 1769 case CODEC_TYPE_AUDIO: 1770 type = "audio"; acdc8520 1771 snprintf(parameters, sizeof(parameters), "%d channel(s), %d Hz", st->codec->channels, st->codec->sample_rate); a6e14edd PG 1772 break; 1773 case CODEC_TYPE_VIDEO: 1774 type = "video"; 01f4895c MN 1775 snprintf(parameters, sizeof(parameters), "%dx%d, q=%d-%d, fps=%d", st->codec->width, st->codec->height, 1776 st->codec->qmin, st->codec->qmax, st->codec->time_base.den / st->codec->time_base.num); a6e14edd PG 1777 break; 1778 default: 0f4e8165 1779 abort(); a6e14edd 1780 } 2effd274 1781 url_fprintf(pb, "<tr><td align=right>%d<td>%s<td align=right>%d<td>%s<td>%s\n", 01f4895c 1782 i, type, st->codec->bit_rate/1000, codec ? codec->name : "", parameters); a6e14edd 1783 } 2effd274 1784 url_fprintf(pb, "</table>\n"); a6e14edd 1785 115329f1 1786 } a6e14edd PG 1787 stream = stream->next; 1788 } 115329f1 1789 85f07f22 FB 1790#if 0 1791 { 1792 float avg; 1793 AVCodecContext *enc; 1794 char buf[1024]; 115329f1 1795 85f07f22 FB 1796 /* feed status */ 1797 stream = first_feed; 1798 while (stream != NULL) { 2effd274 FB 1799 url_fprintf(pb, "<H1>Feed '%s'</H1>\n", stream->filename); 1800 url_fprintf(pb, "<TABLE>\n"); 1801 url_fprintf(pb, "<TR><TD>Parameters<TD>Frame count<TD>Size<TD>Avg bitrate (kbits/s)\n"); 85f07f22 FB 1802 for(i=0;i<stream->nb_streams;i++) { 1803 AVStream *st = stream->streams[i]; 1804 FeedData *fdata = st->priv_data; 01f4895c 1805 enc = st->codec; 115329f1 1806 85f07f22 FB 1807 avcodec_string(buf, sizeof(buf), enc); 1808 avg = fdata->avg_frame_size * (float)enc->rate * 8.0; 1809 if (enc->codec->type == CODEC_TYPE_AUDIO && enc->frame_size > 0) 1810 avg /= enc->frame_size; 949b1a13 1811 url_fprintf(pb, "<TR><TD>%s <TD> %d <TD> %"PRId64" <TD> %0.1f\n", 85f07f22 FB 1812 buf, enc->frame_number, fdata->data_count, avg / 1000.0); 1813 } 2effd274 1814 url_fprintf(pb, "</TABLE>\n"); 85f07f22 FB 1815 stream = stream->next_feed; 1816 } 1817 } 1818#endif 1819 1820 /* connection status */ 2effd274 1821 url_fprintf(pb, "<H2>Connection Status</H2>\n"); 85f07f22 1822 2effd274 1823 url_fprintf(pb, "Number of connections: %d / %d<BR>\n", 85f07f22 FB 1824 nb_connections, nb_max_connections); 1825 2effd274 1826 url_fprintf(pb, "Bandwidth in use: %dk / %dk<BR>\n", 6edd6884 1827 current_bandwidth, max_bandwidth); 42a63c6a 1828 2effd274 FB 1829 url_fprintf(pb, "<TABLE>\n"); 1830 url_fprintf(pb, "<TR><th>#<th>File<th>IP<th>Proto<th>State<th>Target bits/sec<th>Actual bits/sec<th>Bytes transferred\n"); 85f07f22 FB 1831 c1 = first_http_ctx; 1832 i = 0; 2effd274 1833 while (c1 != NULL) { cde25790 PG 1834 int bitrate; 1835 int j; 1836 1837 bitrate = 0; 2effd274 FB 1838 if (c1->stream) { 1839 for (j = 0; j < c1->stream->nb_streams; j++) { 2d563d2f 1840 if (!c1->stream->feed) 01f4895c 1841 bitrate += c1->stream->streams[j]->codec->bit_rate; 2d563d2f AB 1842 else if (c1->feed_streams[j] >= 0) 1843 bitrate += c1->stream->feed->streams[c1->feed_streams[j]]->codec->bit_rate; cde25790 PG 1844 } 1845 } 1846 85f07f22 FB 1847 i++; 1848 p = inet_ntoa(c1->from_addr.sin_addr); 115329f1 DB 1849 url_fprintf(pb, "<TR><TD><B>%d</B><TD>%s%s<TD>%s<TD>%s<TD>%s<td align=right>", 1850 i, 1851 c1->stream ? c1->stream->filename : "", 2effd274 1852 c1->state == HTTPSTATE_RECEIVE_DATA ? "(input)" : "", 115329f1 1853 p, 2effd274 FB 1854 c1->protocol, 1855 http_state[c1->state]); 1856 fmt_bytecount(pb, bitrate); 1857 url_fprintf(pb, "<td align=right>"); 1858 fmt_bytecount(pb, compute_datarate(&c1->datarate, c1->data_count) * 8); 1859 url_fprintf(pb, "<td align=right>"); 1860 fmt_bytecount(pb, c1->data_count); 1861 url_fprintf(pb, "\n"); 85f07f22 FB 1862 c1 = c1->next; 1863 } 2effd274 1864 url_fprintf(pb, "</TABLE>\n"); 115329f1 1865 85f07f22 FB 1866 /* date */ 1867 ti = time(NULL); 1868 p = ctime(&ti); 2effd274 FB 1869 url_fprintf(pb, "<HR size=1 noshade>Generated at %s", p); 1870 url_fprintf(pb, "</BODY>\n</HTML>\n"); 85f07f22 1871 2effd274 FB 1872 len = url_close_dyn_buf(pb, &c->pb_buffer); 1873 c->buffer_ptr = c->pb_buffer; 1874 c->buffer_end = c->pb_buffer + len; 85f07f22 FB 1875} 1876 2effd274 FB 1877/* check if the parser needs to be opened for stream i */ 1878static void open_parser(AVFormatContext *s, int i) 85f07f22 1879{ 2effd274 FB 1880 AVStream *st = s->streams[i]; 1881 AVCodec *codec; 31def229 1882 01f4895c MN 1883 if (!st->codec->codec) { 1884 codec = avcodec_find_decoder(st->codec->codec_id); 2effd274 1885 if (codec && (codec->capabilities & CODEC_CAP_PARSE_ONLY)) { 01f4895c MN 1886 st->codec->parse_only = 1; 1887 if (avcodec_open(st->codec, codec) < 0) { 1888 st->codec->parse_only = 0; 2effd274 1889 } cde25790 PG 1890 } 1891 } 85f07f22 FB 1892} 1893 1894static int open_input_stream(HTTPContext *c, const char *info) 1895{ 1896 char buf[128]; 1897 char input_filename[1024]; 1898 AVFormatContext *s; 2effd274 1899 int buf_size, i; 0c1a9eda 1900 int64_t stream_pos; 85f07f22 FB 1901 1902 /* find file name */ 1903 if (c->stream->feed) { 1904 strcpy(input_filename, c->stream->feed->feed_filename); 1905 buf_size = FFM_PACKET_SIZE; 1906 /* compute position (absolute time) */ 1907 if (find_info_tag(buf, sizeof(buf), "date", info)) { 1908 stream_pos = parse_date(buf, 0); f747e6d3 PG 1909 } else if (find_info_tag(buf, sizeof(buf), "buffer", info)) { 1910 int prebuffer = strtol(buf, 0, 10); 0c1a9eda 1911 stream_pos = av_gettime() - prebuffer * (int64_t)1000000; 85f07f22 1912 } else { 0c1a9eda 1913 stream_pos = av_gettime() - c->stream->prebuffer * (int64_t)1000; 85f07f22 FB 1914 } 1915 } else { 1916 strcpy(input_filename, c->stream->feed_filename); 1917 buf_size = 0; 1918 /* compute position (relative time) */ 1919 if (find_info_tag(buf, sizeof(buf), "date", info)) { 1920 stream_pos = parse_date(buf, 1); 1921 } else { 1922 stream_pos = 0; 1923 } 1924 } 1925 if (input_filename[0] == '\0') 1926 return -1; 1927 8256c0a3 PG 1928#if 0 1929 { time_t when = stream_pos / 1000000; 949b1a13 1930 http_log("Stream pos = %"PRId64", time=%s", stream_pos, ctime(&when)); 8256c0a3 PG 1931 } 1932#endif 1933 85f07f22 1934 /* open stream */ 115329f1 1935 if (av_open_input_file(&s, input_filename, c->stream->ifmt, e240a0bb 1936 buf_size, c->stream->ap_in) < 0) { 2effd274 1937 http_log("%s not found", input_filename); 85f07f22 1938 return -1; 2effd274 1939 } 85f07f22 1940 c->fmt_in = s; 115329f1 1941 2effd274 FB 1942 /* open each parser */ 1943 for(i=0;i<s->nb_streams;i++) 1944 open_parser(s, i); 1945 1946 /* choose stream as clock source (we favorize video stream if 1947 present) for packet sending */ 1948 c->pts_stream_index = 0; 1949 for(i=0;i<c->stream->nb_streams;i++) { 115329f1 1950 if (c->pts_stream_index == 0 && 01f4895c 1951 c->stream->streams[i]->codec->codec_type == CODEC_TYPE_VIDEO) { 2effd274 FB 1952 c->pts_stream_index = i; 1953 } 1954 } 85f07f22 1955 e8d27bc3 1956#if 1 bd7cf6ad 1957 if (c->fmt_in->iformat->read_seek) { e8d27bc3 1958 c->fmt_in->iformat->read_seek(c->fmt_in, 0, stream_pos, 0); 85f07f22 1959 } e240a0bb 1960#endif 2effd274 FB 1961 /* set the start time (needed for maxtime and RTP packet timing) */ 1962 c->start_time = cur_time; 1963 c->first_pts = AV_NOPTS_VALUE; 85f07f22 FB 1964 return 0; 1965} 1966 e240a0bb FB 1967/* return the server clock (in us) */ 1968static int64_t get_server_clock(HTTPContext *c) 2effd274 1969{ e240a0bb 1970 /* compute current pts value from system time */ c3f58185 1971 return (cur_time - c->start_time) * 1000; 2effd274 FB 1972} 1973 e240a0bb FB 1974/* return the estimated time at which the current packet must be sent 1975 (in us) */ 1976static int64_t get_packet_send_clock(HTTPContext *c) 2effd274 1977{ e240a0bb 1978 int bytes_left, bytes_sent, frame_bytes; 115329f1 1979 e240a0bb FB 1980 frame_bytes = c->cur_frame_bytes; 1981 if (frame_bytes <= 0) { 1982 return c->cur_pts; 2effd274 1983 } else { e240a0bb FB 1984 bytes_left = c->buffer_end - c->buffer_ptr; 1985 bytes_sent = frame_bytes - bytes_left; 1986 return c->cur_pts + (c->cur_frame_duration * bytes_sent) / frame_bytes; 2effd274 1987 } 2effd274 1988} 2effd274 1989 2effd274 1990 2effd274 FB 1991static int http_prepare_data(HTTPContext *c) 1992{ 1993 int i, len, ret; 1994 AVFormatContext *ctx; 1995 bc351386 1996 av_freep(&c->pb_buffer); 2effd274 FB 1997 switch(c->state) { 1998 case HTTPSTATE_SEND_DATA_HEADER: 1999 memset(&c->fmt_ctx, 0, sizeof(c->fmt_ctx)); f7d78f36 MR 2000 av_strlcpy(c->fmt_ctx.author, c->stream->author, 2001 sizeof(c->fmt_ctx.author)); 2002 av_strlcpy(c->fmt_ctx.comment, c->stream->comment, 2003 sizeof(c->fmt_ctx.comment)); 2004 av_strlcpy(c->fmt_ctx.copyright, c->stream->copyright, 2005 sizeof(c->fmt_ctx.copyright)); 2006 av_strlcpy(c->fmt_ctx.title, c->stream->title, 2007 sizeof(c->fmt_ctx.title)); 2effd274 FB 2008 2009 /* open output stream by using specified codecs */ 2010 c->fmt_ctx.oformat = c->stream->fmt; 2011 c->fmt_ctx.nb_streams = c->stream->nb_streams; 2012 for(i=0;i<c->fmt_ctx.nb_streams;i++) { 2013 AVStream *st; bb270c08 2014 AVStream *src; 2effd274 2015 st = av_mallocz(sizeof(AVStream)); 8d931070 2016 st->codec= avcodec_alloc_context(); 2effd274 FB 2017 c->fmt_ctx.streams[i] = st; 2018 /* if file or feed, then just take streams from FFStream struct */ 115329f1 2019 if (!c->stream->feed || 2effd274 2020 c->stream->feed == c->stream) 7c054ea7 2021 src = c->stream->streams[i]; 2effd274 2022 else 7c054ea7 PG 2023 src = c->stream->feed->streams[c->stream->feed_streams[i]]; 2024 bb270c08 DB 2025 *st = *src; 2026 st->priv_data = 0; 01f4895c 2027 st->codec->frame_number = 0; /* XXX: should be done in 2effd274 2028 AVStream, not in codec */ a4d70941 PG 2029 /* I'm pretty sure that this is not correct... 2030 * However, without it, we crash 2031 */ 01f4895c 2032 st->codec->coded_frame = &dummy_frame; 2effd274 FB 2033 } 2034 c->got_key_frame = 0; 2035 2036 /* prepare header and save header data in a stream */ 2037 if (url_open_dyn_buf(&c->fmt_ctx.pb) < 0) { 2038 /* XXX: potential leak */ 2039 return -1; 2040 } 2041 c->fmt_ctx.pb.is_streamed = 1; 2042 3c27199b 2043 av_set_parameters(&c->fmt_ctx, NULL); f75cdda7 AB 2044 if (av_write_header(&c->fmt_ctx) < 0) 2045 return -1; 2effd274 FB 2046 2047 len = url_close_dyn_buf(&c->fmt_ctx.pb, &c->pb_buffer); 2048 c->buffer_ptr = c->pb_buffer; 2049 c->buffer_end = c->pb_buffer + len; 2050 2051 c->state = HTTPSTATE_SEND_DATA; 85f07f22 FB 2052 c->last_packet_sent = 0; 2053 break; 2054 case HTTPSTATE_SEND_DATA: 2055 /* find a new packet */ 85f07f22 FB 2056 { 2057 AVPacket pkt; 115329f1 2058 85f07f22 FB 2059 /* read a packet from the input stream */ 2060 if (c->stream->feed) { 115329f1 2061 ffm_set_write_index(c->fmt_in, 85f07f22 FB 2062 c->stream->feed->feed_write_index, 2063 c->stream->feed->feed_size); 2064 } ec3b2232 2065 115329f1 2066 if (c->stream->max_time && 2ac887ba 2067 c->stream->max_time + c->start_time - cur_time < 0) { ec3b2232 PG 2068 /* We have timed out */ 2069 c->state = HTTPSTATE_SEND_DATA_TRAILER; 85f07f22 2070 } else { 6edd6884 2071 redo: 2effd274 FB 2072 if (av_read_frame(c->fmt_in, &pkt) < 0) { 2073 if (c->stream->feed && c->stream->feed->feed_opened) { 2074 /* if coming from feed, it means we reached the end of the 2075 ffm file, so must wait for more data */ 2076 c->state = HTTPSTATE_WAIT_FEED; 2077 return 1; /* state changed */ 2078 } else { 6edd6884 FB 2079 if (c->stream->loop) { 2080 av_close_input_file(c->fmt_in); 2081 c->fmt_in = NULL; 2082 if (open_input_stream(c, "") < 0) 2083 goto no_loop; 2084 goto redo; 2085 } else { 2086 no_loop: 2087 /* must send trailer now because eof or error */ 2088 c->state = HTTPSTATE_SEND_DATA_TRAILER; 2089 } 2effd274 FB 2090 } 2091 } else { 2092 /* update first pts if needed */ 1bc1cfdd 2093 if (c->first_pts == AV_NOPTS_VALUE) { c0df9d75 2094 c->first_pts = av_rescale_q(pkt.dts, c->fmt_in->streams[pkt.stream_index]->time_base, AV_TIME_BASE_Q); 1bc1cfdd GF 2095 c->start_time = cur_time; 2096 } 2effd274 FB 2097 /* send it to the appropriate stream */ 2098 if (c->stream->feed) { 2099 /* if coming from a feed, select the right stream */ 2100 if (c->switch_pending) { 2101 c->switch_pending = 0; 2102 for(i=0;i<c->stream->nb_streams;i++) { 2103 if (c->switch_feed_streams[i] == pkt.stream_index) { 2104 if (pkt.flags & PKT_FLAG_KEY) { 2105 do_switch_stream(c, i); 2106 } 2107 } 2108 if (c->switch_feed_streams[i] >= 0) { 2109 c->switch_pending = 1; 2110 } 2111 } 2112 } cde25790 2113 for(i=0;i<c->stream->nb_streams;i++) { 2effd274 FB 2114 if (c->feed_streams[i] == pkt.stream_index) { 2115 pkt.stream_index = i; cde25790 2116 if (pkt.flags & PKT_FLAG_KEY) { 2effd274 FB 2117 c->got_key_frame |= 1 << i; 2118 } 115329f1 2119 /* See if we have all the key frames, then 2effd274 2120 * we start to send. This logic is not quite 115329f1 2121 * right, but it works for the case of a 2effd274 2122 * single video stream with one or more 115329f1 DB 2123 * audio streams (for which every frame is 2124 * typically a key frame). 2effd274 2125 */ 115329f1 2126 if (!c->stream->send_on_key || 2effd274 FB 2127 ((c->got_key_frame + 1) >> c->stream->nb_streams)) { 2128 goto send_it; cde25790 2129 } cde25790 PG 2130 } 2131 } 2effd274 FB 2132 } else { 2133 AVCodecContext *codec; 115329f1 2134 2effd274 FB 2135 send_it: 2136 /* specific handling for RTP: we use several 2137 output stream (one for each RTP 2138 connection). XXX: need more abstract handling */ 2139 if (c->is_packetized) { e240a0bb FB 2140 AVStream *st; 2141 /* compute send time and duration */ 2142 st = c->fmt_in->streams[pkt.stream_index]; c0df9d75 2143 c->cur_pts = av_rescale_q(pkt.dts, st->time_base, AV_TIME_BASE_Q); e240a0bb 2144 if (st->start_time != AV_NOPTS_VALUE) c0df9d75 MN 2145 c->cur_pts -= av_rescale_q(st->start_time, st->time_base, AV_TIME_BASE_Q); 2146 c->cur_frame_duration = av_rescale_q(pkt.duration, st->time_base, AV_TIME_BASE_Q); e240a0bb FB 2147#if 0 2148 printf("index=%d pts=%0.3f duration=%0.6f\n", 2149 pkt.stream_index, 115329f1 2150 (double)c->cur_pts / e240a0bb 2151 AV_TIME_BASE, 115329f1 2152 (double)c->cur_frame_duration / e240a0bb FB 2153 AV_TIME_BASE); 2154#endif 2155 /* find RTP context */ 2effd274 FB 2156 c->packet_stream_index = pkt.stream_index; 2157 ctx = c->rtp_ctx[c->packet_stream_index]; 1b52b6bd MN 2158 if(!ctx) { 2159 av_free_packet(&pkt); 1bc1cfdd 2160 break; 1b52b6bd 2161 } 01f4895c 2162 codec = ctx->streams[0]->codec; 6edd6884 FB 2163 /* only one stream per RTP connection */ 2164 pkt.stream_index = 0; 2effd274 FB 2165 } else { 2166 ctx = &c->fmt_ctx; 2167 /* Fudge here */ 01f4895c 2168 codec = ctx->streams[pkt.stream_index]->codec; 85f07f22 2169 } 115329f1 2170 492cd3a9 2171 codec->coded_frame->key_frame = ((pkt.flags & PKT_FLAG_KEY) != 0); 2effd274 2172 if (c->is_packetized) { bc351386 FB 2173 int max_packet_size; 2174 if (c->rtp_protocol == RTSP_PROTOCOL_RTP_TCP) 2175 max_packet_size = RTSP_TCP_MAX_PACKET_SIZE; 2176 else 2177 max_packet_size = url_get_max_packet_size(c->rtp_handles[c->packet_stream_index]); 2178 ret = url_open_dyn_packet_buf(&ctx->pb, max_packet_size); 2effd274 FB 2179 } else { 2180 ret = url_open_dyn_buf(&ctx->pb); 2181 } 2182 if (ret < 0) { 2183 /* XXX: potential leak */ 2184 return -1; 2185 } f1debfd0 AB 2186 if (pkt.dts != AV_NOPTS_VALUE) 2187 pkt.dts = av_rescale_q(pkt.dts, 2188 c->fmt_in->streams[pkt.stream_index]->time_base, 2189 ctx->streams[pkt.stream_index]->time_base); 2190 if (pkt.pts != AV_NOPTS_VALUE) 2191 pkt.pts = av_rescale_q(pkt.pts, 2192 c->fmt_in->streams[pkt.stream_index]->time_base, 2193 ctx->streams[pkt.stream_index]->time_base); e928649b 2194 if (av_write_frame(ctx, &pkt)) { 2effd274 FB 2195 c->state = HTTPSTATE_SEND_DATA_TRAILER; 2196 } 115329f1 2197 2effd274 2198 len = url_close_dyn_buf(&ctx->pb, &c->pb_buffer); e240a0bb 2199 c->cur_frame_bytes = len; 2effd274 FB 2200 c->buffer_ptr = c->pb_buffer; 2201 c->buffer_end = c->pb_buffer + len; 115329f1 2202 2effd274 2203 codec->frame_number++; e240a0bb FB 2204 if (len == 0) 2205 goto redo; f747e6d3 2206 } 2effd274 2207 av_free_packet(&pkt); 85f07f22 2208 } 85f07f22 FB 2209 } 2210 } 2211 break; 2212 default: 2213 case HTTPSTATE_SEND_DATA_TRAILER: 2214 /* last packet test ? */ 2effd274 2215 if (c->last_packet_sent || c->is_packetized) 85f07f22 2216 return -1; 2effd274 2217 ctx = &c->fmt_ctx; 85f07f22 2218 /* prepare header */ 2effd274 FB 2219 if (url_open_dyn_buf(&ctx->pb) < 0) { 2220 /* XXX: potential leak */ 2221 return -1; 2222 } 2223 av_write_trailer(ctx); 2224 len = url_close_dyn_buf(&ctx->pb, &c->pb_buffer); 2225 c->buffer_ptr = c->pb_buffer; 2226 c->buffer_end = c->pb_buffer + len; 2227 85f07f22 FB 2228 c->last_packet_sent = 1; 2229 break; 2230 } 2231 return 0; 2232} 2233 2234/* should convert the format at the same time */ bc351386 FB 2235/* send data starting at c->buffer_ptr to the output connection 2236 (either UDP or TCP connection) */ 5eb765ef 2237static int http_send_data(HTTPContext *c) 85f07f22 2238{ e240a0bb 2239 int len, ret; 85f07f22 2240 bc351386 FB 2241 for(;;) { 2242 if (c->buffer_ptr >= c->buffer_end) { 2243 ret = http_prepare_data(c); 2244 if (ret < 0) 2245 return -1; 2246 else if (ret != 0) { 2247 /* state change requested */ 2248 break; f747e6d3 2249 } 2effd274 2250 } else { bc351386 FB 2251 if (c->is_packetized) { 2252 /* RTP data output */ 2253 len = c->buffer_end - c->buffer_ptr; 2254 if (len < 4) { 2255 /* fail safe - should never happen */ 2256 fail1: 2257 c->buffer_ptr = c->buffer_end; 2effd274 FB 2258 return 0; 2259 } bc351386 FB 2260 len = (c->buffer_ptr[0] << 24) | 2261 (c->buffer_ptr[1] << 16) | 2262 (c->buffer_ptr[2] << 8) | 2263 (c->buffer_ptr[3]); 2264 if (len > (c->buffer_end - c->buffer_ptr)) 2265 goto fail1; e240a0bb FB 2266 if ((get_packet_send_clock(c) - get_server_clock(c)) > 0) { 2267 /* nothing to send yet: we can wait */ 2268 return 0; 2269 } 2270 2271 c->data_count += len; 2272 update_datarate(&c->datarate, c->data_count); 2273 if (c->stream) 2274 c->stream->bytes_served += len; 2275 bc351386 FB 2276 if (c->rtp_protocol == RTSP_PROTOCOL_RTP_TCP) { 2277 /* RTP packets are sent inside the RTSP TCP connection */ 2278 ByteIOContext pb1, *pb = &pb1; 2279 int interleaved_index, size; 2280 uint8_t header[4]; 2281 HTTPContext *rtsp_c; 115329f1 2282 bc351386 FB 2283 rtsp_c = c->rtsp_c; 2284 /* if no RTSP connection left, error */ 2285 if (!rtsp_c) 2286 return -1; 2287 /* if already sending something, then wait. */ 2288 if (rtsp_c->state != RTSPSTATE_WAIT_REQUEST) { 2289 break; 2290 } 2291 if (url_open_dyn_buf(pb) < 0) 2292 goto fail1; 2293 interleaved_index = c->packet_stream_index * 2; 2294 /* RTCP packets are sent at odd indexes */ 2295 if (c->buffer_ptr[1] == 200) 2296 interleaved_index++; 2297 /* write RTSP TCP header */ 2298 header[0] = '$'; 2299 header[1] = interleaved_index; 2300 header[2] = len >> 8; 2301 header[3] = len; 2302 put_buffer(pb, header, 4); 2303 /* write RTP packet data */ 2304 c->buffer_ptr += 4; 2305 put_buffer(pb, c->buffer_ptr, len); 2306 size = url_close_dyn_buf(pb, &c->packet_buffer); 2307 /* prepare asynchronous TCP sending */ 2308 rtsp_c->packet_buffer_ptr = c->packet_buffer; 2309 rtsp_c->packet_buffer_end = c->packet_buffer + size; e240a0bb 2310 c->buffer_ptr += len; 115329f1 2311 e240a0bb 2312 /* send everything we can NOW */ c60202df AB 2313 len = send(rtsp_c->fd, rtsp_c->packet_buffer_ptr, 2314 rtsp_c->packet_buffer_end - rtsp_c->packet_buffer_ptr, 0); e240a0bb FB 2315 if (len > 0) { 2316 rtsp_c->packet_buffer_ptr += len; bc351386 2317 } e240a0bb FB 2318 if (rtsp_c->packet_buffer_ptr < rtsp_c->packet_buffer_end) { 2319 /* if we could not send all the data, we will 2320 send it later, so a new state is needed to 2321 "lock" the RTSP TCP connection */ 2322 rtsp_c->state = RTSPSTATE_SEND_PACKET; 2323 break; 2324 } else { 2325 /* all data has been sent */ 2326 av_freep(&c->packet_buffer); 2327 } 2328 } else { 2329 /* send RTP packet directly in UDP */ bc351386 2330 c->buffer_ptr += 4; 115329f1 2331 url_write(c->rtp_handles[c->packet_stream_index], bc351386 2332 c->buffer_ptr, len); e240a0bb FB 2333 c->buffer_ptr += len; 2334 /* here we continue as we can send several packets per 10 ms slot */ bc351386 2335 } bc351386 FB 2336 } else { 2337 /* TCP data output */ c60202df 2338 len = send(c->fd, c->buffer_ptr, c->buffer_end - c->buffer_ptr, 0); bc351386 2339 if (len < 0) { 8da4034f AB 2340 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 2341 ff_neterrno() != FF_NETERROR(EINTR)) { bc351386 FB 2342 /* error : close connection */ 2343 return -1; 2344 } else { 2345 return 0; 2346 } 2347 } else { 2348 c->buffer_ptr += len; 2349 } e240a0bb FB 2350 c->data_count += len; 2351 update_datarate(&c->datarate, c->data_count); 2352 if (c->stream) 2353 c->stream->bytes_served += len; 2354 break; 2effd274 2355 } 85f07f22 2356 } bc351386 2357 } /* for(;;) */ 85f07f22 FB 2358 return 0; 2359} 2360 2361static int http_start_receive_data(HTTPContext *c) 2362{ 2363 int fd; 2364 2365 if (c->stream->feed_opened) 2366 return -1; 2367 e322ea48 PG 2368 /* Don't permit writing to this one */ 2369 if (c->stream->readonly) 2370 return -1; 2371 85f07f22 FB 2372 /* open feed */ 2373 fd = open(c->stream->feed_filename, O_RDWR); 2374 if (fd < 0) 2375 return -1; 2376 c->feed_fd = fd; 115329f1 2377 85f07f22 FB 2378 c->stream->feed_write_index = ffm_read_write_index(fd); 2379 c->stream->feed_size = lseek(fd, 0, SEEK_END); 2380 lseek(fd, 0, SEEK_SET); 2381 2382 /* init buffer input */ 2383 c->buffer_ptr = c->buffer; 2384 c->buffer_end = c->buffer + FFM_PACKET_SIZE; 2385 c->stream->feed_opened = 1; 2386 return 0; 2387} 115329f1 2388 85f07f22 FB 2389static int http_receive_data(HTTPContext *c) 2390{ 85f07f22 FB 2391 HTTPContext *c1; 2392 a6e14edd PG 2393 if (c->buffer_end > c->buffer_ptr) { 2394 int len; 2395 c60202df 2396 len = recv(c->fd, c->buffer_ptr, c->buffer_end - c->buffer_ptr, 0); a6e14edd 2397 if (len < 0) { 8da4034f AB 2398 if (ff_neterrno() != FF_NETERROR(EAGAIN) && 2399 ff_neterrno() != FF_NETERROR(EINTR)) { a6e14edd PG 2400 /* error : close connection */ 2401 goto fail; 2402 } 2403 } else if (len == 0) { 2404 /* end of connection : close it */ 2405 goto fail; 2406 } else { 2407 c->buffer_ptr += len; 2408 c->data_count += len; 5eb765ef 2409 update_datarate(&c->datarate, c->data_count); a6e14edd PG 2410 } 2411 } 2412 d445a7e9 PG 2413 if (c->buffer_ptr - c->buffer >= 2 && c->data_count > FFM_PACKET_SIZE) { 2414 if (c->buffer[0] != 'f' || 2415 c->buffer[1] != 'm') { 2416 http_log("Feed stream has become desynchronized -- disconnecting\n"); 2417 goto fail; 2418 } 2419 } 2420 85f07f22 2421 if (c->buffer_ptr >= c->buffer_end) { f747e6d3 2422 FFStream *feed = c->stream; 85f07f22 FB 2423 /* a packet has been received : write it in the store, except 2424 if header */ 2425 if (c->data_count > FFM_PACKET_SIZE) { 115329f1 2426 949b1a13 2427 // printf("writing pos=0x%"PRIx64" size=0x%"PRIx64"\n", feed->feed_write_index, feed->feed_size); 85f07f22 FB 2428 /* XXX: use llseek or url_seek */ 2429 lseek(c->feed_fd, feed->feed_write_index, SEEK_SET); 2430 write(c->feed_fd, c->buffer, FFM_PACKET_SIZE); 115329f1 2431 85f07f22 FB 2432 feed->feed_write_index += FFM_PACKET_SIZE; 2433 /* update file size */ 2434 if (feed->feed_write_index > c->stream->feed_size) 2435 feed->feed_size = feed->feed_write_index; 2436 2437 /* handle wrap around if max file size reached */ 6b0bdc75 2438 if (c->stream->feed_max_size && feed->feed_write_index >= c->stream->feed_max_size) 85f07f22 FB 2439 feed->feed_write_index = FFM_PACKET_SIZE; 2440 2441 /* write index */ 2442 ffm_write_write_index(c->feed_fd, feed->feed_write_index); 2443 2444 /* wake up any waiting connections */ 2445 for(c1 = first_http_ctx; c1 != NULL; c1 = c1->next) { 115329f1 2446 if (c1->state == HTTPSTATE_WAIT_FEED && 85f07f22 FB 2447 c1->stream->feed == c->stream->feed) { 2448 c1->state = HTTPSTATE_SEND_DATA; 2449 } 2450 } f747e6d3 PG 2451 } else { 2452 /* We have a header in our hands that contains useful data */ 2453 AVFormatContext s; bd7cf6ad 2454 AVInputFormat *fmt_in; f747e6d3 PG 2455 ByteIOContext *pb = &s.pb; 2456 int i; 2457 2458 memset(&s, 0, sizeof(s)); 2459 2460 url_open_buf(pb, c->buffer, c->buffer_end - c->buffer, URL_RDONLY); 2461 pb->buf_end = c->buffer_end; /* ?? */ 2462 pb->is_streamed = 1; 2463 bd7cf6ad FB 2464 /* use feed output format name to find corresponding input format */ 2465 fmt_in = av_find_input_format(feed->fmt->name); 2466 if (!fmt_in) 2467 goto fail; 2468 98486a6b RS 2469 if (fmt_in->priv_data_size > 0) { 2470 s.priv_data = av_mallocz(fmt_in->priv_data_size); 2471 if (!s.priv_data) 2472 goto fail; bb270c08 DB 2473 } else 2474 s.priv_data = NULL; ec3b2232 2475 bd7cf6ad 2476 if (fmt_in->read_header(&s, 0) < 0) { ec3b2232 2477 av_freep(&s.priv_data); f747e6d3 PG 2478 goto fail; 2479 } 2480 2481 /* Now we have the actual streams */ 2482 if (s.nb_streams != feed->nb_streams) { ec3b2232 2483 av_freep(&s.priv_data); f747e6d3 PG 2484 goto fail; 2485 } 2486 for (i = 0; i < s.nb_streams; i++) { 115329f1 2487 memcpy(feed->streams[i]->codec, 01f4895c 2488 s.streams[i]->codec, sizeof(AVCodecContext)); 115329f1 2489 } ec3b2232 2490 av_freep(&s.priv_data); 85f07f22 FB 2491 } 2492 c->buffer_ptr = c->buffer; 2493 } 2494 85f07f22 FB 2495 return 0; 2496 fail: 2497 c->stream->feed_opened = 0; 2498 close(c->feed_fd); 2499 return -1; 2500} 2501 2effd274 FB 2502/********************************************************************/ 2503/* RTSP handling */ 2504 2505static void rtsp_reply_header(HTTPContext *c, enum RTSPStatusCode error_number) 2506{ 2507 const char *str; 2508 time_t ti; 2509 char *p; 2510 char buf2[32]; 2511 2512 switch(error_number) { 7e665cd3 LA 2513 case RTSP_STATUS_OK: 2514 str = "OK"; 2515 break; 2516 case RTSP_STATUS_METHOD: 2517 str = "Method Not Allowed"; 2518 break; 2519 case RTSP_STATUS_BANDWIDTH: 2520 str = "Not Enough Bandwidth"; 2521 break; 2522 case RTSP_STATUS_SESSION: 2523 str = "Session Not Found"; 2524 break; 2525 case RTSP_STATUS_STATE: 2526 str = "Method Not Valid in This State"; 2527 break; 2528 case RTSP_STATUS_AGGREGATE: 2529 str = "Aggregate operation not allowed"; 2530 break; 2531 case RTSP_STATUS_ONLY_AGGREGATE: 2532 str = "Only aggregate operation allowed"; 2533 break; 2534 case RTSP_STATUS_TRANSPORT: 2535 str = "Unsupported transport"; 2536 break; 2537 case RTSP_STATUS_INTERNAL: 2538 str = "Internal Server Error"; 2539 break; 2540 case RTSP_STATUS_SERVICE: 2541 str = "Service Unavailable"; 2542 break; 2543 case RTSP_STATUS_VERSION: 2544 str = "RTSP Version not supported"; 2545 break; 2effd274 FB 2546 default: 2547 str = "Unknown Error"; 2548 break; 2549 } 115329f1 2550 2effd274 FB 2551 url_fprintf(c->pb, "RTSP/1.0 %d %s\r\n", error_number, str); 2552 url_fprintf(c->pb, "CSeq: %d\r\n", c->seq); 2553 2554 /* output GMT time */ 2555 ti = time(NULL); 2556 p = ctime(&ti); 2557 strcpy(buf2, p); 2558 p = buf2 + strlen(p) - 1; 2559 if (*p == '\n') 2560 *p = '\0'; 2561 url_fprintf(c->pb, "Date: %s GMT\r\n", buf2); 2562} 2563 2564static void rtsp_reply_error(HTTPContext *c, enum RTSPStatusCode error_number) 2565{ 2566 rtsp_reply_header(c, error_number); 2567 url_fprintf(c->pb, "\r\n"); 2568} 2569 2570static int rtsp_parse_request(HTTPContext *c) 2571{ 2572 const char *p, *p1, *p2; 2573 char cmd[32]; 2574 char url[1024]; 2575 char protocol[32]; 2576 char line[1024]; 2577 ByteIOContext pb1; 2578 int len; 2579 RTSPHeader header1, *header = &header1; 115329f1 2580 2effd274 FB 2581 c->buffer_ptr[0] = '\0'; 2582 p = c->buffer; 115329f1 2583 2effd274 FB 2584 get_word(cmd, sizeof(cmd), &p); 2585 get_word(url, sizeof(url), &p); 2586 get_word(protocol, sizeof(protocol), &p); 2587 f7d78f36 MR 2588 av_strlcpy(c->method, cmd, sizeof(c->method)); 2589 av_strlcpy(c->url, url, sizeof(c->url)); 2590 av_strlcpy(c->protocol, protocol, sizeof(c->protocol)); 2effd274 FB 2591 2592 c->pb = &pb1; 2593 if (url_open_dyn_buf(c->pb) < 0) { 2594 /* XXX: cannot do more */ 2595 c->pb = NULL; /* safety */ 2596 return -1; 2597 } 2598 2599 /* check version name */ 2600 if (strcmp(protocol, "RTSP/1.0") != 0) { 2601 rtsp_reply_error(c, RTSP_STATUS_VERSION); 2602 goto the_end; 2603 } 2604 2605 /* parse each header line */ 2606 memset(header, 0, sizeof(RTSPHeader)); 2607 /* skip to next line */ 2608 while (*p != '\n' && *p != '\0') 2609 p++; 2610 if (*p == '\n') 2611 p++; 2612 while (*p != '\0') { 2613 p1 = strchr(p, '\n'); 2614 if (!p1) 2615 break; 2616 p2 = p1; 2617 if (p2 > p && p2[-1] == '\r') 2618 p2--; 2619 /* skip empty line */ 2620 if (p2 == p) 2621 break; 2622 len = p2 - p; 2623 if (len > sizeof(line) - 1) 2624 len = sizeof(line) - 1; 2625 memcpy(line, p, len); 2626 line[len] = '\0'; 2627 rtsp_parse_line(header, line); 2628 p = p1 + 1; 2629 } 2630 2631 /* handle sequence number */ 2632 c->seq = header->seq; 2633 2634 if (!strcmp(cmd, "DESCRIBE")) { 2635 rtsp_cmd_describe(c, url); 0df65975 AR 2636 } else if (!strcmp(cmd, "OPTIONS")) { 2637 rtsp_cmd_options(c, url); 2effd274 FB 2638 } else if (!strcmp(cmd, "SETUP")) { 2639 rtsp_cmd_setup(c, url, header); 2640 } else if (!strcmp(cmd, "PLAY")) { 2641 rtsp_cmd_play(c, url, header); 2642 } else if (!strcmp(cmd, "PAUSE")) { 2643 rtsp_cmd_pause(c, url, header); 2644 } else if (!strcmp(cmd, "TEARDOWN")) { 2645 rtsp_cmd_teardown(c, url, header); 2646 } else { 2647 rtsp_reply_error(c, RTSP_STATUS_METHOD); 2648 } 2649 the_end: 2650 len = url_close_dyn_buf(c->pb, &c->pb_buffer); 2651 c->pb = NULL; /* safety */ 2652 if (len < 0) { 2653 /* XXX: cannot do more */ 2654 return -1; 2655 } 2656 c->buffer_ptr = c->pb_buffer; 2657 c->buffer_end = c->pb_buffer + len; 2658 c->state = RTSPSTATE_SEND_REPLY; 2659 return 0; 2660} 2661 829ac53d FB 2662/* XXX: move that to rtsp.c, but would need to replace FFStream by 2663 AVFormatContext */ 115329f1 2664static int prepare_sdp_description(FFStream *stream, uint8_t **pbuffer, 829ac53d 2665 struct in_addr my_ip) 2effd274 FB 2666{ 2667 ByteIOContext pb1, *pb = &pb1; 0fa45e19 2668 int i, payload_type, port, private_payload_type, j; 2effd274 FB 2669 const char *ipstr, *title, *mediatype; 2670 AVStream *st; 115329f1 2671 2effd274 FB 2672 if (url_open_dyn_buf(pb) < 0) 2673 return -1; 115329f1 2674 2effd274 FB 2675 /* general media info */ 2676 2677 url_fprintf(pb, "v=0\n"); 829ac53d 2678 ipstr = inet_ntoa(my_ip); 2effd274 FB 2679 url_fprintf(pb, "o=- 0 0 IN IP4 %s\n", ipstr); 2680 title = stream->title; 2681 if (title[0] == '\0') 2682 title = "No Title"; 2683 url_fprintf(pb, "s=%s\n", title); 2684 if (stream->comment[0] != '\0') 2685 url_fprintf(pb, "i=%s\n", stream->comment); 829ac53d FB 2686 if (stream->is_multicast) { 2687 url_fprintf(pb, "c=IN IP4 %s\n", inet_ntoa(stream->multicast_ip)); 2688 } 2effd274 2689 /* for each stream, we output the necessary info */ e240a0bb 2690 private_payload_type = RTP_PT_PRIVATE; 2effd274 FB 2691 for(i = 0; i < stream->nb_streams; i++) { 2692 st = stream->streams[i]; 01f4895c 2693 if (st->codec->codec_id == CODEC_ID_MPEG2TS) { 2effd274 2694 mediatype = "video"; e240a0bb 2695 } else { 01f4895c 2696 switch(st->codec->codec_type) { e240a0bb FB 2697 case CODEC_TYPE_AUDIO: 2698 mediatype = "audio"; 2699 break; 2700 case CODEC_TYPE_VIDEO: 2701 mediatype = "video"; 2702 break; 2703 default: 2704 mediatype = "application"; 2705 break; 2706 } 2effd274 2707 } 829ac53d FB 2708 /* NOTE: the port indication is not correct in case of 2709 unicast. It is not an issue because RTSP gives it */ 01f4895c 2710 payload_type = rtp_get_payload_type(st->codec); 0fa45e19 FB 2711 if (payload_type < 0) 2712 payload_type = private_payload_type++; 829ac53d FB 2713 if (stream->is_multicast) { 2714 port = stream->multicast_port + 2 * i; 2715 } else { 2716 port = 0; 2717 } 115329f1 2718 url_fprintf(pb, "m=%s %d RTP/AVP %d\n", 829ac53d 2719 mediatype, port, payload_type); e240a0bb 2720 if (payload_type >= RTP_PT_PRIVATE) { 0fa45e19 2721 /* for private payload type, we need to give more info */ 01f4895c 2722 switch(st->codec->codec_id) { 0fa45e19 FB 2723 case CODEC_ID_MPEG4: 2724 { 2725 uint8_t *data; 115329f1 2726 url_fprintf(pb, "a=rtpmap:%d MP4V-ES/%d\n", 0fa45e19 FB 2727 payload_type, 90000); 2728 /* we must also add the mpeg4 header */ 01f4895c 2729 data = st->codec->extradata; 0fa45e19 2730 if (data) { 17705a34 2731 url_fprintf(pb, "a=fmtp:%d config=", payload_type); 01f4895c 2732 for(j=0;j<st->codec->extradata_size;j++) { 0fa45e19 FB 2733 url_fprintf(pb, "%02x", data[j]); 2734 } 2735 url_fprintf(pb, "\n"); 2736 } 2737 } 2738 break; 2739 default: 2740 /* XXX: add other codecs ? */ 2741 goto fail; 2742 } 2743 } 2effd274 FB 2744 url_fprintf(pb, "a=control:streamid=%d\n", i); 2745 } 2746 return url_close_dyn_buf(pb, pbuffer); 0fa45e19 FB 2747 fail: 2748 url_close_dyn_buf(pb, pbuffer); 2749 av_free(*pbuffer); 2750 return -1; 2effd274 FB 2751} 2752 0df65975 AR 2753static void rtsp_cmd_options(HTTPContext *c, const char *url) 2754{ 2755// rtsp_reply_header(c, RTSP_STATUS_OK); 2756 url_fprintf(c->pb, "RTSP/1.0 %d %s\r\n", RTSP_STATUS_OK, "OK"); 2757 url_fprintf(c->pb, "CSeq: %d\r\n", c->seq); 2758 url_fprintf(c->pb, "Public: %s\r\n", "OPTIONS, DESCRIBE, SETUP, TEARDOWN, PLAY, PAUSE"); 2759 url_fprintf(c->pb, "\r\n"); 2760} 2761 2effd274 FB 2762static void rtsp_cmd_describe(HTTPContext *c, const char *url) 2763{ 2764 FFStream *stream; 2765 char path1[1024]; 2766 const char *path; 0c1a9eda 2767 uint8_t *content; 829ac53d FB 2768 int content_length, len; 2769 struct sockaddr_in my_addr; 115329f1 2770 2effd274 2771 /* find which url is asked */ 6ba5cbc6 2772 url_split(NULL, 0, NULL, 0, NULL, 0, NULL, path1, sizeof(path1), url); 2effd274 FB 2773 path = path1; 2774 if (*path == '/') 2775 path++; 2776 2777 for(stream = first_stream; stream != NULL; stream = stream->next) { d2a067d1 2778 if (!stream->is_feed && stream->fmt == &rtp_muxer && 2effd274 FB 2779 !strcmp(path, stream->filename)) { 2780 goto found; 2781 } 2782 } 2783 /* no stream found */ 2784 rtsp_reply_error(c, RTSP_STATUS_SERVICE); /* XXX: right error ? */ 2785 return; 2786 2787 found: 2788 /* prepare the media description in sdp format */ 829ac53d FB 2789 2790 /* get the host IP */ 2791 len = sizeof(my_addr); 2792 getsockname(c->fd, (struct sockaddr *)&my_addr, &len); 829ac53d 2793 content_length = prepare_sdp_description(stream, &content, my_addr.sin_addr); 2effd274 FB 2794 if (content_length < 0) { 2795 rtsp_reply_error(c, RTSP_STATUS_INTERNAL); 2796 return; 2797 } 2798 rtsp_reply_header(c, RTSP_STATUS_OK); 2799 url_fprintf(c->pb, "Content-Type: application/sdp\r\n"); 2800 url_fprintf(c->pb, "Content-Length: %d\r\n", content_length); 2801 url_fprintf(c->pb, "\r\n"); 2802 put_buffer(c->pb, content, content_length); 2803} 2804 2805static HTTPContext *find_rtp_session(const char *session_id) 2806{ 2807 HTTPContext *c; 2808 2809 if (session_id[0] == '\0') 2810 return NULL; 2811 2812 for(c = first_http_ctx; c != NULL; c = c->next) { 2813 if (!strcmp(c->session_id, session_id)) 2814 return c; 2815 } 2816 return NULL; 2817} 2818 b29f97d1 2819static RTSPTransportField *find_transport(RTSPHeader *h, enum RTSPProtocol protocol) 2effd274 FB 2820{ 2821 RTSPTransportField *th; 2822 int i; 2823 2824 for(i=0;i<h->nb_transports;i++) { 2825 th = &h->transports[i]; 2826 if (th->protocol == protocol) 2827 return th; 2828 } 2829 return NULL; 2830} 2831 115329f1 2832static void rtsp_cmd_setup(HTTPContext *c, const char *url, 2effd274 FB 2833 RTSPHeader *h) 2834{ 2835 FFStream *stream; 2836 int stream_index, port; 2837 char buf[1024]; 2838 char path1[1024]; 2839 const char *path; 2840 HTTPContext *rtp_c; 2841 RTSPTransportField *th; 2842 struct sockaddr_in dest_addr; 2843 RTSPActionServerSetup setup; 115329f1 2844 2effd274 2845 /* find which url is asked */ 6ba5cbc6 2846 url_split(NULL, 0, NULL, 0, NULL, 0, NULL, path1, sizeof(path1), url); 2effd274 FB 2847 path = path1; 2848 if (*path == '/') 2849 path++; 2850 2851 /* now check each stream */ 2852 for(stream = first_stream; stream != NULL; stream = stream->next) { d2a067d1 2853 if (!stream->is_feed && stream->fmt == &rtp_muxer) { 2effd274 FB 2854 /* accept aggregate filenames only if single stream */ 2855 if (!strcmp(path, stream->filename)) { 2856 if (stream->nb_streams != 1) { 2857 rtsp_reply_error(c, RTSP_STATUS_AGGREGATE); 2858 return; 2859 } 2860 stream_index = 0; 2861 goto found; 2862 } 115329f1 2863 2effd274 FB 2864 for(stream_index = 0; stream_index < stream->nb_streams; 2865 stream_index++) { 115329f1 2866 snprintf(buf, sizeof(buf), "%s/streamid=%d", 2effd274 FB 2867 stream->filename, stream_index); 2868 if (!strcmp(path, buf)) 2869 goto found; 2870 } 2871 } 2872 } 2873 /* no stream found */ 2874 rtsp_reply_error(c, RTSP_STATUS_SERVICE); /* XXX: right error ? */ 2875 return; 2876 found: 2877 2878 /* generate session id if needed */ 2879 if (h->session_id[0] == '\0') { 1df93ae9 AB 2880 snprintf(h->session_id, sizeof(h->session_id), "%08x%08x", 2881 av_random(&random_state), av_random(&random_state)); 2effd274 FB 2882 } 2883 2884 /* find rtp session, and create it if none found */ 2885 rtp_c = find_rtp_session(h->session_id); 2886 if (!rtp_c) { bc351386 FB 2887 /* always prefer UDP */ 2888 th = find_transport(h, RTSP_PROTOCOL_RTP_UDP); 2889 if (!th) { 2890 th = find_transport(h, RTSP_PROTOCOL_RTP_TCP); 2891 if (!th) { 2892 rtsp_reply_error(c, RTSP_STATUS_TRANSPORT); 2893 return; 2894 } 2895 } 2896 2897 rtp_c = rtp_new_connection(&c->from_addr, stream, h->session_id, 2898 th->protocol); 2effd274 FB 2899 if (!rtp_c) { 2900 rtsp_reply_error(c, RTSP_STATUS_BANDWIDTH); 2901 return; 2902 } 2903 2904 /* open input stream */ 2905 if (open_input_stream(rtp_c, "") < 0) { 2906 rtsp_reply_error(c, RTSP_STATUS_INTERNAL); 2907 return; 2908 } 2effd274 2909 } 115329f1 2910 2effd274 FB 2911 /* test if stream is OK (test needed because several SETUP needs 2912 to be done for a given file) */ 2913 if (rtp_c->stream != stream) { 2914 rtsp_reply_error(c, RTSP_STATUS_SERVICE); 2915 return; 2916 } 115329f1 2917 2effd274 FB 2918 /* test if stream is already set up */ 2919 if (rtp_c->rtp_ctx[stream_index]) { 2920 rtsp_reply_error(c, RTSP_STATUS_STATE); 2921 return; 2922 } 2923 2924 /* check transport */ 2925 th = find_transport(h, rtp_c->rtp_protocol); 115329f1 2926 if (!th || (th->protocol == RTSP_PROTOCOL_RTP_UDP && 2effd274 FB 2927 th->client_port_min <= 0)) { 2928 rtsp_reply_error(c, RTSP_STATUS_TRANSPORT); 2929 return; 2930 } 2931 2932 /* setup default options */ 2933 setup.transport_option[0] = '\0'; 2934 dest_addr = rtp_c->from_addr; 2935 dest_addr.sin_port = htons(th->client_port_min); 115329f1 2936 2effd274 2937 /* setup stream */ bc351386 2938 if (rtp_new_av_stream(rtp_c, stream_index, &dest_addr, c) < 0) { 2effd274 FB 2939 rtsp_reply_error(c, RTSP_STATUS_TRANSPORT); 2940 return; 2941 } 2942 2943 /* now everything is OK, so we can send the connection parameters */ 2944 rtsp_reply_header(c, RTSP_STATUS_OK); 2945 /* session ID */ 2946 url_fprintf(c->pb, "Session: %s\r\n", rtp_c->session_id); 2947 2948 switch(rtp_c->rtp_protocol) { 2949 case RTSP_PROTOCOL_RTP_UDP: 2950 port = rtp_get_local_port(rtp_c->rtp_handles[stream_index]); 2951 url_fprintf(c->pb, "Transport: RTP/AVP/UDP;unicast;" 2952 "client_port=%d-%d;server_port=%d-%d", 2953 th->client_port_min, th->client_port_min + 1, 2954 port, port + 1); 2955 break; 2956 case RTSP_PROTOCOL_RTP_TCP: 2957 url_fprintf(c->pb, "Transport: RTP/AVP/TCP;interleaved=%d-%d", 2958 stream_index * 2, stream_index * 2 + 1); 2959 break; 2960 default: 2961 break; 2962 } 2963 if (setup.transport_option[0] != '\0') { 2964 url_fprintf(c->pb, ";%s", setup.transport_option); 2965 } 2966 url_fprintf(c->pb, "\r\n"); 115329f1 2967 2effd274 FB 2968 2969 url_fprintf(c->pb, "\r\n"); 2970} 2971 2972 2973/* find an rtp connection by using the session ID. Check consistency 2974 with filename */ 115329f1 2975static HTTPContext *find_rtp_session_with_url(const char *url, 2effd274 FB 2976 const char *session_id) 2977{ 2978 HTTPContext *rtp_c; 2979 char path1[1024]; 2980 const char *path; 94d9ad5f GF 2981 char buf[1024]; 2982 int s; 2effd274 FB 2983 2984 rtp_c = find_rtp_session(session_id); 2985 if (!rtp_c) 2986 return NULL; 2987 2988 /* find which url is asked */ 6ba5cbc6 2989 url_split(NULL, 0, NULL, 0, NULL, 0, NULL, path1, sizeof(path1), url); 2effd274 FB 2990 path = path1; 2991 if (*path == '/') 2992 path++; 94d9ad5f GF 2993 if(!strcmp(path, rtp_c->stream->filename)) return rtp_c; 2994 for(s=0; s<rtp_c->stream->nb_streams; ++s) { 2995 snprintf(buf, sizeof(buf), "%s/streamid=%d", 2996 rtp_c->stream->filename, s); 2997 if(!strncmp(path, buf, sizeof(buf))) { 2998 // XXX: Should we reply with RTSP_STATUS_ONLY_AGGREGATE if nb_streams>1? 2999 return rtp_c; 3000 } 3001 } 3002 return NULL; 2effd274 FB 3003} 3004 3005static void rtsp_cmd_play(HTTPContext *c, const char *url, RTSPHeader *h) 3006{ 3007 HTTPContext *rtp_c; 3008 3009 rtp_c = find_rtp_session_with_url(url, h->session_id); 3010 if (!rtp_c) { 3011 rtsp_reply_error(c, RTSP_STATUS_SESSION); 3012 return; 3013 } 115329f1 3014 2effd274 FB 3015 if (rtp_c->state != HTTPSTATE_SEND_DATA && 3016 rtp_c->state != HTTPSTATE_WAIT_FEED && 3017 rtp_c->state != HTTPSTATE_READY) { 3018 rtsp_reply_error(c, RTSP_STATUS_STATE); 3019 return; 3020 } 3021 e240a0bb FB 3022#if 0 3023 /* XXX: seek in stream */ 3024 if (h->range_start != AV_NOPTS_VALUE) { 3025 printf("range_start=%0.3f\n", (double)h->range_start / AV_TIME_BASE); 3026 av_seek_frame(rtp_c->fmt_in, -1, h->range_start); 3027 } 3028#endif 3029 2effd274 3030 rtp_c->state = HTTPSTATE_SEND_DATA; 115329f1 3031 2effd274 FB 3032 /* now everything is OK, so we can send the connection parameters */ 3033 rtsp_reply_header(c, RTSP_STATUS_OK); 3034 /* session ID */ 3035 url_fprintf(c->pb, "Session: %s\r\n", rtp_c->session_id); 3036 url_fprintf(c->pb, "\r\n"); 3037} 3038 3039static void rtsp_cmd_pause(HTTPContext *c, const char *url, RTSPHeader *h) 3040{ 3041 HTTPContext *rtp_c; 3042 3043 rtp_c = find_rtp_session_with_url(url, h->session_id); 3044 if (!rtp_c) { 3045 rtsp_reply_error(c, RTSP_STATUS_SESSION); 3046 return; 3047 } 115329f1 3048 2effd274 FB 3049 if (rtp_c->state != HTTPSTATE_SEND_DATA && 3050 rtp_c->state != HTTPSTATE_WAIT_FEED) { 3051 rtsp_reply_error(c, RTSP_STATUS_STATE); 3052 return; 3053 } 115329f1 3054 2effd274 3055 rtp_c->state = HTTPSTATE_READY; 1bc1cfdd 3056 rtp_c->first_pts = AV_NOPTS_VALUE; 2effd274 FB 3057 /* now everything is OK, so we can send the connection parameters */ 3058 rtsp_reply_header(c, RTSP_STATUS_OK); 3059 /* session ID */ 3060 url_fprintf(c->pb, "Session: %s\r\n", rtp_c->session_id); 3061 url_fprintf(c->pb, "\r\n"); 3062} 3063 3064static void rtsp_cmd_teardown(HTTPContext *c, const char *url, RTSPHeader *h) 3065{ 3066 HTTPContext *rtp_c; b0b2faa7 3067 char session_id[32]; 2effd274 FB 3068 3069 rtp_c = find_rtp_session_with_url(url, h->session_id); 3070 if (!rtp_c) { 3071 rtsp_reply_error(c, RTSP_STATUS_SESSION); 3072 return; 3073 } 115329f1 3074 f7d78f36 3075 av_strlcpy(session_id, rtp_c->session_id, sizeof(session_id)); b0b2faa7 3076 2effd274 FB 3077 /* abort the session */ 3078 close_connection(rtp_c); 3079 2effd274 FB 3080 /* now everything is OK, so we can send the connection parameters */ 3081 rtsp_reply_header(c, RTSP_STATUS_OK); 3082 /* session ID */ b0b2faa7 3083 url_fprintf(c->pb, "Session: %s\r\n", session_id); 2effd274 FB 3084 url_fprintf(c->pb, "\r\n"); 3085} 3086 3087 3088/********************************************************************/ 3089/* RTP handling */ 3090 115329f1 3091static HTTPContext *rtp_new_connection(struct sockaddr_in *from_addr, bc351386 FB 3092 FFStream *stream, const char *session_id, 3093 enum RTSPProtocol rtp_protocol) 2effd274 FB 3094{ 3095 HTTPContext *c = NULL; bc351386 3096 const char *proto_str; 115329f1 3097 2effd274 FB 3098 /* XXX: should output a warning page when coming 3099 close to the connection limit */ 3100 if (nb_connections >= nb_max_connections) 3101 goto fail; 115329f1 3102 2effd274 FB 3103 /* add a new connection */ 3104 c = av_mallocz(sizeof(HTTPContext)); 3105 if (!c) 3106 goto fail; 115329f1 3107 2effd274 FB 3108 c->fd = -1; 3109 c->poll_entry = NULL; 6edd6884 3110 c->from_addr = *from_addr; 2effd274 FB 3111 c->buffer_size = IOBUFFER_INIT_SIZE; 3112 c->buffer = av_malloc(c->buffer_size); 3113 if (!c->buffer) 3114 goto fail; 3115 nb_connections++; 3116 c->stream = stream; f7d78f36 3117 av_strlcpy(c->session_id, session_id, sizeof(c->session_id)); 2effd274 FB 3118 c->state = HTTPSTATE_READY; 3119 c->is_packetized = 1; bc351386 FB 3120 c->rtp_protocol = rtp_protocol; 3121 2effd274 3122 /* protocol is shown in statistics */ bc351386 FB 3123 switch(c->rtp_protocol) { 3124 case RTSP_PROTOCOL_RTP_UDP_MULTICAST: 3125 proto_str = "MCAST"; 3126 break; 3127 case RTSP_PROTOCOL_RTP_UDP: 3128 proto_str = "UDP"; 3129 break; 3130 case RTSP_PROTOCOL_RTP_TCP: 3131 proto_str = "TCP"; 3132 break; 3133 default: 3134 proto_str = "???"; 3135 break; 3136 } f7d78f36 MR 3137 av_strlcpy(c->protocol, "RTP/", sizeof(c->protocol)); 3138 av_strlcat(c->protocol, proto_str, sizeof(c->protocol)); 2effd274 3139 6edd6884 FB 3140 current_bandwidth += stream->bandwidth; 3141 2effd274 FB 3142 c->next = first_http_ctx; 3143 first_http_ctx = c; 3144 return c; 115329f1 3145 2effd274 FB 3146 fail: 3147 if (c) { 3148 av_free(c->buffer); 3149 av_free(c); 3150 } 3151 return NULL; 3152} 3153 3154/* add a new RTP stream in an RTP connection (used in RTSP SETUP bc351386 3155 command). If RTP/TCP protocol is used, TCP connection 'rtsp_c' is 2effd274 3156 used. */ 115329f1 3157static int rtp_new_av_stream(HTTPContext *c, bc351386 FB 3158 int stream_index, struct sockaddr_in *dest_addr, 3159 HTTPContext *rtsp_c) 2effd274 FB 3160{ 3161 AVFormatContext *ctx; 3162 AVStream *st; 3163 char *ipaddr; 3164 URLContext *h; 0c1a9eda 3165 uint8_t *dummy_buf; 6edd6884 3166 char buf2[32]; bc351386 3167 int max_packet_size; 115329f1 3168 2effd274 3169 /* now we can open the relevant output stream */ bc874dae 3170 ctx = av_alloc_format_context(); 2effd274 FB 3171 if (!ctx) 3172 return -1; d2a067d1 3173 ctx->oformat = &rtp_muxer; 2effd274 FB 3174 3175 st = av_mallocz(sizeof(AVStream)); 3176 if (!st) 3177 goto fail; 8d931070 3178 st->codec= avcodec_alloc_context(); 2effd274 FB 3179 ctx->nb_streams = 1; 3180 ctx->streams[0] = st; 3181 115329f1 3182 if (!c->stream->feed || 2effd274 FB 3183 c->stream->feed == c->stream) { 3184 memcpy(st, c->stream->streams[stream_index], sizeof(AVStream)); 3185 } else { 115329f1 3186 memcpy(st, 2effd274 FB 3187 c->stream->feed->streams[c->stream->feed_streams[stream_index]], 3188 sizeof(AVStream)); 3189 } 57dbe08b 3190 st->priv_data = NULL; 115329f1 3191 bc351386 FB 3192 /* build destination RTP address */ 3193 ipaddr = inet_ntoa(dest_addr->sin_addr); 3194 3195 switch(c->rtp_protocol) { 3196 case RTSP_PROTOCOL_RTP_UDP: 3197 case RTSP_PROTOCOL_RTP_UDP_MULTICAST: 3198 /* RTP/UDP case */ 115329f1 3199 6edd6884 FB 3200 /* XXX: also pass as parameter to function ? */ 3201 if (c->stream->is_multicast) { 3202 int ttl; 3203 ttl = c->stream->multicast_ttl; 3204 if (!ttl) 3205 ttl = 16; 3206 snprintf(ctx->filename, sizeof(ctx->filename), 115329f1 3207 "rtp://%s:%d?multicast=1&ttl=%d", 6edd6884 FB 3208 ipaddr, ntohs(dest_addr->sin_port), ttl); 3209 } else { 3210 snprintf(ctx->filename, sizeof(ctx->filename), 3211 "rtp://%s:%d", ipaddr, ntohs(dest_addr->sin_port)); 3212 } 2effd274 FB 3213 3214 if (url_open(&h, ctx->filename, URL_WRONLY) < 0) 3215 goto fail; 3216 c->rtp_handles[stream_index] = h; bc351386 FB 3217 max_packet_size = url_get_max_packet_size(h); 3218 break; 3219 case RTSP_PROTOCOL_RTP_TCP: 3220 /* RTP/TCP case */ 3221 c->rtsp_c = rtsp_c; 3222 max_packet_size = RTSP_TCP_MAX_PACKET_SIZE; 3223 break; 3224 default: 2effd274 FB 3225 goto fail; 3226 } 3227 bc351386 3228 http_log("%s:%d - - [%s] \"PLAY %s/streamid=%d %s\"\n", 115329f1 DB 3229 ipaddr, ntohs(dest_addr->sin_port), 3230 ctime1(buf2), bc351386 3231 c->stream->filename, stream_index, c->protocol); 6edd6884 3232 2effd274 3233 /* normally, no packets should be output here, but the packet size may be checked */ bc351386 3234 if (url_open_dyn_packet_buf(&ctx->pb, max_packet_size) < 0) { 2effd274 FB 3235 /* XXX: close stream */ 3236 goto fail; 3237 } 3c27199b 3238 av_set_parameters(ctx, NULL); 2effd274 FB 3239 if (av_write_header(ctx) < 0) { 3240 fail: 3241 if (h) 3242 url_close(h); 3243 av_free(ctx); 3244 return -1; 3245 } 3246 url_close_dyn_buf(&ctx->pb, &dummy_buf); 3247 av_free(dummy_buf); 115329f1 3248 2effd274 FB 3249 c->rtp_ctx[stream_index] = ctx; 3250 return 0; 3251} 3252 3253/********************************************************************/ 3254/* ffserver initialization */ 3255 b29f97d1 3256static AVStream *add_av_stream1(FFStream *stream, AVCodecContext *codec) 2effd274 FB 3257{ 3258 AVStream *fst; 3259 3260 fst = av_mallocz(sizeof(AVStream)); 3261 if (!fst) 3262 return NULL; 8d931070 3263 fst->codec= avcodec_alloc_context(); 2effd274 3264 fst->priv_data = av_mallocz(sizeof(FeedData)); 01f4895c MN 3265 memcpy(fst->codec, codec, sizeof(AVCodecContext)); 3266 fst->codec->coded_frame = &dummy_frame; d445a7e9 3267 fst->index = stream->nb_streams; 7c054ea7 3268 av_set_pts_info(fst, 33, 1, 90000); 2effd274 FB 3269 stream->streams[stream->nb_streams++] = fst; 3270 return fst; 3271} 3272 85f07f22 3273/* return the stream number in the feed */ b29f97d1 3274static int add_av_stream(FFStream *feed, AVStream *st) 85f07f22 FB 3275{ 3276 AVStream *fst; 3277 AVCodecContext *av, *av1; 3278 int i; 3279 01f4895c 3280 av = st->codec; 85f07f22 FB 3281 for(i=0;i<feed->nb_streams;i++) { 3282 st = feed->streams[i]; 01f4895c 3283 av1 = st->codec; f747e6d3 PG 3284 if (av1->codec_id == av->codec_id && 3285 av1->codec_type == av->codec_type && 85f07f22 FB 3286 av1->bit_rate == av->bit_rate) { 3287 3288 switch(av->codec_type) { 3289 case CODEC_TYPE_AUDIO: 3290 if (av1->channels == av->channels && 3291 av1->sample_rate == av->sample_rate) 3292 goto found; 3293 break; 3294 case CODEC_TYPE_VIDEO: 3295 if (av1->width == av->width && 3296 av1->height == av->height && c0df9d75 MN 3297 av1->time_base.den == av->time_base.den && 3298 av1->time_base.num == av->time_base.num && 85f07f22 FB 3299 av1->gop_size == av->gop_size) 3300 goto found; 3301 break; f747e6d3 3302 default: 0f4e8165 3303 abort(); 85f07f22 FB 3304 } 3305 } 3306 } 115329f1 3307 2effd274 3308 fst = add_av_stream1(feed, av); 85f07f22 FB 3309 if (!fst) 3310 return -1; 85f07f22 FB 3311 return feed->nb_streams - 1; 3312 found: 3313 return i; 3314} 3315 b29f97d1 3316static void remove_stream(FFStream *stream) 2effd274 FB 3317{ 3318 FFStream **ps; 3319 ps = &first_stream; 3320 while (*ps != NULL) { 3321 if (*ps == stream) { 3322 *ps = (*ps)->next; 3323 } else { 3324 ps = &(*ps)->next; 3325 } 3326 } 3327} 3328 0fa45e19 3329/* specific mpeg4 handling : we extract the raw parameters */ b29f97d1 3330static void extract_mpeg4_header(AVFormatContext *infile) 0fa45e19 FB 3331{ 3332 int mpeg4_count, i, size; 3333 AVPacket pkt; 3334 AVStream *st; 0c1a9eda 3335 const uint8_t *p; 0fa45e19 FB 3336 3337 mpeg4_count = 0; 3338 for(i=0;i<infile->nb_streams;i++) { 3339 st = infile->streams[i]; 01f4895c MN 3340 if (st->codec->codec_id == CODEC_ID_MPEG4 && 3341 st->codec->extradata_size == 0) { 0fa45e19 FB 3342 mpeg4_count++; 3343 } 3344 } 3345 if (!mpeg4_count) 3346 return; 3347 d445a7e9 3348 printf("MPEG4 without extra data: trying to find header in %s\n", infile->filename); 0fa45e19 FB 3349 while (mpeg4_count > 0) { 3350 if (av_read_packet(infile, &pkt) < 0) 3351 break; 3352 st = infile->streams[pkt.stream_index]; 01f4895c MN 3353 if (st->codec->codec_id == CODEC_ID_MPEG4 && 3354 st->codec->extradata_size == 0) { 3355 av_freep(&st->codec->extradata); 0fa45e19 FB 3356 /* fill extradata with the header */ 3357 /* XXX: we make hard suppositions here ! */ 3358 p = pkt.data; 3359 while (p < pkt.data + pkt.size - 4) { 3360 /* stop when vop header is found */ 115329f1 3361 if (p[0] == 0x00 && p[1] == 0x00 && 0fa45e19 FB 3362 p[2] == 0x01 && p[3] == 0xb6) { 3363 size = p - pkt.data; 750f0e1f 3364 // av_hex_dump_log(infile, AV_LOG_DEBUG, pkt.data, size); 01f4895c MN 3365 st->codec->extradata = av_malloc(size); 3366 st->codec->extradata_size = size; 3367 memcpy(st->codec->extradata, pkt.data, size); 0fa45e19 FB 3368 break; 3369 } 3370 p++; 3371 } 3372 mpeg4_count--; 3373 } 3374 av_free_packet(&pkt); 3375 } 3376} 3377 2effd274 3378/* compute the needed AVStream for each file */ b29f97d1 3379static void build_file_streams(void) 2effd274 FB 3380{ 3381 FFStream *stream, *stream_next; 3382 AVFormatContext *infile; 3383 int i; 3384 3385 /* gather all streams */ 3386 for(stream = first_stream; stream != NULL; stream = stream_next) { 3387 stream_next = stream->next; 3388 if (stream->stream_type == STREAM_TYPE_LIVE && 3389 !stream->feed) { 3390 /* the stream comes from a file */ 3391 /* try to open the file */ 3392 /* open stream */ e240a0bb 3393 stream->ap_in = av_mallocz(sizeof(AVFormatParameters)); d2a067d1 3394 if (stream->fmt == &rtp_muxer) { e240a0bb FB 3395 /* specific case : if transport stream output to RTP, 3396 we use a raw transport stream reader */ 3397 stream->ap_in->mpeg2ts_raw = 1; 3398 stream->ap_in->mpeg2ts_compute_pcr = 1; 3399 } 115329f1 DB 3400 3401 if (av_open_input_file(&infile, stream->feed_filename, e240a0bb 3402 stream->ifmt, 0, stream->ap_in) < 0) { 2effd274 FB 3403 http_log("%s not found", stream->feed_filename); 3404 /* remove stream (no need to spend more time on it) */ 3405 fail: 3406 remove_stream(stream); 3407 } else { 3408 /* find all the AVStreams inside and reference them in 3409 'stream' */ 3410 if (av_find_stream_info(infile) < 0) { 115329f1 3411 http_log("Could not find codec parameters from '%s'", 2effd274 FB 3412 stream->feed_filename); 3413 av_close_input_file(infile); 3414 goto fail; 3415 } 0fa45e19 FB 3416 extract_mpeg4_header(infile); 3417 2effd274 3418 for(i=0;i<infile->nb_streams;i++) { 01f4895c 3419 add_av_stream1(stream, infile->streams[i]->codec); 2effd274 FB 3420 } 3421 av_close_input_file(infile); 3422 } 3423 } 3424 } 3425} 3426 85f07f22 3427/* compute the needed AVStream for each feed */ b29f97d1 3428static void build_feed_streams(void) 85f07f22 FB 3429{ 3430 FFStream *stream, *feed; 3431 int i; 3432 3433 /* gather all streams */ 3434 for(stream = first_stream; stream != NULL; stream = stream->next) { 3435 feed = stream->feed; 3436 if (feed) { 3437 if (!stream->is_feed) { 2effd274 3438 /* we handle a stream coming from a feed */ 85f07f22 FB 3439 for(i=0;i<stream->nb_streams;i++) { 3440 stream->feed_streams[i] = add_av_stream(feed, stream->streams[i]); 3441 } cde25790 PG 3442 } 3443 } 3444 } 3445 3446 /* gather all streams */ 3447 for(stream = first_stream; stream != NULL; stream = stream->next) { 3448 feed = stream->feed; 3449 if (feed) { 3450 if (stream->is_feed) { 85f07f22 FB 3451 for(i=0;i<stream->nb_streams;i++) { 3452 stream->feed_streams[i] = i; 3453 } 3454 }
__label__pos
0.945595
The Boyer-Moore string matching algorithm in C This C program implements the Boyer-Moore string searching algorithm. #include <stdlib.h> #include <stdio.h> #include <string.h> #include <stdint.h> #include <stdarg.h> #include "bm.h" #ifdef HEADER typedef struct bm { /* The string to search for. */ const unsigned char * needle; /* The length of "needle" in bytes. */ int len; /* The offset of the first byte of where "needle" was found in the searched string. */ int offset; /* True if "needle" was found within the searched string, false otherwise. */ unsigned int found : 1; /* True if initialised correctly. */ unsigned int init : 1; /* The length into "needle" where each byte is last found. */ int delta1[0x100]; int * delta2; } bm_t; typedef enum { bm_status_ok, bm_status_memory_failure, /* Over or under flow. */ bm_status_bounds, /* Uninitialised structure. */ bm_status_uninit, } bm_status_t; #endif /* def HEADER */ /* https://www.lemoda.net/c/die/index.html */ static void error (const char * format, ...) { va_list vargs; va_start (vargs, format); vfprintf (stderr, format, vargs); fprintf (stderr, ".\n"); va_end (vargs); exit (EXIT_FAILURE); } /* Print messages if necessary to debug. */ #if 0 || defined(TEST) #define MESSAGE(...) { \ fprintf (stderr, "# %s:%d: ", \ __FILE__, __LINE__); \ fprintf (stderr, __VA_ARGS__); \ fprintf (stderr, ".\n"); \ } #else #define MESSAGE(...) #endif /* 0 */ /* Given "word" of length "wordlen", is word[0..pos-1] the same as word[wordlen - pos - 1 .. wordlen - 1], in other words is the end of the string from pos onwards a prefix of the string. */ static int is_prefix (const unsigned char *word, int wordlen, int pos) { int i; int suffixlen = wordlen - pos; for (i = 0; i < suffixlen; i++) { if (word[i] != word[pos + i]) { MESSAGE ("%d: %c != %c", i, word[i], word[pos + i]); return 0; } } MESSAGE ("%s %d %.*s It's a prefix", word, suffixlen, suffixlen, word); return 1; } /* Find the longest suffix of "bm->needle" which is the same as "bm->needle[pos]". */ static bm_status_t suffix_length (const bm_t * bm, int pos, int * length_ptr) { int i; int end; if (pos >= bm->len) { error ("bad position %d >= %d", pos, bm->len); return bm_status_bounds; } if (pos <= 0) { error ("meaningless position %d <= 0", pos); return bm_status_bounds; } MESSAGE ("suffix_length (%s, %d, %d)", bm->needle, bm->len, pos); end = bm->len - 1; for (i = 0; i < pos; i++) { if (bm->needle[pos - i] != bm->needle[end - i]) { MESSAGE ("%d %c != %d %c", pos - i, bm->needle[pos - i], end - i, bm->needle[end - i]); break; } } MESSAGE ("%d %.*s suffix %.*s", i, i, bm->needle + pos - i, i, bm->needle + bm->len - 1 - i); * length_ptr = i; return bm_status_ok; } /* Fill in the delta2 table of "bm" using the given search string. */ static bm_status_t make_delta2 (bm_t * bm) { int p; int last_prefix_index; /* Allocate memory for bm->delta2. */ bm->delta2 = calloc (bm->len, sizeof (* bm->delta2)); if (! bm->delta2) { error ("No memory"); return bm_status_memory_failure; } /* "bm->needle" is always a prefix of itself, so "last_prefix_index" starts at "bm->len". */ last_prefix_index = bm->len; for (p = bm->len - 2; p >= 0; p--) { /* If the first p+1 characters of bm->needle are the same as the final p+1 characters of bm->needle, the last prefix index is p+1, in other words we can shift the string down by p+1 characters. */ if (is_prefix (bm->needle, bm->len, p + 1)) { last_prefix_index = p + 1; } /* In the case of a mismatch at character "p", we can jump over this many characters to find the next match. */ bm->delta2[p] = last_prefix_index + bm->len - 1 - p; } /* Search for common substrings which are not prefixes. */ for (p = 1; p < bm->len - 1; p++) { int slen; int offset; bm_status_t status; status = suffix_length (bm, p, & slen); if (status != bm_status_ok) { error ("Bad status %d getting suffix length (%s, %d, %d)", status, bm->needle, bm->len, p); return status; } offset = bm->len - 1 - slen; if (bm->needle[p - slen] != bm->needle[offset]) { MESSAGE ("%s: Mismatch %c, %c at %d %d", bm->needle, bm->needle[p - slen], bm->needle[offset], p-slen, offset); bm->delta2[offset] = bm->len - 1 - p + slen; } } bm->init = 1; return bm_status_ok; } /* Quiet some compiler warnings. */ static int ustrlen (const unsigned char * u) { return strlen ((const char *) u); } /* Given "needle", allocate the values of "bm". */ bm_status_t bm_prepare (bm_t * bm, const unsigned char * needle) { int i; bm->found = 0; bm->init = 0; bm->needle = needle; bm->len = ustrlen (needle); for (i = 0; i < 0x100; i++) { /* 'If "char" does not occur in "pat", "delta_1" is "patlen".' [BM p.763] */ bm->delta1[i] = bm->len; } for (i = 0; i < bm->len; i++) { /* 'If "char" does occur in "pat", "delta_1" is the difference between "patlen" and the position of the rightmost occurrence of "char" in "pat".' [BM p.763] */ bm->delta1[needle[i]] = bm->len - 1 - i; } return make_delta2 (bm); } bm_status_t bm_free (bm_t * bm) { free (bm->delta2); bm->delta2 = 0; return bm_status_ok; } static int delta2 (bm_t * bm, int j) { if (j >= bm->len) { error ("index %d >= pattern length %d", j, bm->len); } return bm->delta2[j]; } /* Search for bm->needle in "haystack". */ bm_status_t bm_search (bm_t * bm, const unsigned char * haystack) { int stringlen; int i; int j; if (! bm->init) { error ("Uninitialised structure"); return bm_status_uninit; } stringlen = ustrlen (haystack); i = bm->len - 1; while (1) { int d1; int d2; if (i > stringlen) { bm->found = 0; return bm_status_ok; } MESSAGE ("pointer = %d, comparing %.*s", i, bm->len, haystack + i - (bm->len - 1)); j = bm->len - 1; while (bm->needle[j] == haystack[i]) { /* 'we have matched all of "pat" (and thus have succeeded in finding a match)' [BM p.763] */ if (j == 0) { bm->found = 1; bm->offset = i; return bm_status_ok; } j--; i--; } /* 'we come to a mismatch at some new "char" after matching the last "m" characters of "pat".' [BM p.763] */ d1 = bm->delta1[haystack[i]]; d2 = delta2 (bm, j); if (d1 > d2) { /* Move the pointer forward to the next possible place which could match, disregarding suffix/prefix. */ MESSAGE ("delta_1 > delta_2: jumping by %d for char %c", d1, haystack[i]); i += d1; } else { MESSAGE ("delta_2 >= delta_1: jumping ahead by %d", d2); i += d2; } } return bm_status_ok; } #ifdef TEST /* _____ _ _ |_ _|__ ___| |_(_)_ __ __ _ | |/ _ \/ __| __| | '_ \ / _` | | | __/\__ \ |_| | | | | (_| | |_|\___||___/\__|_|_| |_|\__, | |___/ */ #include "c-tap-test.h" #define CALL_OK(X) { \ bm_status_t status; \ status = X; \ TAP_TEST_EQUAL (status, bm_status_ok); \ } static void is_prefix_test () { const unsigned char * test; test = (const unsigned char *) "ABCDABC"; TAP_TEST_EQUAL (is_prefix (test, 7, 7), 1); TAP_TEST_EQUAL (is_prefix (test, 7, 3), 0); TAP_TEST_EQUAL (is_prefix (test, 7, 4), 1); } static void suffix_length_test () { const unsigned char * test; int slen; bm_t bm = {0}; test = (const unsigned char *) "ABCDABC"; bm.len = ustrlen (test); bm.needle = test; MESSAGE ("suffix length test for %s", test); CALL_OK (suffix_length (& bm, 6, & slen)); TAP_TEST_EQUAL (slen, 6); CALL_OK (suffix_length (& bm, 2, & slen)); TAP_TEST_EQUAL (slen, 2); CALL_OK (suffix_length (& bm, 4, & slen)); /* DABC is not the same as ABCD. */ TAP_TEST_EQUAL (slen, 0); } /* This tests whether delta_2 contains the same values as those of [BM p. 765]. */ static void bm_prepare_test () { bm_t bm; int d2[] = {14, 13, 12, 11, 10, 9, 11, 10, 1}; int d3[] = {17, 16, 15, 14, 13, 12, 7, 10, 1}; int i; bm_prepare (& bm, (unsigned char *) "ABCXXXABC"); for (i = 0; i < sizeof (d2) / sizeof (int); i++) { TAP_TEST_EQUAL (bm.delta2[i], d2[i]); } bm_free (& bm); bm_prepare (& bm, (unsigned char *) "ABYXCDEYX"); for (i = 0; i < sizeof (d3) / sizeof (int); i++) { TAP_TEST_EQUAL (bm.delta2[i], d3[i]); } bm_free (& bm); /* This very simple case is not in Boyer-Moore's original paper. */ bm_prepare (& bm, (unsigned char *) "ABCDEFGHIJK"); for (i = 0; i < bm.len - 1; i++) { TAP_TEST_EQUAL (bm.delta2[i], 2 * bm.len - i - 1); } bm_free (& bm); } int main () { bm_t bm; /* [BM p.764]. */ const unsigned char * needle = (unsigned char *) "AT-THAT"; const unsigned char * haystack = (unsigned char *) "WHICH-FINALLY-HALTS.--AT-THAT-POINT"; const unsigned char * needle1 = (unsigned char *) "needle"; const unsigned char * haystack1 = (unsigned char *) "For your own ladies and pale-visag'd maids,\n" "Like Amazons, come tripping after drums,\n" "Their thimbles into armed gauntlets change,\n" "Their needles to lances, and their gentle hearts\n" "To fierce and bloody inclination.\n"; is_prefix_test (); suffix_length_test (); exit (0); bm_prepare_test (); bm_prepare (& bm, needle1); bm_search (& bm, haystack1); TAP_TEST (bm.found); bm_free (& bm); bm_prepare (& bm, needle); bm_search (& bm, haystack); TAP_TEST (bm.found); bm_free (& bm); TAP_PLAN; return 0; } #endif /* def TEST */ (download) The header file looks like this. /* This file was generated by the following command: cfunctions bm.c */ #ifndef CFH_BM_H #define CFH_BM_H #line 7 "bm.c" typedef struct bm { /* The string to search for. */ const unsigned char * needle; /* The length of "needle" in bytes. */ int len; /* The offset of the first byte of where "needle" was found in the searched string. */ int offset; /* True if "needle" was found within the searched string, false otherwise. */ unsigned int found : 1; /* True if initialised correctly. */ unsigned int init : 1; /* The length into "needle" where each byte is last found. */ int delta1[0x100]; int * delta2; } bm_t; typedef enum { bm_status_ok, bm_status_memory_failure, /* Over or under flow. */ bm_status_bounds, /* Uninitialised structure. */ bm_status_uninit, } bm_status_t; #line 153 "bm.c" bm_status_t bm_prepare (bm_t* bm, const unsigned char* needle); #line 173 "bm.c" bm_status_t bm_free (bm_t* bm); #line 192 "bm.c" bm_status_t bm_search (bm_t* bm, const unsigned char* haystack); #endif /* CFH_BM_H */ (download) The makefile looks like this. CFLAGS=-Wall -g -O all: bm.txt bm.txt: bm-example ./bm-example > $@ bm-example: bm-example.c bm.c bm.h $(CC) $(CFLAGS) -o $@ bm.c bm-example.c test: bm-test prove ./bm-test bm-test: bm.c bm.h c-tap-test.h $(CC) -DTEST $(CFLAGS) -o $@ bm.c bm.h: bm.c cfunctions bm.c CTT=/home/ben/projects/c-tap-test c-tap-test.h: $(CTT)/$@ cp $(CTT)/$@ $@ clean: -rm -f bm-test bm.txt c-tap-test.h bm-example *.o bm.h purge (download) Here is a simple example of its use. #include <stdio.h> #include "bm.h" #define CALL(X) { \ bm_status_t status; \ status = X; \ if (status != bm_status_ok) { \ fprintf (stderr, "Bad status %d from %s.\n", \ status, #X); \ return -1; \ } \ } int main () { bm_t bm; const unsigned char * needle1 = (unsigned char *) "needle"; const unsigned char * haystack1 = (unsigned char *) "For your own ladies and pale-visag'd maids,\n" "Like Amazons, come tripping after drums,\n" "Their thimbles into armed gauntlets change,\n" "Their needles to lances, and their gentle hearts\n" "To fierce and bloody inclination.\n"; CALL (bm_prepare (& bm, needle1)); CALL (bm_search (& bm, haystack1)); if (bm.found) { printf ("Found at byte %d.\n", bm.offset); } else { printf ("Not found.\n"); } CALL (bm_free (& bm)); return 0; } (download) The output of the example looks like this: Found at byte 135. The [BM p.123] quotes in the source code above refer to the location of the quote in Boyer and Moore's paper, The Boyer-Moore Fast String Searching Algorithm. Copyright © Ben Bullock 2009-2021. All rights reserved. For comments, questions, and corrections, please email Ben Bullock ([email protected]) or use the discussion group at Google Groups. News about the site. / Privacy / Disclaimer
__label__pos
0.994595
8 $\begingroup$ Let $A\in \mathbb{R}^{n\times n}$, symmetric and positive definite. Suppose it takes $m$ units of work to multiply a vector by $A$. It is well known that performing the CG algorithm on $A$ with condition number $\kappa$ requires $\mathcal{O} (m\sqrt{\kappa})$, units of work. Now, of course, being a $\mathcal{O}$ statement this is an upper-bound. And the CG algorithm can always terminate in zero steps with a lucky initial guess. Do we know if there exists a RHS and an initial (unlucky) guess that will require $\mathcal{\Theta}(\sqrt{\kappa})$ steps? Put another way, is worst-case work-complexity of CG really $\Theta( m \sqrt{\kappa})$? This question arises when I tried to determine if the benefit of a preconditioner (lower $\kappa$) outweighed its cost (higher $m$). Right now, I am working with toy problems and would like to have a better idea before I implement anything in a compiled language. $\endgroup$ • 5 $\begingroup$ You could presumably construct a pessimal initial guess by running the CG algorithm "backwards" and putting suitable energy into each of the $A$-orthogonal search directions that the algorithm requires all the steps. $\endgroup$ – origimbo Mar 17 '16 at 16:11 8 $\begingroup$ The answer is a resounding yes. The convergence rate bound of $(\sqrt{\kappa}-1) / (\sqrt{\kappa}+1)$ is sharp over the set of symmetric positive definite matrices with condition number $\kappa$. In other words, knowing nothing more about $A$ than its condition number, CG really can take $\sim\sqrt{\kappa}$ iterations to converge. Loosely speaking, the upper-bound is attained if the eigenvalues of $A$ are uniformly distributed (i.e. "peppered") within an interval of condition number $\kappa$. Here's a more rigorous statement. Deterministic versions are more involved but work using the same principles. Theorem (Worst-case choice of $A$). Pick any random orthogonal matrix $U$, let $\lambda_1,\ldots,\lambda_n$ be $n$ real numbers uniformly sampled from the real interval $[1,\kappa]$, and let $b=[b_1;\ldots;b_n]$ be $n$ real numbers sampled i.i.d. from the standard Gaussian. Define $$A=U\mathrm{diag}(\lambda_1,\ldots,\lambda_n)U^T.$$ Then in the limit $n\to\infty$, conjugate gradients will convergence with probability one to an $\epsilon$ accurate solution of $Ax=b$ in no less than $\Omega(\sqrt{\kappa}\log\epsilon^{-1})$ iterations. Proof. The standard proof is based on optimal Chebyshev polynomial approximations, using techniques found in a number of places, such as Greenbaum's book or Saad's book. $\endgroup$ • 1 $\begingroup$ The bound is not sharp, as the answer explains later, If the eigenvalues are not uniformly distributed, cg converges faster, since it is not a stationalry iteration. Thus, we need to know more about the matrix. $\endgroup$ – Guido Kanschat Mar 20 '16 at 3:39 • $\begingroup$ @GuidoKanschat: Good point, and I've fixed the statement to clarify that sharpness is attained over all $A$ with condition $\kappa$. $\endgroup$ – Richard Zhang Mar 21 '16 at 15:50 • $\begingroup$ The proof boils down to minimizing $\|p(A)\|$ in the space of order-$k$ polynomials satisfying $p(0)=1$. Equivalently this is $\min_p \max_{\lambda\in\Lambda(A)} |p(\lambda)|$. In the stated limit, $\Lambda(A)\to[1,\kappa]$, and the solution for the minimax problem is then the Chebyshev polynomial, whose error converges as $\sim\sqrt{\kappa}$ $\endgroup$ – Richard Zhang Mar 21 '16 at 21:27 0 $\begingroup$ Taking this as my original question: Do we know if there exists a RHS and an initial (unlucky) guess that will require $\Theta(\sqrt{\kappa})$ steps? The answer to the question is "no". The idea of this answer comes from the comment from Guido Kanschat. Claim: For any given condition number $k$, there exists a matrix $A$, with that condition number for which the CG algorithm will terminate in at most two steps (for any given RHS and initial guess). Consider $A\in \mathbb{R}^{n\times n} $ where $A=\mathrm{diag}(1,\kappa,\kappa,\ldots, \kappa)$. Then the condition number of $A$ is $\kappa$. Let $b\in \mathbb{R}^n$ be the RHS, and denote the eigenvalues of $A$ as $\lambda_i$ where $$\lambda_i = \left\{\begin{array}{ll}1 & i=1\\ \kappa & i\not= 1 \end{array} \right. . $$ We first consider the case where $x^{(0)} \in \mathbb{R}^n$, the initial guess, is zero. Denote $x^{(2)}\in \mathbb{R}^n$ as the second estimate of $A^{-1}b$ from the CG algorithm. We show that $x^{(2)} =A^{-1}b$ by showing $\langle x^{(2)}-A^{-1}b, A(x^{(2)}-A^{-1}b)\rangle =0$. Indeed, we have \begin{align*} \langle x^{(2)}-A^{-1}b, A(x^{(2)}-A^{-1}b)\rangle &= \left\| x^{(2)}-A^{-1}b \right\|_A^2 \\ &=\min_{p\in \mathrm{poly}_{1} } \left\| (p(A)-A^{-1}) b \right\|_A^2\\ &=\min_{p\in \mathrm{poly}_{1} } \sum_{i=1}^n (p(\lambda_i) - \lambda_i^{-1})^2 \lambda_i b_i^2 \\ &\le \sum_{i=1}^n (\widehat{p}(\lambda_i) - \lambda_i^{-1})^2 \lambda_i b_i^2 = 0 \end{align*} Where we use the first order polynomial $\widehat{p}$ defined as $\widehat{p}(x)= (1+\kappa-x)/\kappa$. So we proven the case for $x^{(0)}= 0$. If $x^{(0)} \not = 0$, then $x^{(2)}= \overline{x^{(2)}}+ x^{(0)}$ where $\overline{x^{(2)} }$ is the second estimate of the CG algorithm with $b$ replaced with $\overline{b} = b-A x^{(0)}$. So we have reduced this case to the previous one. $\endgroup$ • $\begingroup$ How much of this is robust to finite precision arithmetic? $\endgroup$ – origimbo Mar 21 '16 at 16:28 • $\begingroup$ @origimbo If your question was directed to me, the answer is, "I don't know." $\endgroup$ – fred Mar 21 '16 at 16:36 Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.996979
NRT score&audio rendering issues Hi, I am relatively new to SuperCollider & having some difficulty setting up non-realtime score & audio rendering. My primary problem, I think, is in figuring out how to load my synthdef and audio buffer onto the NRT server. I have prepared an example (three .scd files and an example .wav to read into buffer) that presents my issue: http://petervanhaaften.net/files/nrt-example.zip In ‘main.scd’, I give two examples: 1) working pattern playback with synthdef and buffer on the default server, 2) my first attempt to convert that to work in NRT. In this example, I have a Pdef pattern (‘score.scd’) controlling a granular Synthdef (‘gransyn.scd’) which operates on a stored buffer (check ‘main.scd’). I know for sure I am incorrectly allocating my buffer and synthdef on the NRT server in my second example in ‘main.scd’, however I have tried a lot of iterations of this without success. I have also made some experiments with the CTK library to resolve that, due to info found in old threads like: NRT mode and soundfile buffers …which I could not get working and didn’t include for reasons of conciseness. But I am also open to a solution using CTK. Might anyone be able to offer any pointers on this topic? & thank you in advance. Hi, I have attempted to simplify my problem, so that it could be clearly shown in a single post. Below is a working example with default server playback. A sample buffer is loaded, a granular synthdef is loaded, and a pattern performs the synthdef: ( b = Buffer.read(s, Platform.resourceDir +/+ "sounds/a11wlk01-44_1.aiff"); SynthDef(\buf_grain_test, { |out, gate = 1, amp = 1, pan = 0, sndbuf, envbuf| var env, freqdev; env = EnvGen.kr( Env([0, 1, 0], [1, 1], \sin, 1), gate, levelScale: amp, doneAction: Done.freeSelf); Out.ar(out, GrainBuf.ar(2, Impulse.kr(20), 0.1, sndbuf, LFNoise1.kr.range(0.5, 2), LFNoise2.kr(0.1).range(0, 1), 2, pan, envbuf) * env) }).add; ~pattern = ( Pdef(\scene0, Pdef(\part1, Ppar([ Pmono( \buf_grain_test, \sndbuf, b, \envbuf, Pseq([-1, 0, -0.5, 0.5], 1), \dur, Pseq([2, 2, 2, 2], 1), \pan, Pseq([-1, 1, -1, 1], 1), ), ]) )).play; ); ) Below is a non-working example of converting this to NRT. It throws me the error: “ERROR: Message ‘at’ not understood”. I am guessing this is because my synthdef and/or audio buffer is not correctly called in the NRT server. ( var score, sndbuf, options; SynthDef(\buf_grain_test, { |out, gate = 1, amp = 1, pan = 0, sndbuf, envbuf| var env, freqdev; env = EnvGen.kr( Env([0, 1, 0], [1, 1], \sin, 1), gate, levelScale: amp, doneAction: Done.freeSelf); Out.ar(out, GrainBuf.ar(2, Impulse.kr(20), 0.1, sndbuf, LFNoise1.kr.range(0.5, 2), LFNoise2.kr(0.1).range(0, 1), 2, pan, envbuf) * env) }).load(s); ~score = Score.new; // create a Buffer object for adding to the Score sndbuf = Buffer.new; // for NRT rendering, the buffer messages must be added to the Score ~score.add([0, sndbuf.allocReadMsg(Platform.resourceDir +/+ "sounds/a11wlk01-44_1.aiff")]); ~pattern = Pdef(\scene0, Pdef(\part1, Ppar([ Pmono( \buf_grain_test, \sndbuf, b, \envbuf, Pseq([-1, 0, -0.5, 0.5], 1), \dur, Pseq([2, 2, 2, 2], 1), \pan, Pseq([-1, 1, -1, 1], 1), ), ]) )); ~pattern = ~pattern.asScore(60); ~score.add(~pattern); // the ServerOptions for rendering the soundfile options = ServerOptions.new.numOutputBusChannels_(2); ( // Destination path and file name ~outFile = "~/test.wav"; // Render the score as wav file ~score.recordNRT(outputFilePath: ~outFile.asAbsolutePath, headerFormat: "WAV", options: options); ); ) So, maybe someone has an idea where I’m making a mistake. I think it is something simple! Thank you. Hi, You’re using global variable b instead of local variable sndbuf in Pmono. And there’s a catch which the .add function. // ~score.add(~pattern); ~pattern.score.do{|e| ~score.add(e)}; // I found myself using it this way ~score.score.do{|e| e.postln}; // Posts each OSC-msg as a line. Use this to see the difference. Here’s the full code: ( var sndbuf, options; SynthDef(\buf_grain_test, { |out, gate = 1, amp = 1, pan = 0, sndbuf, envbuf| var env, freqdev; env = EnvGen.kr( Env([0, 1, 0], [1, 1], \sin, 1), gate, levelScale: amp, doneAction: Done.freeSelf); Out.ar(out, GrainBuf.ar(2, Impulse.kr(20), 0.1, sndbuf, LFNoise1.kr.range(0.5, 2), LFNoise2.kr(0.1).range(0, 1), 2, pan, envbuf) * env) }).load(s); ~score = Score.new; // create a Buffer object for adding to the Score sndbuf = Buffer.new; // for NRT rendering, the buffer messages must be added to the Score ~score.add([0, sndbuf.allocReadMsg(Platform.resourceDir +/+ "sounds/a11wlk01-44_1.aiff")]); ~pattern = Pdef(\scene0, Pdef(\part1, Ppar([ Pmono( \buf_grain_test, \sndbuf, sndbuf, \envbuf, Pseq([-1, 0, -0.5, 0.5], 1), \dur, Pseq([2, 2, 2, 2], 1), \pan, Pseq([-1, 1, -1, 1], 1), ), ]) )); ~pattern = ~pattern.asScore(60); // ~score.add(~pattern); ~pattern.score.do{|e| ~score.add(e)}; // I found myself using it this way // the ServerOptions for rendering the soundfile options = ServerOptions.new.numOutputBusChannels_(2); ( // Destination path and file name ~outFile = "~/test.wav"; // Render the score as wav file ~score.recordNRT(outputFilePath: ~outFile.asAbsolutePath, headerFormat: "WAV", options: options); ); ) Hi Jildert, Thank you for your solution! I have tested and it is working. However my simplified explanation of my problem has slightly complicated my problem. By chance do you have a moment to look at the following, which is my attempt to resolve my original issue, using my actual instruments & pattern? Again, I think it is something very simple, inside of the “main.scd” code (bottom of this post). I have a score [“score.scd”]: ~pattern = Pdef(\scene0, Pdef(\part1, Ppar([ Pmono( \gransyn, \soundBuf, ~sndbuf, \dur, Pseq([1, 5, 4,5], 1), \attack, Pseq([0, 0, 0, 1], 1), \release, Pseq([0, 0, 0, 1], 1), \posLo, Pseq([0, 0, 0, 1], 1), \posHi, Pseq([1, 1, 1, 1], 1), \posRateE, Pseq([0, 0, 0, 0], 1), \posRateM, Pseq([1, 0.5, 1, 1], 1), \posRateMLag, Pseq([0, 0, 0, 0], 1), \posRateMCurve, Pseq([0, 0, 0, 0], 1), \overlap, Pseq([1, 2, 1, 1], 1), \overlapLag, Pseq([0, 2, 1, 1], 1), \overlapCurve, Pseq([0, 0, 0, 0], 1), \trigRate, Pseq([0, 10, 2, 2], 1), \trigRateLag, Pseq([0, 0, 3, 3], 1), \trigRateCurve, Pseq([0, 0, 0, 0], 1), \rate, Pseq([1, 3, 1, 1], 1), \rateLag, Pseq([0, 3, 2, 0], 1), \rateCurve, Pseq([0, 0, 0, 0], 1), \lpFreq, Pseq([20000,20000, 20000, 20000], 1), \lpLag, Pseq([0, 0, 0, 0], 1), \lpCurve, Pseq([0, 0, 0, 0], 1), \hpFreq,Pseq([10, 10, 10, 10], 1), \hpLag, Pseq([0, 0, 0, 0], 1), \hpCurve, Pseq([0, 0, 0, 0], 1), \panMax, Pseq([0, 0.2, 0, 0], 1), \amp, Pseq([0, 1, 0, 0], 1), \ampLag, Pseq([3, 1, 1/2, 4], 1), \ampCurve, Pseq([0, 0, 0, 0], 1), ), ]) )); That controls a synthdef [“gransyn.scd”]: ( ~m = 5; ~n = 2 * ~m; SynthDef(\gransyn, { |out = 0, soundBuf, gate = 1, attack = 0.01, release = 0.5, posLo = 0.1, posHi = 0.9, posRateE = 0, posRateM= 1, posRateMLag = 0, posRateMCurve = 0, rate= 1, rateLag = 0, rateCurve = 0, panMax = 0, bpRQ = 0.1, bpRQLag = 0, bpRQCurve = 0, bpLo = 50, bpLoLag = 0, bpLoCurve = 0, bpHi = 5000, bpHiLag = 0, bpHiCurve = 0, amp = 1, ampLag = 0, ampCurve = 0 overlap = 2, overlapLag = 0, overlapCurve = 0, trigRate = 1, trigRateLag = 0, trigRateCurve = 0, interp = 2, posRate = 0, lpFreq = 20000, lpLag = 3, lpCurve = 0, hpFreq = 20, hpLag = 0, hpCurve = 0, rateRandomness = 0, rateRandomnessLag = 0, rateRandomnessCurve = 0, overlapRandomness = 0, overlapRandomnessLag = 0, overlapRandomnessCurve = 0, verbMix = 0, verbMixLag = 0, verbMixCurve = 0, verbRoom = 0, verbDamp =0| var sig, sigL, sigR, sigOut, sigLimiter, sigCompressor, env, bpFreq, chan, dUgen, trig, trigs, bufDur, pos, lpfSig, rateNoiseSig, overlapNoiseSig, verbSig; //trigger for grains trig = Impulse.ar(trigRate); //randomness for rate rateNoiseSig = PinkNoise.kr(mul: VarLag.kr(rateRandomness, rateRandomnessLag, rateRandomnessCurve), add: 0); //rateNoiseSig = rateNoiseSig / 2 //randomness for overlap overlapNoiseSig = PinkNoise.kr(mul: VarLag.kr(overlapRandomness, overlapRandomnessLag, overlapRandomnessCurve), add: 0); //define all of your VarLag controlled values here trigRate = Demand.ar(trig, 0, VarLag.ar(Demand.ar(trig, 0, trigRate, inf), trigRateLag, trigRateCurve), inf); lpFreq = VarLag.kr(lpFreq, lpLag, lpCurve); hpFreq = VarLag.kr(hpFreq, hpLag, hpCurve); rate = VarLag.ar(Demand.ar(trig, 0, rate + rateNoiseSig, inf), rateLag, rateCurve); overlap = VarLag.ar(Demand.ar(trig, 0, overlap + overlapNoiseSig, inf), overlapLag, overlapCurve); //trigRate = VarLag.kr(trigRate, trigRateLag, trigRateCurve); posRateM = VarLag.kr(posRateM, posRateMLag, posRateMCurve); amp = VarLag.kr(amp, ampLag, ampCurve); verbMix = VarLag.ar(Demand.ar(trig, 0, verbMix, inf), verbMixLag, verbMixCurve); // we need a multichannel trigger that steps through all consecutive channels trigs = { |i| PulseDivider.ar(trig, ~n, ~n-1-i) } ! ~n; chan = Demand.ar(trig, 0, Dseq((0..~n-1), inf)); env = Linen.kr(gate, attack, 1, release, 2) * amp; posRate = 10 ** posRateE * posRateM; bufDur = BufDur.kr(soundBuf); pos = Phasor.ar(0, BufRateScale.kr(soundBuf) * posRate * SampleDur.ir / bufDur, posLo, posHi); sig = TGrains.ar(~n, trig, soundBuf, Demand.ar(trig, 0, rate, inf), pos * bufDur, Demand.ar(trig, 0, overlap, inf) / Demand.ar(trig, 0, trigRate, inf), // Panning convention is that from PanAz, // speakers should be from 0 to 2, but (orientation) // 1/n has to be substracted for n speakers. // If this isn't done correctly grains are spread onto more than one channel // and per-grain application of fxs fails. chan.linlin(0, ~n-1, -1/~n, (2*~n - 3)/~n), 1, interp); dUgen = Dwhite(0.0, 1); sig = sig.collect { |ch, i| // this is the place to define fxs per channel/grain lpfSig = LPF.ar(in: ch, freq: lpFreq, mul: 1, add: 0); HPF.ar(in: lpfSig, freq: hpFreq, mul: 1, add: 0); }; // routing to two channels ... sigL = Mix(((0..(~m-1)) * 2).collect(sig[_])); sigR = Mix(((0..(~m-1)) * 2 + 1).collect(sig[_])); //route stereo sig thru cheap verb verbSig = FreeVerb2.ar(sigL, sigR, mix: Demand.ar(trig, 0, verbMix, inf), room: verbRoom, damp: verbDamp); //output Out.ar(out, Pan2.ar(verbSig[0], panMax.neg) + Pan2.ar(verbSig[1], panMax) * env); }).store; ) And this code [“main.scd”] strings them together to render in NRT, and is based on the example working code you provided me. However this version is not working, it prints an empty file: ( var options; // create new score ~score = Score.new; // call external score Require("score.scd"); // create a Buffer object for adding to the Score ~sndbuf = Buffer.new; //call synthdef Require("gransyn.scd"); //~pattern.play; // for NRT rendering, the buffer messages must be added to the Score ~score.add([0, ~sndbuf.allocReadMsg(Platform.resourceDir +/+ "sounds/a11wlk01-44_1.aiff")]); ~pattern = ~pattern.asScore(100); // ~score.add(~pattern); ~pattern.score.do{|e| ~score.add(e)}; // I found myself using it this way // the ServerOptions for rendering the soundfile options = ServerOptions.new.numOutputBusChannels_(2); ( // Destination path and file name ~outFile = "~/test.wav"; // Render the score as wav file ~score.recordNRT(outputFilePath: ~outFile.asAbsolutePath, headerFormat: "WAV", options: options); ); ) Otherwise. . . I will keep hacking away at this issue. Thank you again for your help. Here’s a simpler version of the problem: SynthDef(\bufGrainPan, { |start, time, bufnum, pan, rate = 1, amp = 1, attack = 0.001, decay = 0.02, outbus| var sig; sig = PlayBuf.ar(1, bufnum, rate * BufRateScale.kr(bufnum), 1, start, 0) * EnvGen.kr(Env.linen(attack, time, decay), doneAction:2); OffsetOut.ar(outbus, Pan2.ar(sig, pan, amp)); }).add; p = Pbind( \instrument, \bufGrainPan, \bufnum, b, \start, 15000, \time, 0.1, \dur, 0.2 ); b = Buffer.read(s, Platform.resourceDir +/+ "sounds/a11wlk01.wav"); q = p.trace.play; // nothing Many many many users get confused about variables in patterns. Variables are resolved to their values immediately, but in many pattern cases, users seem to expect that the variable will be resolved only at the moment of getting a value or event from the stream. In the above, at the moment of creating the Pbind, b is nil. So the Pbind really means: p = Pbind( \instrument, \bufGrainPan, \bufnum, nil, \start, 15000, \time, 0.1, \dur, 0.2 ); nil in Pbind means to stop the stream – so this pattern, by definition, will produce zero events. It seems that users often expect \bufnum, b to mean “create a reference to either a current or future value of b” – in reality, when you use a variable name, it behaves right now as the variable’s current value, only. (If you want the “reference to future,” you have to write something explicit for it, e.g., \bufnum, Pfunc { b }.) In your example, you’re Require-ing the pattern first, and then populating ~sndbuf after that. You should do ~sndbuf = first, and then Require. hjh 1 Like HI jamshark70, thank you very much for your super detailed explanation! So, with the benefit of your example, I now have this partially working. I am still having a small issue that is likely due to syntax, or my method of calling ~pattern.asScore. I am using Ppar() to run two Pmono instances in parallel, like this: ( Pdef(\scene0, Pdef(\part1, Ppar([ Pmono( \gransyn, \soundBuf, ~sndbuf, \dur, Pseq([1, 5, 4,5], 1), \attack, Pseq([0, 0, 0, 1], 1), \release, Pseq([0, 0, 0, 1], 1), \posLo, Pseq([0, 0, 0, 1], 1), \posHi, Pseq([1, 1, 1, 1], 1), \posRateE, Pseq([0, 0, 0, 0], 1), \posRateM, Pseq([1, 0.5, 1, 1], 1), \posRateMLag, Pseq([0, 0, 0, 0], 1), \posRateMCurve, Pseq([0, 0, 0, 0], 1), \overlap, Pseq([1, 2, 1, 1], 1), \overlapLag, Pseq([0, 2, 1, 1], 1), \overlapCurve, Pseq([0, 0, 0, 0], 1), \trigRate, Pseq([0, 10, 2, 2], 1), \trigRateLag, Pseq([0, 0, 3, 3], 1), \trigRateCurve, Pseq([0, 0, 0, 0], 1), \rate, Pseq([1, 3, 1, 1], 1), \rateLag, Pseq([0, 3, 2, 0], 1), \rateCurve, Pseq([0, 0, 0, 0], 1), \lpFreq, Pseq([20000,20000, 20000, 20000], 1), \lpLag, Pseq([0, 0, 0, 0], 1), \lpCurve, Pseq([0, 0, 0, 0], 1), \hpFreq,Pseq([10, 10, 10, 10], 1), \hpLag, Pseq([0, 0, 0, 0], 1), \hpCurve, Pseq([0, 0, 0, 0], 1), \amp, Pseq([0, 1, 0, 0], 1), \ampLag, Pseq([3, 1, 1/2, 4], 1), \ampCurve, Pseq([0, 0, 0, 0], 1), ), Pmono( \gransyn, \soundBuf, ~sndbuf, \dur, Pseq([1, 5, 4, 7], 1), \attack, Pseq([0, 0, 0, 0], 1), \release, Pseq([0, 0, 0, 0], 1), \posLo, Pseq([0, 0, 0.4, 0.4], 1), \posHi, Pseq([1, 1, 0.5, 0.5], 1), \posRateE, Pseq([0, 0, 0, 0], 1), \posRateM, Pseq([1, 0.5, -1, 1], 1), \posRateMLag, Pseq([0, 0, 0, 0], 1), \posRateMCurve, Pseq([0, 0, 0, 0], 1), \overlap, Pseq([1, 2, 2, 3], 1), \overlapLag, Pseq([0, 2, 0, 0], 1), \overlapCurve, Pseq([0, 0, 0, 0], 1), \trigRate, Pseq([0, 10, 10, 20], 1), \trigRateLag, Pseq([0, 0, 0, 2], 1), \trigRateCurve, Pseq([0, 0, 0, 0], 1), \rate, Pseq([1, 1, 0.75, 0.6], 1), \rateLag, Pseq([0, 3, 0, 0], 1), \rateCurve, Pseq([0, 0, 0, 0], 1), \lpFreq, Pseq([20000,20000, 2000, 1000], 1), \lpLag, Pseq([0, 0, 3, 3], 1), \lpCurve, Pseq([0, 0, 0, 0], 1), \hpFreq,Pseq([10, 10, 10, 10], 1), \hpLag, Pseq([0, 0, 0, 0], 1), \hpCurve, Pseq([0, 0, 0, 0], 1), \amp, Pseq([0, 1, 1, 0], 1), \ampLag, Pseq([0, 3, 2, 3], 1), \ampCurve, Pseq([0, 0, 0, 0.5], 1), ), ]) )); ); In this example, only the the second Pmono() stream is included in my NRT recorded audio file. For reference, my current code for recording to NRT is as follows. I use a Pdef(), Pseq(), and Pfindur() to run through the above score (which is actually in 3 parts): ( var options; // create new score ~score = Score.new; // create a Buffer object for adding to the Score ~sndbuf = Buffer.new; // for NRT rendering, the buffer messages must be added to the Score ~score.add([0, ~sndbuf.allocReadMsg(thisProcess.nowExecutingPath.dirname +/+ "voice.wav")]); //call synthdef Require("gransyn.scd"); // call external score Require("score.scd"); ~pattern = Pdef(\scene0, Pdef(\main, Pseq([ //start score Pfindur(17, Pdef(\part1)), Pfindur(20, Pdef(\part2)), Pfindur(22, Pdef(\part3)), ],1), )); ~pattern = ~pattern.asScore(59); // ~score.add(~pattern); ~pattern.score.do{|e| ~score.add(e)}; // the ServerOptions for rendering the soundfile options = ServerOptions.new.numOutputBusChannels_(2); ~score.recordNRT( outputFilePath: "~/test.wav".standardizePath, headerFormat: "wav", sampleFormat: "int24", options: options, duration: 59, action: { "done".postln } ); ) Is there anything obvious I am doing here which would cause .recordNRT to record only one part of two parallel processes in my score? Thank you sincerely for your assistance!
__label__pos
0.89461
Custom Movement Rotation help Hello! I am currently making a custom movement system for my game, and ive made so much progress so far. I make this post because I’m unsure how I should make the players character turn to face the direction they are walking to. Any help would greatly be appreciated. Well, you have to be getting a vector with the direction you are wanting to move, I know this because the character does move in the direction. So you just make the character’s body (primary part) rotate (with an align orientation) to be facing the vector of where you are moving. 1 Like I tried this yes, but unfortunately it didnt work. Im not sure if im doing it right tho, could you show me an example? Try this, maybe this can help. (Just know that I don’t actually have a custom character controller, so I threw this one together in the last hour or so, but maybe its enough to demonstrate how to do the rotation) CustomCharacterMovement.rbxl (44.0 KB) 2 Likes You can calculate the unit vector between last frames hrp cframe and current frame hrp cframe, based on that you can orientate the character on the direction they’re moving. 1 Like Im sorry for the late response! How could i accomplish this? Hello! so sorry for the late response. I looked into the code here, and i tried to impliment it into my system but it still didnt work. Im not exactly sure what method you used for this tho, could you go into detail? so here is the script script.Parent.PrimaryPart.Anchored = true local custom = workspace.Custom.PrimaryPart local alignPosition = custom.AlignPosition local alignOrientation = custom.AlignOrientation workspace.CurrentCamera.CameraSubject = custom local speed = 16 local hipHeight = custom.HipHeight.Value+(custom.Size.Y/2) alignPosition.Position = custom.CFrame.Position alignOrientation.CFrame = custom.CFrame local lastState = "" function DoMove(dir) alignPosition.MaxVelocity = speed local src = Vector3.new(custom.CFrame.Position.X,hipHeight,custom.CFrame.Position.Z) local dst = ((src + (dir*speed))*Vector3.new(1,0,1))+Vector3.new(0,hipHeight,0) local rot = CFrame.lookAt(src,dst)-src alignOrientation.CFrame = rot alignPosition.Position = dst end while true do local dir = script.Parent.Humanoid.MoveDirection if dir.Magnitude<.5 then if lastState ~= "idle" then lastState = "idle" alignPosition.Position = Vector3.new(custom.CFrame.Position.X,hipHeight,custom.CFrame.Position.Z) end else lastState = "moving" DoMove(dir) end wait(.1) end The main thing to have is a direction of which the character is moving… image Then, once you have that direction, you rotate the characters body to face that direction The above code gets the characters position (src) and a position to where the character is trying to move (dst) then gets a cframe rotation (rot) that will be used to rotate the character Then in my case, (using an anignOrientation) I supply it the rotation cframe image There is also another code in the place, that is server side, and it sets the NetworkOwnership of the object to the player (so the player has control from the client of the custom character) game.Players.PlayerAdded:Connect(function(player) player.CharacterAdded:Connect(function(character) character.PrimaryPart.Anchored = true workspace.Custom.PrimaryPart:SetNetworkOwner(player) end) end)
__label__pos
0.946068
• facebook • whatsapp • telegram Circles Definition: A circle is the locus of a point which moves in a plane such that the distancem from a fixed point in that plane is always constant. * The fixed point is called centre of the circle. * The fixed distance is called radius of the circle.   EQUATION OF CIRCLE WITH CENTRE (h, k) AND RADIUS 'r'. The equation of the circle with centre (h, k) and radius 'r' is (x - h)2 + (y - k)2 = r2. * The general equation of the circle is x+ y2 + 2gx + 2fy + c = 0; where g, f, c constants and (-g, -f) is centre of the circle and  (g2 + f2 - c 0) is radius of the circle. * A second degree equation ax2 + 2hxy + by2 + 2gx + 2fy + c = 0 represents a circle if (i) a = b; coefficient of x = coefficient of y (ii) h = 0; coefficient of xy = 0 (iii) ∆ = abc + 2fgh − af2 − bg2 − ch2 ≠ 0 (for point circle ∆ = 0) (iv) g2 + f2 - c 0 * If two circles have same centre and different radius, then they are called concentric circles. * The equations of concentric circles differ by only the constant term.   Equation of Circle on a given Diameter Equation of circle having (x1, y1) and (x2, y2) as the extremities of the diameter is  (x - x1)(x - x2) + (y - y1)(y - y2) = 0  Different forms of the Equations of a Circle   1) Equation of the circle having centre (h, k) and which touches X - axis is (x - h)2 + (y - k)2 = k2   2) Equation of the circle having centre (h, k) and which touches Y-axis is (x - h)2 + (y - k)2 = h2   3) Equation of the circle having centre (h, h) and which touches both the axis is (x - h)2 + (y - h)2 = h2  4) Equation of the circle passing through origin (0, 0) and which cuts X - axis at (h, 0) and Y - axis at (0, k) is x2 + y2 - hx - ky = 0    Intercepts made by a Circle on the axes Let the circle x2 + y2 + 2gx + 2fy + c = 0, then   X - intercept = AB =   and Y - intercept = CD =   Note: Let the circle equation S = x2 + y2 + 2gx + 2fy + c = 0 and P(x1, y1) be a point, then we define S11 = x12 + y12 + 2gx1 + 2fy1 + c S1 = xx1 + yy1 + g(x + x1) + f(y + y1) + c   POSITION OF A POINT WITH RESPECT TO A CIRCLE Let the circle S = x2 + y2 + 2gx + 2fy + c = 0 and P(x1, y1) be any point, then (i) if S11 > 0 then 'P' lies out side the circle. (ii) if S11 < 0 then 'P' lies inside the circle. (iii) if S11 = 0 then 'P' lies on the circle.   PARAMETRIC FORM OF CIRCLE The parametric coordinates of any point on the circle (x - h)2 + (y - k)2 = r2 is (h + r cos , k + r sin ) where θ is a parameter and              * Let 'P' be any point. If we draw a secant line through the point 'P' and it cuts the circle at two different points A and B, then PA.PB is always constant and PA.PB = S11   INTERSECTION OF A LINE AND A CIRCLE Let the equation of circle be x2 + y2 + 2gx + 2fy + c = 0 Equation of line be y = mx + c Let the length of perpendicular from centre to the line = d and radius  = r, then (i) if d > r  the line does not intersect the circle (ii) if d < r  the line intersect the circle in two different points (iii) if d = r  the line touches the circle at single point            i.e., the line is tangent to the circle THE LENGTH OF THE INTERCEPT CUT OFF FROM A LINE BY A CIRCLE The length of the intercept cut off from the line y = mx + c by the circle   The equation of tangent to the circle (x - h)2 + (y - k)2 = r2 at the point (h + r cos , k + r sin ) is (x - h)cos + (y - k)sin  = r   Pairs of Tangents  If a point 'P' lies out side the circle, then we can draw two tangents through the point 'P' to the circle.  Equation of pairs of tangents drawn from a point P (x1, y1) to the circle   LENGTH OF THE TANGENTS The length of the tangent from the point P(x1, y1) to the circle S ≡ x2 + y2 + 2gx + 2fy + c = 0 is    DIRECTOR CIRCLE: The locus of point of intersection of perpendicular tangents to a given circle is known as director circle. The equation of the director circle of the circle (x - h)2 + (y - k)2 = a2 is (x - h)2 + (y - k)2 = 2a2   CHORD OF CONTACT: The chord joining the points of contact of the two tangents to a circle drawn from a given out side point is called the chord of contact of tangents.   TANGENT TO A CIRCLE The tangent at a point 'P' is the limiting position of a secant PQ when 'Q' tends to 'P' along the circle and point 'P' is called the point of contact of the tangent.     DIFFERENT FORMS OF THE EQUATIONS OF TANGENTS 1) Slope form: The equation of tangent of slope 'm' to the circle x2 + y2 = a2 is   and the point of contact is    * The equation of tangent of slope 'm' to the circle x2 + y2 + 2gx + 2fy + c = 0 are    2) Point form: The equation of tangent to the circle x2 + y2 + 2gx + 2fy + c = 0 at the point (x1, y1) is S1 = xx1 + yy1 + g(x + x1) + f(y + y1) + c = 0 3) Parametric form: The equation of the tangent to the circle x2 + y2 = a2 at the point(a cos , a sin ) is x cos  + y sin = a. The equation of the chord contact of tangent drawn from a point (x1, y1) to the circle S ≡ x2 + y2 + 2gx + 2fy + c = 0 is S1 = 0  xx1 + yy1 + g(x + x1) + f(y + y1) + c = 0 CHORD BISECTED AT A GIVEN POINT The equation of the chord of the circle S ≡ x2 + y2 + 2gx + 2fy + c = 0 bisected at a point (x1, y1) is given by S1 = S11.   NORMAL TO A CIRCLE AT A GIVEN POINT The normal of a circle is a straight line passing through the given point and perpendicular to the tangent at the point of contact. The normal always passing through the centre of the circle.   COMMON TANGENTS TO TWO CIRCLES  Direct common tangents: If the two circles on same side of the given tangent, then it is called direct common tangent.  Transverse common tangents: If the two circles on opposite side of given tangent then it is called transverse common tangent.   COMMON TANGENTS Let the circles (x - h1)2 + (y - k1)2 = r12  ................... (1) (x - h2)2 + (y - k2)2 = r2    .................... (2) with centres C1(h1, k1) and C2 (h2, k2) and radii 'r1' and 'r2' respectively. CASE - I : If  C1C2 > r1 + r2 In this case two circles do not intersect. Then two transverse common tangents and two direct common tangents exist and point of intersection of direct common tangents (B) and transverse common tangents (A) lies on line joining 'C1' and 'C2' The point 'A' divides C1 and C2 in the ratio r1 : r2 internally. The point 'B' divides the line segment joining 'C1' and 'C2' in the ratio r1 : r2 externally. CASE II: If C1C2 = r1 + r2 In this case two circles touching externally, then two direct common tangents and one transverse common tangent exists. CASE III: If C1C2 < r1 + r2 In this case two circles are intersect at two different points, then only two direct common tangents exist.   CASE IV: If C1C2 = r1 − r2 In this case two circles touch internally (One lies inside other and touches it), then only one direct common tangent exists.    CASE V: If C1C1 < r1 − r2 In this case two circles are such that one lies inside other and do not touch one another, then no common tangent exists.    COMMON CHORD The chord joining the points of intersection of two given circles is called their common chord. * The equation of the common chord of two circles S1 ≡ x2 + y2 + 2g1x + 2f1y + c1 = 0 and S1 ≡ x2 + y2 + 2g2x + 2f2y + c2 = 0 is S1 - S2 = 0 i.e., 2(g1 − g2) x + 2(f1 − f2)y + c1 − c2 = 0   Length of Common Chord Length of Common Chord of two circles S1 ≡ 0 and S2 ≡ 0 is  where r1 is radius of the circle S≡ 0 and d1 is perpendicular distance from the centre of the circle S≡ 0 to the common chord. * If the length of the common chord is zero, then the two circles is touch each other and the common chord becomes common tangent at their point of contact. * The length of common chord of two circles will be of  maximum length if it is a diameter of the smaller of the two circles.   ANGLE OF INTERSECTION OF TWO CIRCLES  The angle of intersection of two circles is defined as the angle between the tangents to the two circles at their point of intersection.  If 'θ' is angle of intersection of two circles  S1 ≡ x2 + y2 + 2g1x + 2f1y + c1 = 0 and S2 ≡ x2 + y2 + 2g2x + 2f2y + c2 = 0, then      Where r1 = radius of the circle S1 = 0              r2 = radius of the circle S2 = 0 and d = distance between the centres of the two circles.   ORTHOGONAL CIRCLES Two circles are said to intersect orthogonally if their angle of intersection is a right angle.  The condition for orthogonality of two circles         S1 ≡ x2 + y2 + 2g1x + 2f1y + c1 = 0 and S2 ≡ x2 + y2 + 2g2x + 2f2y + c2 = 0 is 2[g1g2 + f1f2] = c1 + c2   RADICAL AXIS The radical axis of two circles is the locus of a point which moves in such a way that the lengths of the tangents drawn from it to the two circles are equal. The equation of Radical axis of two circles. S1 ≡ x2 + y2 + 2g1x + 2f1y + c1 = 0 and S2 ≡ x2 + y2 + 2g2x + 2f2y + c2= 0 is S1 - S2 = 0 i.e., 2(g1 − g2)x + 2(f1− f2)y + c1− c2 = 0   PROPERTIES OF RADICAL AXIS When two circles are intersect ( ... r2 − r1 < C1C2 < r1 + r2), then the common chord coincide with the Radical axis ( ... The points lying on common chord lies on the Radical axis) When one circle lies inside the other ( ... C1C2 < r1 − r2) In this case Radical axis and common chord both do not exist. * The Radical axis of two circles is always perpendicular to the line joining the centres of the circles. * The Radical axis of three circles whose centres are non-collinear taken in pairs are concurrent. * The locus of the centre of the circle cutting two given circles orthogonally is their Radical axis. RADICAL CENTRE      The point of intersection of Radical axis of three circles whose centres are non collinear taken in pairs is called their Radical centre.  The circles with centre at the Radical centre and radius equal to the length of the tangent from it to one of the circles intersects all the three circles orthogonally.   FAMILY OF CIRCLES The equation of family of circles passing through the intersection of a circle S ≡ x2 + y2 + 2gx + 2fy + c = 0 and line L = lx + my + n = 0 is x2 + y2 + 2gx + 2fy + c + λ (lx + my + n) = 0 i.e., S + λL = 0 where λ  R. The equation of the family of circles passing through the points A(x1, y1), B(x2, y2) is    The equation of the family of circles touching the circle S ≡ x2 + y2 + 2gx + 2fy + c = 0 at a point (x1, y1) is   i.e., x2 + y2 + 2gx + 2fy + c + λ[xx1 + yy1 + g(x + x1) + f(y + y1) + c] = 0  The equation of the family of circles passing through the intersection of the circles S1 ≡ 0 and S2 ≡ 0 is   (Where λ(≠ −1)  R)   CO - AXIAL SYSTEM OF CIRCLES *  A system of circles is said to be a coaxial system of circles if each pair of the circles in the system has the same radical axis. * Since the line joining the centres of two circles is perpendicular to their radical axis. * Then the centres of all circles of a coaxial system lie on a straight line which is perpendicular to the common radical axis. *  Circles passing through two fixed points 'P' and 'Q' form a co-axial system because every pair of circles has the same common chord PQ and therefore the same radical axis which is perpendicular bisector of 'PQ'.  If S ≡ 0 is the equation of a member of a system of coaxial circles and the equation of common radical axis L ≡ 0, then the equation S + λL = 0 (λ  R) represents the coaxial system of circles. LIMITING POINTS      Limiting points of a coaxial system of circles are the members of the system which have zero radius.  Posted Date : 17-02-2021 గమనిక : ప్రతిభ.ఈనాడు.నెట్లో వచ్చే ప్రకటనలు అనేక దేశాల నుండి, వ్యాపారస్తులు లేదా వ్యక్తుల నుండి వివిధ పద్ధతులలో సేకరించబడతాయి. ఆయా ప్రకటనకర్తల ఉత్పత్తులు లేదా సేవల గురించి ఈనాడు యాజమాన్యానికీ, ఉద్యోగస్తులకూ ఎటువంటి అవగాహనా ఉండదు. కొన్ని ప్రకటనలు పాఠకుల అభిరుచిననుసరించి కృత్రిమ మేధస్సు సాంకేతికతతో పంపబడతాయి. ఏ ప్రకటనని అయినా పాఠకులు తగినంత జాగ్రత్త వహించి, ఉత్పత్తులు లేదా సేవల గురించి తగిన విచారణ చేసి, తగిన జాగ్రత్తలు తీసుకొని కొనుగోలు చేయాలి. ఉత్పత్తులు / సేవలపై ఈనాడు యాజమాన్యానికి ఎటువంటి నియంత్రణ ఉండదు. కనుక ఉత్పత్తులు లేదా సేవల నాణ్యత లేదా లోపాల విషయంలో ఈనాడు యాజమాన్యం ఎటువంటి బాధ్యత వహించదు. ఈ విషయంలో ఎటువంటి ఉత్తర ప్రత్యుత్తరాలకీ తావు లేదు. ఫిర్యాదులు తీసుకోబడవు. స్టడీమెటీరియల్ పాత ప్రశ్నప‌త్రాలు   విద్యా ఉద్యోగ సమాచారం   నమూనా ప్రశ్నపత్రాలు   లేటెస్ట్ నోటిఫికేష‌న్స్‌  
__label__pos
0.975383