id
int32
0
167k
repo
stringlengths
5
54
path
stringlengths
4
155
func_name
stringlengths
1
118
original_string
stringlengths
52
85.5k
language
stringclasses
1 value
code
stringlengths
52
85.5k
code_tokens
sequencelengths
21
1.41k
docstring
stringlengths
6
2.61k
docstring_tokens
sequencelengths
3
215
sha
stringlengths
40
40
url
stringlengths
85
252
5,500
kotakanbe/go-cve-dictionary
log/log.go
SetLogger
func SetLogger(logDir string, quiet, debug, logJSON bool) { stderrHundler := logger.StderrHandler logFormat := logger.LogfmtFormat() if logJSON { logFormat = logger.JsonFormatEx(false, true) stderrHundler = logger.StreamHandler(os.Stderr, logFormat) } lvlHundler := logger.LvlFilterHandler(logger.LvlInfo, stderrHundler) if debug { lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, stderrHundler) } if quiet { lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, logger.DiscardHandler()) pp.SetDefaultOutput(ioutil.Discard) } if _, err := os.Stat(logDir); os.IsNotExist(err) { if err := os.Mkdir(logDir, 0700); err != nil { logger.Error("Failed to create log directory", "err", err) } } var hundler logger.Handler if _, err := os.Stat(logDir); err == nil { logPath := filepath.Join(logDir, "cve-dictionary.log") hundler = logger.MultiHandler( logger.Must.FileHandler(logPath, logFormat), lvlHundler, ) } else { hundler = lvlHundler } logger.Root().SetHandler(hundler) }
go
func SetLogger(logDir string, quiet, debug, logJSON bool) { stderrHundler := logger.StderrHandler logFormat := logger.LogfmtFormat() if logJSON { logFormat = logger.JsonFormatEx(false, true) stderrHundler = logger.StreamHandler(os.Stderr, logFormat) } lvlHundler := logger.LvlFilterHandler(logger.LvlInfo, stderrHundler) if debug { lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, stderrHundler) } if quiet { lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, logger.DiscardHandler()) pp.SetDefaultOutput(ioutil.Discard) } if _, err := os.Stat(logDir); os.IsNotExist(err) { if err := os.Mkdir(logDir, 0700); err != nil { logger.Error("Failed to create log directory", "err", err) } } var hundler logger.Handler if _, err := os.Stat(logDir); err == nil { logPath := filepath.Join(logDir, "cve-dictionary.log") hundler = logger.MultiHandler( logger.Must.FileHandler(logPath, logFormat), lvlHundler, ) } else { hundler = lvlHundler } logger.Root().SetHandler(hundler) }
[ "func", "SetLogger", "(", "logDir", "string", ",", "quiet", ",", "debug", ",", "logJSON", "bool", ")", "{", "stderrHundler", ":=", "logger", ".", "StderrHandler", "\n", "logFormat", ":=", "logger", ".", "LogfmtFormat", "(", ")", "\n", "if", "logJSON", "{", "logFormat", "=", "logger", ".", "JsonFormatEx", "(", "false", ",", "true", ")", "\n", "stderrHundler", "=", "logger", ".", "StreamHandler", "(", "os", ".", "Stderr", ",", "logFormat", ")", "\n", "}", "\n\n", "lvlHundler", ":=", "logger", ".", "LvlFilterHandler", "(", "logger", ".", "LvlInfo", ",", "stderrHundler", ")", "\n", "if", "debug", "{", "lvlHundler", "=", "logger", ".", "LvlFilterHandler", "(", "logger", ".", "LvlDebug", ",", "stderrHundler", ")", "\n", "}", "\n", "if", "quiet", "{", "lvlHundler", "=", "logger", ".", "LvlFilterHandler", "(", "logger", ".", "LvlDebug", ",", "logger", ".", "DiscardHandler", "(", ")", ")", "\n", "pp", ".", "SetDefaultOutput", "(", "ioutil", ".", "Discard", ")", "\n", "}", "\n\n", "if", "_", ",", "err", ":=", "os", ".", "Stat", "(", "logDir", ")", ";", "os", ".", "IsNotExist", "(", "err", ")", "{", "if", "err", ":=", "os", ".", "Mkdir", "(", "logDir", ",", "0700", ")", ";", "err", "!=", "nil", "{", "logger", ".", "Error", "(", "\"", "\"", ",", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "}", "\n", "var", "hundler", "logger", ".", "Handler", "\n", "if", "_", ",", "err", ":=", "os", ".", "Stat", "(", "logDir", ")", ";", "err", "==", "nil", "{", "logPath", ":=", "filepath", ".", "Join", "(", "logDir", ",", "\"", "\"", ")", "\n", "hundler", "=", "logger", ".", "MultiHandler", "(", "logger", ".", "Must", ".", "FileHandler", "(", "logPath", ",", "logFormat", ")", ",", "lvlHundler", ",", ")", "\n", "}", "else", "{", "hundler", "=", "lvlHundler", "\n", "}", "\n", "logger", ".", "Root", "(", ")", ".", "SetHandler", "(", "hundler", ")", "\n", "}" ]
// SetLogger set logger
[ "SetLogger", "set", "logger" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/log/log.go#L14-L47
5,501
kotakanbe/go-cve-dictionary
log/log.go
Fatalf
func Fatalf(format string, args ...interface{}) { logger.Crit(fmt.Sprintf(format, args...)) }
go
func Fatalf(format string, args ...interface{}) { logger.Crit(fmt.Sprintf(format, args...)) }
[ "func", "Fatalf", "(", "format", "string", ",", "args", "...", "interface", "{", "}", ")", "{", "logger", ".", "Crit", "(", "fmt", ".", "Sprintf", "(", "format", ",", "args", "...", ")", ")", "\n", "}" ]
// Fatalf is wrapper function
[ "Fatalf", "is", "wrapper", "function" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/log/log.go#L70-L72
5,502
kotakanbe/go-cve-dictionary
server/server.go
Start
func Start(logDir string, driver db.DB) error { e := echo.New() e.Debug = config.Conf.Debug // Middleware e.Use(middleware.Logger()) e.Use(middleware.Recover()) // setup access logger logPath := filepath.Join(logDir, "access.log") if _, err := os.Stat(logPath); os.IsNotExist(err) { if _, err := os.Create(logPath); err != nil { return err } } f, err := os.OpenFile(logPath, os.O_APPEND|os.O_WRONLY, 0600) if err != nil { return err } defer f.Close() e.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{ Output: f, })) // Routes e.GET("/health", health()) e.GET("/cves/:id", getCve(driver)) e.POST("/cpes", getCveByCpeName(driver)) bindURL := fmt.Sprintf("%s:%s", config.Conf.Bind, config.Conf.Port) log.Infof("Listening on %s", bindURL) e.Start(bindURL) return nil }
go
func Start(logDir string, driver db.DB) error { e := echo.New() e.Debug = config.Conf.Debug // Middleware e.Use(middleware.Logger()) e.Use(middleware.Recover()) // setup access logger logPath := filepath.Join(logDir, "access.log") if _, err := os.Stat(logPath); os.IsNotExist(err) { if _, err := os.Create(logPath); err != nil { return err } } f, err := os.OpenFile(logPath, os.O_APPEND|os.O_WRONLY, 0600) if err != nil { return err } defer f.Close() e.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{ Output: f, })) // Routes e.GET("/health", health()) e.GET("/cves/:id", getCve(driver)) e.POST("/cpes", getCveByCpeName(driver)) bindURL := fmt.Sprintf("%s:%s", config.Conf.Bind, config.Conf.Port) log.Infof("Listening on %s", bindURL) e.Start(bindURL) return nil }
[ "func", "Start", "(", "logDir", "string", ",", "driver", "db", ".", "DB", ")", "error", "{", "e", ":=", "echo", ".", "New", "(", ")", "\n", "e", ".", "Debug", "=", "config", ".", "Conf", ".", "Debug", "\n\n", "// Middleware", "e", ".", "Use", "(", "middleware", ".", "Logger", "(", ")", ")", "\n", "e", ".", "Use", "(", "middleware", ".", "Recover", "(", ")", ")", "\n\n", "// setup access logger", "logPath", ":=", "filepath", ".", "Join", "(", "logDir", ",", "\"", "\"", ")", "\n", "if", "_", ",", "err", ":=", "os", ".", "Stat", "(", "logPath", ")", ";", "os", ".", "IsNotExist", "(", "err", ")", "{", "if", "_", ",", "err", ":=", "os", ".", "Create", "(", "logPath", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "\n", "f", ",", "err", ":=", "os", ".", "OpenFile", "(", "logPath", ",", "os", ".", "O_APPEND", "|", "os", ".", "O_WRONLY", ",", "0600", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "defer", "f", ".", "Close", "(", ")", "\n", "e", ".", "Use", "(", "middleware", ".", "LoggerWithConfig", "(", "middleware", ".", "LoggerConfig", "{", "Output", ":", "f", ",", "}", ")", ")", "\n\n", "// Routes", "e", ".", "GET", "(", "\"", "\"", ",", "health", "(", ")", ")", "\n", "e", ".", "GET", "(", "\"", "\"", ",", "getCve", "(", "driver", ")", ")", "\n", "e", ".", "POST", "(", "\"", "\"", ",", "getCveByCpeName", "(", "driver", ")", ")", "\n\n", "bindURL", ":=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "config", ".", "Conf", ".", "Bind", ",", "config", ".", "Conf", ".", "Port", ")", "\n", "log", ".", "Infof", "(", "\"", "\"", ",", "bindURL", ")", "\n\n", "e", ".", "Start", "(", "bindURL", ")", "\n", "return", "nil", "\n", "}" ]
// Start starts CVE dictionary HTTP Server.
[ "Start", "starts", "CVE", "dictionary", "HTTP", "Server", "." ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/server/server.go#L17-L51
5,503
kotakanbe/go-cve-dictionary
db/redis.go
NewRedis
func NewRedis(dbType, dbpath string, debugSQL bool) (driver *RedisDriver, locked bool, err error) { driver = &RedisDriver{ name: dbType, } log.Debugf("Opening DB (%s).", driver.Name()) if err = driver.OpenDB(dbType, dbpath, debugSQL); err != nil { return } return }
go
func NewRedis(dbType, dbpath string, debugSQL bool) (driver *RedisDriver, locked bool, err error) { driver = &RedisDriver{ name: dbType, } log.Debugf("Opening DB (%s).", driver.Name()) if err = driver.OpenDB(dbType, dbpath, debugSQL); err != nil { return } return }
[ "func", "NewRedis", "(", "dbType", ",", "dbpath", "string", ",", "debugSQL", "bool", ")", "(", "driver", "*", "RedisDriver", ",", "locked", "bool", ",", "err", "error", ")", "{", "driver", "=", "&", "RedisDriver", "{", "name", ":", "dbType", ",", "}", "\n", "log", ".", "Debugf", "(", "\"", "\"", ",", "driver", ".", "Name", "(", ")", ")", "\n", "if", "err", "=", "driver", ".", "OpenDB", "(", "dbType", ",", "dbpath", ",", "debugSQL", ")", ";", "err", "!=", "nil", "{", "return", "\n", "}", "\n\n", "return", "\n", "}" ]
// NewRedis return Redis driver
[ "NewRedis", "return", "Redis", "driver" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L61-L71
5,504
kotakanbe/go-cve-dictionary
db/redis.go
InsertJvn
func (r *RedisDriver) InsertJvn(cves []models.CveDetail) error { log.Infof("Inserting fetched CVEs...") var err error var refreshedJvns []string bar := pb.New(len(cves)) if c.Conf.Quiet { bar.SetWriter(ioutil.Discard) } else { bar.SetWriter(os.Stderr) } bar.Start() for chunked := range chunkSlice(cves, 10) { var pipe redis.Pipeliner pipe = r.conn.Pipeline() for _, c := range chunked { bar.Increment() cpes := make([]models.Cpe, len(c.Jvn.Cpes)) copy(cpes, c.Jvn.Cpes) c.Jvn.Cpes = nil var jj []byte if jj, err = json.Marshal(c.Jvn); err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } refreshedJvns = append(refreshedJvns, c.CveID) if result := pipe.HSet(hashKeyPrefix+c.CveID, "Jvn", string(jj)); result.Err() != nil { return fmt.Errorf("Failed to HSet CVE. err: %s", result.Err()) } for _, cpe := range cpes { if result := pipe.ZAdd( fmt.Sprintf("%s%s::%s", hashKeyPrefix, cpe.Vendor, cpe.Product), redis.Z{Score: 0, Member: c.CveID}, ); result.Err() != nil { return fmt.Errorf("Failed to ZAdd cpe. err: %s", result.Err()) } } var jc []byte if jc, err = json.Marshal(cpes); err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } if result := pipe.HSet(cpeHashKeyPrefix+c.CveID, "Jvn", string(jc)); result.Err() != nil { return fmt.Errorf("Failed to HSet CPE. err: %s", result.Err()) } } if _, err = pipe.Exec(); err != nil { return fmt.Errorf("Failed to exec pipeline. err: %s", err) } } bar.Finish() log.Infof("Refreshed %d Jvns.", len(refreshedJvns)) // log.Debugf("%v", refreshedJvns) return nil }
go
func (r *RedisDriver) InsertJvn(cves []models.CveDetail) error { log.Infof("Inserting fetched CVEs...") var err error var refreshedJvns []string bar := pb.New(len(cves)) if c.Conf.Quiet { bar.SetWriter(ioutil.Discard) } else { bar.SetWriter(os.Stderr) } bar.Start() for chunked := range chunkSlice(cves, 10) { var pipe redis.Pipeliner pipe = r.conn.Pipeline() for _, c := range chunked { bar.Increment() cpes := make([]models.Cpe, len(c.Jvn.Cpes)) copy(cpes, c.Jvn.Cpes) c.Jvn.Cpes = nil var jj []byte if jj, err = json.Marshal(c.Jvn); err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } refreshedJvns = append(refreshedJvns, c.CveID) if result := pipe.HSet(hashKeyPrefix+c.CveID, "Jvn", string(jj)); result.Err() != nil { return fmt.Errorf("Failed to HSet CVE. err: %s", result.Err()) } for _, cpe := range cpes { if result := pipe.ZAdd( fmt.Sprintf("%s%s::%s", hashKeyPrefix, cpe.Vendor, cpe.Product), redis.Z{Score: 0, Member: c.CveID}, ); result.Err() != nil { return fmt.Errorf("Failed to ZAdd cpe. err: %s", result.Err()) } } var jc []byte if jc, err = json.Marshal(cpes); err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } if result := pipe.HSet(cpeHashKeyPrefix+c.CveID, "Jvn", string(jc)); result.Err() != nil { return fmt.Errorf("Failed to HSet CPE. err: %s", result.Err()) } } if _, err = pipe.Exec(); err != nil { return fmt.Errorf("Failed to exec pipeline. err: %s", err) } } bar.Finish() log.Infof("Refreshed %d Jvns.", len(refreshedJvns)) // log.Debugf("%v", refreshedJvns) return nil }
[ "func", "(", "r", "*", "RedisDriver", ")", "InsertJvn", "(", "cves", "[", "]", "models", ".", "CveDetail", ")", "error", "{", "log", ".", "Infof", "(", "\"", "\"", ")", "\n", "var", "err", "error", "\n", "var", "refreshedJvns", "[", "]", "string", "\n", "bar", ":=", "pb", ".", "New", "(", "len", "(", "cves", ")", ")", "\n", "if", "c", ".", "Conf", ".", "Quiet", "{", "bar", ".", "SetWriter", "(", "ioutil", ".", "Discard", ")", "\n", "}", "else", "{", "bar", ".", "SetWriter", "(", "os", ".", "Stderr", ")", "\n", "}", "\n", "bar", ".", "Start", "(", ")", "\n\n", "for", "chunked", ":=", "range", "chunkSlice", "(", "cves", ",", "10", ")", "{", "var", "pipe", "redis", ".", "Pipeliner", "\n", "pipe", "=", "r", ".", "conn", ".", "Pipeline", "(", ")", "\n", "for", "_", ",", "c", ":=", "range", "chunked", "{", "bar", ".", "Increment", "(", ")", "\n\n", "cpes", ":=", "make", "(", "[", "]", "models", ".", "Cpe", ",", "len", "(", "c", ".", "Jvn", ".", "Cpes", ")", ")", "\n", "copy", "(", "cpes", ",", "c", ".", "Jvn", ".", "Cpes", ")", "\n", "c", ".", "Jvn", ".", "Cpes", "=", "nil", "\n\n", "var", "jj", "[", "]", "byte", "\n", "if", "jj", ",", "err", "=", "json", ".", "Marshal", "(", "c", ".", "Jvn", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "refreshedJvns", "=", "append", "(", "refreshedJvns", ",", "c", ".", "CveID", ")", "\n", "if", "result", ":=", "pipe", ".", "HSet", "(", "hashKeyPrefix", "+", "c", ".", "CveID", ",", "\"", "\"", ",", "string", "(", "jj", ")", ")", ";", "result", ".", "Err", "(", ")", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "result", ".", "Err", "(", ")", ")", "\n", "}", "\n\n", "for", "_", ",", "cpe", ":=", "range", "cpes", "{", "if", "result", ":=", "pipe", ".", "ZAdd", "(", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "hashKeyPrefix", ",", "cpe", ".", "Vendor", ",", "cpe", ".", "Product", ")", ",", "redis", ".", "Z", "{", "Score", ":", "0", ",", "Member", ":", "c", ".", "CveID", "}", ",", ")", ";", "result", ".", "Err", "(", ")", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "result", ".", "Err", "(", ")", ")", "\n", "}", "\n", "}", "\n", "var", "jc", "[", "]", "byte", "\n", "if", "jc", ",", "err", "=", "json", ".", "Marshal", "(", "cpes", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "if", "result", ":=", "pipe", ".", "HSet", "(", "cpeHashKeyPrefix", "+", "c", ".", "CveID", ",", "\"", "\"", ",", "string", "(", "jc", ")", ")", ";", "result", ".", "Err", "(", ")", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "result", ".", "Err", "(", ")", ")", "\n", "}", "\n", "}", "\n", "if", "_", ",", "err", "=", "pipe", ".", "Exec", "(", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "}", "\n", "bar", ".", "Finish", "(", ")", "\n", "log", ".", "Infof", "(", "\"", "\"", ",", "len", "(", "refreshedJvns", ")", ")", "\n", "// log.Debugf(\"%v\", refreshedJvns)", "return", "nil", "\n", "}" ]
// InsertJvn insert items fetched from JVN.
[ "InsertJvn", "insert", "items", "fetched", "from", "JVN", "." ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L274-L329
5,505
kotakanbe/go-cve-dictionary
db/redis.go
GetFetchedFeedMeta
func (r *RedisDriver) GetFetchedFeedMeta(url string) (*models.FeedMeta, error) { var result *redis.StringStringMapCmd if result = r.conn.HGetAll(hashKeyPrefix + "Meta"); result.Err() != nil { return nil, result.Err() } meta := &models.FeedMeta{} if s, ok := result.Val()[url]; ok { if err := json.Unmarshal([]byte(s), meta); err != nil { return nil, err } return meta, nil } return meta, nil }
go
func (r *RedisDriver) GetFetchedFeedMeta(url string) (*models.FeedMeta, error) { var result *redis.StringStringMapCmd if result = r.conn.HGetAll(hashKeyPrefix + "Meta"); result.Err() != nil { return nil, result.Err() } meta := &models.FeedMeta{} if s, ok := result.Val()[url]; ok { if err := json.Unmarshal([]byte(s), meta); err != nil { return nil, err } return meta, nil } return meta, nil }
[ "func", "(", "r", "*", "RedisDriver", ")", "GetFetchedFeedMeta", "(", "url", "string", ")", "(", "*", "models", ".", "FeedMeta", ",", "error", ")", "{", "var", "result", "*", "redis", ".", "StringStringMapCmd", "\n", "if", "result", "=", "r", ".", "conn", ".", "HGetAll", "(", "hashKeyPrefix", "+", "\"", "\"", ")", ";", "result", ".", "Err", "(", ")", "!=", "nil", "{", "return", "nil", ",", "result", ".", "Err", "(", ")", "\n", "}", "\n", "meta", ":=", "&", "models", ".", "FeedMeta", "{", "}", "\n", "if", "s", ",", "ok", ":=", "result", ".", "Val", "(", ")", "[", "url", "]", ";", "ok", "{", "if", "err", ":=", "json", ".", "Unmarshal", "(", "[", "]", "byte", "(", "s", ")", ",", "meta", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "return", "meta", ",", "nil", "\n", "}", "\n", "return", "meta", ",", "nil", "\n", "}" ]
// GetFetchedFeedMeta selects hash in metafile of the year
[ "GetFetchedFeedMeta", "selects", "hash", "in", "metafile", "of", "the", "year" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L458-L471
5,506
kotakanbe/go-cve-dictionary
db/redis.go
UpsertFeedHash
func (r *RedisDriver) UpsertFeedHash(m models.FeedMeta) error { jn, err := json.Marshal(m) if err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } var pipe redis.Pipeliner pipe = r.conn.Pipeline() if result := pipe.HSet(hashKeyPrefix+"Meta", m.URL, jn); result.Err() != nil { return fmt.Errorf("Failed to HSet META. err: %s", result.Err()) } if _, err := pipe.Exec(); err != nil { return fmt.Errorf("Failed to exec pipeline. err: %s", err) } return nil }
go
func (r *RedisDriver) UpsertFeedHash(m models.FeedMeta) error { jn, err := json.Marshal(m) if err != nil { return fmt.Errorf("Failed to marshal json. err: %s", err) } var pipe redis.Pipeliner pipe = r.conn.Pipeline() if result := pipe.HSet(hashKeyPrefix+"Meta", m.URL, jn); result.Err() != nil { return fmt.Errorf("Failed to HSet META. err: %s", result.Err()) } if _, err := pipe.Exec(); err != nil { return fmt.Errorf("Failed to exec pipeline. err: %s", err) } return nil }
[ "func", "(", "r", "*", "RedisDriver", ")", "UpsertFeedHash", "(", "m", "models", ".", "FeedMeta", ")", "error", "{", "jn", ",", "err", ":=", "json", ".", "Marshal", "(", "m", ")", "\n", "if", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n\n", "var", "pipe", "redis", ".", "Pipeliner", "\n", "pipe", "=", "r", ".", "conn", ".", "Pipeline", "(", ")", "\n", "if", "result", ":=", "pipe", ".", "HSet", "(", "hashKeyPrefix", "+", "\"", "\"", ",", "m", ".", "URL", ",", "jn", ")", ";", "result", ".", "Err", "(", ")", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "result", ".", "Err", "(", ")", ")", "\n", "}", "\n", "if", "_", ",", "err", ":=", "pipe", ".", "Exec", "(", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// UpsertFeedHash selects hash in metafile of the year
[ "UpsertFeedHash", "selects", "hash", "in", "metafile", "of", "the", "year" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L474-L489
5,507
kotakanbe/go-cve-dictionary
models/models.go
UpToDate
func (f FeedMeta) UpToDate() bool { return !f.Newly() && f.Hash == f.LatestHash }
go
func (f FeedMeta) UpToDate() bool { return !f.Newly() && f.Hash == f.LatestHash }
[ "func", "(", "f", "FeedMeta", ")", "UpToDate", "(", ")", "bool", "{", "return", "!", "f", ".", "Newly", "(", ")", "&&", "f", ".", "Hash", "==", "f", ".", "LatestHash", "\n", "}" ]
// UpToDate checks whether last fetched feed is up to date
[ "UpToDate", "checks", "whether", "last", "fetched", "feed", "is", "up", "to", "date" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L25-L27
5,508
kotakanbe/go-cve-dictionary
models/models.go
OutDated
func (f FeedMeta) OutDated() bool { return !f.Newly() && f.Hash != f.LatestHash }
go
func (f FeedMeta) OutDated() bool { return !f.Newly() && f.Hash != f.LatestHash }
[ "func", "(", "f", "FeedMeta", ")", "OutDated", "(", ")", "bool", "{", "return", "!", "f", ".", "Newly", "(", ")", "&&", "f", ".", "Hash", "!=", "f", ".", "LatestHash", "\n", "}" ]
// OutDated checks whether last fetched feed is out dated
[ "OutDated", "checks", "whether", "last", "fetched", "feed", "is", "out", "dated" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L30-L32
5,509
kotakanbe/go-cve-dictionary
models/models.go
StatusForStdout
func (f FeedMeta) StatusForStdout() string { if f.Newly() { return "Newly" } else if f.OutDated() { red := color.New(color.FgRed, color.Bold).SprintFunc() return red("Out-Dated") } else if f.UpToDate() { return color.GreenString("Up-to-Date") } return "Unknown" }
go
func (f FeedMeta) StatusForStdout() string { if f.Newly() { return "Newly" } else if f.OutDated() { red := color.New(color.FgRed, color.Bold).SprintFunc() return red("Out-Dated") } else if f.UpToDate() { return color.GreenString("Up-to-Date") } return "Unknown" }
[ "func", "(", "f", "FeedMeta", ")", "StatusForStdout", "(", ")", "string", "{", "if", "f", ".", "Newly", "(", ")", "{", "return", "\"", "\"", "\n", "}", "else", "if", "f", ".", "OutDated", "(", ")", "{", "red", ":=", "color", ".", "New", "(", "color", ".", "FgRed", ",", "color", ".", "Bold", ")", ".", "SprintFunc", "(", ")", "\n", "return", "red", "(", "\"", "\"", ")", "\n", "}", "else", "if", "f", ".", "UpToDate", "(", ")", "{", "return", "color", ".", "GreenString", "(", "\"", "\"", ")", "\n", "}", "\n", "return", "\"", "\"", "\n", "}" ]
// StatusForStdout returns a status of fetched feed
[ "StatusForStdout", "returns", "a", "status", "of", "fetched", "feed" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L40-L50
5,510
kotakanbe/go-cve-dictionary
models/models.go
Year
func (f FeedMeta) Year() (year string, xml bool, err error) { switch f.source() { case nvdxml: return strings.TrimSuffix( strings.Split(f.URL, "nvdcve-2.0-")[1], ".xml.gz"), true, nil case nvdjson: return strings.TrimSuffix( strings.Split(f.URL, "nvdcve-1.0-")[1], ".json.gz"), false, nil case jvn: if strings.HasSuffix(f.URL, "jvndb.rdf") { return "modified", true, nil } else if strings.HasSuffix(f.URL, "jvndb_new.rdf") { return "recent", true, nil } else { return strings.TrimSuffix( strings.Split(f.URL, "jvndb_")[1], ".rdf"), true, nil } default: return "", false, fmt.Errorf("Failed to parse URL: %s", f.URL) } }
go
func (f FeedMeta) Year() (year string, xml bool, err error) { switch f.source() { case nvdxml: return strings.TrimSuffix( strings.Split(f.URL, "nvdcve-2.0-")[1], ".xml.gz"), true, nil case nvdjson: return strings.TrimSuffix( strings.Split(f.URL, "nvdcve-1.0-")[1], ".json.gz"), false, nil case jvn: if strings.HasSuffix(f.URL, "jvndb.rdf") { return "modified", true, nil } else if strings.HasSuffix(f.URL, "jvndb_new.rdf") { return "recent", true, nil } else { return strings.TrimSuffix( strings.Split(f.URL, "jvndb_")[1], ".rdf"), true, nil } default: return "", false, fmt.Errorf("Failed to parse URL: %s", f.URL) } }
[ "func", "(", "f", "FeedMeta", ")", "Year", "(", ")", "(", "year", "string", ",", "xml", "bool", ",", "err", "error", ")", "{", "switch", "f", ".", "source", "(", ")", "{", "case", "nvdxml", ":", "return", "strings", ".", "TrimSuffix", "(", "strings", ".", "Split", "(", "f", ".", "URL", ",", "\"", "\"", ")", "[", "1", "]", ",", "\"", "\"", ")", ",", "true", ",", "nil", "\n", "case", "nvdjson", ":", "return", "strings", ".", "TrimSuffix", "(", "strings", ".", "Split", "(", "f", ".", "URL", ",", "\"", "\"", ")", "[", "1", "]", ",", "\"", "\"", ")", ",", "false", ",", "nil", "\n", "case", "jvn", ":", "if", "strings", ".", "HasSuffix", "(", "f", ".", "URL", ",", "\"", "\"", ")", "{", "return", "\"", "\"", ",", "true", ",", "nil", "\n", "}", "else", "if", "strings", ".", "HasSuffix", "(", "f", ".", "URL", ",", "\"", "\"", ")", "{", "return", "\"", "\"", ",", "true", ",", "nil", "\n", "}", "else", "{", "return", "strings", ".", "TrimSuffix", "(", "strings", ".", "Split", "(", "f", ".", "URL", ",", "\"", "\"", ")", "[", "1", "]", ",", "\"", "\"", ")", ",", "true", ",", "nil", "\n", "}", "\n", "default", ":", "return", "\"", "\"", ",", "false", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "f", ".", "URL", ")", "\n", "}", "\n", "}" ]
// Year returns year, whether xml or not of the feed
[ "Year", "returns", "year", "whether", "xml", "or", "not", "of", "the", "feed" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L93-L113
5,511
kotakanbe/go-cve-dictionary
models/models.go
ToTableWriterRow
func (f FeedMeta) ToTableWriterRow() []string { y, _, _ := f.Year() fetched, latest := f.modifiedTimesToStrs() return []string{ f.color(f.source()), f.color(y), f.StatusForStdout(), f.color(fetched), f.color(latest), } }
go
func (f FeedMeta) ToTableWriterRow() []string { y, _, _ := f.Year() fetched, latest := f.modifiedTimesToStrs() return []string{ f.color(f.source()), f.color(y), f.StatusForStdout(), f.color(fetched), f.color(latest), } }
[ "func", "(", "f", "FeedMeta", ")", "ToTableWriterRow", "(", ")", "[", "]", "string", "{", "y", ",", "_", ",", "_", ":=", "f", ".", "Year", "(", ")", "\n", "fetched", ",", "latest", ":=", "f", ".", "modifiedTimesToStrs", "(", ")", "\n", "return", "[", "]", "string", "{", "f", ".", "color", "(", "f", ".", "source", "(", ")", ")", ",", "f", ".", "color", "(", "y", ")", ",", "f", ".", "StatusForStdout", "(", ")", ",", "f", ".", "color", "(", "fetched", ")", ",", "f", ".", "color", "(", "latest", ")", ",", "}", "\n", "}" ]
// ToTableWriterRow generate data for table writer
[ "ToTableWriterRow", "generate", "data", "for", "table", "writer" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L135-L145
5,512
kotakanbe/go-cve-dictionary
fetcher/nvd/json/nvd.go
FetchConvert
func FetchConvert(metas []models.FeedMeta) (cves []models.CveDetail, err error) { reqs := []fetcher.FetchRequest{} for _, meta := range metas { reqs = append(reqs, fetcher.FetchRequest{ URL: meta.URL, GZIP: true, }) } results, err := fetcher.FetchFeedFiles(reqs) if err != nil { return nil, fmt.Errorf("Failed to fetch. err: %s", err) } for _, res := range results { nvd := NvdJSON{} if err = json.Unmarshal(res.Body, &nvd); err != nil { return nil, fmt.Errorf( "Failed to unmarshal. url: %s, err: %s", res.URL, err) } for _, item := range nvd.CveItems { cve, err := convertToModel(&item) if err != nil { return nil, fmt.Errorf("Failed to convert to model. cve: %s, err: %s", item.Cve.CveDataMeta.ID, err) } cves = append(cves, *cve) } } return }
go
func FetchConvert(metas []models.FeedMeta) (cves []models.CveDetail, err error) { reqs := []fetcher.FetchRequest{} for _, meta := range metas { reqs = append(reqs, fetcher.FetchRequest{ URL: meta.URL, GZIP: true, }) } results, err := fetcher.FetchFeedFiles(reqs) if err != nil { return nil, fmt.Errorf("Failed to fetch. err: %s", err) } for _, res := range results { nvd := NvdJSON{} if err = json.Unmarshal(res.Body, &nvd); err != nil { return nil, fmt.Errorf( "Failed to unmarshal. url: %s, err: %s", res.URL, err) } for _, item := range nvd.CveItems { cve, err := convertToModel(&item) if err != nil { return nil, fmt.Errorf("Failed to convert to model. cve: %s, err: %s", item.Cve.CveDataMeta.ID, err) } cves = append(cves, *cve) } } return }
[ "func", "FetchConvert", "(", "metas", "[", "]", "models", ".", "FeedMeta", ")", "(", "cves", "[", "]", "models", ".", "CveDetail", ",", "err", "error", ")", "{", "reqs", ":=", "[", "]", "fetcher", ".", "FetchRequest", "{", "}", "\n", "for", "_", ",", "meta", ":=", "range", "metas", "{", "reqs", "=", "append", "(", "reqs", ",", "fetcher", ".", "FetchRequest", "{", "URL", ":", "meta", ".", "URL", ",", "GZIP", ":", "true", ",", "}", ")", "\n", "}", "\n\n", "results", ",", "err", ":=", "fetcher", ".", "FetchFeedFiles", "(", "reqs", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n\n", "for", "_", ",", "res", ":=", "range", "results", "{", "nvd", ":=", "NvdJSON", "{", "}", "\n", "if", "err", "=", "json", ".", "Unmarshal", "(", "res", ".", "Body", ",", "&", "nvd", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "res", ".", "URL", ",", "err", ")", "\n", "}", "\n", "for", "_", ",", "item", ":=", "range", "nvd", ".", "CveItems", "{", "cve", ",", "err", ":=", "convertToModel", "(", "&", "item", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "item", ".", "Cve", ".", "CveDataMeta", ".", "ID", ",", "err", ")", "\n", "}", "\n", "cves", "=", "append", "(", "cves", ",", "*", "cve", ")", "\n", "}", "\n", "}", "\n", "return", "\n", "}" ]
// FetchConvert Fetch CVE vulnerability information from NVD
[ "FetchConvert", "Fetch", "CVE", "vulnerability", "information", "from", "NVD" ]
5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a
https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/fetcher/nvd/json/nvd.go#L17-L49
5,513
go-pg/migrations
collection.go
Register
func (c *Collection) Register(fns ...func(DB) error) error { return c.register(false, fns...) }
go
func (c *Collection) Register(fns ...func(DB) error) error { return c.register(false, fns...) }
[ "func", "(", "c", "*", "Collection", ")", "Register", "(", "fns", "...", "func", "(", "DB", ")", "error", ")", "error", "{", "return", "c", ".", "register", "(", "false", ",", "fns", "...", ")", "\n", "}" ]
// Register registers new database migration. Must be called // from a file with name like "1_initialize_db.go".
[ "Register", "registers", "new", "database", "migration", ".", "Must", "be", "called", "from", "a", "file", "with", "name", "like", "1_initialize_db", ".", "go", "." ]
42c5771cbefbd8c12aa3e935e082b388186764bb
https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L74-L76
5,514
go-pg/migrations
collection.go
RegisterTx
func (c *Collection) RegisterTx(fns ...func(DB) error) error { return c.register(true, fns...) }
go
func (c *Collection) RegisterTx(fns ...func(DB) error) error { return c.register(true, fns...) }
[ "func", "(", "c", "*", "Collection", ")", "RegisterTx", "(", "fns", "...", "func", "(", "DB", ")", "error", ")", "error", "{", "return", "c", ".", "register", "(", "true", ",", "fns", "...", ")", "\n", "}" ]
// RegisterTx is like Register, but migration will be run in a transaction.
[ "RegisterTx", "is", "like", "Register", "but", "migration", "will", "be", "run", "in", "a", "transaction", "." ]
42c5771cbefbd8c12aa3e935e082b388186764bb
https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L79-L81
5,515
go-pg/migrations
collection.go
DiscoverSQLMigrations
func (c *Collection) DiscoverSQLMigrations(dir string) error { dir, err := filepath.Abs(dir) if err != nil { return err } if c.isVisitedDir(dir) { return nil } if _, err := os.Stat(dir); os.IsNotExist(err) { return nil } var ms []*Migration newMigration := func(version int64) *Migration { for i := range ms { m := ms[i] if m.Version == version { return m } } ms = append(ms, &Migration{ Version: version, }) return ms[len(ms)-1] } files, err := ioutil.ReadDir(dir) if err != nil { return err } for _, f := range files { if f.IsDir() { continue } fileName := f.Name() if !strings.HasSuffix(fileName, ".sql") { continue } idx := strings.IndexByte(fileName, '_') if idx == -1 { err := fmt.Errorf( "file=%q must have name in format version_comment, e.g. 1_initial", fileName) return err } version, err := strconv.ParseInt(fileName[:idx], 10, 64) if err != nil { return err } m := newMigration(version) filePath := filepath.Join(dir, fileName) if strings.HasSuffix(fileName, ".up.sql") { if m.Up != nil { return fmt.Errorf("migration=%d already has Up func", version) } m.UpTx = strings.HasSuffix(fileName, ".tx.up.sql") m.Up = newSQLMigration(filePath) continue } if strings.HasSuffix(fileName, ".down.sql") { if m.Down != nil { return fmt.Errorf("migration=%d already has Down func", version) } m.DownTx = strings.HasSuffix(fileName, ".tx.down.sql") m.Down = newSQLMigration(filePath) continue } return fmt.Errorf( "file=%q must have extension .up.sql or .down.sql", fileName) } for _, m := range ms { c.addMigration(m) } return nil }
go
func (c *Collection) DiscoverSQLMigrations(dir string) error { dir, err := filepath.Abs(dir) if err != nil { return err } if c.isVisitedDir(dir) { return nil } if _, err := os.Stat(dir); os.IsNotExist(err) { return nil } var ms []*Migration newMigration := func(version int64) *Migration { for i := range ms { m := ms[i] if m.Version == version { return m } } ms = append(ms, &Migration{ Version: version, }) return ms[len(ms)-1] } files, err := ioutil.ReadDir(dir) if err != nil { return err } for _, f := range files { if f.IsDir() { continue } fileName := f.Name() if !strings.HasSuffix(fileName, ".sql") { continue } idx := strings.IndexByte(fileName, '_') if idx == -1 { err := fmt.Errorf( "file=%q must have name in format version_comment, e.g. 1_initial", fileName) return err } version, err := strconv.ParseInt(fileName[:idx], 10, 64) if err != nil { return err } m := newMigration(version) filePath := filepath.Join(dir, fileName) if strings.HasSuffix(fileName, ".up.sql") { if m.Up != nil { return fmt.Errorf("migration=%d already has Up func", version) } m.UpTx = strings.HasSuffix(fileName, ".tx.up.sql") m.Up = newSQLMigration(filePath) continue } if strings.HasSuffix(fileName, ".down.sql") { if m.Down != nil { return fmt.Errorf("migration=%d already has Down func", version) } m.DownTx = strings.HasSuffix(fileName, ".tx.down.sql") m.Down = newSQLMigration(filePath) continue } return fmt.Errorf( "file=%q must have extension .up.sql or .down.sql", fileName) } for _, m := range ms { c.addMigration(m) } return nil }
[ "func", "(", "c", "*", "Collection", ")", "DiscoverSQLMigrations", "(", "dir", "string", ")", "error", "{", "dir", ",", "err", ":=", "filepath", ".", "Abs", "(", "dir", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "if", "c", ".", "isVisitedDir", "(", "dir", ")", "{", "return", "nil", "\n", "}", "\n\n", "if", "_", ",", "err", ":=", "os", ".", "Stat", "(", "dir", ")", ";", "os", ".", "IsNotExist", "(", "err", ")", "{", "return", "nil", "\n", "}", "\n\n", "var", "ms", "[", "]", "*", "Migration", "\n", "newMigration", ":=", "func", "(", "version", "int64", ")", "*", "Migration", "{", "for", "i", ":=", "range", "ms", "{", "m", ":=", "ms", "[", "i", "]", "\n", "if", "m", ".", "Version", "==", "version", "{", "return", "m", "\n", "}", "\n", "}", "\n\n", "ms", "=", "append", "(", "ms", ",", "&", "Migration", "{", "Version", ":", "version", ",", "}", ")", "\n", "return", "ms", "[", "len", "(", "ms", ")", "-", "1", "]", "\n", "}", "\n\n", "files", ",", "err", ":=", "ioutil", ".", "ReadDir", "(", "dir", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "for", "_", ",", "f", ":=", "range", "files", "{", "if", "f", ".", "IsDir", "(", ")", "{", "continue", "\n", "}", "\n\n", "fileName", ":=", "f", ".", "Name", "(", ")", "\n", "if", "!", "strings", ".", "HasSuffix", "(", "fileName", ",", "\"", "\"", ")", "{", "continue", "\n", "}", "\n\n", "idx", ":=", "strings", ".", "IndexByte", "(", "fileName", ",", "'_'", ")", "\n", "if", "idx", "==", "-", "1", "{", "err", ":=", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "fileName", ")", "\n", "return", "err", "\n", "}", "\n\n", "version", ",", "err", ":=", "strconv", ".", "ParseInt", "(", "fileName", "[", ":", "idx", "]", ",", "10", ",", "64", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "m", ":=", "newMigration", "(", "version", ")", "\n", "filePath", ":=", "filepath", ".", "Join", "(", "dir", ",", "fileName", ")", "\n\n", "if", "strings", ".", "HasSuffix", "(", "fileName", ",", "\"", "\"", ")", "{", "if", "m", ".", "Up", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "version", ")", "\n", "}", "\n", "m", ".", "UpTx", "=", "strings", ".", "HasSuffix", "(", "fileName", ",", "\"", "\"", ")", "\n", "m", ".", "Up", "=", "newSQLMigration", "(", "filePath", ")", "\n", "continue", "\n", "}", "\n\n", "if", "strings", ".", "HasSuffix", "(", "fileName", ",", "\"", "\"", ")", "{", "if", "m", ".", "Down", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "version", ")", "\n", "}", "\n", "m", ".", "DownTx", "=", "strings", ".", "HasSuffix", "(", "fileName", ",", "\"", "\"", ")", "\n", "m", ".", "Down", "=", "newSQLMigration", "(", "filePath", ")", "\n", "continue", "\n", "}", "\n\n", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "fileName", ")", "\n", "}", "\n\n", "for", "_", ",", "m", ":=", "range", "ms", "{", "c", ".", "addMigration", "(", "m", ")", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// DiscoverSQLMigrations scan the dir for files with .sql extension // and adds discovered SQL migrations to the collection.
[ "DiscoverSQLMigrations", "scan", "the", "dir", "for", "files", "with", ".", "sql", "extension", "and", "adds", "discovered", "SQL", "migrations", "to", "the", "collection", "." ]
42c5771cbefbd8c12aa3e935e082b388186764bb
https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L144-L231
5,516
tcnksm/go-httpstat
httpstat.go
Format
func (r Result) Format(s fmt.State, verb rune) { switch verb { case 'v': if s.Flag('+') { var buf bytes.Buffer fmt.Fprintf(&buf, "DNS lookup: %4d ms\n", int(r.DNSLookup/time.Millisecond)) fmt.Fprintf(&buf, "TCP connection: %4d ms\n", int(r.TCPConnection/time.Millisecond)) fmt.Fprintf(&buf, "TLS handshake: %4d ms\n", int(r.TLSHandshake/time.Millisecond)) fmt.Fprintf(&buf, "Server processing: %4d ms\n", int(r.ServerProcessing/time.Millisecond)) if r.total > 0 { fmt.Fprintf(&buf, "Content transfer: %4d ms\n\n", int(r.contentTransfer/time.Millisecond)) } else { fmt.Fprintf(&buf, "Content transfer: %4s ms\n\n", "-") } fmt.Fprintf(&buf, "Name Lookup: %4d ms\n", int(r.NameLookup/time.Millisecond)) fmt.Fprintf(&buf, "Connect: %4d ms\n", int(r.Connect/time.Millisecond)) fmt.Fprintf(&buf, "Pre Transfer: %4d ms\n", int(r.Pretransfer/time.Millisecond)) fmt.Fprintf(&buf, "Start Transfer: %4d ms\n", int(r.StartTransfer/time.Millisecond)) if r.total > 0 { fmt.Fprintf(&buf, "Total: %4d ms\n", int(r.total/time.Millisecond)) } else { fmt.Fprintf(&buf, "Total: %4s ms\n", "-") } io.WriteString(s, buf.String()) return } fallthrough case 's', 'q': d := r.durations() list := make([]string, 0, len(d)) for k, v := range d { // Handle when End function is not called if (k == "ContentTransfer" || k == "Total") && r.t5.IsZero() { list = append(list, fmt.Sprintf("%s: - ms", k)) continue } list = append(list, fmt.Sprintf("%s: %d ms", k, v/time.Millisecond)) } io.WriteString(s, strings.Join(list, ", ")) } }
go
func (r Result) Format(s fmt.State, verb rune) { switch verb { case 'v': if s.Flag('+') { var buf bytes.Buffer fmt.Fprintf(&buf, "DNS lookup: %4d ms\n", int(r.DNSLookup/time.Millisecond)) fmt.Fprintf(&buf, "TCP connection: %4d ms\n", int(r.TCPConnection/time.Millisecond)) fmt.Fprintf(&buf, "TLS handshake: %4d ms\n", int(r.TLSHandshake/time.Millisecond)) fmt.Fprintf(&buf, "Server processing: %4d ms\n", int(r.ServerProcessing/time.Millisecond)) if r.total > 0 { fmt.Fprintf(&buf, "Content transfer: %4d ms\n\n", int(r.contentTransfer/time.Millisecond)) } else { fmt.Fprintf(&buf, "Content transfer: %4s ms\n\n", "-") } fmt.Fprintf(&buf, "Name Lookup: %4d ms\n", int(r.NameLookup/time.Millisecond)) fmt.Fprintf(&buf, "Connect: %4d ms\n", int(r.Connect/time.Millisecond)) fmt.Fprintf(&buf, "Pre Transfer: %4d ms\n", int(r.Pretransfer/time.Millisecond)) fmt.Fprintf(&buf, "Start Transfer: %4d ms\n", int(r.StartTransfer/time.Millisecond)) if r.total > 0 { fmt.Fprintf(&buf, "Total: %4d ms\n", int(r.total/time.Millisecond)) } else { fmt.Fprintf(&buf, "Total: %4s ms\n", "-") } io.WriteString(s, buf.String()) return } fallthrough case 's', 'q': d := r.durations() list := make([]string, 0, len(d)) for k, v := range d { // Handle when End function is not called if (k == "ContentTransfer" || k == "Total") && r.t5.IsZero() { list = append(list, fmt.Sprintf("%s: - ms", k)) continue } list = append(list, fmt.Sprintf("%s: %d ms", k, v/time.Millisecond)) } io.WriteString(s, strings.Join(list, ", ")) } }
[ "func", "(", "r", "Result", ")", "Format", "(", "s", "fmt", ".", "State", ",", "verb", "rune", ")", "{", "switch", "verb", "{", "case", "'v'", ":", "if", "s", ".", "Flag", "(", "'+'", ")", "{", "var", "buf", "bytes", ".", "Buffer", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "DNSLookup", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "TCPConnection", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "TLSHandshake", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "ServerProcessing", "/", "time", ".", "Millisecond", ")", ")", "\n\n", "if", "r", ".", "total", ">", "0", "{", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\\n", "\"", ",", "int", "(", "r", ".", "contentTransfer", "/", "time", ".", "Millisecond", ")", ")", "\n", "}", "else", "{", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\\n", "\"", ",", "\"", "\"", ")", "\n", "}", "\n\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "NameLookup", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "Connect", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "Pretransfer", "/", "time", ".", "Millisecond", ")", ")", "\n", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "StartTransfer", "/", "time", ".", "Millisecond", ")", ")", "\n\n", "if", "r", ".", "total", ">", "0", "{", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "int", "(", "r", ".", "total", "/", "time", ".", "Millisecond", ")", ")", "\n", "}", "else", "{", "fmt", ".", "Fprintf", "(", "&", "buf", ",", "\"", "\\n", "\"", ",", "\"", "\"", ")", "\n", "}", "\n", "io", ".", "WriteString", "(", "s", ",", "buf", ".", "String", "(", ")", ")", "\n", "return", "\n", "}", "\n\n", "fallthrough", "\n", "case", "'s'", ",", "'q'", ":", "d", ":=", "r", ".", "durations", "(", ")", "\n", "list", ":=", "make", "(", "[", "]", "string", ",", "0", ",", "len", "(", "d", ")", ")", "\n", "for", "k", ",", "v", ":=", "range", "d", "{", "// Handle when End function is not called", "if", "(", "k", "==", "\"", "\"", "||", "k", "==", "\"", "\"", ")", "&&", "r", ".", "t5", ".", "IsZero", "(", ")", "{", "list", "=", "append", "(", "list", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "k", ")", ")", "\n", "continue", "\n", "}", "\n", "list", "=", "append", "(", "list", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "k", ",", "v", "/", "time", ".", "Millisecond", ")", ")", "\n", "}", "\n", "io", ".", "WriteString", "(", "s", ",", "strings", ".", "Join", "(", "list", ",", "\"", "\"", ")", ")", "\n", "}", "\n\n", "}" ]
// Format formats stats result.
[ "Format", "formats", "stats", "result", "." ]
6962beb23919c3de82c788a87b36f4f901ab514d
https://github.com/tcnksm/go-httpstat/blob/6962beb23919c3de82c788a87b36f4f901ab514d/httpstat.go#L72-L127
5,517
tcnksm/go-httpstat
httpstat.go
WithHTTPStat
func WithHTTPStat(ctx context.Context, r *Result) context.Context { return withClientTrace(ctx, r) }
go
func WithHTTPStat(ctx context.Context, r *Result) context.Context { return withClientTrace(ctx, r) }
[ "func", "WithHTTPStat", "(", "ctx", "context", ".", "Context", ",", "r", "*", "Result", ")", "context", ".", "Context", "{", "return", "withClientTrace", "(", "ctx", ",", "r", ")", "\n", "}" ]
// WithHTTPStat is a wrapper of httptrace.WithClientTrace. It records the // time of each httptrace hooks.
[ "WithHTTPStat", "is", "a", "wrapper", "of", "httptrace", ".", "WithClientTrace", ".", "It", "records", "the", "time", "of", "each", "httptrace", "hooks", "." ]
6962beb23919c3de82c788a87b36f4f901ab514d
https://github.com/tcnksm/go-httpstat/blob/6962beb23919c3de82c788a87b36f4f901ab514d/httpstat.go#L131-L133
5,518
PuerkitoBio/gocrawl
worker.go
run
func (w *worker) run() { defer func() { w.logFunc(LogInfo, "worker done.") w.wg.Done() }() // Enter loop to process URLs until stop signal is received for { var idleChan <-chan time.Time w.logFunc(LogInfo, "waiting for pop...") // Initialize the idle timeout channel, if required if w.opts.WorkerIdleTTL > 0 { idleChan = time.After(w.opts.WorkerIdleTTL) } select { case <-w.stop: w.logFunc(LogInfo, "stop signal received.") return case <-idleChan: w.logFunc(LogInfo, "idle timeout received.") w.sendResponse(nil, false, nil, true) return case batch := <-w.pop: // Got a batch of urls to crawl, loop and check at each iteration if a stop // is received. for _, ctx := range batch { w.logFunc(LogInfo, "popped: %s", ctx.url) if ctx.IsRobotsURL() { w.requestRobotsTxt(ctx) } else if w.isAllowedPerRobotsPolicies(ctx.url) { w.requestURL(ctx, ctx.HeadBeforeGet) } else { // Must still notify Crawler that this URL was processed, although not visited w.opts.Extender.Disallowed(ctx) w.sendResponse(ctx, false, nil, false) } // No need to check for idle timeout here, no idling while looping through // a batch of URLs. select { case <-w.stop: w.logFunc(LogInfo, "stop signal received.") return default: // Nothing, just continue... } } } } }
go
func (w *worker) run() { defer func() { w.logFunc(LogInfo, "worker done.") w.wg.Done() }() // Enter loop to process URLs until stop signal is received for { var idleChan <-chan time.Time w.logFunc(LogInfo, "waiting for pop...") // Initialize the idle timeout channel, if required if w.opts.WorkerIdleTTL > 0 { idleChan = time.After(w.opts.WorkerIdleTTL) } select { case <-w.stop: w.logFunc(LogInfo, "stop signal received.") return case <-idleChan: w.logFunc(LogInfo, "idle timeout received.") w.sendResponse(nil, false, nil, true) return case batch := <-w.pop: // Got a batch of urls to crawl, loop and check at each iteration if a stop // is received. for _, ctx := range batch { w.logFunc(LogInfo, "popped: %s", ctx.url) if ctx.IsRobotsURL() { w.requestRobotsTxt(ctx) } else if w.isAllowedPerRobotsPolicies(ctx.url) { w.requestURL(ctx, ctx.HeadBeforeGet) } else { // Must still notify Crawler that this URL was processed, although not visited w.opts.Extender.Disallowed(ctx) w.sendResponse(ctx, false, nil, false) } // No need to check for idle timeout here, no idling while looping through // a batch of URLs. select { case <-w.stop: w.logFunc(LogInfo, "stop signal received.") return default: // Nothing, just continue... } } } } }
[ "func", "(", "w", "*", "worker", ")", "run", "(", ")", "{", "defer", "func", "(", ")", "{", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "w", ".", "wg", ".", "Done", "(", ")", "\n", "}", "(", ")", "\n\n", "// Enter loop to process URLs until stop signal is received", "for", "{", "var", "idleChan", "<-", "chan", "time", ".", "Time", "\n\n", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n\n", "// Initialize the idle timeout channel, if required", "if", "w", ".", "opts", ".", "WorkerIdleTTL", ">", "0", "{", "idleChan", "=", "time", ".", "After", "(", "w", ".", "opts", ".", "WorkerIdleTTL", ")", "\n", "}", "\n\n", "select", "{", "case", "<-", "w", ".", "stop", ":", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n\n", "case", "<-", "idleChan", ":", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "w", ".", "sendResponse", "(", "nil", ",", "false", ",", "nil", ",", "true", ")", "\n", "return", "\n\n", "case", "batch", ":=", "<-", "w", ".", "pop", ":", "// Got a batch of urls to crawl, loop and check at each iteration if a stop", "// is received.", "for", "_", ",", "ctx", ":=", "range", "batch", "{", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ",", "ctx", ".", "url", ")", "\n\n", "if", "ctx", ".", "IsRobotsURL", "(", ")", "{", "w", ".", "requestRobotsTxt", "(", "ctx", ")", "\n", "}", "else", "if", "w", ".", "isAllowedPerRobotsPolicies", "(", "ctx", ".", "url", ")", "{", "w", ".", "requestURL", "(", "ctx", ",", "ctx", ".", "HeadBeforeGet", ")", "\n", "}", "else", "{", "// Must still notify Crawler that this URL was processed, although not visited", "w", ".", "opts", ".", "Extender", ".", "Disallowed", "(", "ctx", ")", "\n", "w", ".", "sendResponse", "(", "ctx", ",", "false", ",", "nil", ",", "false", ")", "\n", "}", "\n\n", "// No need to check for idle timeout here, no idling while looping through", "// a batch of URLs.", "select", "{", "case", "<-", "w", ".", "stop", ":", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "default", ":", "// Nothing, just continue...", "}", "\n", "}", "\n", "}", "\n", "}", "\n", "}" ]
// Start crawling the host.
[ "Start", "crawling", "the", "host", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L48-L104
5,519
PuerkitoBio/gocrawl
worker.go
isAllowedPerRobotsPolicies
func (w *worker) isAllowedPerRobotsPolicies(u *url.URL) bool { if w.robotsGroup != nil { // Is this URL allowed per robots.txt policy? ok := w.robotsGroup.Test(u.Path) if !ok { w.logFunc(LogIgnored, "ignored on robots.txt policy: %s", u.String()) } return ok } // No robots.txt = everything is allowed return true }
go
func (w *worker) isAllowedPerRobotsPolicies(u *url.URL) bool { if w.robotsGroup != nil { // Is this URL allowed per robots.txt policy? ok := w.robotsGroup.Test(u.Path) if !ok { w.logFunc(LogIgnored, "ignored on robots.txt policy: %s", u.String()) } return ok } // No robots.txt = everything is allowed return true }
[ "func", "(", "w", "*", "worker", ")", "isAllowedPerRobotsPolicies", "(", "u", "*", "url", ".", "URL", ")", "bool", "{", "if", "w", ".", "robotsGroup", "!=", "nil", "{", "// Is this URL allowed per robots.txt policy?", "ok", ":=", "w", ".", "robotsGroup", ".", "Test", "(", "u", ".", "Path", ")", "\n", "if", "!", "ok", "{", "w", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "u", ".", "String", "(", ")", ")", "\n", "}", "\n", "return", "ok", "\n", "}", "\n\n", "// No robots.txt = everything is allowed", "return", "true", "\n", "}" ]
// Checks if the given URL can be fetched based on robots.txt policies.
[ "Checks", "if", "the", "given", "URL", "can", "be", "fetched", "based", "on", "robots", ".", "txt", "policies", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L107-L119
5,520
PuerkitoBio/gocrawl
worker.go
requestURL
func (w *worker) requestURL(ctx *URLContext, headRequest bool) { if res, ok := w.fetchURL(ctx, w.opts.UserAgent, headRequest); ok { var harvested interface{} var visited bool // Close the body on function end defer res.Body.Close() // Any 2xx status code is good to go if res.StatusCode >= 200 && res.StatusCode < 300 { // Success, visit the URL harvested = w.visitURL(ctx, res) visited = true } else { // Error based on status code received w.opts.Extender.Error(newCrawlErrorMessage(ctx, res.Status, CekHttpStatusCode)) w.logFunc(LogError, "ERROR status code for %s: %s", ctx.url, res.Status) } w.sendResponse(ctx, visited, harvested, false) } }
go
func (w *worker) requestURL(ctx *URLContext, headRequest bool) { if res, ok := w.fetchURL(ctx, w.opts.UserAgent, headRequest); ok { var harvested interface{} var visited bool // Close the body on function end defer res.Body.Close() // Any 2xx status code is good to go if res.StatusCode >= 200 && res.StatusCode < 300 { // Success, visit the URL harvested = w.visitURL(ctx, res) visited = true } else { // Error based on status code received w.opts.Extender.Error(newCrawlErrorMessage(ctx, res.Status, CekHttpStatusCode)) w.logFunc(LogError, "ERROR status code for %s: %s", ctx.url, res.Status) } w.sendResponse(ctx, visited, harvested, false) } }
[ "func", "(", "w", "*", "worker", ")", "requestURL", "(", "ctx", "*", "URLContext", ",", "headRequest", "bool", ")", "{", "if", "res", ",", "ok", ":=", "w", ".", "fetchURL", "(", "ctx", ",", "w", ".", "opts", ".", "UserAgent", ",", "headRequest", ")", ";", "ok", "{", "var", "harvested", "interface", "{", "}", "\n", "var", "visited", "bool", "\n\n", "// Close the body on function end", "defer", "res", ".", "Body", ".", "Close", "(", ")", "\n\n", "// Any 2xx status code is good to go", "if", "res", ".", "StatusCode", ">=", "200", "&&", "res", ".", "StatusCode", "<", "300", "{", "// Success, visit the URL", "harvested", "=", "w", ".", "visitURL", "(", "ctx", ",", "res", ")", "\n", "visited", "=", "true", "\n", "}", "else", "{", "// Error based on status code received", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlErrorMessage", "(", "ctx", ",", "res", ".", "Status", ",", "CekHttpStatusCode", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "url", ",", "res", ".", "Status", ")", "\n", "}", "\n", "w", ".", "sendResponse", "(", "ctx", ",", "visited", ",", "harvested", ",", "false", ")", "\n", "}", "\n", "}" ]
// Process the specified URL.
[ "Process", "the", "specified", "URL", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L122-L142
5,521
PuerkitoBio/gocrawl
worker.go
requestRobotsTxt
func (w *worker) requestRobotsTxt(ctx *URLContext) { // Ask if it should be fetched if robData, reqRob := w.opts.Extender.RequestRobots(ctx, w.opts.RobotUserAgent); !reqRob { w.logFunc(LogInfo, "using robots.txt from cache") w.robotsGroup = w.getRobotsTxtGroup(ctx, robData, nil) } else if res, ok := w.fetchURL(ctx, w.opts.UserAgent, false); ok { // Close the body on function end defer res.Body.Close() w.robotsGroup = w.getRobotsTxtGroup(ctx, nil, res) } }
go
func (w *worker) requestRobotsTxt(ctx *URLContext) { // Ask if it should be fetched if robData, reqRob := w.opts.Extender.RequestRobots(ctx, w.opts.RobotUserAgent); !reqRob { w.logFunc(LogInfo, "using robots.txt from cache") w.robotsGroup = w.getRobotsTxtGroup(ctx, robData, nil) } else if res, ok := w.fetchURL(ctx, w.opts.UserAgent, false); ok { // Close the body on function end defer res.Body.Close() w.robotsGroup = w.getRobotsTxtGroup(ctx, nil, res) } }
[ "func", "(", "w", "*", "worker", ")", "requestRobotsTxt", "(", "ctx", "*", "URLContext", ")", "{", "// Ask if it should be fetched", "if", "robData", ",", "reqRob", ":=", "w", ".", "opts", ".", "Extender", ".", "RequestRobots", "(", "ctx", ",", "w", ".", "opts", ".", "RobotUserAgent", ")", ";", "!", "reqRob", "{", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "w", ".", "robotsGroup", "=", "w", ".", "getRobotsTxtGroup", "(", "ctx", ",", "robData", ",", "nil", ")", "\n\n", "}", "else", "if", "res", ",", "ok", ":=", "w", ".", "fetchURL", "(", "ctx", ",", "w", ".", "opts", ".", "UserAgent", ",", "false", ")", ";", "ok", "{", "// Close the body on function end", "defer", "res", ".", "Body", ".", "Close", "(", ")", "\n", "w", ".", "robotsGroup", "=", "w", ".", "getRobotsTxtGroup", "(", "ctx", ",", "nil", ",", "res", ")", "\n", "}", "\n", "}" ]
// Process the robots.txt URL.
[ "Process", "the", "robots", ".", "txt", "URL", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L145-L156
5,522
PuerkitoBio/gocrawl
worker.go
getRobotsTxtGroup
func (w *worker) getRobotsTxtGroup(ctx *URLContext, b []byte, res *http.Response) (g *robotstxt.Group) { var data *robotstxt.RobotsData var e error if res != nil { var buf bytes.Buffer io.Copy(&buf, res.Body) res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes())) data, e = robotstxt.FromResponse(res) // Rewind the res.Body (by re-creating it from the bytes) res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes())) // Error or not, the robots.txt has been fetched, so notify w.opts.Extender.FetchedRobots(ctx, res) } else { data, e = robotstxt.FromBytes(b) } // If robots data cannot be parsed, will return nil, which will allow access by default. // Reasonable, since by default no robots.txt means full access, so invalid // robots.txt is similar behavior. if e != nil { w.opts.Extender.Error(newCrawlError(nil, e, CekParseRobots)) w.logFunc(LogError, "ERROR parsing robots.txt for host %s: %s", w.host, e) } else { g = data.FindGroup(w.opts.RobotUserAgent) } return g }
go
func (w *worker) getRobotsTxtGroup(ctx *URLContext, b []byte, res *http.Response) (g *robotstxt.Group) { var data *robotstxt.RobotsData var e error if res != nil { var buf bytes.Buffer io.Copy(&buf, res.Body) res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes())) data, e = robotstxt.FromResponse(res) // Rewind the res.Body (by re-creating it from the bytes) res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes())) // Error or not, the robots.txt has been fetched, so notify w.opts.Extender.FetchedRobots(ctx, res) } else { data, e = robotstxt.FromBytes(b) } // If robots data cannot be parsed, will return nil, which will allow access by default. // Reasonable, since by default no robots.txt means full access, so invalid // robots.txt is similar behavior. if e != nil { w.opts.Extender.Error(newCrawlError(nil, e, CekParseRobots)) w.logFunc(LogError, "ERROR parsing robots.txt for host %s: %s", w.host, e) } else { g = data.FindGroup(w.opts.RobotUserAgent) } return g }
[ "func", "(", "w", "*", "worker", ")", "getRobotsTxtGroup", "(", "ctx", "*", "URLContext", ",", "b", "[", "]", "byte", ",", "res", "*", "http", ".", "Response", ")", "(", "g", "*", "robotstxt", ".", "Group", ")", "{", "var", "data", "*", "robotstxt", ".", "RobotsData", "\n", "var", "e", "error", "\n\n", "if", "res", "!=", "nil", "{", "var", "buf", "bytes", ".", "Buffer", "\n", "io", ".", "Copy", "(", "&", "buf", ",", "res", ".", "Body", ")", "\n", "res", ".", "Body", "=", "ioutil", ".", "NopCloser", "(", "bytes", ".", "NewReader", "(", "buf", ".", "Bytes", "(", ")", ")", ")", "\n", "data", ",", "e", "=", "robotstxt", ".", "FromResponse", "(", "res", ")", "\n", "// Rewind the res.Body (by re-creating it from the bytes)", "res", ".", "Body", "=", "ioutil", ".", "NopCloser", "(", "bytes", ".", "NewReader", "(", "buf", ".", "Bytes", "(", ")", ")", ")", "\n", "// Error or not, the robots.txt has been fetched, so notify", "w", ".", "opts", ".", "Extender", ".", "FetchedRobots", "(", "ctx", ",", "res", ")", "\n", "}", "else", "{", "data", ",", "e", "=", "robotstxt", ".", "FromBytes", "(", "b", ")", "\n", "}", "\n\n", "// If robots data cannot be parsed, will return nil, which will allow access by default.", "// Reasonable, since by default no robots.txt means full access, so invalid", "// robots.txt is similar behavior.", "if", "e", "!=", "nil", "{", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "nil", ",", "e", ",", "CekParseRobots", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "w", ".", "host", ",", "e", ")", "\n", "}", "else", "{", "g", "=", "data", ".", "FindGroup", "(", "w", ".", "opts", ".", "RobotUserAgent", ")", "\n", "}", "\n", "return", "g", "\n", "}" ]
// Get the robots.txt group for this crawler.
[ "Get", "the", "robots", ".", "txt", "group", "for", "this", "crawler", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L159-L186
5,523
PuerkitoBio/gocrawl
worker.go
setCrawlDelay
func (w *worker) setCrawlDelay() { var robDelay time.Duration if w.robotsGroup != nil { robDelay = w.robotsGroup.CrawlDelay } w.lastCrawlDelay = w.opts.Extender.ComputeDelay(w.host, &DelayInfo{ w.opts.CrawlDelay, robDelay, w.lastCrawlDelay, }, w.lastFetch) w.logFunc(LogInfo, "using crawl-delay: %v", w.lastCrawlDelay) }
go
func (w *worker) setCrawlDelay() { var robDelay time.Duration if w.robotsGroup != nil { robDelay = w.robotsGroup.CrawlDelay } w.lastCrawlDelay = w.opts.Extender.ComputeDelay(w.host, &DelayInfo{ w.opts.CrawlDelay, robDelay, w.lastCrawlDelay, }, w.lastFetch) w.logFunc(LogInfo, "using crawl-delay: %v", w.lastCrawlDelay) }
[ "func", "(", "w", "*", "worker", ")", "setCrawlDelay", "(", ")", "{", "var", "robDelay", "time", ".", "Duration", "\n\n", "if", "w", ".", "robotsGroup", "!=", "nil", "{", "robDelay", "=", "w", ".", "robotsGroup", ".", "CrawlDelay", "\n", "}", "\n", "w", ".", "lastCrawlDelay", "=", "w", ".", "opts", ".", "Extender", ".", "ComputeDelay", "(", "w", ".", "host", ",", "&", "DelayInfo", "{", "w", ".", "opts", ".", "CrawlDelay", ",", "robDelay", ",", "w", ".", "lastCrawlDelay", ",", "}", ",", "w", ".", "lastFetch", ")", "\n", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ",", "w", ".", "lastCrawlDelay", ")", "\n", "}" ]
// Set the crawl delay between this request and the next.
[ "Set", "the", "crawl", "delay", "between", "this", "request", "and", "the", "next", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L189-L203
5,524
PuerkitoBio/gocrawl
worker.go
fetchURL
func (w *worker) fetchURL(ctx *URLContext, agent string, headRequest bool) (res *http.Response, ok bool) { var e error var silent bool for { // Wait for crawl delay, if one is pending. w.logFunc(LogTrace, "waiting for crawl delay") if w.wait != nil { <-w.wait w.wait = nil } // Compute the next delay w.setCrawlDelay() // Compute the fetch duration now := time.Now() // Request the URL if res, e = w.opts.Extender.Fetch(ctx, agent, headRequest); e != nil { // Check if this is an ErrEnqueueRedirect, in which case we will enqueue // the redirect-to URL. if ue, ok := e.(*url.Error); ok { // We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect if ue.Err == ErrEnqueueRedirect { // Do not notify this error outside of this if block, this is not a // "real" error. We either enqueue the new URL, or fail to parse it, // and then stop processing the current URL. silent = true // Parse the URL in the context of the original URL (so that relative URLs are ok). // Absolute URLs that point to another host are ok too. if ur, e := ctx.url.Parse(ue.URL); e != nil { // Notify error w.opts.Extender.Error(newCrawlError(nil, e, CekParseRedirectURL)) w.logFunc(LogError, "ERROR parsing redirect URL %s: %s", ue.URL, e) } else { w.logFunc(LogTrace, "redirect to %s from %s, linked from %s", ur, ctx.URL(), ctx.SourceURL()) // Enqueue the redirect-to URL with the original source rCtx := ctx.cloneForRedirect(ur, w.opts.URLNormalizationFlags) w.enqueue <- rCtx } } } // No fetch, so set to nil w.lastFetch = nil if !silent { // Notify error w.opts.Extender.Error(newCrawlError(ctx, e, CekFetch)) w.logFunc(LogError, "ERROR fetching %s: %s", ctx.url, e) } // Return from this URL crawl w.sendResponse(ctx, false, nil, false) return nil, false } // Get the fetch duration fetchDuration := time.Now().Sub(now) // Crawl delay starts now. w.wait = time.After(w.lastCrawlDelay) // Keep trace of this last fetch info w.lastFetch = &FetchInfo{ ctx, fetchDuration, res.StatusCode, headRequest, } if headRequest { // Close the HEAD request's body defer res.Body.Close() // Next up is GET request, maybe headRequest = false // Ask caller if we should proceed with a GET if !w.opts.Extender.RequestGet(ctx, res) { w.logFunc(LogIgnored, "ignored on HEAD filter policy: %s", ctx.url) w.sendResponse(ctx, false, nil, false) ok = false break } } else { ok = true break } } return }
go
func (w *worker) fetchURL(ctx *URLContext, agent string, headRequest bool) (res *http.Response, ok bool) { var e error var silent bool for { // Wait for crawl delay, if one is pending. w.logFunc(LogTrace, "waiting for crawl delay") if w.wait != nil { <-w.wait w.wait = nil } // Compute the next delay w.setCrawlDelay() // Compute the fetch duration now := time.Now() // Request the URL if res, e = w.opts.Extender.Fetch(ctx, agent, headRequest); e != nil { // Check if this is an ErrEnqueueRedirect, in which case we will enqueue // the redirect-to URL. if ue, ok := e.(*url.Error); ok { // We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect if ue.Err == ErrEnqueueRedirect { // Do not notify this error outside of this if block, this is not a // "real" error. We either enqueue the new URL, or fail to parse it, // and then stop processing the current URL. silent = true // Parse the URL in the context of the original URL (so that relative URLs are ok). // Absolute URLs that point to another host are ok too. if ur, e := ctx.url.Parse(ue.URL); e != nil { // Notify error w.opts.Extender.Error(newCrawlError(nil, e, CekParseRedirectURL)) w.logFunc(LogError, "ERROR parsing redirect URL %s: %s", ue.URL, e) } else { w.logFunc(LogTrace, "redirect to %s from %s, linked from %s", ur, ctx.URL(), ctx.SourceURL()) // Enqueue the redirect-to URL with the original source rCtx := ctx.cloneForRedirect(ur, w.opts.URLNormalizationFlags) w.enqueue <- rCtx } } } // No fetch, so set to nil w.lastFetch = nil if !silent { // Notify error w.opts.Extender.Error(newCrawlError(ctx, e, CekFetch)) w.logFunc(LogError, "ERROR fetching %s: %s", ctx.url, e) } // Return from this URL crawl w.sendResponse(ctx, false, nil, false) return nil, false } // Get the fetch duration fetchDuration := time.Now().Sub(now) // Crawl delay starts now. w.wait = time.After(w.lastCrawlDelay) // Keep trace of this last fetch info w.lastFetch = &FetchInfo{ ctx, fetchDuration, res.StatusCode, headRequest, } if headRequest { // Close the HEAD request's body defer res.Body.Close() // Next up is GET request, maybe headRequest = false // Ask caller if we should proceed with a GET if !w.opts.Extender.RequestGet(ctx, res) { w.logFunc(LogIgnored, "ignored on HEAD filter policy: %s", ctx.url) w.sendResponse(ctx, false, nil, false) ok = false break } } else { ok = true break } } return }
[ "func", "(", "w", "*", "worker", ")", "fetchURL", "(", "ctx", "*", "URLContext", ",", "agent", "string", ",", "headRequest", "bool", ")", "(", "res", "*", "http", ".", "Response", ",", "ok", "bool", ")", "{", "var", "e", "error", "\n", "var", "silent", "bool", "\n\n", "for", "{", "// Wait for crawl delay, if one is pending.", "w", ".", "logFunc", "(", "LogTrace", ",", "\"", "\"", ")", "\n", "if", "w", ".", "wait", "!=", "nil", "{", "<-", "w", ".", "wait", "\n", "w", ".", "wait", "=", "nil", "\n", "}", "\n\n", "// Compute the next delay", "w", ".", "setCrawlDelay", "(", ")", "\n\n", "// Compute the fetch duration", "now", ":=", "time", ".", "Now", "(", ")", "\n\n", "// Request the URL", "if", "res", ",", "e", "=", "w", ".", "opts", ".", "Extender", ".", "Fetch", "(", "ctx", ",", "agent", ",", "headRequest", ")", ";", "e", "!=", "nil", "{", "// Check if this is an ErrEnqueueRedirect, in which case we will enqueue", "// the redirect-to URL.", "if", "ue", ",", "ok", ":=", "e", ".", "(", "*", "url", ".", "Error", ")", ";", "ok", "{", "// We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect", "if", "ue", ".", "Err", "==", "ErrEnqueueRedirect", "{", "// Do not notify this error outside of this if block, this is not a", "// \"real\" error. We either enqueue the new URL, or fail to parse it,", "// and then stop processing the current URL.", "silent", "=", "true", "\n", "// Parse the URL in the context of the original URL (so that relative URLs are ok).", "// Absolute URLs that point to another host are ok too.", "if", "ur", ",", "e", ":=", "ctx", ".", "url", ".", "Parse", "(", "ue", ".", "URL", ")", ";", "e", "!=", "nil", "{", "// Notify error", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "nil", ",", "e", ",", "CekParseRedirectURL", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ue", ".", "URL", ",", "e", ")", "\n", "}", "else", "{", "w", ".", "logFunc", "(", "LogTrace", ",", "\"", "\"", ",", "ur", ",", "ctx", ".", "URL", "(", ")", ",", "ctx", ".", "SourceURL", "(", ")", ")", "\n", "// Enqueue the redirect-to URL with the original source", "rCtx", ":=", "ctx", ".", "cloneForRedirect", "(", "ur", ",", "w", ".", "opts", ".", "URLNormalizationFlags", ")", "\n", "w", ".", "enqueue", "<-", "rCtx", "\n", "}", "\n", "}", "\n", "}", "\n\n", "// No fetch, so set to nil", "w", ".", "lastFetch", "=", "nil", "\n\n", "if", "!", "silent", "{", "// Notify error", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "ctx", ",", "e", ",", "CekFetch", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "url", ",", "e", ")", "\n", "}", "\n\n", "// Return from this URL crawl", "w", ".", "sendResponse", "(", "ctx", ",", "false", ",", "nil", ",", "false", ")", "\n", "return", "nil", ",", "false", "\n\n", "}", "\n", "// Get the fetch duration", "fetchDuration", ":=", "time", ".", "Now", "(", ")", ".", "Sub", "(", "now", ")", "\n", "// Crawl delay starts now.", "w", ".", "wait", "=", "time", ".", "After", "(", "w", ".", "lastCrawlDelay", ")", "\n\n", "// Keep trace of this last fetch info", "w", ".", "lastFetch", "=", "&", "FetchInfo", "{", "ctx", ",", "fetchDuration", ",", "res", ".", "StatusCode", ",", "headRequest", ",", "}", "\n\n", "if", "headRequest", "{", "// Close the HEAD request's body", "defer", "res", ".", "Body", ".", "Close", "(", ")", "\n", "// Next up is GET request, maybe", "headRequest", "=", "false", "\n", "// Ask caller if we should proceed with a GET", "if", "!", "w", ".", "opts", ".", "Extender", ".", "RequestGet", "(", "ctx", ",", "res", ")", "{", "w", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "ctx", ".", "url", ")", "\n", "w", ".", "sendResponse", "(", "ctx", ",", "false", ",", "nil", ",", "false", ")", "\n", "ok", "=", "false", "\n", "break", "\n", "}", "\n", "}", "else", "{", "ok", "=", "true", "\n", "break", "\n", "}", "\n", "}", "\n", "return", "\n", "}" ]
// Request the specified URL and return the response.
[ "Request", "the", "specified", "URL", "and", "return", "the", "response", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L206-L295
5,525
PuerkitoBio/gocrawl
worker.go
sendResponse
func (w *worker) sendResponse(ctx *URLContext, visited bool, harvested interface{}, idleDeath bool) { // Push harvested urls back to crawler, even if empty (uses the channel communication // to decrement reference count of pending URLs) if ctx == nil || !isRobotsURL(ctx.url) { // If a stop signal has been received, ignore the response, since the push // channel may be full and could block indefinitely. select { case <-w.stop: w.logFunc(LogInfo, "ignoring send response, will stop.") return default: // Nothing, just continue... } // No stop signal, send the response res := &workerResponse{ ctx, visited, harvested, w.host, idleDeath, } w.push <- res } }
go
func (w *worker) sendResponse(ctx *URLContext, visited bool, harvested interface{}, idleDeath bool) { // Push harvested urls back to crawler, even if empty (uses the channel communication // to decrement reference count of pending URLs) if ctx == nil || !isRobotsURL(ctx.url) { // If a stop signal has been received, ignore the response, since the push // channel may be full and could block indefinitely. select { case <-w.stop: w.logFunc(LogInfo, "ignoring send response, will stop.") return default: // Nothing, just continue... } // No stop signal, send the response res := &workerResponse{ ctx, visited, harvested, w.host, idleDeath, } w.push <- res } }
[ "func", "(", "w", "*", "worker", ")", "sendResponse", "(", "ctx", "*", "URLContext", ",", "visited", "bool", ",", "harvested", "interface", "{", "}", ",", "idleDeath", "bool", ")", "{", "// Push harvested urls back to crawler, even if empty (uses the channel communication", "// to decrement reference count of pending URLs)", "if", "ctx", "==", "nil", "||", "!", "isRobotsURL", "(", "ctx", ".", "url", ")", "{", "// If a stop signal has been received, ignore the response, since the push", "// channel may be full and could block indefinitely.", "select", "{", "case", "<-", "w", ".", "stop", ":", "w", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "default", ":", "// Nothing, just continue...", "}", "\n\n", "// No stop signal, send the response", "res", ":=", "&", "workerResponse", "{", "ctx", ",", "visited", ",", "harvested", ",", "w", ".", "host", ",", "idleDeath", ",", "}", "\n", "w", ".", "push", "<-", "res", "\n", "}", "\n", "}" ]
// Send a response to the crawler.
[ "Send", "a", "response", "to", "the", "crawler", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L298-L322
5,526
PuerkitoBio/gocrawl
worker.go
visitURL
func (w *worker) visitURL(ctx *URLContext, res *http.Response) interface{} { var doc *goquery.Document var harvested interface{} var doLinks bool // Load a goquery document and call the visitor function if bd, e := ioutil.ReadAll(res.Body); e != nil { w.opts.Extender.Error(newCrawlError(ctx, e, CekReadBody)) w.logFunc(LogError, "ERROR reading body %s: %s", ctx.url, e) } else { if node, e := html.Parse(bytes.NewBuffer(bd)); e != nil { w.opts.Extender.Error(newCrawlError(ctx, e, CekParseBody)) w.logFunc(LogError, "ERROR parsing %s: %s", ctx.url, e) } else { doc = goquery.NewDocumentFromNode(node) doc.Url = res.Request.URL } // Re-assign the body so it can be consumed by the visitor function res.Body = ioutil.NopCloser(bytes.NewBuffer(bd)) } // Visit the document (with nil goquery doc if failed to load) if harvested, doLinks = w.opts.Extender.Visit(ctx, res, doc); doLinks { // Links were not processed by the visitor, so process links if doc != nil { harvested = w.processLinks(doc) } else { w.opts.Extender.Error(newCrawlErrorMessage(ctx, "No goquery document to process links.", CekProcessLinks)) w.logFunc(LogError, "ERROR processing links %s", ctx.url) } } // Notify that this URL has been visited w.opts.Extender.Visited(ctx, harvested) return harvested }
go
func (w *worker) visitURL(ctx *URLContext, res *http.Response) interface{} { var doc *goquery.Document var harvested interface{} var doLinks bool // Load a goquery document and call the visitor function if bd, e := ioutil.ReadAll(res.Body); e != nil { w.opts.Extender.Error(newCrawlError(ctx, e, CekReadBody)) w.logFunc(LogError, "ERROR reading body %s: %s", ctx.url, e) } else { if node, e := html.Parse(bytes.NewBuffer(bd)); e != nil { w.opts.Extender.Error(newCrawlError(ctx, e, CekParseBody)) w.logFunc(LogError, "ERROR parsing %s: %s", ctx.url, e) } else { doc = goquery.NewDocumentFromNode(node) doc.Url = res.Request.URL } // Re-assign the body so it can be consumed by the visitor function res.Body = ioutil.NopCloser(bytes.NewBuffer(bd)) } // Visit the document (with nil goquery doc if failed to load) if harvested, doLinks = w.opts.Extender.Visit(ctx, res, doc); doLinks { // Links were not processed by the visitor, so process links if doc != nil { harvested = w.processLinks(doc) } else { w.opts.Extender.Error(newCrawlErrorMessage(ctx, "No goquery document to process links.", CekProcessLinks)) w.logFunc(LogError, "ERROR processing links %s", ctx.url) } } // Notify that this URL has been visited w.opts.Extender.Visited(ctx, harvested) return harvested }
[ "func", "(", "w", "*", "worker", ")", "visitURL", "(", "ctx", "*", "URLContext", ",", "res", "*", "http", ".", "Response", ")", "interface", "{", "}", "{", "var", "doc", "*", "goquery", ".", "Document", "\n", "var", "harvested", "interface", "{", "}", "\n", "var", "doLinks", "bool", "\n\n", "// Load a goquery document and call the visitor function", "if", "bd", ",", "e", ":=", "ioutil", ".", "ReadAll", "(", "res", ".", "Body", ")", ";", "e", "!=", "nil", "{", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "ctx", ",", "e", ",", "CekReadBody", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "url", ",", "e", ")", "\n", "}", "else", "{", "if", "node", ",", "e", ":=", "html", ".", "Parse", "(", "bytes", ".", "NewBuffer", "(", "bd", ")", ")", ";", "e", "!=", "nil", "{", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "ctx", ",", "e", ",", "CekParseBody", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "url", ",", "e", ")", "\n", "}", "else", "{", "doc", "=", "goquery", ".", "NewDocumentFromNode", "(", "node", ")", "\n", "doc", ".", "Url", "=", "res", ".", "Request", ".", "URL", "\n", "}", "\n", "// Re-assign the body so it can be consumed by the visitor function", "res", ".", "Body", "=", "ioutil", ".", "NopCloser", "(", "bytes", ".", "NewBuffer", "(", "bd", ")", ")", "\n", "}", "\n\n", "// Visit the document (with nil goquery doc if failed to load)", "if", "harvested", ",", "doLinks", "=", "w", ".", "opts", ".", "Extender", ".", "Visit", "(", "ctx", ",", "res", ",", "doc", ")", ";", "doLinks", "{", "// Links were not processed by the visitor, so process links", "if", "doc", "!=", "nil", "{", "harvested", "=", "w", ".", "processLinks", "(", "doc", ")", "\n", "}", "else", "{", "w", ".", "opts", ".", "Extender", ".", "Error", "(", "newCrawlErrorMessage", "(", "ctx", ",", "\"", "\"", ",", "CekProcessLinks", ")", ")", "\n", "w", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "url", ")", "\n", "}", "\n", "}", "\n", "// Notify that this URL has been visited", "w", ".", "opts", ".", "Extender", ".", "Visited", "(", "ctx", ",", "harvested", ")", "\n\n", "return", "harvested", "\n", "}" ]
// Process the response for a URL.
[ "Process", "the", "response", "for", "a", "URL", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L325-L360
5,527
PuerkitoBio/gocrawl
worker.go
processLinks
func (w *worker) processLinks(doc *goquery.Document) (result []*url.URL) { baseURL, _ := doc.Find("base[href]").Attr("href") urls := doc.Find("a[href]").Map(func(_ int, s *goquery.Selection) string { val, _ := s.Attr("href") if baseURL != "" { val = handleBaseTag(doc.Url, baseURL, val) } return val }) for _, s := range urls { // If href starts with "#", then it points to this same exact URL, ignore (will fail to parse anyway) if len(s) > 0 && !strings.HasPrefix(s, "#") { if parsed, e := url.Parse(s); e == nil { parsed = doc.Url.ResolveReference(parsed) result = append(result, parsed) } else { w.logFunc(LogIgnored, "ignore on unparsable policy %s: %s", s, e.Error()) } } } return }
go
func (w *worker) processLinks(doc *goquery.Document) (result []*url.URL) { baseURL, _ := doc.Find("base[href]").Attr("href") urls := doc.Find("a[href]").Map(func(_ int, s *goquery.Selection) string { val, _ := s.Attr("href") if baseURL != "" { val = handleBaseTag(doc.Url, baseURL, val) } return val }) for _, s := range urls { // If href starts with "#", then it points to this same exact URL, ignore (will fail to parse anyway) if len(s) > 0 && !strings.HasPrefix(s, "#") { if parsed, e := url.Parse(s); e == nil { parsed = doc.Url.ResolveReference(parsed) result = append(result, parsed) } else { w.logFunc(LogIgnored, "ignore on unparsable policy %s: %s", s, e.Error()) } } } return }
[ "func", "(", "w", "*", "worker", ")", "processLinks", "(", "doc", "*", "goquery", ".", "Document", ")", "(", "result", "[", "]", "*", "url", ".", "URL", ")", "{", "baseURL", ",", "_", ":=", "doc", ".", "Find", "(", "\"", "\"", ")", ".", "Attr", "(", "\"", "\"", ")", "\n", "urls", ":=", "doc", ".", "Find", "(", "\"", "\"", ")", ".", "Map", "(", "func", "(", "_", "int", ",", "s", "*", "goquery", ".", "Selection", ")", "string", "{", "val", ",", "_", ":=", "s", ".", "Attr", "(", "\"", "\"", ")", "\n", "if", "baseURL", "!=", "\"", "\"", "{", "val", "=", "handleBaseTag", "(", "doc", ".", "Url", ",", "baseURL", ",", "val", ")", "\n", "}", "\n", "return", "val", "\n", "}", ")", "\n", "for", "_", ",", "s", ":=", "range", "urls", "{", "// If href starts with \"#\", then it points to this same exact URL, ignore (will fail to parse anyway)", "if", "len", "(", "s", ")", ">", "0", "&&", "!", "strings", ".", "HasPrefix", "(", "s", ",", "\"", "\"", ")", "{", "if", "parsed", ",", "e", ":=", "url", ".", "Parse", "(", "s", ")", ";", "e", "==", "nil", "{", "parsed", "=", "doc", ".", "Url", ".", "ResolveReference", "(", "parsed", ")", "\n", "result", "=", "append", "(", "result", ",", "parsed", ")", "\n", "}", "else", "{", "w", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "s", ",", "e", ".", "Error", "(", ")", ")", "\n", "}", "\n", "}", "\n", "}", "\n", "return", "\n", "}" ]
// Scrape the document's content to gather all links
[ "Scrape", "the", "document", "s", "content", "to", "gather", "all", "links" ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L385-L406
5,528
PuerkitoBio/gocrawl
ext.go
Log
func (de *DefaultExtender) Log(logFlags LogFlags, msgLevel LogFlags, msg string) { if logFlags&msgLevel == msgLevel { log.Println(msg) } }
go
func (de *DefaultExtender) Log(logFlags LogFlags, msgLevel LogFlags, msg string) { if logFlags&msgLevel == msgLevel { log.Println(msg) } }
[ "func", "(", "de", "*", "DefaultExtender", ")", "Log", "(", "logFlags", "LogFlags", ",", "msgLevel", "LogFlags", ",", "msg", "string", ")", "{", "if", "logFlags", "&", "msgLevel", "==", "msgLevel", "{", "log", ".", "Println", "(", "msg", ")", "\n", "}", "\n", "}" ]
// Log prints to the standard error by default, based on the requested log verbosity.
[ "Log", "prints", "to", "the", "standard", "error", "by", "default", "based", "on", "the", "requested", "log", "verbosity", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L101-L105
5,529
PuerkitoBio/gocrawl
ext.go
ComputeDelay
func (de *DefaultExtender) ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration { if di.RobotsDelay > 0 { return di.RobotsDelay } return di.OptsDelay }
go
func (de *DefaultExtender) ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration { if di.RobotsDelay > 0 { return di.RobotsDelay } return di.OptsDelay }
[ "func", "(", "de", "*", "DefaultExtender", ")", "ComputeDelay", "(", "host", "string", ",", "di", "*", "DelayInfo", ",", "lastFetch", "*", "FetchInfo", ")", "time", ".", "Duration", "{", "if", "di", ".", "RobotsDelay", ">", "0", "{", "return", "di", ".", "RobotsDelay", "\n", "}", "\n", "return", "di", ".", "OptsDelay", "\n", "}" ]
// ComputeDelay returns the delay specified in the Crawler's Options, unless a // crawl-delay is specified in the robots.txt file, which has precedence.
[ "ComputeDelay", "returns", "the", "delay", "specified", "in", "the", "Crawler", "s", "Options", "unless", "a", "crawl", "-", "delay", "is", "specified", "in", "the", "robots", ".", "txt", "file", "which", "has", "precedence", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L109-L114
5,530
PuerkitoBio/gocrawl
ext.go
Visit
func (de *DefaultExtender) Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool) { return nil, true }
go
func (de *DefaultExtender) Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool) { return nil, true }
[ "func", "(", "de", "*", "DefaultExtender", ")", "Visit", "(", "ctx", "*", "URLContext", ",", "res", "*", "http", ".", "Response", ",", "doc", "*", "goquery", ".", "Document", ")", "(", "harvested", "interface", "{", "}", ",", "findLinks", "bool", ")", "{", "return", "nil", ",", "true", "\n", "}" ]
// Visit asks the worker to harvest the links in this page.
[ "Visit", "asks", "the", "worker", "to", "harvest", "the", "links", "in", "this", "page", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L180-L182
5,531
PuerkitoBio/gocrawl
crawler.go
NewCrawlerWithOptions
func NewCrawlerWithOptions(opts *Options) *Crawler { ret := new(Crawler) ret.Options = opts return ret }
go
func NewCrawlerWithOptions(opts *Options) *Crawler { ret := new(Crawler) ret.Options = opts return ret }
[ "func", "NewCrawlerWithOptions", "(", "opts", "*", "Options", ")", "*", "Crawler", "{", "ret", ":=", "new", "(", "Crawler", ")", "\n", "ret", ".", "Options", "=", "opts", "\n", "return", "ret", "\n", "}" ]
// NewCrawlerWithOptions returns a Crawler initialized with the // provided Options.
[ "NewCrawlerWithOptions", "returns", "a", "Crawler", "initialized", "with", "the", "provided", "Options", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L42-L46
5,532
PuerkitoBio/gocrawl
crawler.go
init
func (c *Crawler) init(ctxs []*URLContext) { // Initialize the internal hosts map c.hosts = make(map[string]struct{}, len(ctxs)) for _, ctx := range ctxs { // Add this normalized URL's host if it is not already there. if _, ok := c.hosts[ctx.normalizedURL.Host]; !ok { c.hosts[ctx.normalizedURL.Host] = struct{}{} } } hostCount := len(c.hosts) l := len(ctxs) c.logFunc(LogTrace, "init() - seeds length: %d", l) c.logFunc(LogTrace, "init() - host count: %d", hostCount) c.logFunc(LogInfo, "robot user-agent: %s", c.Options.RobotUserAgent) // Create a shiny new WaitGroup c.wg = new(sync.WaitGroup) // Initialize the visits fields c.visited = make(map[string]struct{}, l) c.pushPopRefCount, c.visits = 0, 0 // Create the workers map and the push channel (the channel used by workers // to communicate back to the crawler) c.stop = make(chan struct{}) if c.Options.SameHostOnly { c.workers, c.push = make(map[string]*worker, hostCount), make(chan *workerResponse, hostCount) } else { c.workers, c.push = make(map[string]*worker, c.Options.HostBufferFactor*hostCount), make(chan *workerResponse, c.Options.HostBufferFactor*hostCount) } // Create and pass the enqueue channel c.enqueue = make(chan interface{}, c.Options.EnqueueChanBuffer) c.setExtenderEnqueueChan() }
go
func (c *Crawler) init(ctxs []*URLContext) { // Initialize the internal hosts map c.hosts = make(map[string]struct{}, len(ctxs)) for _, ctx := range ctxs { // Add this normalized URL's host if it is not already there. if _, ok := c.hosts[ctx.normalizedURL.Host]; !ok { c.hosts[ctx.normalizedURL.Host] = struct{}{} } } hostCount := len(c.hosts) l := len(ctxs) c.logFunc(LogTrace, "init() - seeds length: %d", l) c.logFunc(LogTrace, "init() - host count: %d", hostCount) c.logFunc(LogInfo, "robot user-agent: %s", c.Options.RobotUserAgent) // Create a shiny new WaitGroup c.wg = new(sync.WaitGroup) // Initialize the visits fields c.visited = make(map[string]struct{}, l) c.pushPopRefCount, c.visits = 0, 0 // Create the workers map and the push channel (the channel used by workers // to communicate back to the crawler) c.stop = make(chan struct{}) if c.Options.SameHostOnly { c.workers, c.push = make(map[string]*worker, hostCount), make(chan *workerResponse, hostCount) } else { c.workers, c.push = make(map[string]*worker, c.Options.HostBufferFactor*hostCount), make(chan *workerResponse, c.Options.HostBufferFactor*hostCount) } // Create and pass the enqueue channel c.enqueue = make(chan interface{}, c.Options.EnqueueChanBuffer) c.setExtenderEnqueueChan() }
[ "func", "(", "c", "*", "Crawler", ")", "init", "(", "ctxs", "[", "]", "*", "URLContext", ")", "{", "// Initialize the internal hosts map", "c", ".", "hosts", "=", "make", "(", "map", "[", "string", "]", "struct", "{", "}", ",", "len", "(", "ctxs", ")", ")", "\n", "for", "_", ",", "ctx", ":=", "range", "ctxs", "{", "// Add this normalized URL's host if it is not already there.", "if", "_", ",", "ok", ":=", "c", ".", "hosts", "[", "ctx", ".", "normalizedURL", ".", "Host", "]", ";", "!", "ok", "{", "c", ".", "hosts", "[", "ctx", ".", "normalizedURL", ".", "Host", "]", "=", "struct", "{", "}", "{", "}", "\n", "}", "\n", "}", "\n\n", "hostCount", ":=", "len", "(", "c", ".", "hosts", ")", "\n", "l", ":=", "len", "(", "ctxs", ")", "\n", "c", ".", "logFunc", "(", "LogTrace", ",", "\"", "\"", ",", "l", ")", "\n", "c", ".", "logFunc", "(", "LogTrace", ",", "\"", "\"", ",", "hostCount", ")", "\n", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ",", "c", ".", "Options", ".", "RobotUserAgent", ")", "\n\n", "// Create a shiny new WaitGroup", "c", ".", "wg", "=", "new", "(", "sync", ".", "WaitGroup", ")", "\n\n", "// Initialize the visits fields", "c", ".", "visited", "=", "make", "(", "map", "[", "string", "]", "struct", "{", "}", ",", "l", ")", "\n", "c", ".", "pushPopRefCount", ",", "c", ".", "visits", "=", "0", ",", "0", "\n\n", "// Create the workers map and the push channel (the channel used by workers", "// to communicate back to the crawler)", "c", ".", "stop", "=", "make", "(", "chan", "struct", "{", "}", ")", "\n", "if", "c", ".", "Options", ".", "SameHostOnly", "{", "c", ".", "workers", ",", "c", ".", "push", "=", "make", "(", "map", "[", "string", "]", "*", "worker", ",", "hostCount", ")", ",", "make", "(", "chan", "*", "workerResponse", ",", "hostCount", ")", "\n", "}", "else", "{", "c", ".", "workers", ",", "c", ".", "push", "=", "make", "(", "map", "[", "string", "]", "*", "worker", ",", "c", ".", "Options", ".", "HostBufferFactor", "*", "hostCount", ")", ",", "make", "(", "chan", "*", "workerResponse", ",", "c", ".", "Options", ".", "HostBufferFactor", "*", "hostCount", ")", "\n", "}", "\n", "// Create and pass the enqueue channel", "c", ".", "enqueue", "=", "make", "(", "chan", "interface", "{", "}", ",", "c", ".", "Options", ".", "EnqueueChanBuffer", ")", "\n", "c", ".", "setExtenderEnqueueChan", "(", ")", "\n", "}" ]
// Initialize the Crawler's internal fields before a crawling execution.
[ "Initialize", "the", "Crawler", "s", "internal", "fields", "before", "a", "crawling", "execution", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L78-L114
5,533
PuerkitoBio/gocrawl
crawler.go
setExtenderEnqueueChan
func (c *Crawler) setExtenderEnqueueChan() { defer func() { if err := recover(); err != nil { // Panic can happen if the field exists on a pointer struct, but that // pointer is nil. c.logFunc(LogError, "cannot set the enqueue channel: %s", err) } }() // Using reflection, check if the extender has a `EnqueueChan` field // of type `chan<- interface{}`. If it does, set it to the crawler's // enqueue channel. v := reflect.ValueOf(c.Options.Extender) el := v.Elem() if el.Kind() != reflect.Struct { c.logFunc(LogInfo, "extender is not a struct, cannot set the enqueue channel") return } ec := el.FieldByName("EnqueueChan") if !ec.IsValid() { c.logFunc(LogInfo, "extender.EnqueueChan does not exist, cannot set the enqueue channel") return } t := ec.Type() if t.Kind() != reflect.Chan || t.ChanDir() != reflect.SendDir { c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel") return } tt := t.Elem() if tt.Kind() != reflect.Interface || tt.NumMethod() != 0 { c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel") return } src := reflect.ValueOf(c.enqueue) ec.Set(src) }
go
func (c *Crawler) setExtenderEnqueueChan() { defer func() { if err := recover(); err != nil { // Panic can happen if the field exists on a pointer struct, but that // pointer is nil. c.logFunc(LogError, "cannot set the enqueue channel: %s", err) } }() // Using reflection, check if the extender has a `EnqueueChan` field // of type `chan<- interface{}`. If it does, set it to the crawler's // enqueue channel. v := reflect.ValueOf(c.Options.Extender) el := v.Elem() if el.Kind() != reflect.Struct { c.logFunc(LogInfo, "extender is not a struct, cannot set the enqueue channel") return } ec := el.FieldByName("EnqueueChan") if !ec.IsValid() { c.logFunc(LogInfo, "extender.EnqueueChan does not exist, cannot set the enqueue channel") return } t := ec.Type() if t.Kind() != reflect.Chan || t.ChanDir() != reflect.SendDir { c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel") return } tt := t.Elem() if tt.Kind() != reflect.Interface || tt.NumMethod() != 0 { c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel") return } src := reflect.ValueOf(c.enqueue) ec.Set(src) }
[ "func", "(", "c", "*", "Crawler", ")", "setExtenderEnqueueChan", "(", ")", "{", "defer", "func", "(", ")", "{", "if", "err", ":=", "recover", "(", ")", ";", "err", "!=", "nil", "{", "// Panic can happen if the field exists on a pointer struct, but that", "// pointer is nil.", "c", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "}", "(", ")", "\n\n", "// Using reflection, check if the extender has a `EnqueueChan` field", "// of type `chan<- interface{}`. If it does, set it to the crawler's", "// enqueue channel.", "v", ":=", "reflect", ".", "ValueOf", "(", "c", ".", "Options", ".", "Extender", ")", "\n", "el", ":=", "v", ".", "Elem", "(", ")", "\n", "if", "el", ".", "Kind", "(", ")", "!=", "reflect", ".", "Struct", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "}", "\n", "ec", ":=", "el", ".", "FieldByName", "(", "\"", "\"", ")", "\n", "if", "!", "ec", ".", "IsValid", "(", ")", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "}", "\n", "t", ":=", "ec", ".", "Type", "(", ")", "\n", "if", "t", ".", "Kind", "(", ")", "!=", "reflect", ".", "Chan", "||", "t", ".", "ChanDir", "(", ")", "!=", "reflect", ".", "SendDir", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "}", "\n", "tt", ":=", "t", ".", "Elem", "(", ")", "\n", "if", "tt", ".", "Kind", "(", ")", "!=", "reflect", ".", "Interface", "||", "tt", ".", "NumMethod", "(", ")", "!=", "0", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "return", "\n", "}", "\n", "src", ":=", "reflect", ".", "ValueOf", "(", "c", ".", "enqueue", ")", "\n", "ec", ".", "Set", "(", "src", ")", "\n", "}" ]
// Set the Enqueue channel on the extender, based on the naming convention.
[ "Set", "the", "Enqueue", "channel", "on", "the", "extender", "based", "on", "the", "naming", "convention", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L117-L152
5,534
PuerkitoBio/gocrawl
crawler.go
launchWorker
func (c *Crawler) launchWorker(ctx *URLContext) *worker { // Initialize index and channels i := len(c.workers) + 1 pop := newPopChannel() // Create the worker w := &worker{ host: ctx.normalizedURL.Host, index: i, push: c.push, pop: pop, stop: c.stop, enqueue: c.enqueue, wg: c.wg, logFunc: getLogFunc(c.Options.Extender, c.Options.LogFlags, i), opts: c.Options, } // Increment wait group count c.wg.Add(1) // Launch worker go w.run() c.logFunc(LogInfo, "worker %d launched for host %s", i, w.host) c.workers[w.host] = w return w }
go
func (c *Crawler) launchWorker(ctx *URLContext) *worker { // Initialize index and channels i := len(c.workers) + 1 pop := newPopChannel() // Create the worker w := &worker{ host: ctx.normalizedURL.Host, index: i, push: c.push, pop: pop, stop: c.stop, enqueue: c.enqueue, wg: c.wg, logFunc: getLogFunc(c.Options.Extender, c.Options.LogFlags, i), opts: c.Options, } // Increment wait group count c.wg.Add(1) // Launch worker go w.run() c.logFunc(LogInfo, "worker %d launched for host %s", i, w.host) c.workers[w.host] = w return w }
[ "func", "(", "c", "*", "Crawler", ")", "launchWorker", "(", "ctx", "*", "URLContext", ")", "*", "worker", "{", "// Initialize index and channels", "i", ":=", "len", "(", "c", ".", "workers", ")", "+", "1", "\n", "pop", ":=", "newPopChannel", "(", ")", "\n\n", "// Create the worker", "w", ":=", "&", "worker", "{", "host", ":", "ctx", ".", "normalizedURL", ".", "Host", ",", "index", ":", "i", ",", "push", ":", "c", ".", "push", ",", "pop", ":", "pop", ",", "stop", ":", "c", ".", "stop", ",", "enqueue", ":", "c", ".", "enqueue", ",", "wg", ":", "c", ".", "wg", ",", "logFunc", ":", "getLogFunc", "(", "c", ".", "Options", ".", "Extender", ",", "c", ".", "Options", ".", "LogFlags", ",", "i", ")", ",", "opts", ":", "c", ".", "Options", ",", "}", "\n\n", "// Increment wait group count", "c", ".", "wg", ".", "Add", "(", "1", ")", "\n\n", "// Launch worker", "go", "w", ".", "run", "(", ")", "\n", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ",", "i", ",", "w", ".", "host", ")", "\n", "c", ".", "workers", "[", "w", ".", "host", "]", "=", "w", "\n\n", "return", "w", "\n", "}" ]
// Launch a new worker goroutine for a given host.
[ "Launch", "a", "new", "worker", "goroutine", "for", "a", "given", "host", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L155-L182
5,535
PuerkitoBio/gocrawl
crawler.go
isSameHost
func (c *Crawler) isSameHost(ctx *URLContext) bool { // If there is a source URL, then just check if the new URL is from the same host if ctx.normalizedSourceURL != nil { return ctx.normalizedURL.Host == ctx.normalizedSourceURL.Host } // Otherwise, check if the URL is from one of the seed hosts _, ok := c.hosts[ctx.normalizedURL.Host] return ok }
go
func (c *Crawler) isSameHost(ctx *URLContext) bool { // If there is a source URL, then just check if the new URL is from the same host if ctx.normalizedSourceURL != nil { return ctx.normalizedURL.Host == ctx.normalizedSourceURL.Host } // Otherwise, check if the URL is from one of the seed hosts _, ok := c.hosts[ctx.normalizedURL.Host] return ok }
[ "func", "(", "c", "*", "Crawler", ")", "isSameHost", "(", "ctx", "*", "URLContext", ")", "bool", "{", "// If there is a source URL, then just check if the new URL is from the same host", "if", "ctx", ".", "normalizedSourceURL", "!=", "nil", "{", "return", "ctx", ".", "normalizedURL", ".", "Host", "==", "ctx", ".", "normalizedSourceURL", ".", "Host", "\n", "}", "\n\n", "// Otherwise, check if the URL is from one of the seed hosts", "_", ",", "ok", ":=", "c", ".", "hosts", "[", "ctx", ".", "normalizedURL", ".", "Host", "]", "\n", "return", "ok", "\n", "}" ]
// Check if the specified URL is from the same host as its source URL, or if // nil, from the same host as one of the seed URLs.
[ "Check", "if", "the", "specified", "URL", "is", "from", "the", "same", "host", "as", "its", "source", "URL", "or", "if", "nil", "from", "the", "same", "host", "as", "one", "of", "the", "seed", "URLs", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L186-L195
5,536
PuerkitoBio/gocrawl
crawler.go
enqueueUrls
func (c *Crawler) enqueueUrls(ctxs []*URLContext) (cnt int) { for _, ctx := range ctxs { var isVisited, enqueue bool // Cannot directly enqueue a robots.txt URL, since it is managed as a special case // in the worker (doesn't return a response to crawler). if ctx.IsRobotsURL() { continue } // Check if it has been visited before, using the normalized URL _, isVisited = c.visited[ctx.normalizedURL.String()] // Filter the URL if enqueue = c.Options.Extender.Filter(ctx, isVisited); !enqueue { // Filter said NOT to use this url, so continue with next c.logFunc(LogIgnored, "ignore on filter policy: %s", ctx.normalizedURL) continue } // Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed, // and comply with the same host policy if requested. if !ctx.normalizedURL.IsAbs() { // Only absolute URLs are processed, so ignore c.logFunc(LogIgnored, "ignore on absolute policy: %s", ctx.normalizedURL) } else if !strings.HasPrefix(ctx.normalizedURL.Scheme, "http") { c.logFunc(LogIgnored, "ignore on scheme policy: %s", ctx.normalizedURL) } else if c.Options.SameHostOnly && !c.isSameHost(ctx) { // Only allow URLs coming from the same host c.logFunc(LogIgnored, "ignore on same host policy: %s", ctx.normalizedURL) } else { // All is good, visit this URL (robots.txt verification is done by worker) // Possible caveat: if the normalization changes the host, it is possible // that the robots.txt fetched for this host would differ from the one for // the unnormalized host. However, this should be rare, and is a weird // behaviour from the host (i.e. why would site.com differ in its rules // from www.site.com) and can be fixed by using a different normalization // flag. So this is an acceptable behaviour for gocrawl. // Launch worker if required, based on the host of the normalized URL w, ok := c.workers[ctx.normalizedURL.Host] if !ok { // No worker exists for this host, launch a new one w = c.launchWorker(ctx) // Automatically enqueue the robots.txt URL as first in line if robCtx, e := ctx.getRobotsURLCtx(); e != nil { c.Options.Extender.Error(newCrawlError(ctx, e, CekParseRobots)) c.logFunc(LogError, "ERROR parsing robots.txt from %s: %s", ctx.normalizedURL, e) } else { c.logFunc(LogEnqueued, "enqueue: %s", robCtx.url) c.Options.Extender.Enqueued(robCtx) w.pop.stack(robCtx) } } cnt++ c.logFunc(LogEnqueued, "enqueue: %s", ctx.url) c.Options.Extender.Enqueued(ctx) w.pop.stack(ctx) c.pushPopRefCount++ // Once it is stacked, it WILL be visited eventually, so add it to the visited slice // (unless denied by robots.txt, but this is out of our hands, for all we // care, it is visited). if !isVisited { // The visited map works with the normalized URL c.visited[ctx.normalizedURL.String()] = struct{}{} } } } return }
go
func (c *Crawler) enqueueUrls(ctxs []*URLContext) (cnt int) { for _, ctx := range ctxs { var isVisited, enqueue bool // Cannot directly enqueue a robots.txt URL, since it is managed as a special case // in the worker (doesn't return a response to crawler). if ctx.IsRobotsURL() { continue } // Check if it has been visited before, using the normalized URL _, isVisited = c.visited[ctx.normalizedURL.String()] // Filter the URL if enqueue = c.Options.Extender.Filter(ctx, isVisited); !enqueue { // Filter said NOT to use this url, so continue with next c.logFunc(LogIgnored, "ignore on filter policy: %s", ctx.normalizedURL) continue } // Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed, // and comply with the same host policy if requested. if !ctx.normalizedURL.IsAbs() { // Only absolute URLs are processed, so ignore c.logFunc(LogIgnored, "ignore on absolute policy: %s", ctx.normalizedURL) } else if !strings.HasPrefix(ctx.normalizedURL.Scheme, "http") { c.logFunc(LogIgnored, "ignore on scheme policy: %s", ctx.normalizedURL) } else if c.Options.SameHostOnly && !c.isSameHost(ctx) { // Only allow URLs coming from the same host c.logFunc(LogIgnored, "ignore on same host policy: %s", ctx.normalizedURL) } else { // All is good, visit this URL (robots.txt verification is done by worker) // Possible caveat: if the normalization changes the host, it is possible // that the robots.txt fetched for this host would differ from the one for // the unnormalized host. However, this should be rare, and is a weird // behaviour from the host (i.e. why would site.com differ in its rules // from www.site.com) and can be fixed by using a different normalization // flag. So this is an acceptable behaviour for gocrawl. // Launch worker if required, based on the host of the normalized URL w, ok := c.workers[ctx.normalizedURL.Host] if !ok { // No worker exists for this host, launch a new one w = c.launchWorker(ctx) // Automatically enqueue the robots.txt URL as first in line if robCtx, e := ctx.getRobotsURLCtx(); e != nil { c.Options.Extender.Error(newCrawlError(ctx, e, CekParseRobots)) c.logFunc(LogError, "ERROR parsing robots.txt from %s: %s", ctx.normalizedURL, e) } else { c.logFunc(LogEnqueued, "enqueue: %s", robCtx.url) c.Options.Extender.Enqueued(robCtx) w.pop.stack(robCtx) } } cnt++ c.logFunc(LogEnqueued, "enqueue: %s", ctx.url) c.Options.Extender.Enqueued(ctx) w.pop.stack(ctx) c.pushPopRefCount++ // Once it is stacked, it WILL be visited eventually, so add it to the visited slice // (unless denied by robots.txt, but this is out of our hands, for all we // care, it is visited). if !isVisited { // The visited map works with the normalized URL c.visited[ctx.normalizedURL.String()] = struct{}{} } } } return }
[ "func", "(", "c", "*", "Crawler", ")", "enqueueUrls", "(", "ctxs", "[", "]", "*", "URLContext", ")", "(", "cnt", "int", ")", "{", "for", "_", ",", "ctx", ":=", "range", "ctxs", "{", "var", "isVisited", ",", "enqueue", "bool", "\n\n", "// Cannot directly enqueue a robots.txt URL, since it is managed as a special case", "// in the worker (doesn't return a response to crawler).", "if", "ctx", ".", "IsRobotsURL", "(", ")", "{", "continue", "\n", "}", "\n", "// Check if it has been visited before, using the normalized URL", "_", ",", "isVisited", "=", "c", ".", "visited", "[", "ctx", ".", "normalizedURL", ".", "String", "(", ")", "]", "\n\n", "// Filter the URL", "if", "enqueue", "=", "c", ".", "Options", ".", "Extender", ".", "Filter", "(", "ctx", ",", "isVisited", ")", ";", "!", "enqueue", "{", "// Filter said NOT to use this url, so continue with next", "c", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "ctx", ".", "normalizedURL", ")", "\n", "continue", "\n", "}", "\n\n", "// Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed,", "// and comply with the same host policy if requested.", "if", "!", "ctx", ".", "normalizedURL", ".", "IsAbs", "(", ")", "{", "// Only absolute URLs are processed, so ignore", "c", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "ctx", ".", "normalizedURL", ")", "\n\n", "}", "else", "if", "!", "strings", ".", "HasPrefix", "(", "ctx", ".", "normalizedURL", ".", "Scheme", ",", "\"", "\"", ")", "{", "c", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "ctx", ".", "normalizedURL", ")", "\n\n", "}", "else", "if", "c", ".", "Options", ".", "SameHostOnly", "&&", "!", "c", ".", "isSameHost", "(", "ctx", ")", "{", "// Only allow URLs coming from the same host", "c", ".", "logFunc", "(", "LogIgnored", ",", "\"", "\"", ",", "ctx", ".", "normalizedURL", ")", "\n\n", "}", "else", "{", "// All is good, visit this URL (robots.txt verification is done by worker)", "// Possible caveat: if the normalization changes the host, it is possible", "// that the robots.txt fetched for this host would differ from the one for", "// the unnormalized host. However, this should be rare, and is a weird", "// behaviour from the host (i.e. why would site.com differ in its rules", "// from www.site.com) and can be fixed by using a different normalization", "// flag. So this is an acceptable behaviour for gocrawl.", "// Launch worker if required, based on the host of the normalized URL", "w", ",", "ok", ":=", "c", ".", "workers", "[", "ctx", ".", "normalizedURL", ".", "Host", "]", "\n", "if", "!", "ok", "{", "// No worker exists for this host, launch a new one", "w", "=", "c", ".", "launchWorker", "(", "ctx", ")", "\n", "// Automatically enqueue the robots.txt URL as first in line", "if", "robCtx", ",", "e", ":=", "ctx", ".", "getRobotsURLCtx", "(", ")", ";", "e", "!=", "nil", "{", "c", ".", "Options", ".", "Extender", ".", "Error", "(", "newCrawlError", "(", "ctx", ",", "e", ",", "CekParseRobots", ")", ")", "\n", "c", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "ctx", ".", "normalizedURL", ",", "e", ")", "\n", "}", "else", "{", "c", ".", "logFunc", "(", "LogEnqueued", ",", "\"", "\"", ",", "robCtx", ".", "url", ")", "\n", "c", ".", "Options", ".", "Extender", ".", "Enqueued", "(", "robCtx", ")", "\n", "w", ".", "pop", ".", "stack", "(", "robCtx", ")", "\n", "}", "\n", "}", "\n\n", "cnt", "++", "\n", "c", ".", "logFunc", "(", "LogEnqueued", ",", "\"", "\"", ",", "ctx", ".", "url", ")", "\n", "c", ".", "Options", ".", "Extender", ".", "Enqueued", "(", "ctx", ")", "\n", "w", ".", "pop", ".", "stack", "(", "ctx", ")", "\n", "c", ".", "pushPopRefCount", "++", "\n\n", "// Once it is stacked, it WILL be visited eventually, so add it to the visited slice", "// (unless denied by robots.txt, but this is out of our hands, for all we", "// care, it is visited).", "if", "!", "isVisited", "{", "// The visited map works with the normalized URL", "c", ".", "visited", "[", "ctx", ".", "normalizedURL", ".", "String", "(", ")", "]", "=", "struct", "{", "}", "{", "}", "\n", "}", "\n", "}", "\n", "}", "\n", "return", "\n", "}" ]
// Enqueue the URLs returned from the worker, as long as it complies with the // selection policies.
[ "Enqueue", "the", "URLs", "returned", "from", "the", "worker", "as", "long", "as", "it", "complies", "with", "the", "selection", "policies", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L199-L273
5,537
PuerkitoBio/gocrawl
crawler.go
collectUrls
func (c *Crawler) collectUrls() error { defer func() { c.logFunc(LogInfo, "waiting for goroutines to complete...") c.wg.Wait() c.logFunc(LogInfo, "crawler done.") }() for { // By checking this after each channel reception, there is a bug if the worker // wants to reenqueue following an error or a redirection. The pushPopRefCount // temporarily gets to zero before the new URL is enqueued. Check the length // of the enqueue channel to see if this is really over, or just this temporary // state. // // Check if refcount is zero - MUST be before the select statement, so that if // no valid seeds are enqueued, the crawler stops. if c.pushPopRefCount == 0 && len(c.enqueue) == 0 { c.logFunc(LogInfo, "sending STOP signals...") close(c.stop) return nil } select { case res := <-c.push: // Received a response, check if it contains URLs to enqueue if res.visited { c.visits++ if c.Options.MaxVisits > 0 && c.visits >= c.Options.MaxVisits { // Limit reached, request workers to stop c.logFunc(LogInfo, "sending STOP signals...") close(c.stop) return ErrMaxVisits } } if res.idleDeath { // The worker timed out from its Idle TTL delay, remove from active workers delete(c.workers, res.host) c.logFunc(LogInfo, "worker for host %s cleared on idle policy", res.host) } else { c.enqueueUrls(c.toURLContexts(res.harvestedURLs, res.ctx.url)) c.pushPopRefCount-- } case enq := <-c.enqueue: // Received a command to enqueue a URL, proceed ctxs := c.toURLContexts(enq, nil) c.logFunc(LogTrace, "receive url(s) to enqueue %v", toStringArrayContextURL(ctxs)) c.enqueueUrls(ctxs) case <-c.stop: return ErrInterrupted } } }
go
func (c *Crawler) collectUrls() error { defer func() { c.logFunc(LogInfo, "waiting for goroutines to complete...") c.wg.Wait() c.logFunc(LogInfo, "crawler done.") }() for { // By checking this after each channel reception, there is a bug if the worker // wants to reenqueue following an error or a redirection. The pushPopRefCount // temporarily gets to zero before the new URL is enqueued. Check the length // of the enqueue channel to see if this is really over, or just this temporary // state. // // Check if refcount is zero - MUST be before the select statement, so that if // no valid seeds are enqueued, the crawler stops. if c.pushPopRefCount == 0 && len(c.enqueue) == 0 { c.logFunc(LogInfo, "sending STOP signals...") close(c.stop) return nil } select { case res := <-c.push: // Received a response, check if it contains URLs to enqueue if res.visited { c.visits++ if c.Options.MaxVisits > 0 && c.visits >= c.Options.MaxVisits { // Limit reached, request workers to stop c.logFunc(LogInfo, "sending STOP signals...") close(c.stop) return ErrMaxVisits } } if res.idleDeath { // The worker timed out from its Idle TTL delay, remove from active workers delete(c.workers, res.host) c.logFunc(LogInfo, "worker for host %s cleared on idle policy", res.host) } else { c.enqueueUrls(c.toURLContexts(res.harvestedURLs, res.ctx.url)) c.pushPopRefCount-- } case enq := <-c.enqueue: // Received a command to enqueue a URL, proceed ctxs := c.toURLContexts(enq, nil) c.logFunc(LogTrace, "receive url(s) to enqueue %v", toStringArrayContextURL(ctxs)) c.enqueueUrls(ctxs) case <-c.stop: return ErrInterrupted } } }
[ "func", "(", "c", "*", "Crawler", ")", "collectUrls", "(", ")", "error", "{", "defer", "func", "(", ")", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "c", ".", "wg", ".", "Wait", "(", ")", "\n", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "}", "(", ")", "\n\n", "for", "{", "// By checking this after each channel reception, there is a bug if the worker", "// wants to reenqueue following an error or a redirection. The pushPopRefCount", "// temporarily gets to zero before the new URL is enqueued. Check the length", "// of the enqueue channel to see if this is really over, or just this temporary", "// state.", "//", "// Check if refcount is zero - MUST be before the select statement, so that if", "// no valid seeds are enqueued, the crawler stops.", "if", "c", ".", "pushPopRefCount", "==", "0", "&&", "len", "(", "c", ".", "enqueue", ")", "==", "0", "{", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "close", "(", "c", ".", "stop", ")", "\n", "return", "nil", "\n", "}", "\n\n", "select", "{", "case", "res", ":=", "<-", "c", ".", "push", ":", "// Received a response, check if it contains URLs to enqueue", "if", "res", ".", "visited", "{", "c", ".", "visits", "++", "\n", "if", "c", ".", "Options", ".", "MaxVisits", ">", "0", "&&", "c", ".", "visits", ">=", "c", ".", "Options", ".", "MaxVisits", "{", "// Limit reached, request workers to stop", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ")", "\n", "close", "(", "c", ".", "stop", ")", "\n", "return", "ErrMaxVisits", "\n", "}", "\n", "}", "\n", "if", "res", ".", "idleDeath", "{", "// The worker timed out from its Idle TTL delay, remove from active workers", "delete", "(", "c", ".", "workers", ",", "res", ".", "host", ")", "\n", "c", ".", "logFunc", "(", "LogInfo", ",", "\"", "\"", ",", "res", ".", "host", ")", "\n", "}", "else", "{", "c", ".", "enqueueUrls", "(", "c", ".", "toURLContexts", "(", "res", ".", "harvestedURLs", ",", "res", ".", "ctx", ".", "url", ")", ")", "\n", "c", ".", "pushPopRefCount", "--", "\n", "}", "\n\n", "case", "enq", ":=", "<-", "c", ".", "enqueue", ":", "// Received a command to enqueue a URL, proceed", "ctxs", ":=", "c", ".", "toURLContexts", "(", "enq", ",", "nil", ")", "\n", "c", ".", "logFunc", "(", "LogTrace", ",", "\"", "\"", ",", "toStringArrayContextURL", "(", "ctxs", ")", ")", "\n", "c", ".", "enqueueUrls", "(", "ctxs", ")", "\n", "case", "<-", "c", ".", "stop", ":", "return", "ErrInterrupted", "\n", "}", "\n", "}", "\n", "}" ]
// This is the main loop of the crawler, waiting for responses from the workers // and processing these responses.
[ "This", "is", "the", "main", "loop", "of", "the", "crawler", "waiting", "for", "responses", "from", "the", "workers", "and", "processing", "these", "responses", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L277-L329
5,538
PuerkitoBio/gocrawl
crawler.go
Stop
func (c *Crawler) Stop() { defer func() { if err := recover(); err != nil { c.logFunc(LogError, "error when manually stopping crawler: %s", err) } }() // this channel may be closed already close(c.stop) }
go
func (c *Crawler) Stop() { defer func() { if err := recover(); err != nil { c.logFunc(LogError, "error when manually stopping crawler: %s", err) } }() // this channel may be closed already close(c.stop) }
[ "func", "(", "c", "*", "Crawler", ")", "Stop", "(", ")", "{", "defer", "func", "(", ")", "{", "if", "err", ":=", "recover", "(", ")", ";", "err", "!=", "nil", "{", "c", ".", "logFunc", "(", "LogError", ",", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "}", "(", ")", "\n\n", "// this channel may be closed already", "close", "(", "c", ".", "stop", ")", "\n", "}" ]
// Stop terminates the crawler.
[ "Stop", "terminates", "the", "crawler", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L332-L341
5,539
PuerkitoBio/gocrawl
errors.go
Error
func (ce CrawlError) Error() string { if ce.Err != nil { return ce.Err.Error() } return ce.msg }
go
func (ce CrawlError) Error() string { if ce.Err != nil { return ce.Err.Error() } return ce.msg }
[ "func", "(", "ce", "CrawlError", ")", "Error", "(", ")", "string", "{", "if", "ce", ".", "Err", "!=", "nil", "{", "return", "ce", ".", "Err", ".", "Error", "(", ")", "\n", "}", "\n", "return", "ce", ".", "msg", "\n", "}" ]
// Error implements of the error interface for CrawlError.
[ "Error", "implements", "of", "the", "error", "interface", "for", "CrawlError", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L70-L75
5,540
PuerkitoBio/gocrawl
errors.go
newCrawlError
func newCrawlError(ctx *URLContext, e error, kind CrawlErrorKind) *CrawlError { return &CrawlError{ctx, e, kind, ""} }
go
func newCrawlError(ctx *URLContext, e error, kind CrawlErrorKind) *CrawlError { return &CrawlError{ctx, e, kind, ""} }
[ "func", "newCrawlError", "(", "ctx", "*", "URLContext", ",", "e", "error", ",", "kind", "CrawlErrorKind", ")", "*", "CrawlError", "{", "return", "&", "CrawlError", "{", "ctx", ",", "e", ",", "kind", ",", "\"", "\"", "}", "\n", "}" ]
// Create a new CrawlError based on a source error.
[ "Create", "a", "new", "CrawlError", "based", "on", "a", "source", "error", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L78-L80
5,541
PuerkitoBio/gocrawl
errors.go
newCrawlErrorMessage
func newCrawlErrorMessage(ctx *URLContext, msg string, kind CrawlErrorKind) *CrawlError { return &CrawlError{ctx, nil, kind, msg} }
go
func newCrawlErrorMessage(ctx *URLContext, msg string, kind CrawlErrorKind) *CrawlError { return &CrawlError{ctx, nil, kind, msg} }
[ "func", "newCrawlErrorMessage", "(", "ctx", "*", "URLContext", ",", "msg", "string", ",", "kind", "CrawlErrorKind", ")", "*", "CrawlError", "{", "return", "&", "CrawlError", "{", "ctx", ",", "nil", ",", "kind", ",", "msg", "}", "\n", "}" ]
// Create a new CrawlError with the specified message.
[ "Create", "a", "new", "CrawlError", "with", "the", "specified", "message", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L83-L85
5,542
PuerkitoBio/gocrawl
urlcontext.go
cloneForRedirect
func (uc *URLContext) cloneForRedirect(dst *url.URL, normFlags purell.NormalizationFlags) *URLContext { var src, normalizedSrc *url.URL if uc.sourceURL != nil { src = &url.URL{} *src = *uc.sourceURL } if src == nil && uc.url != nil { // if the current context doesn't have a source URL, use its URL as // source (e.g. for a seed URL that triggers a redirect) src = &url.URL{} *src = *uc.url } if uc.normalizedSourceURL != nil { normalizedSrc = &url.URL{} *normalizedSrc = *uc.normalizedSourceURL } if normalizedSrc == nil { normalizedSrc = &url.URL{} *normalizedSrc = *uc.normalizedURL } rawDst := &url.URL{} *rawDst = *dst purell.NormalizeURL(dst, normFlags) return &URLContext{ HeadBeforeGet: uc.HeadBeforeGet, State: uc.State, url: rawDst, normalizedURL: dst, sourceURL: src, normalizedSourceURL: normalizedSrc, } }
go
func (uc *URLContext) cloneForRedirect(dst *url.URL, normFlags purell.NormalizationFlags) *URLContext { var src, normalizedSrc *url.URL if uc.sourceURL != nil { src = &url.URL{} *src = *uc.sourceURL } if src == nil && uc.url != nil { // if the current context doesn't have a source URL, use its URL as // source (e.g. for a seed URL that triggers a redirect) src = &url.URL{} *src = *uc.url } if uc.normalizedSourceURL != nil { normalizedSrc = &url.URL{} *normalizedSrc = *uc.normalizedSourceURL } if normalizedSrc == nil { normalizedSrc = &url.URL{} *normalizedSrc = *uc.normalizedURL } rawDst := &url.URL{} *rawDst = *dst purell.NormalizeURL(dst, normFlags) return &URLContext{ HeadBeforeGet: uc.HeadBeforeGet, State: uc.State, url: rawDst, normalizedURL: dst, sourceURL: src, normalizedSourceURL: normalizedSrc, } }
[ "func", "(", "uc", "*", "URLContext", ")", "cloneForRedirect", "(", "dst", "*", "url", ".", "URL", ",", "normFlags", "purell", ".", "NormalizationFlags", ")", "*", "URLContext", "{", "var", "src", ",", "normalizedSrc", "*", "url", ".", "URL", "\n", "if", "uc", ".", "sourceURL", "!=", "nil", "{", "src", "=", "&", "url", ".", "URL", "{", "}", "\n", "*", "src", "=", "*", "uc", ".", "sourceURL", "\n", "}", "\n", "if", "src", "==", "nil", "&&", "uc", ".", "url", "!=", "nil", "{", "// if the current context doesn't have a source URL, use its URL as", "// source (e.g. for a seed URL that triggers a redirect)", "src", "=", "&", "url", ".", "URL", "{", "}", "\n", "*", "src", "=", "*", "uc", ".", "url", "\n", "}", "\n\n", "if", "uc", ".", "normalizedSourceURL", "!=", "nil", "{", "normalizedSrc", "=", "&", "url", ".", "URL", "{", "}", "\n", "*", "normalizedSrc", "=", "*", "uc", ".", "normalizedSourceURL", "\n", "}", "\n", "if", "normalizedSrc", "==", "nil", "{", "normalizedSrc", "=", "&", "url", ".", "URL", "{", "}", "\n", "*", "normalizedSrc", "=", "*", "uc", ".", "normalizedURL", "\n", "}", "\n\n", "rawDst", ":=", "&", "url", ".", "URL", "{", "}", "\n", "*", "rawDst", "=", "*", "dst", "\n", "purell", ".", "NormalizeURL", "(", "dst", ",", "normFlags", ")", "\n", "return", "&", "URLContext", "{", "HeadBeforeGet", ":", "uc", ".", "HeadBeforeGet", ",", "State", ":", "uc", ".", "State", ",", "url", ":", "rawDst", ",", "normalizedURL", ":", "dst", ",", "sourceURL", ":", "src", ",", "normalizedSourceURL", ":", "normalizedSrc", ",", "}", "\n", "}" ]
// cloneForRedirect returns a new URLContext with the given // destination URL with the same sourceURL and normalizedSourceURL.
[ "cloneForRedirect", "returns", "a", "new", "URLContext", "with", "the", "given", "destination", "URL", "with", "the", "same", "sourceURL", "and", "normalizedSourceURL", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/urlcontext.go#L65-L98
5,543
PuerkitoBio/gocrawl
options.go
NewOptions
func NewOptions(ext Extender) *Options { // Use defaults except for Extender return &Options{ DefaultUserAgent, DefaultRobotUserAgent, 0, DefaultEnqueueChanBuffer, DefaultHostBufferFactor, DefaultCrawlDelay, DefaultIdleTTL, true, false, DefaultNormalizationFlags, LogError, ext, } }
go
func NewOptions(ext Extender) *Options { // Use defaults except for Extender return &Options{ DefaultUserAgent, DefaultRobotUserAgent, 0, DefaultEnqueueChanBuffer, DefaultHostBufferFactor, DefaultCrawlDelay, DefaultIdleTTL, true, false, DefaultNormalizationFlags, LogError, ext, } }
[ "func", "NewOptions", "(", "ext", "Extender", ")", "*", "Options", "{", "// Use defaults except for Extender", "return", "&", "Options", "{", "DefaultUserAgent", ",", "DefaultRobotUserAgent", ",", "0", ",", "DefaultEnqueueChanBuffer", ",", "DefaultHostBufferFactor", ",", "DefaultCrawlDelay", ",", "DefaultIdleTTL", ",", "true", ",", "false", ",", "DefaultNormalizationFlags", ",", "LogError", ",", "ext", ",", "}", "\n", "}" ]
// NewOptions creates a new set of Options with default values // using the provided Extender. The RobotUserAgent option should // be set to the name of your crawler, it is used to find the matching // entry in the robots.txt file.
[ "NewOptions", "creates", "a", "new", "set", "of", "Options", "with", "default", "values", "using", "the", "provided", "Extender", ".", "The", "RobotUserAgent", "option", "should", "be", "set", "to", "the", "name", "of", "your", "crawler", "it", "is", "used", "to", "find", "the", "matching", "entry", "in", "the", "robots", ".", "txt", "file", "." ]
0aa0507f0cf9eee3726aeabf9848501a5ca4d960
https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/options.go#L84-L100
5,544
iancoleman/strcase
camel.go
toCamelInitCase
func toCamelInitCase(s string, initCase bool) string { s = addWordBoundariesToNumbers(s) s = strings.Trim(s, " ") n := "" capNext := initCase for _, v := range s { if v >= 'A' && v <= 'Z' { n += string(v) } if v >= '0' && v <= '9' { n += string(v) } if v >= 'a' && v <= 'z' { if capNext { n += strings.ToUpper(string(v)) } else { n += string(v) } } if v == '_' || v == ' ' || v == '-' { capNext = true } else { capNext = false } } return n }
go
func toCamelInitCase(s string, initCase bool) string { s = addWordBoundariesToNumbers(s) s = strings.Trim(s, " ") n := "" capNext := initCase for _, v := range s { if v >= 'A' && v <= 'Z' { n += string(v) } if v >= '0' && v <= '9' { n += string(v) } if v >= 'a' && v <= 'z' { if capNext { n += strings.ToUpper(string(v)) } else { n += string(v) } } if v == '_' || v == ' ' || v == '-' { capNext = true } else { capNext = false } } return n }
[ "func", "toCamelInitCase", "(", "s", "string", ",", "initCase", "bool", ")", "string", "{", "s", "=", "addWordBoundariesToNumbers", "(", "s", ")", "\n", "s", "=", "strings", ".", "Trim", "(", "s", ",", "\"", "\"", ")", "\n", "n", ":=", "\"", "\"", "\n", "capNext", ":=", "initCase", "\n", "for", "_", ",", "v", ":=", "range", "s", "{", "if", "v", ">=", "'A'", "&&", "v", "<=", "'Z'", "{", "n", "+=", "string", "(", "v", ")", "\n", "}", "\n", "if", "v", ">=", "'0'", "&&", "v", "<=", "'9'", "{", "n", "+=", "string", "(", "v", ")", "\n", "}", "\n", "if", "v", ">=", "'a'", "&&", "v", "<=", "'z'", "{", "if", "capNext", "{", "n", "+=", "strings", ".", "ToUpper", "(", "string", "(", "v", ")", ")", "\n", "}", "else", "{", "n", "+=", "string", "(", "v", ")", "\n", "}", "\n", "}", "\n", "if", "v", "==", "'_'", "||", "v", "==", "' '", "||", "v", "==", "'-'", "{", "capNext", "=", "true", "\n", "}", "else", "{", "capNext", "=", "false", "\n", "}", "\n", "}", "\n", "return", "n", "\n", "}" ]
// Converts a string to CamelCase
[ "Converts", "a", "string", "to", "CamelCase" ]
e506e3ef73653e84c592ba44aab577a46678f68c
https://github.com/iancoleman/strcase/blob/e506e3ef73653e84c592ba44aab577a46678f68c/camel.go#L33-L59
5,545
iancoleman/strcase
camel.go
ToLowerCamel
func ToLowerCamel(s string) string { if s == "" { return s } if r := rune(s[0]); r >= 'A' && r <= 'Z' { s = strings.ToLower(string(r)) + s[1:] } return toCamelInitCase(s, false) }
go
func ToLowerCamel(s string) string { if s == "" { return s } if r := rune(s[0]); r >= 'A' && r <= 'Z' { s = strings.ToLower(string(r)) + s[1:] } return toCamelInitCase(s, false) }
[ "func", "ToLowerCamel", "(", "s", "string", ")", "string", "{", "if", "s", "==", "\"", "\"", "{", "return", "s", "\n", "}", "\n", "if", "r", ":=", "rune", "(", "s", "[", "0", "]", ")", ";", "r", ">=", "'A'", "&&", "r", "<=", "'Z'", "{", "s", "=", "strings", ".", "ToLower", "(", "string", "(", "r", ")", ")", "+", "s", "[", "1", ":", "]", "\n", "}", "\n", "return", "toCamelInitCase", "(", "s", ",", "false", ")", "\n", "}" ]
// ToLowerCamel converts a string to lowerCamelCase
[ "ToLowerCamel", "converts", "a", "string", "to", "lowerCamelCase" ]
e506e3ef73653e84c592ba44aab577a46678f68c
https://github.com/iancoleman/strcase/blob/e506e3ef73653e84c592ba44aab577a46678f68c/camel.go#L67-L75
5,546
codeskyblue/go-sh
sh.go
Call
func (s *Session) Call(name string, a ...interface{}) error { return s.Command(name, a...).Run() }
go
func (s *Session) Call(name string, a ...interface{}) error { return s.Command(name, a...).Run() }
[ "func", "(", "s", "*", "Session", ")", "Call", "(", "name", "string", ",", "a", "...", "interface", "{", "}", ")", "error", "{", "return", "s", ".", "Command", "(", "name", ",", "a", "...", ")", ".", "Run", "(", ")", "\n", "}" ]
// combine Command and Run
[ "combine", "Command", "and", "Run" ]
76bd3d59ff27c0518096e0af6c165bf626b809ec
https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/sh.go#L130-L132
5,547
codeskyblue/go-sh
pipe.go
UnmarshalJSON
func (s *Session) UnmarshalJSON(data interface{}) (err error) { bufrw := bytes.NewBuffer(nil) s.Stdout = bufrw if err = s.Run(); err != nil { return } return json.NewDecoder(bufrw).Decode(data) }
go
func (s *Session) UnmarshalJSON(data interface{}) (err error) { bufrw := bytes.NewBuffer(nil) s.Stdout = bufrw if err = s.Run(); err != nil { return } return json.NewDecoder(bufrw).Decode(data) }
[ "func", "(", "s", "*", "Session", ")", "UnmarshalJSON", "(", "data", "interface", "{", "}", ")", "(", "err", "error", ")", "{", "bufrw", ":=", "bytes", ".", "NewBuffer", "(", "nil", ")", "\n", "s", ".", "Stdout", "=", "bufrw", "\n", "if", "err", "=", "s", ".", "Run", "(", ")", ";", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "return", "json", ".", "NewDecoder", "(", "bufrw", ")", ".", "Decode", "(", "data", ")", "\n", "}" ]
// unmarshal shell output to decode json
[ "unmarshal", "shell", "output", "to", "decode", "json" ]
76bd3d59ff27c0518096e0af6c165bf626b809ec
https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/pipe.go#L18-L25
5,548
codeskyblue/go-sh
pipe.go
UnmarshalXML
func (s *Session) UnmarshalXML(data interface{}) (err error) { bufrw := bytes.NewBuffer(nil) s.Stdout = bufrw if err = s.Run(); err != nil { return } return xml.NewDecoder(bufrw).Decode(data) }
go
func (s *Session) UnmarshalXML(data interface{}) (err error) { bufrw := bytes.NewBuffer(nil) s.Stdout = bufrw if err = s.Run(); err != nil { return } return xml.NewDecoder(bufrw).Decode(data) }
[ "func", "(", "s", "*", "Session", ")", "UnmarshalXML", "(", "data", "interface", "{", "}", ")", "(", "err", "error", ")", "{", "bufrw", ":=", "bytes", ".", "NewBuffer", "(", "nil", ")", "\n", "s", ".", "Stdout", "=", "bufrw", "\n", "if", "err", "=", "s", ".", "Run", "(", ")", ";", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "return", "xml", ".", "NewDecoder", "(", "bufrw", ")", ".", "Decode", "(", "data", ")", "\n", "}" ]
// unmarshal command output into xml
[ "unmarshal", "command", "output", "into", "xml" ]
76bd3d59ff27c0518096e0af6c165bf626b809ec
https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/pipe.go#L28-L35
5,549
go-audio/wav
encoder.go
NewEncoder
func NewEncoder(w io.WriteSeeker, sampleRate, bitDepth, numChans, audioFormat int) *Encoder { return &Encoder{ w: w, SampleRate: sampleRate, BitDepth: bitDepth, NumChans: numChans, WavAudioFormat: audioFormat, } }
go
func NewEncoder(w io.WriteSeeker, sampleRate, bitDepth, numChans, audioFormat int) *Encoder { return &Encoder{ w: w, SampleRate: sampleRate, BitDepth: bitDepth, NumChans: numChans, WavAudioFormat: audioFormat, } }
[ "func", "NewEncoder", "(", "w", "io", ".", "WriteSeeker", ",", "sampleRate", ",", "bitDepth", ",", "numChans", ",", "audioFormat", "int", ")", "*", "Encoder", "{", "return", "&", "Encoder", "{", "w", ":", "w", ",", "SampleRate", ":", "sampleRate", ",", "BitDepth", ":", "bitDepth", ",", "NumChans", ":", "numChans", ",", "WavAudioFormat", ":", "audioFormat", ",", "}", "\n", "}" ]
// NewEncoder creates a new encoder to create a new wav file. // Don't forget to add Frames to the encoder before writing.
[ "NewEncoder", "creates", "a", "new", "encoder", "to", "create", "a", "new", "wav", "file", ".", "Don", "t", "forget", "to", "add", "Frames", "to", "the", "encoder", "before", "writing", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L41-L49
5,550
go-audio/wav
encoder.go
AddLE
func (e *Encoder) AddLE(src interface{}) error { e.WrittenBytes += binary.Size(src) return binary.Write(e.w, binary.LittleEndian, src) }
go
func (e *Encoder) AddLE(src interface{}) error { e.WrittenBytes += binary.Size(src) return binary.Write(e.w, binary.LittleEndian, src) }
[ "func", "(", "e", "*", "Encoder", ")", "AddLE", "(", "src", "interface", "{", "}", ")", "error", "{", "e", ".", "WrittenBytes", "+=", "binary", ".", "Size", "(", "src", ")", "\n", "return", "binary", ".", "Write", "(", "e", ".", "w", ",", "binary", ".", "LittleEndian", ",", "src", ")", "\n", "}" ]
// AddLE serializes and adds the passed value using little endian
[ "AddLE", "serializes", "and", "adds", "the", "passed", "value", "using", "little", "endian" ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L52-L55
5,551
go-audio/wav
encoder.go
AddBE
func (e *Encoder) AddBE(src interface{}) error { e.WrittenBytes += binary.Size(src) return binary.Write(e.w, binary.BigEndian, src) }
go
func (e *Encoder) AddBE(src interface{}) error { e.WrittenBytes += binary.Size(src) return binary.Write(e.w, binary.BigEndian, src) }
[ "func", "(", "e", "*", "Encoder", ")", "AddBE", "(", "src", "interface", "{", "}", ")", "error", "{", "e", ".", "WrittenBytes", "+=", "binary", ".", "Size", "(", "src", ")", "\n", "return", "binary", ".", "Write", "(", "e", ".", "w", ",", "binary", ".", "BigEndian", ",", "src", ")", "\n", "}" ]
// AddBE serializes and adds the passed value using big endian
[ "AddBE", "serializes", "and", "adds", "the", "passed", "value", "using", "big", "endian" ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L58-L61
5,552
go-audio/wav
encoder.go
WriteFrame
func (e *Encoder) WriteFrame(value interface{}) error { if !e.wroteHeader { e.writeHeader() } if !e.pcmChunkStarted { // sound header if err := e.AddLE(riff.DataFormatID); err != nil { return fmt.Errorf("error encoding sound header %v", err) } e.pcmChunkStarted = true // write a temporary chunksize e.pcmChunkSizePos = e.WrittenBytes if err := e.AddLE(uint32(42)); err != nil { return fmt.Errorf("%v when writing wav data chunk size header", err) } } e.frames++ return e.AddLE(value) }
go
func (e *Encoder) WriteFrame(value interface{}) error { if !e.wroteHeader { e.writeHeader() } if !e.pcmChunkStarted { // sound header if err := e.AddLE(riff.DataFormatID); err != nil { return fmt.Errorf("error encoding sound header %v", err) } e.pcmChunkStarted = true // write a temporary chunksize e.pcmChunkSizePos = e.WrittenBytes if err := e.AddLE(uint32(42)); err != nil { return fmt.Errorf("%v when writing wav data chunk size header", err) } } e.frames++ return e.AddLE(value) }
[ "func", "(", "e", "*", "Encoder", ")", "WriteFrame", "(", "value", "interface", "{", "}", ")", "error", "{", "if", "!", "e", ".", "wroteHeader", "{", "e", ".", "writeHeader", "(", ")", "\n", "}", "\n", "if", "!", "e", ".", "pcmChunkStarted", "{", "// sound header", "if", "err", ":=", "e", ".", "AddLE", "(", "riff", ".", "DataFormatID", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "e", ".", "pcmChunkStarted", "=", "true", "\n\n", "// write a temporary chunksize", "e", ".", "pcmChunkSizePos", "=", "e", ".", "WrittenBytes", "\n", "if", "err", ":=", "e", ".", "AddLE", "(", "uint32", "(", "42", ")", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "}", "\n\n", "e", ".", "frames", "++", "\n", "return", "e", ".", "AddLE", "(", "value", ")", "\n", "}" ]
// WriteFrame writes a single frame of data to the underlying writer.
[ "WriteFrame", "writes", "a", "single", "frame", "of", "data", "to", "the", "underlying", "writer", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L199-L219
5,553
go-audio/wav
cue_chunk.go
DecodeCueChunk
func DecodeCueChunk(d *Decoder, ch *riff.Chunk) error { if ch == nil { return fmt.Errorf("can't decode a nil chunk") } if d == nil { return fmt.Errorf("nil decoder") } if ch.ID == CIDCue { // read the entire chunk in memory buf := make([]byte, ch.Size) var err error if _, err = ch.Read(buf); err != nil { return fmt.Errorf("failed to read the CUE chunk - %v", err) } r := bytes.NewReader(buf) var nbrCues uint32 if err := binary.Read(r, binary.LittleEndian, &nbrCues); err != nil { return fmt.Errorf("failed to read the number of cues - %v", err) } if nbrCues > 0 { if d.Metadata == nil { d.Metadata = &Metadata{} } d.Metadata.CuePoints = []*CuePoint{} scratch := make([]byte, 4) for i := uint32(0); i < nbrCues; i++ { c := &CuePoint{} if _, err = r.Read(scratch); err != nil { return fmt.Errorf("failed to read the cue point ID") } copy(c.ID[:], scratch[:4]) if err := binary.Read(r, binary.LittleEndian, &c.Position); err != nil { return err } if _, err = r.Read(scratch); err != nil { return fmt.Errorf("failed to read the data chunk id") } copy(c.DataChunkID[:], scratch[:4]) if err := binary.Read(r, binary.LittleEndian, &c.ChunkStart); err != nil { return err } if err := binary.Read(r, binary.LittleEndian, &c.BlockStart); err != nil { return err } if err := binary.Read(r, binary.LittleEndian, &c.SampleOffset); err != nil { return err } d.Metadata.CuePoints = append(d.Metadata.CuePoints, c) } } } return nil }
go
func DecodeCueChunk(d *Decoder, ch *riff.Chunk) error { if ch == nil { return fmt.Errorf("can't decode a nil chunk") } if d == nil { return fmt.Errorf("nil decoder") } if ch.ID == CIDCue { // read the entire chunk in memory buf := make([]byte, ch.Size) var err error if _, err = ch.Read(buf); err != nil { return fmt.Errorf("failed to read the CUE chunk - %v", err) } r := bytes.NewReader(buf) var nbrCues uint32 if err := binary.Read(r, binary.LittleEndian, &nbrCues); err != nil { return fmt.Errorf("failed to read the number of cues - %v", err) } if nbrCues > 0 { if d.Metadata == nil { d.Metadata = &Metadata{} } d.Metadata.CuePoints = []*CuePoint{} scratch := make([]byte, 4) for i := uint32(0); i < nbrCues; i++ { c := &CuePoint{} if _, err = r.Read(scratch); err != nil { return fmt.Errorf("failed to read the cue point ID") } copy(c.ID[:], scratch[:4]) if err := binary.Read(r, binary.LittleEndian, &c.Position); err != nil { return err } if _, err = r.Read(scratch); err != nil { return fmt.Errorf("failed to read the data chunk id") } copy(c.DataChunkID[:], scratch[:4]) if err := binary.Read(r, binary.LittleEndian, &c.ChunkStart); err != nil { return err } if err := binary.Read(r, binary.LittleEndian, &c.BlockStart); err != nil { return err } if err := binary.Read(r, binary.LittleEndian, &c.SampleOffset); err != nil { return err } d.Metadata.CuePoints = append(d.Metadata.CuePoints, c) } } } return nil }
[ "func", "DecodeCueChunk", "(", "d", "*", "Decoder", ",", "ch", "*", "riff", ".", "Chunk", ")", "error", "{", "if", "ch", "==", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n", "if", "d", "==", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n", "if", "ch", ".", "ID", "==", "CIDCue", "{", "// read the entire chunk in memory", "buf", ":=", "make", "(", "[", "]", "byte", ",", "ch", ".", "Size", ")", "\n", "var", "err", "error", "\n", "if", "_", ",", "err", "=", "ch", ".", "Read", "(", "buf", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "r", ":=", "bytes", ".", "NewReader", "(", "buf", ")", "\n", "var", "nbrCues", "uint32", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "r", ",", "binary", ".", "LittleEndian", ",", "&", "nbrCues", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "err", ")", "\n", "}", "\n", "if", "nbrCues", ">", "0", "{", "if", "d", ".", "Metadata", "==", "nil", "{", "d", ".", "Metadata", "=", "&", "Metadata", "{", "}", "\n", "}", "\n", "d", ".", "Metadata", ".", "CuePoints", "=", "[", "]", "*", "CuePoint", "{", "}", "\n", "scratch", ":=", "make", "(", "[", "]", "byte", ",", "4", ")", "\n", "for", "i", ":=", "uint32", "(", "0", ")", ";", "i", "<", "nbrCues", ";", "i", "++", "{", "c", ":=", "&", "CuePoint", "{", "}", "\n", "if", "_", ",", "err", "=", "r", ".", "Read", "(", "scratch", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n", "copy", "(", "c", ".", "ID", "[", ":", "]", ",", "scratch", "[", ":", "4", "]", ")", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "r", ",", "binary", ".", "LittleEndian", ",", "&", "c", ".", "Position", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "_", ",", "err", "=", "r", ".", "Read", "(", "scratch", ")", ";", "err", "!=", "nil", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ")", "\n", "}", "\n", "copy", "(", "c", ".", "DataChunkID", "[", ":", "]", ",", "scratch", "[", ":", "4", "]", ")", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "r", ",", "binary", ".", "LittleEndian", ",", "&", "c", ".", "ChunkStart", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "r", ",", "binary", ".", "LittleEndian", ",", "&", "c", ".", "BlockStart", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "r", ",", "binary", ".", "LittleEndian", ",", "&", "c", ".", "SampleOffset", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "d", ".", "Metadata", ".", "CuePoints", "=", "append", "(", "d", ".", "Metadata", ".", "CuePoints", ",", "c", ")", "\n", "}", "\n", "}", "\n\n", "}", "\n", "return", "nil", "\n", "}" ]
// DecodeCueChunk decodes the optional cue chunk and extracts cue points.
[ "DecodeCueChunk", "decodes", "the", "optional", "cue", "chunk", "and", "extracts", "cue", "points", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/cue_chunk.go#L53-L106
5,554
go-audio/wav
decoder.go
NewDecoder
func NewDecoder(r io.ReadSeeker) *Decoder { return &Decoder{ r: r, parser: riff.New(r), } }
go
func NewDecoder(r io.ReadSeeker) *Decoder { return &Decoder{ r: r, parser: riff.New(r), } }
[ "func", "NewDecoder", "(", "r", "io", ".", "ReadSeeker", ")", "*", "Decoder", "{", "return", "&", "Decoder", "{", "r", ":", "r", ",", "parser", ":", "riff", ".", "New", "(", "r", ")", ",", "}", "\n", "}" ]
// NewDecoder creates a decoder for the passed wav reader. // Note that the reader doesn't get rewinded as the container is processed.
[ "NewDecoder", "creates", "a", "decoder", "for", "the", "passed", "wav", "reader", ".", "Note", "that", "the", "reader", "doesn", "t", "get", "rewinded", "as", "the", "container", "is", "processed", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L50-L55
5,555
go-audio/wav
decoder.go
Seek
func (d *Decoder) Seek(offset int64, whence int) (int64, error) { return d.r.Seek(offset, whence) }
go
func (d *Decoder) Seek(offset int64, whence int) (int64, error) { return d.r.Seek(offset, whence) }
[ "func", "(", "d", "*", "Decoder", ")", "Seek", "(", "offset", "int64", ",", "whence", "int", ")", "(", "int64", ",", "error", ")", "{", "return", "d", ".", "r", ".", "Seek", "(", "offset", ",", "whence", ")", "\n", "}" ]
// Seek provides access to the cursor position in the PCM data
[ "Seek", "provides", "access", "to", "the", "cursor", "position", "in", "the", "PCM", "data" ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L58-L60
5,556
go-audio/wav
decoder.go
Err
func (d *Decoder) Err() error { if d.err == io.EOF { return nil } return d.err }
go
func (d *Decoder) Err() error { if d.err == io.EOF { return nil } return d.err }
[ "func", "(", "d", "*", "Decoder", ")", "Err", "(", ")", "error", "{", "if", "d", ".", "err", "==", "io", ".", "EOF", "{", "return", "nil", "\n", "}", "\n", "return", "d", ".", "err", "\n", "}" ]
// Err returns the first non-EOF error that was encountered by the Decoder.
[ "Err", "returns", "the", "first", "non", "-", "EOF", "error", "that", "was", "encountered", "by", "the", "Decoder", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L79-L84
5,557
go-audio/wav
decoder.go
EOF
func (d *Decoder) EOF() bool { if d == nil || d.err == io.EOF { return true } return false }
go
func (d *Decoder) EOF() bool { if d == nil || d.err == io.EOF { return true } return false }
[ "func", "(", "d", "*", "Decoder", ")", "EOF", "(", ")", "bool", "{", "if", "d", "==", "nil", "||", "d", ".", "err", "==", "io", ".", "EOF", "{", "return", "true", "\n", "}", "\n", "return", "false", "\n", "}" ]
// EOF returns positively if the underlying reader reached the end of file.
[ "EOF", "returns", "positively", "if", "the", "underlying", "reader", "reached", "the", "end", "of", "file", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L87-L92
5,558
go-audio/wav
decoder.go
ReadMetadata
func (d *Decoder) ReadMetadata() { if d.Metadata != nil { return } d.ReadInfo() if d.Err() != nil || d.Metadata != nil { return } var ( chunk *riff.Chunk err error ) for err == nil { chunk, err = d.parser.NextChunk() if err != nil { break } switch chunk.ID { case CIDList: if err = DecodeListChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } if d.Metadata != nil && d.Metadata.SamplerInfo != nil { // we got everything we were looking for break } case CIDSmpl: if err = DecodeSamplerChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } case CIDCue: if err = DecodeCueChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } default: // fmt.Println(string(chunk.ID[:])) chunk.Drain() } } }
go
func (d *Decoder) ReadMetadata() { if d.Metadata != nil { return } d.ReadInfo() if d.Err() != nil || d.Metadata != nil { return } var ( chunk *riff.Chunk err error ) for err == nil { chunk, err = d.parser.NextChunk() if err != nil { break } switch chunk.ID { case CIDList: if err = DecodeListChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } if d.Metadata != nil && d.Metadata.SamplerInfo != nil { // we got everything we were looking for break } case CIDSmpl: if err = DecodeSamplerChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } case CIDCue: if err = DecodeCueChunk(d, chunk); err != nil { if err != io.EOF { d.err = err } } default: // fmt.Println(string(chunk.ID[:])) chunk.Drain() } } }
[ "func", "(", "d", "*", "Decoder", ")", "ReadMetadata", "(", ")", "{", "if", "d", ".", "Metadata", "!=", "nil", "{", "return", "\n", "}", "\n", "d", ".", "ReadInfo", "(", ")", "\n", "if", "d", ".", "Err", "(", ")", "!=", "nil", "||", "d", ".", "Metadata", "!=", "nil", "{", "return", "\n", "}", "\n", "var", "(", "chunk", "*", "riff", ".", "Chunk", "\n", "err", "error", "\n", ")", "\n", "for", "err", "==", "nil", "{", "chunk", ",", "err", "=", "d", ".", "parser", ".", "NextChunk", "(", ")", "\n", "if", "err", "!=", "nil", "{", "break", "\n", "}", "\n\n", "switch", "chunk", ".", "ID", "{", "case", "CIDList", ":", "if", "err", "=", "DecodeListChunk", "(", "d", ",", "chunk", ")", ";", "err", "!=", "nil", "{", "if", "err", "!=", "io", ".", "EOF", "{", "d", ".", "err", "=", "err", "\n", "}", "\n", "}", "\n", "if", "d", ".", "Metadata", "!=", "nil", "&&", "d", ".", "Metadata", ".", "SamplerInfo", "!=", "nil", "{", "// we got everything we were looking for", "break", "\n", "}", "\n", "case", "CIDSmpl", ":", "if", "err", "=", "DecodeSamplerChunk", "(", "d", ",", "chunk", ")", ";", "err", "!=", "nil", "{", "if", "err", "!=", "io", ".", "EOF", "{", "d", ".", "err", "=", "err", "\n", "}", "\n", "}", "\n", "case", "CIDCue", ":", "if", "err", "=", "DecodeCueChunk", "(", "d", ",", "chunk", ")", ";", "err", "!=", "nil", "{", "if", "err", "!=", "io", ".", "EOF", "{", "d", ".", "err", "=", "err", "\n", "}", "\n", "}", "\n", "default", ":", "// fmt.Println(string(chunk.ID[:]))", "chunk", ".", "Drain", "(", ")", "\n", "}", "\n", "}", "\n\n", "}" ]
// ReadMetadata parses the file for extra metadata such as the INFO list chunk. // The entire file will be read and should be rewinded if more data must be // accessed.
[ "ReadMetadata", "parses", "the", "file", "for", "extra", "metadata", "such", "as", "the", "INFO", "list", "chunk", ".", "The", "entire", "file", "will", "be", "read", "and", "should", "be", "rewinded", "if", "more", "data", "must", "be", "accessed", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L122-L169
5,559
go-audio/wav
decoder.go
Duration
func (d *Decoder) Duration() (time.Duration, error) { if d == nil || d.parser == nil { return 0, errors.New("can't calculate the duration of a nil pointer") } return d.parser.Duration() }
go
func (d *Decoder) Duration() (time.Duration, error) { if d == nil || d.parser == nil { return 0, errors.New("can't calculate the duration of a nil pointer") } return d.parser.Duration() }
[ "func", "(", "d", "*", "Decoder", ")", "Duration", "(", ")", "(", "time", ".", "Duration", ",", "error", ")", "{", "if", "d", "==", "nil", "||", "d", ".", "parser", "==", "nil", "{", "return", "0", ",", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n", "return", "d", ".", "parser", ".", "Duration", "(", ")", "\n", "}" ]
// Duration returns the time duration for the current audio container
[ "Duration", "returns", "the", "time", "duration", "for", "the", "current", "audio", "container" ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L380-L385
5,560
go-audio/wav
decoder.go
readHeaders
func (d *Decoder) readHeaders() error { if d == nil || d.NumChans > 0 { return nil } id, size, err := d.parser.IDnSize() if err != nil { return err } d.parser.ID = id if d.parser.ID != riff.RiffID { return fmt.Errorf("%s - %s", d.parser.ID, riff.ErrFmtNotSupported) } d.parser.Size = size if err := binary.Read(d.r, binary.BigEndian, &d.parser.Format); err != nil { return err } var chunk *riff.Chunk var rewindBytes int64 for err == nil { chunk, err = d.parser.NextChunk() if err != nil { break } if chunk.ID == riff.FmtID { chunk.DecodeWavHeader(d.parser) d.NumChans = d.parser.NumChannels d.BitDepth = d.parser.BitsPerSample d.SampleRate = d.parser.SampleRate d.WavAudioFormat = d.parser.WavAudioFormat d.AvgBytesPerSec = d.parser.AvgBytesPerSec if rewindBytes > 0 { d.r.Seek(-(rewindBytes + int64(chunk.Size) + 8), 1) } break } else if chunk.ID == CIDList { // The list chunk can be in the header or footer // because so many players don't support that chunk properly // it is recommended to have it at the end of the file. DecodeListChunk(d, chunk) // unexpected chunk order, might be a bext chunk rewindBytes += int64(chunk.Size) + 8 } else if chunk.ID == CIDSmpl { DecodeSamplerChunk(d, chunk) rewindBytes += int64(chunk.Size) + 8 } else { // unexpected chunk order, might be a bext chunk rewindBytes += int64(chunk.Size) + 8 // drain the chunk io.CopyN(ioutil.Discard, d.r, int64(chunk.Size)) } } return d.err }
go
func (d *Decoder) readHeaders() error { if d == nil || d.NumChans > 0 { return nil } id, size, err := d.parser.IDnSize() if err != nil { return err } d.parser.ID = id if d.parser.ID != riff.RiffID { return fmt.Errorf("%s - %s", d.parser.ID, riff.ErrFmtNotSupported) } d.parser.Size = size if err := binary.Read(d.r, binary.BigEndian, &d.parser.Format); err != nil { return err } var chunk *riff.Chunk var rewindBytes int64 for err == nil { chunk, err = d.parser.NextChunk() if err != nil { break } if chunk.ID == riff.FmtID { chunk.DecodeWavHeader(d.parser) d.NumChans = d.parser.NumChannels d.BitDepth = d.parser.BitsPerSample d.SampleRate = d.parser.SampleRate d.WavAudioFormat = d.parser.WavAudioFormat d.AvgBytesPerSec = d.parser.AvgBytesPerSec if rewindBytes > 0 { d.r.Seek(-(rewindBytes + int64(chunk.Size) + 8), 1) } break } else if chunk.ID == CIDList { // The list chunk can be in the header or footer // because so many players don't support that chunk properly // it is recommended to have it at the end of the file. DecodeListChunk(d, chunk) // unexpected chunk order, might be a bext chunk rewindBytes += int64(chunk.Size) + 8 } else if chunk.ID == CIDSmpl { DecodeSamplerChunk(d, chunk) rewindBytes += int64(chunk.Size) + 8 } else { // unexpected chunk order, might be a bext chunk rewindBytes += int64(chunk.Size) + 8 // drain the chunk io.CopyN(ioutil.Discard, d.r, int64(chunk.Size)) } } return d.err }
[ "func", "(", "d", "*", "Decoder", ")", "readHeaders", "(", ")", "error", "{", "if", "d", "==", "nil", "||", "d", ".", "NumChans", ">", "0", "{", "return", "nil", "\n", "}", "\n\n", "id", ",", "size", ",", "err", ":=", "d", ".", "parser", ".", "IDnSize", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "d", ".", "parser", ".", "ID", "=", "id", "\n", "if", "d", ".", "parser", ".", "ID", "!=", "riff", ".", "RiffID", "{", "return", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "d", ".", "parser", ".", "ID", ",", "riff", ".", "ErrFmtNotSupported", ")", "\n", "}", "\n", "d", ".", "parser", ".", "Size", "=", "size", "\n", "if", "err", ":=", "binary", ".", "Read", "(", "d", ".", "r", ",", "binary", ".", "BigEndian", ",", "&", "d", ".", "parser", ".", "Format", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "var", "chunk", "*", "riff", ".", "Chunk", "\n", "var", "rewindBytes", "int64", "\n\n", "for", "err", "==", "nil", "{", "chunk", ",", "err", "=", "d", ".", "parser", ".", "NextChunk", "(", ")", "\n", "if", "err", "!=", "nil", "{", "break", "\n", "}", "\n\n", "if", "chunk", ".", "ID", "==", "riff", ".", "FmtID", "{", "chunk", ".", "DecodeWavHeader", "(", "d", ".", "parser", ")", "\n", "d", ".", "NumChans", "=", "d", ".", "parser", ".", "NumChannels", "\n", "d", ".", "BitDepth", "=", "d", ".", "parser", ".", "BitsPerSample", "\n", "d", ".", "SampleRate", "=", "d", ".", "parser", ".", "SampleRate", "\n", "d", ".", "WavAudioFormat", "=", "d", ".", "parser", ".", "WavAudioFormat", "\n", "d", ".", "AvgBytesPerSec", "=", "d", ".", "parser", ".", "AvgBytesPerSec", "\n\n", "if", "rewindBytes", ">", "0", "{", "d", ".", "r", ".", "Seek", "(", "-", "(", "rewindBytes", "+", "int64", "(", "chunk", ".", "Size", ")", "+", "8", ")", ",", "1", ")", "\n", "}", "\n", "break", "\n", "}", "else", "if", "chunk", ".", "ID", "==", "CIDList", "{", "// The list chunk can be in the header or footer", "// because so many players don't support that chunk properly", "// it is recommended to have it at the end of the file.", "DecodeListChunk", "(", "d", ",", "chunk", ")", "\n", "// unexpected chunk order, might be a bext chunk", "rewindBytes", "+=", "int64", "(", "chunk", ".", "Size", ")", "+", "8", "\n", "}", "else", "if", "chunk", ".", "ID", "==", "CIDSmpl", "{", "DecodeSamplerChunk", "(", "d", ",", "chunk", ")", "\n", "rewindBytes", "+=", "int64", "(", "chunk", ".", "Size", ")", "+", "8", "\n", "}", "else", "{", "// unexpected chunk order, might be a bext chunk", "rewindBytes", "+=", "int64", "(", "chunk", ".", "Size", ")", "+", "8", "\n", "// drain the chunk", "io", ".", "CopyN", "(", "ioutil", ".", "Discard", ",", "d", ".", "r", ",", "int64", "(", "chunk", ".", "Size", ")", ")", "\n", "}", "\n", "}", "\n\n", "return", "d", ".", "err", "\n", "}" ]
// readHeaders is safe to call multiple times
[ "readHeaders", "is", "safe", "to", "call", "multiple", "times" ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L393-L451
5,561
go-audio/wav
decoder.go
sampleFloat64DecodeFunc
func sampleFloat64DecodeFunc(bitsPerSample int) (func([]byte) float64, error) { bytesPerSample := bitsPerSample / 8 switch bytesPerSample { case 1: // 8bit values are unsigned return func(s []byte) float64 { return float64(uint8(s[0])) }, nil case 2: return func(s []byte) float64 { return float64(int(s[0]) + int(s[1])<<8) }, nil case 3: return func(s []byte) float64 { var output int32 output |= int32(s[2]) << 0 output |= int32(s[1]) << 8 output |= int32(s[0]) << 16 return float64(output) }, nil case 4: // TODO: fix the float64 conversion (current int implementation) return func(s []byte) float64 { return float64(int(s[0]) + int(s[1])<<8 + int(s[2])<<16 + int(s[3])<<24) }, nil default: return nil, fmt.Errorf("unhandled byte depth:%d", bitsPerSample) } }
go
func sampleFloat64DecodeFunc(bitsPerSample int) (func([]byte) float64, error) { bytesPerSample := bitsPerSample / 8 switch bytesPerSample { case 1: // 8bit values are unsigned return func(s []byte) float64 { return float64(uint8(s[0])) }, nil case 2: return func(s []byte) float64 { return float64(int(s[0]) + int(s[1])<<8) }, nil case 3: return func(s []byte) float64 { var output int32 output |= int32(s[2]) << 0 output |= int32(s[1]) << 8 output |= int32(s[0]) << 16 return float64(output) }, nil case 4: // TODO: fix the float64 conversion (current int implementation) return func(s []byte) float64 { return float64(int(s[0]) + int(s[1])<<8 + int(s[2])<<16 + int(s[3])<<24) }, nil default: return nil, fmt.Errorf("unhandled byte depth:%d", bitsPerSample) } }
[ "func", "sampleFloat64DecodeFunc", "(", "bitsPerSample", "int", ")", "(", "func", "(", "[", "]", "byte", ")", "float64", ",", "error", ")", "{", "bytesPerSample", ":=", "bitsPerSample", "/", "8", "\n", "switch", "bytesPerSample", "{", "case", "1", ":", "// 8bit values are unsigned", "return", "func", "(", "s", "[", "]", "byte", ")", "float64", "{", "return", "float64", "(", "uint8", "(", "s", "[", "0", "]", ")", ")", "\n", "}", ",", "nil", "\n", "case", "2", ":", "return", "func", "(", "s", "[", "]", "byte", ")", "float64", "{", "return", "float64", "(", "int", "(", "s", "[", "0", "]", ")", "+", "int", "(", "s", "[", "1", "]", ")", "<<", "8", ")", "\n", "}", ",", "nil", "\n", "case", "3", ":", "return", "func", "(", "s", "[", "]", "byte", ")", "float64", "{", "var", "output", "int32", "\n", "output", "|=", "int32", "(", "s", "[", "2", "]", ")", "<<", "0", "\n", "output", "|=", "int32", "(", "s", "[", "1", "]", ")", "<<", "8", "\n", "output", "|=", "int32", "(", "s", "[", "0", "]", ")", "<<", "16", "\n", "return", "float64", "(", "output", ")", "\n", "}", ",", "nil", "\n", "case", "4", ":", "// TODO: fix the float64 conversion (current int implementation)", "return", "func", "(", "s", "[", "]", "byte", ")", "float64", "{", "return", "float64", "(", "int", "(", "s", "[", "0", "]", ")", "+", "int", "(", "s", "[", "1", "]", ")", "<<", "8", "+", "int", "(", "s", "[", "2", "]", ")", "<<", "16", "+", "int", "(", "s", "[", "3", "]", ")", "<<", "24", ")", "\n", "}", ",", "nil", "\n", "default", ":", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "bitsPerSample", ")", "\n", "}", "\n", "}" ]
// sampleDecodeFloat64Func returns a function that can be used to convert // a byte range into a float64 value based on the amount of bits used per sample.
[ "sampleDecodeFloat64Func", "returns", "a", "function", "that", "can", "be", "used", "to", "convert", "a", "byte", "range", "into", "a", "float64", "value", "based", "on", "the", "amount", "of", "bits", "used", "per", "sample", "." ]
cc323235893be1cbdb031cb52acfc0184b366870
https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L495-L523
5,562
kr/pretty
diff.go
Diff
func Diff(a, b interface{}) (desc []string) { Pdiff((*sbuf)(&desc), a, b) return desc }
go
func Diff(a, b interface{}) (desc []string) { Pdiff((*sbuf)(&desc), a, b) return desc }
[ "func", "Diff", "(", "a", ",", "b", "interface", "{", "}", ")", "(", "desc", "[", "]", "string", ")", "{", "Pdiff", "(", "(", "*", "sbuf", ")", "(", "&", "desc", ")", ",", "a", ",", "b", ")", "\n", "return", "desc", "\n", "}" ]
// Diff returns a slice where each element describes // a difference between a and b.
[ "Diff", "returns", "a", "slice", "where", "each", "element", "describes", "a", "difference", "between", "a", "and", "b", "." ]
73f6ac0b30a98e433b289500d779f50c1a6f0712
https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L18-L21
5,563
kr/pretty
diff.go
Fdiff
func Fdiff(w io.Writer, a, b interface{}) { Pdiff(&wprintfer{w}, a, b) }
go
func Fdiff(w io.Writer, a, b interface{}) { Pdiff(&wprintfer{w}, a, b) }
[ "func", "Fdiff", "(", "w", "io", ".", "Writer", ",", "a", ",", "b", "interface", "{", "}", ")", "{", "Pdiff", "(", "&", "wprintfer", "{", "w", "}", ",", "a", ",", "b", ")", "\n", "}" ]
// Fdiff writes to w a description of the differences between a and b.
[ "Fdiff", "writes", "to", "w", "a", "description", "of", "the", "differences", "between", "a", "and", "b", "." ]
73f6ac0b30a98e433b289500d779f50c1a6f0712
https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L32-L34
5,564
kr/pretty
diff.go
Pdiff
func Pdiff(p Printfer, a, b interface{}) { diffPrinter{w: p}.diff(reflect.ValueOf(a), reflect.ValueOf(b)) }
go
func Pdiff(p Printfer, a, b interface{}) { diffPrinter{w: p}.diff(reflect.ValueOf(a), reflect.ValueOf(b)) }
[ "func", "Pdiff", "(", "p", "Printfer", ",", "a", ",", "b", "interface", "{", "}", ")", "{", "diffPrinter", "{", "w", ":", "p", "}", ".", "diff", "(", "reflect", ".", "ValueOf", "(", "a", ")", ",", "reflect", ".", "ValueOf", "(", "b", ")", ")", "\n", "}" ]
// Pdiff prints to p a description of the differences between a and b. // It calls Printf once for each difference, with no trailing newline. // The standard library log.Logger is a Printfer.
[ "Pdiff", "prints", "to", "p", "a", "description", "of", "the", "differences", "between", "a", "and", "b", ".", "It", "calls", "Printf", "once", "for", "each", "difference", "with", "no", "trailing", "newline", ".", "The", "standard", "library", "log", ".", "Logger", "is", "a", "Printfer", "." ]
73f6ac0b30a98e433b289500d779f50c1a6f0712
https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L43-L45
5,565
kr/pretty
diff.go
Ldiff
func Ldiff(l Logfer, a, b interface{}) { Pdiff(&logprintfer{l}, a, b) }
go
func Ldiff(l Logfer, a, b interface{}) { Pdiff(&logprintfer{l}, a, b) }
[ "func", "Ldiff", "(", "l", "Logfer", ",", "a", ",", "b", "interface", "{", "}", ")", "{", "Pdiff", "(", "&", "logprintfer", "{", "l", "}", ",", "a", ",", "b", ")", "\n", "}" ]
// Ldiff prints to l a description of the differences between a and b. // It calls Logf once for each difference, with no trailing newline. // The standard library testing.T and testing.B are Logfers.
[ "Ldiff", "prints", "to", "l", "a", "description", "of", "the", "differences", "between", "a", "and", "b", ".", "It", "calls", "Logf", "once", "for", "each", "difference", "with", "no", "trailing", "newline", ".", "The", "standard", "library", "testing", ".", "T", "and", "testing", ".", "B", "are", "Logfers", "." ]
73f6ac0b30a98e433b289500d779f50c1a6f0712
https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L62-L64
5,566
kr/pretty
diff.go
keyEqual
func keyEqual(av, bv reflect.Value) bool { if !av.IsValid() && !bv.IsValid() { return true } if !av.IsValid() || !bv.IsValid() || av.Type() != bv.Type() { return false } switch kind := av.Kind(); kind { case reflect.Bool: a, b := av.Bool(), bv.Bool() return a == b case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: a, b := av.Int(), bv.Int() return a == b case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: a, b := av.Uint(), bv.Uint() return a == b case reflect.Float32, reflect.Float64: a, b := av.Float(), bv.Float() return a == b case reflect.Complex64, reflect.Complex128: a, b := av.Complex(), bv.Complex() return a == b case reflect.Array: for i := 0; i < av.Len(); i++ { if !keyEqual(av.Index(i), bv.Index(i)) { return false } } return true case reflect.Chan, reflect.UnsafePointer, reflect.Ptr: a, b := av.Pointer(), bv.Pointer() return a == b case reflect.Interface: return keyEqual(av.Elem(), bv.Elem()) case reflect.String: a, b := av.String(), bv.String() return a == b case reflect.Struct: for i := 0; i < av.NumField(); i++ { if !keyEqual(av.Field(i), bv.Field(i)) { return false } } return true default: panic("invalid map key type " + av.Type().String()) } }
go
func keyEqual(av, bv reflect.Value) bool { if !av.IsValid() && !bv.IsValid() { return true } if !av.IsValid() || !bv.IsValid() || av.Type() != bv.Type() { return false } switch kind := av.Kind(); kind { case reflect.Bool: a, b := av.Bool(), bv.Bool() return a == b case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: a, b := av.Int(), bv.Int() return a == b case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr: a, b := av.Uint(), bv.Uint() return a == b case reflect.Float32, reflect.Float64: a, b := av.Float(), bv.Float() return a == b case reflect.Complex64, reflect.Complex128: a, b := av.Complex(), bv.Complex() return a == b case reflect.Array: for i := 0; i < av.Len(); i++ { if !keyEqual(av.Index(i), bv.Index(i)) { return false } } return true case reflect.Chan, reflect.UnsafePointer, reflect.Ptr: a, b := av.Pointer(), bv.Pointer() return a == b case reflect.Interface: return keyEqual(av.Elem(), bv.Elem()) case reflect.String: a, b := av.String(), bv.String() return a == b case reflect.Struct: for i := 0; i < av.NumField(); i++ { if !keyEqual(av.Field(i), bv.Field(i)) { return false } } return true default: panic("invalid map key type " + av.Type().String()) } }
[ "func", "keyEqual", "(", "av", ",", "bv", "reflect", ".", "Value", ")", "bool", "{", "if", "!", "av", ".", "IsValid", "(", ")", "&&", "!", "bv", ".", "IsValid", "(", ")", "{", "return", "true", "\n", "}", "\n", "if", "!", "av", ".", "IsValid", "(", ")", "||", "!", "bv", ".", "IsValid", "(", ")", "||", "av", ".", "Type", "(", ")", "!=", "bv", ".", "Type", "(", ")", "{", "return", "false", "\n", "}", "\n", "switch", "kind", ":=", "av", ".", "Kind", "(", ")", ";", "kind", "{", "case", "reflect", ".", "Bool", ":", "a", ",", "b", ":=", "av", ".", "Bool", "(", ")", ",", "bv", ".", "Bool", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Int", ",", "reflect", ".", "Int8", ",", "reflect", ".", "Int16", ",", "reflect", ".", "Int32", ",", "reflect", ".", "Int64", ":", "a", ",", "b", ":=", "av", ".", "Int", "(", ")", ",", "bv", ".", "Int", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Uint", ",", "reflect", ".", "Uint8", ",", "reflect", ".", "Uint16", ",", "reflect", ".", "Uint32", ",", "reflect", ".", "Uint64", ",", "reflect", ".", "Uintptr", ":", "a", ",", "b", ":=", "av", ".", "Uint", "(", ")", ",", "bv", ".", "Uint", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Float32", ",", "reflect", ".", "Float64", ":", "a", ",", "b", ":=", "av", ".", "Float", "(", ")", ",", "bv", ".", "Float", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Complex64", ",", "reflect", ".", "Complex128", ":", "a", ",", "b", ":=", "av", ".", "Complex", "(", ")", ",", "bv", ".", "Complex", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Array", ":", "for", "i", ":=", "0", ";", "i", "<", "av", ".", "Len", "(", ")", ";", "i", "++", "{", "if", "!", "keyEqual", "(", "av", ".", "Index", "(", "i", ")", ",", "bv", ".", "Index", "(", "i", ")", ")", "{", "return", "false", "\n", "}", "\n", "}", "\n", "return", "true", "\n", "case", "reflect", ".", "Chan", ",", "reflect", ".", "UnsafePointer", ",", "reflect", ".", "Ptr", ":", "a", ",", "b", ":=", "av", ".", "Pointer", "(", ")", ",", "bv", ".", "Pointer", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Interface", ":", "return", "keyEqual", "(", "av", ".", "Elem", "(", ")", ",", "bv", ".", "Elem", "(", ")", ")", "\n", "case", "reflect", ".", "String", ":", "a", ",", "b", ":=", "av", ".", "String", "(", ")", ",", "bv", ".", "String", "(", ")", "\n", "return", "a", "==", "b", "\n", "case", "reflect", ".", "Struct", ":", "for", "i", ":=", "0", ";", "i", "<", "av", ".", "NumField", "(", ")", ";", "i", "++", "{", "if", "!", "keyEqual", "(", "av", ".", "Field", "(", "i", ")", ",", "bv", ".", "Field", "(", "i", ")", ")", "{", "return", "false", "\n", "}", "\n", "}", "\n", "return", "true", "\n", "default", ":", "panic", "(", "\"", "\"", "+", "av", ".", "Type", "(", ")", ".", "String", "(", ")", ")", "\n", "}", "\n", "}" ]
// keyEqual compares a and b for equality. // Both a and b must be valid map keys.
[ "keyEqual", "compares", "a", "and", "b", "for", "equality", ".", "Both", "a", "and", "b", "must", "be", "valid", "map", "keys", "." ]
73f6ac0b30a98e433b289500d779f50c1a6f0712
https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L188-L236
5,567
hashicorp/hil
parser/error.go
TokenErrorf
func TokenErrorf(token *scanner.Token, format string, args ...interface{}) error { return Errorf(token.Pos, format, args...) }
go
func TokenErrorf(token *scanner.Token, format string, args ...interface{}) error { return Errorf(token.Pos, format, args...) }
[ "func", "TokenErrorf", "(", "token", "*", "scanner", ".", "Token", ",", "format", "string", ",", "args", "...", "interface", "{", "}", ")", "error", "{", "return", "Errorf", "(", "token", ".", "Pos", ",", "format", ",", "args", "...", ")", "\n", "}" ]
// TokenErrorf is a convenient wrapper around Errorf that uses the // position of the given token.
[ "TokenErrorf", "is", "a", "convenient", "wrapper", "around", "Errorf", "that", "uses", "the", "position", "of", "the", "given", "token", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/error.go#L24-L26
5,568
hashicorp/hil
parse.go
ParseWithPosition
func ParseWithPosition(v string, pos ast.Pos) (ast.Node, error) { ch := scanner.Scan(v, pos) return parser.Parse(ch) }
go
func ParseWithPosition(v string, pos ast.Pos) (ast.Node, error) { ch := scanner.Scan(v, pos) return parser.Parse(ch) }
[ "func", "ParseWithPosition", "(", "v", "string", ",", "pos", "ast", ".", "Pos", ")", "(", "ast", ".", "Node", ",", "error", ")", "{", "ch", ":=", "scanner", ".", "Scan", "(", "v", ",", "pos", ")", "\n", "return", "parser", ".", "Parse", "(", "ch", ")", "\n", "}" ]
// ParseWithPosition is like Parse except that it overrides the source // row and column position of the first character in the string, which should // be 1-based. // // This can be used when HIL is embedded in another language and the outer // parser knows the row and column where the HIL expression started within // the overall source file.
[ "ParseWithPosition", "is", "like", "Parse", "except", "that", "it", "overrides", "the", "source", "row", "and", "column", "position", "of", "the", "first", "character", "in", "the", "string", "which", "should", "be", "1", "-", "based", ".", "This", "can", "be", "used", "when", "HIL", "is", "embedded", "in", "another", "language", "and", "the", "outer", "parser", "knows", "the", "row", "and", "column", "where", "the", "HIL", "expression", "started", "within", "the", "overall", "source", "file", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parse.go#L26-L29
5,569
hashicorp/hil
scanner/peeker.go
Peek
func (p *Peeker) Peek() *Token { if p.peeked == nil { p.peeked = <-p.ch } return p.peeked }
go
func (p *Peeker) Peek() *Token { if p.peeked == nil { p.peeked = <-p.ch } return p.peeked }
[ "func", "(", "p", "*", "Peeker", ")", "Peek", "(", ")", "*", "Token", "{", "if", "p", ".", "peeked", "==", "nil", "{", "p", ".", "peeked", "=", "<-", "p", ".", "ch", "\n", "}", "\n", "return", "p", ".", "peeked", "\n", "}" ]
// Peek returns the next token in the stream without consuming it. A // subsequent call to Read will return the same token.
[ "Peek", "returns", "the", "next", "token", "in", "the", "stream", "without", "consuming", "it", ".", "A", "subsequent", "call", "to", "Read", "will", "return", "the", "same", "token", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L20-L25
5,570
hashicorp/hil
scanner/peeker.go
Read
func (p *Peeker) Read() *Token { token := p.Peek() // As a special case, we will produce the EOF token forever once // it is reached. if token.Type != EOF { p.peeked = nil } return token }
go
func (p *Peeker) Read() *Token { token := p.Peek() // As a special case, we will produce the EOF token forever once // it is reached. if token.Type != EOF { p.peeked = nil } return token }
[ "func", "(", "p", "*", "Peeker", ")", "Read", "(", ")", "*", "Token", "{", "token", ":=", "p", ".", "Peek", "(", ")", "\n\n", "// As a special case, we will produce the EOF token forever once", "// it is reached.", "if", "token", ".", "Type", "!=", "EOF", "{", "p", ".", "peeked", "=", "nil", "\n", "}", "\n\n", "return", "token", "\n", "}" ]
// Read consumes the next token in the stream and returns it.
[ "Read", "consumes", "the", "next", "token", "in", "the", "stream", "and", "returns", "it", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L28-L38
5,571
hashicorp/hil
scanner/peeker.go
Close
func (p *Peeker) Close() { for _ = range p.ch { // discard } // Install a synthetic EOF token in 'peeked' in case someone // erroneously calls Peek() or Read() after we've closed. p.peeked = &Token{ Type: EOF, Content: "", } }
go
func (p *Peeker) Close() { for _ = range p.ch { // discard } // Install a synthetic EOF token in 'peeked' in case someone // erroneously calls Peek() or Read() after we've closed. p.peeked = &Token{ Type: EOF, Content: "", } }
[ "func", "(", "p", "*", "Peeker", ")", "Close", "(", ")", "{", "for", "_", "=", "range", "p", ".", "ch", "{", "// discard", "}", "\n", "// Install a synthetic EOF token in 'peeked' in case someone", "// erroneously calls Peek() or Read() after we've closed.", "p", ".", "peeked", "=", "&", "Token", "{", "Type", ":", "EOF", ",", "Content", ":", "\"", "\"", ",", "}", "\n", "}" ]
// Close ensures that the token stream has been exhausted, to prevent // the goroutine in the underlying scanner from leaking. // // It's not necessary to call this if the caller reads the token stream // to EOF, since that implicitly closes the scanner.
[ "Close", "ensures", "that", "the", "token", "stream", "has", "been", "exhausted", "to", "prevent", "the", "goroutine", "in", "the", "underlying", "scanner", "from", "leaking", ".", "It", "s", "not", "necessary", "to", "call", "this", "if", "the", "caller", "reads", "the", "token", "stream", "to", "EOF", "since", "that", "implicitly", "closes", "the", "scanner", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L45-L55
5,572
hashicorp/hil
parser/parser.go
parseInterpolationSeq
func (p *parser) parseInterpolationSeq(quoted bool) (ast.Node, error) { literalType := scanner.LITERAL endType := scanner.EOF if quoted { // exceptions for quoted sequences literalType = scanner.STRING endType = scanner.CQUOTE } startPos := p.peeker.Peek().Pos if quoted { tok := p.peeker.Read() if tok.Type != scanner.OQUOTE { return nil, ExpectationError("open quote", tok) } } var exprs []ast.Node for { tok := p.peeker.Read() if tok.Type == endType { break } switch tok.Type { case literalType: val, err := p.parseStringToken(tok) if err != nil { return nil, err } exprs = append(exprs, &ast.LiteralNode{ Value: val, Typex: ast.TypeString, Posx: tok.Pos, }) case scanner.BEGIN: expr, err := p.ParseInterpolation() if err != nil { return nil, err } exprs = append(exprs, expr) default: return nil, ExpectationError(`"${"`, tok) } } if len(exprs) == 0 { // If we have no parts at all then the input must've // been an empty string. exprs = append(exprs, &ast.LiteralNode{ Value: "", Typex: ast.TypeString, Posx: startPos, }) } // As a special case, if our "Output" contains only one expression // and it's a literal string then we'll hoist it up to be our // direct return value, so callers can easily recognize a string // that has no interpolations at all. if len(exprs) == 1 { if lit, ok := exprs[0].(*ast.LiteralNode); ok { if lit.Typex == ast.TypeString { return lit, nil } } } return &ast.Output{ Exprs: exprs, Posx: startPos, }, nil }
go
func (p *parser) parseInterpolationSeq(quoted bool) (ast.Node, error) { literalType := scanner.LITERAL endType := scanner.EOF if quoted { // exceptions for quoted sequences literalType = scanner.STRING endType = scanner.CQUOTE } startPos := p.peeker.Peek().Pos if quoted { tok := p.peeker.Read() if tok.Type != scanner.OQUOTE { return nil, ExpectationError("open quote", tok) } } var exprs []ast.Node for { tok := p.peeker.Read() if tok.Type == endType { break } switch tok.Type { case literalType: val, err := p.parseStringToken(tok) if err != nil { return nil, err } exprs = append(exprs, &ast.LiteralNode{ Value: val, Typex: ast.TypeString, Posx: tok.Pos, }) case scanner.BEGIN: expr, err := p.ParseInterpolation() if err != nil { return nil, err } exprs = append(exprs, expr) default: return nil, ExpectationError(`"${"`, tok) } } if len(exprs) == 0 { // If we have no parts at all then the input must've // been an empty string. exprs = append(exprs, &ast.LiteralNode{ Value: "", Typex: ast.TypeString, Posx: startPos, }) } // As a special case, if our "Output" contains only one expression // and it's a literal string then we'll hoist it up to be our // direct return value, so callers can easily recognize a string // that has no interpolations at all. if len(exprs) == 1 { if lit, ok := exprs[0].(*ast.LiteralNode); ok { if lit.Typex == ast.TypeString { return lit, nil } } } return &ast.Output{ Exprs: exprs, Posx: startPos, }, nil }
[ "func", "(", "p", "*", "parser", ")", "parseInterpolationSeq", "(", "quoted", "bool", ")", "(", "ast", ".", "Node", ",", "error", ")", "{", "literalType", ":=", "scanner", ".", "LITERAL", "\n", "endType", ":=", "scanner", ".", "EOF", "\n", "if", "quoted", "{", "// exceptions for quoted sequences", "literalType", "=", "scanner", ".", "STRING", "\n", "endType", "=", "scanner", ".", "CQUOTE", "\n", "}", "\n\n", "startPos", ":=", "p", ".", "peeker", ".", "Peek", "(", ")", ".", "Pos", "\n\n", "if", "quoted", "{", "tok", ":=", "p", ".", "peeker", ".", "Read", "(", ")", "\n", "if", "tok", ".", "Type", "!=", "scanner", ".", "OQUOTE", "{", "return", "nil", ",", "ExpectationError", "(", "\"", "\"", ",", "tok", ")", "\n", "}", "\n", "}", "\n\n", "var", "exprs", "[", "]", "ast", ".", "Node", "\n", "for", "{", "tok", ":=", "p", ".", "peeker", ".", "Read", "(", ")", "\n\n", "if", "tok", ".", "Type", "==", "endType", "{", "break", "\n", "}", "\n\n", "switch", "tok", ".", "Type", "{", "case", "literalType", ":", "val", ",", "err", ":=", "p", ".", "parseStringToken", "(", "tok", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "exprs", "=", "append", "(", "exprs", ",", "&", "ast", ".", "LiteralNode", "{", "Value", ":", "val", ",", "Typex", ":", "ast", ".", "TypeString", ",", "Posx", ":", "tok", ".", "Pos", ",", "}", ")", "\n", "case", "scanner", ".", "BEGIN", ":", "expr", ",", "err", ":=", "p", ".", "ParseInterpolation", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "exprs", "=", "append", "(", "exprs", ",", "expr", ")", "\n", "default", ":", "return", "nil", ",", "ExpectationError", "(", "`\"${\"`", ",", "tok", ")", "\n", "}", "\n", "}", "\n\n", "if", "len", "(", "exprs", ")", "==", "0", "{", "// If we have no parts at all then the input must've", "// been an empty string.", "exprs", "=", "append", "(", "exprs", ",", "&", "ast", ".", "LiteralNode", "{", "Value", ":", "\"", "\"", ",", "Typex", ":", "ast", ".", "TypeString", ",", "Posx", ":", "startPos", ",", "}", ")", "\n", "}", "\n\n", "// As a special case, if our \"Output\" contains only one expression", "// and it's a literal string then we'll hoist it up to be our", "// direct return value, so callers can easily recognize a string", "// that has no interpolations at all.", "if", "len", "(", "exprs", ")", "==", "1", "{", "if", "lit", ",", "ok", ":=", "exprs", "[", "0", "]", ".", "(", "*", "ast", ".", "LiteralNode", ")", ";", "ok", "{", "if", "lit", ".", "Typex", "==", "ast", ".", "TypeString", "{", "return", "lit", ",", "nil", "\n", "}", "\n", "}", "\n", "}", "\n\n", "return", "&", "ast", ".", "Output", "{", "Exprs", ":", "exprs", ",", "Posx", ":", "startPos", ",", "}", ",", "nil", "\n", "}" ]
// parseInterpolationSeq parses either the top-level sequence of literals // and interpolation expressions or a similar sequence within a quoted // string inside an interpolation expression. The latter case is requested // by setting 'quoted' to true.
[ "parseInterpolationSeq", "parses", "either", "the", "top", "-", "level", "sequence", "of", "literals", "and", "interpolation", "expressions", "or", "a", "similar", "sequence", "within", "a", "quoted", "string", "inside", "an", "interpolation", "expression", ".", "The", "latter", "case", "is", "requested", "by", "setting", "quoted", "to", "true", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L35-L109
5,573
hashicorp/hil
parser/parser.go
parseStringToken
func (p *parser) parseStringToken(tok *scanner.Token) (string, error) { var backslashes bool switch tok.Type { case scanner.LITERAL: backslashes = false case scanner.STRING: backslashes = true default: panic("unsupported string token type") } raw := []byte(tok.Content) buf := make([]byte, 0, len(raw)) for i := 0; i < len(raw); i++ { b := raw[i] more := len(raw) > (i + 1) if b == '$' { if more && raw[i+1] == '$' { // skip over the second dollar sign i++ } } else if backslashes && b == '\\' { if !more { return "", Errorf( ast.Pos{ Column: tok.Pos.Column + utf8.RuneCount(raw[:i]), Line: tok.Pos.Line, }, `unfinished backslash escape sequence`, ) } escapeType := raw[i+1] switch escapeType { case '\\': // skip over the second slash i++ case 'n': b = '\n' i++ case '"': b = '"' i++ default: return "", Errorf( ast.Pos{ Column: tok.Pos.Column + utf8.RuneCount(raw[:i]), Line: tok.Pos.Line, }, `invalid backslash escape sequence`, ) } } buf = append(buf, b) } return string(buf), nil }
go
func (p *parser) parseStringToken(tok *scanner.Token) (string, error) { var backslashes bool switch tok.Type { case scanner.LITERAL: backslashes = false case scanner.STRING: backslashes = true default: panic("unsupported string token type") } raw := []byte(tok.Content) buf := make([]byte, 0, len(raw)) for i := 0; i < len(raw); i++ { b := raw[i] more := len(raw) > (i + 1) if b == '$' { if more && raw[i+1] == '$' { // skip over the second dollar sign i++ } } else if backslashes && b == '\\' { if !more { return "", Errorf( ast.Pos{ Column: tok.Pos.Column + utf8.RuneCount(raw[:i]), Line: tok.Pos.Line, }, `unfinished backslash escape sequence`, ) } escapeType := raw[i+1] switch escapeType { case '\\': // skip over the second slash i++ case 'n': b = '\n' i++ case '"': b = '"' i++ default: return "", Errorf( ast.Pos{ Column: tok.Pos.Column + utf8.RuneCount(raw[:i]), Line: tok.Pos.Line, }, `invalid backslash escape sequence`, ) } } buf = append(buf, b) } return string(buf), nil }
[ "func", "(", "p", "*", "parser", ")", "parseStringToken", "(", "tok", "*", "scanner", ".", "Token", ")", "(", "string", ",", "error", ")", "{", "var", "backslashes", "bool", "\n", "switch", "tok", ".", "Type", "{", "case", "scanner", ".", "LITERAL", ":", "backslashes", "=", "false", "\n", "case", "scanner", ".", "STRING", ":", "backslashes", "=", "true", "\n", "default", ":", "panic", "(", "\"", "\"", ")", "\n", "}", "\n\n", "raw", ":=", "[", "]", "byte", "(", "tok", ".", "Content", ")", "\n", "buf", ":=", "make", "(", "[", "]", "byte", ",", "0", ",", "len", "(", "raw", ")", ")", "\n\n", "for", "i", ":=", "0", ";", "i", "<", "len", "(", "raw", ")", ";", "i", "++", "{", "b", ":=", "raw", "[", "i", "]", "\n", "more", ":=", "len", "(", "raw", ")", ">", "(", "i", "+", "1", ")", "\n\n", "if", "b", "==", "'$'", "{", "if", "more", "&&", "raw", "[", "i", "+", "1", "]", "==", "'$'", "{", "// skip over the second dollar sign", "i", "++", "\n", "}", "\n", "}", "else", "if", "backslashes", "&&", "b", "==", "'\\\\'", "{", "if", "!", "more", "{", "return", "\"", "\"", ",", "Errorf", "(", "ast", ".", "Pos", "{", "Column", ":", "tok", ".", "Pos", ".", "Column", "+", "utf8", ".", "RuneCount", "(", "raw", "[", ":", "i", "]", ")", ",", "Line", ":", "tok", ".", "Pos", ".", "Line", ",", "}", ",", "`unfinished backslash escape sequence`", ",", ")", "\n", "}", "\n", "escapeType", ":=", "raw", "[", "i", "+", "1", "]", "\n", "switch", "escapeType", "{", "case", "'\\\\'", ":", "// skip over the second slash", "i", "++", "\n", "case", "'n'", ":", "b", "=", "'\\n'", "\n", "i", "++", "\n", "case", "'\"'", ":", "b", "=", "'\"'", "\n", "i", "++", "\n", "default", ":", "return", "\"", "\"", ",", "Errorf", "(", "ast", ".", "Pos", "{", "Column", ":", "tok", ".", "Pos", ".", "Column", "+", "utf8", ".", "RuneCount", "(", "raw", "[", ":", "i", "]", ")", ",", "Line", ":", "tok", ".", "Pos", ".", "Line", ",", "}", ",", "`invalid backslash escape sequence`", ",", ")", "\n", "}", "\n", "}", "\n\n", "buf", "=", "append", "(", "buf", ",", "b", ")", "\n", "}", "\n\n", "return", "string", "(", "buf", ")", ",", "nil", "\n", "}" ]
// parseStringToken takes a token of either LITERAL or STRING type and // returns the interpreted string, after processing any relevant // escape sequences.
[ "parseStringToken", "takes", "a", "token", "of", "either", "LITERAL", "or", "STRING", "type", "and", "returns", "the", "interpreted", "string", "after", "processing", "any", "relevant", "escape", "sequences", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L114-L173
5,574
hashicorp/hil
parser/parser.go
parseBinaryOps
func (p *parser) parseBinaryOps(ops []map[scanner.TokenType]ast.ArithmeticOp) (ast.Node, error) { if len(ops) == 0 { // We've run out of operators, so now we'll just try to parse a term. return p.ParseExpressionTerm() } thisLevel := ops[0] remaining := ops[1:] startPos := p.peeker.Peek().Pos var lhs, rhs ast.Node operator := ast.ArithmeticOpInvalid var err error // parse a term that might be the first operand of a binary // expression or it might just be a standalone term, but // we won't know until we've parsed it and can look ahead // to see if there's an operator token. lhs, err = p.parseBinaryOps(remaining) if err != nil { return nil, err } // We'll keep eating up arithmetic operators until we run // out, so that operators with the same precedence will combine in a // left-associative manner: // a+b+c => (a+b)+c, not a+(b+c) // // Should we later want to have right-associative operators, a way // to achieve that would be to call back up to ParseExpression here // instead of iteratively parsing only the remaining operators. for { next := p.peeker.Peek() var newOperator ast.ArithmeticOp var ok bool if newOperator, ok = thisLevel[next.Type]; !ok { break } // Are we extending an expression started on // the previous iteration? if operator != ast.ArithmeticOpInvalid { lhs = &ast.Arithmetic{ Op: operator, Exprs: []ast.Node{lhs, rhs}, Posx: startPos, } } operator = newOperator p.peeker.Read() // eat operator token rhs, err = p.parseBinaryOps(remaining) if err != nil { return nil, err } } if operator != ast.ArithmeticOpInvalid { return &ast.Arithmetic{ Op: operator, Exprs: []ast.Node{lhs, rhs}, Posx: startPos, }, nil } else { return lhs, nil } }
go
func (p *parser) parseBinaryOps(ops []map[scanner.TokenType]ast.ArithmeticOp) (ast.Node, error) { if len(ops) == 0 { // We've run out of operators, so now we'll just try to parse a term. return p.ParseExpressionTerm() } thisLevel := ops[0] remaining := ops[1:] startPos := p.peeker.Peek().Pos var lhs, rhs ast.Node operator := ast.ArithmeticOpInvalid var err error // parse a term that might be the first operand of a binary // expression or it might just be a standalone term, but // we won't know until we've parsed it and can look ahead // to see if there's an operator token. lhs, err = p.parseBinaryOps(remaining) if err != nil { return nil, err } // We'll keep eating up arithmetic operators until we run // out, so that operators with the same precedence will combine in a // left-associative manner: // a+b+c => (a+b)+c, not a+(b+c) // // Should we later want to have right-associative operators, a way // to achieve that would be to call back up to ParseExpression here // instead of iteratively parsing only the remaining operators. for { next := p.peeker.Peek() var newOperator ast.ArithmeticOp var ok bool if newOperator, ok = thisLevel[next.Type]; !ok { break } // Are we extending an expression started on // the previous iteration? if operator != ast.ArithmeticOpInvalid { lhs = &ast.Arithmetic{ Op: operator, Exprs: []ast.Node{lhs, rhs}, Posx: startPos, } } operator = newOperator p.peeker.Read() // eat operator token rhs, err = p.parseBinaryOps(remaining) if err != nil { return nil, err } } if operator != ast.ArithmeticOpInvalid { return &ast.Arithmetic{ Op: operator, Exprs: []ast.Node{lhs, rhs}, Posx: startPos, }, nil } else { return lhs, nil } }
[ "func", "(", "p", "*", "parser", ")", "parseBinaryOps", "(", "ops", "[", "]", "map", "[", "scanner", ".", "TokenType", "]", "ast", ".", "ArithmeticOp", ")", "(", "ast", ".", "Node", ",", "error", ")", "{", "if", "len", "(", "ops", ")", "==", "0", "{", "// We've run out of operators, so now we'll just try to parse a term.", "return", "p", ".", "ParseExpressionTerm", "(", ")", "\n", "}", "\n\n", "thisLevel", ":=", "ops", "[", "0", "]", "\n", "remaining", ":=", "ops", "[", "1", ":", "]", "\n\n", "startPos", ":=", "p", ".", "peeker", ".", "Peek", "(", ")", ".", "Pos", "\n\n", "var", "lhs", ",", "rhs", "ast", ".", "Node", "\n", "operator", ":=", "ast", ".", "ArithmeticOpInvalid", "\n", "var", "err", "error", "\n\n", "// parse a term that might be the first operand of a binary", "// expression or it might just be a standalone term, but", "// we won't know until we've parsed it and can look ahead", "// to see if there's an operator token.", "lhs", ",", "err", "=", "p", ".", "parseBinaryOps", "(", "remaining", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "// We'll keep eating up arithmetic operators until we run", "// out, so that operators with the same precedence will combine in a", "// left-associative manner:", "// a+b+c => (a+b)+c, not a+(b+c)", "//", "// Should we later want to have right-associative operators, a way", "// to achieve that would be to call back up to ParseExpression here", "// instead of iteratively parsing only the remaining operators.", "for", "{", "next", ":=", "p", ".", "peeker", ".", "Peek", "(", ")", "\n", "var", "newOperator", "ast", ".", "ArithmeticOp", "\n", "var", "ok", "bool", "\n", "if", "newOperator", ",", "ok", "=", "thisLevel", "[", "next", ".", "Type", "]", ";", "!", "ok", "{", "break", "\n", "}", "\n\n", "// Are we extending an expression started on", "// the previous iteration?", "if", "operator", "!=", "ast", ".", "ArithmeticOpInvalid", "{", "lhs", "=", "&", "ast", ".", "Arithmetic", "{", "Op", ":", "operator", ",", "Exprs", ":", "[", "]", "ast", ".", "Node", "{", "lhs", ",", "rhs", "}", ",", "Posx", ":", "startPos", ",", "}", "\n", "}", "\n\n", "operator", "=", "newOperator", "\n", "p", ".", "peeker", ".", "Read", "(", ")", "// eat operator token", "\n", "rhs", ",", "err", "=", "p", ".", "parseBinaryOps", "(", "remaining", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "}", "\n\n", "if", "operator", "!=", "ast", ".", "ArithmeticOpInvalid", "{", "return", "&", "ast", ".", "Arithmetic", "{", "Op", ":", "operator", ",", "Exprs", ":", "[", "]", "ast", ".", "Node", "{", "lhs", ",", "rhs", "}", ",", "Posx", ":", "startPos", ",", "}", ",", "nil", "\n", "}", "else", "{", "return", "lhs", ",", "nil", "\n", "}", "\n", "}" ]
// parseBinaryOps calls itself recursively to work through all of the // operator precedence groups, and then eventually calls ParseExpressionTerm // for each operand.
[ "parseBinaryOps", "calls", "itself", "recursively", "to", "work", "through", "all", "of", "the", "operator", "precedence", "groups", "and", "then", "eventually", "calls", "ParseExpressionTerm", "for", "each", "operand", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L248-L315
5,575
hashicorp/hil
eval.go
internalEval
func internalEval(root ast.Node, config *EvalConfig) (interface{}, ast.Type, error) { // Copy the scope so we can add our builtins if config == nil { config = new(EvalConfig) } scope := registerBuiltins(config.GlobalScope) implicitMap := map[ast.Type]map[ast.Type]string{ ast.TypeFloat: { ast.TypeInt: "__builtin_FloatToInt", ast.TypeString: "__builtin_FloatToString", }, ast.TypeInt: { ast.TypeFloat: "__builtin_IntToFloat", ast.TypeString: "__builtin_IntToString", }, ast.TypeString: { ast.TypeInt: "__builtin_StringToInt", ast.TypeFloat: "__builtin_StringToFloat", ast.TypeBool: "__builtin_StringToBool", }, ast.TypeBool: { ast.TypeString: "__builtin_BoolToString", }, } // Build our own semantic checks that we always run tv := &TypeCheck{Scope: scope, Implicit: implicitMap} ic := &IdentifierCheck{Scope: scope} // Build up the semantic checks for execution checks := make( []SemanticChecker, len(config.SemanticChecks), len(config.SemanticChecks)+2) copy(checks, config.SemanticChecks) checks = append(checks, ic.Visit) checks = append(checks, tv.Visit) // Run the semantic checks for _, check := range checks { if err := check(root); err != nil { return nil, ast.TypeInvalid, err } } // Execute v := &evalVisitor{Scope: scope} return v.Visit(root) }
go
func internalEval(root ast.Node, config *EvalConfig) (interface{}, ast.Type, error) { // Copy the scope so we can add our builtins if config == nil { config = new(EvalConfig) } scope := registerBuiltins(config.GlobalScope) implicitMap := map[ast.Type]map[ast.Type]string{ ast.TypeFloat: { ast.TypeInt: "__builtin_FloatToInt", ast.TypeString: "__builtin_FloatToString", }, ast.TypeInt: { ast.TypeFloat: "__builtin_IntToFloat", ast.TypeString: "__builtin_IntToString", }, ast.TypeString: { ast.TypeInt: "__builtin_StringToInt", ast.TypeFloat: "__builtin_StringToFloat", ast.TypeBool: "__builtin_StringToBool", }, ast.TypeBool: { ast.TypeString: "__builtin_BoolToString", }, } // Build our own semantic checks that we always run tv := &TypeCheck{Scope: scope, Implicit: implicitMap} ic := &IdentifierCheck{Scope: scope} // Build up the semantic checks for execution checks := make( []SemanticChecker, len(config.SemanticChecks), len(config.SemanticChecks)+2) copy(checks, config.SemanticChecks) checks = append(checks, ic.Visit) checks = append(checks, tv.Visit) // Run the semantic checks for _, check := range checks { if err := check(root); err != nil { return nil, ast.TypeInvalid, err } } // Execute v := &evalVisitor{Scope: scope} return v.Visit(root) }
[ "func", "internalEval", "(", "root", "ast", ".", "Node", ",", "config", "*", "EvalConfig", ")", "(", "interface", "{", "}", ",", "ast", ".", "Type", ",", "error", ")", "{", "// Copy the scope so we can add our builtins", "if", "config", "==", "nil", "{", "config", "=", "new", "(", "EvalConfig", ")", "\n", "}", "\n", "scope", ":=", "registerBuiltins", "(", "config", ".", "GlobalScope", ")", "\n", "implicitMap", ":=", "map", "[", "ast", ".", "Type", "]", "map", "[", "ast", ".", "Type", "]", "string", "{", "ast", ".", "TypeFloat", ":", "{", "ast", ".", "TypeInt", ":", "\"", "\"", ",", "ast", ".", "TypeString", ":", "\"", "\"", ",", "}", ",", "ast", ".", "TypeInt", ":", "{", "ast", ".", "TypeFloat", ":", "\"", "\"", ",", "ast", ".", "TypeString", ":", "\"", "\"", ",", "}", ",", "ast", ".", "TypeString", ":", "{", "ast", ".", "TypeInt", ":", "\"", "\"", ",", "ast", ".", "TypeFloat", ":", "\"", "\"", ",", "ast", ".", "TypeBool", ":", "\"", "\"", ",", "}", ",", "ast", ".", "TypeBool", ":", "{", "ast", ".", "TypeString", ":", "\"", "\"", ",", "}", ",", "}", "\n\n", "// Build our own semantic checks that we always run", "tv", ":=", "&", "TypeCheck", "{", "Scope", ":", "scope", ",", "Implicit", ":", "implicitMap", "}", "\n", "ic", ":=", "&", "IdentifierCheck", "{", "Scope", ":", "scope", "}", "\n\n", "// Build up the semantic checks for execution", "checks", ":=", "make", "(", "[", "]", "SemanticChecker", ",", "len", "(", "config", ".", "SemanticChecks", ")", ",", "len", "(", "config", ".", "SemanticChecks", ")", "+", "2", ")", "\n", "copy", "(", "checks", ",", "config", ".", "SemanticChecks", ")", "\n", "checks", "=", "append", "(", "checks", ",", "ic", ".", "Visit", ")", "\n", "checks", "=", "append", "(", "checks", ",", "tv", ".", "Visit", ")", "\n\n", "// Run the semantic checks", "for", "_", ",", "check", ":=", "range", "checks", "{", "if", "err", ":=", "check", "(", "root", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "ast", ".", "TypeInvalid", ",", "err", "\n", "}", "\n", "}", "\n\n", "// Execute", "v", ":=", "&", "evalVisitor", "{", "Scope", ":", "scope", "}", "\n", "return", "v", ".", "Visit", "(", "root", ")", "\n", "}" ]
// Eval evaluates the given AST tree and returns its output value, the type // of the output, and any error that occurred.
[ "Eval", "evaluates", "the", "given", "AST", "tree", "and", "returns", "its", "output", "value", "the", "type", "of", "the", "output", "and", "any", "error", "that", "occurred", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/eval.go#L106-L154
5,576
hashicorp/hil
eval.go
evalNode
func evalNode(raw ast.Node) (EvalNode, error) { switch n := raw.(type) { case *ast.Index: return &evalIndex{n}, nil case *ast.Call: return &evalCall{n}, nil case *ast.Conditional: return &evalConditional{n}, nil case *ast.Output: return &evalOutput{n}, nil case *ast.LiteralNode: return &evalLiteralNode{n}, nil case *ast.VariableAccess: return &evalVariableAccess{n}, nil default: en, ok := n.(EvalNode) if !ok { return nil, fmt.Errorf("node doesn't support evaluation: %#v", raw) } return en, nil } }
go
func evalNode(raw ast.Node) (EvalNode, error) { switch n := raw.(type) { case *ast.Index: return &evalIndex{n}, nil case *ast.Call: return &evalCall{n}, nil case *ast.Conditional: return &evalConditional{n}, nil case *ast.Output: return &evalOutput{n}, nil case *ast.LiteralNode: return &evalLiteralNode{n}, nil case *ast.VariableAccess: return &evalVariableAccess{n}, nil default: en, ok := n.(EvalNode) if !ok { return nil, fmt.Errorf("node doesn't support evaluation: %#v", raw) } return en, nil } }
[ "func", "evalNode", "(", "raw", "ast", ".", "Node", ")", "(", "EvalNode", ",", "error", ")", "{", "switch", "n", ":=", "raw", ".", "(", "type", ")", "{", "case", "*", "ast", ".", "Index", ":", "return", "&", "evalIndex", "{", "n", "}", ",", "nil", "\n", "case", "*", "ast", ".", "Call", ":", "return", "&", "evalCall", "{", "n", "}", ",", "nil", "\n", "case", "*", "ast", ".", "Conditional", ":", "return", "&", "evalConditional", "{", "n", "}", ",", "nil", "\n", "case", "*", "ast", ".", "Output", ":", "return", "&", "evalOutput", "{", "n", "}", ",", "nil", "\n", "case", "*", "ast", ".", "LiteralNode", ":", "return", "&", "evalLiteralNode", "{", "n", "}", ",", "nil", "\n", "case", "*", "ast", ".", "VariableAccess", ":", "return", "&", "evalVariableAccess", "{", "n", "}", ",", "nil", "\n", "default", ":", "en", ",", "ok", ":=", "n", ".", "(", "EvalNode", ")", "\n", "if", "!", "ok", "{", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "raw", ")", "\n", "}", "\n\n", "return", "en", ",", "nil", "\n", "}", "\n", "}" ]
// evalNode is a private function that returns an EvalNode for built-in // types as well as any other EvalNode implementations.
[ "evalNode", "is", "a", "private", "function", "that", "returns", "an", "EvalNode", "for", "built", "-", "in", "types", "as", "well", "as", "any", "other", "EvalNode", "implementations", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/eval.go#L237-L259
5,577
hashicorp/hil
ast/literal.go
NewLiteralNode
func NewLiteralNode(value interface{}, pos Pos) (*LiteralNode, error) { goType := reflect.TypeOf(value) var hilType Type switch goType.Kind() { case reflect.Bool: hilType = TypeBool case reflect.Int: hilType = TypeInt case reflect.Float64: hilType = TypeFloat case reflect.String: hilType = TypeString default: return nil, fmt.Errorf("unsupported literal node type: %T", value) } return &LiteralNode{ Value: value, Typex: hilType, Posx: pos, }, nil }
go
func NewLiteralNode(value interface{}, pos Pos) (*LiteralNode, error) { goType := reflect.TypeOf(value) var hilType Type switch goType.Kind() { case reflect.Bool: hilType = TypeBool case reflect.Int: hilType = TypeInt case reflect.Float64: hilType = TypeFloat case reflect.String: hilType = TypeString default: return nil, fmt.Errorf("unsupported literal node type: %T", value) } return &LiteralNode{ Value: value, Typex: hilType, Posx: pos, }, nil }
[ "func", "NewLiteralNode", "(", "value", "interface", "{", "}", ",", "pos", "Pos", ")", "(", "*", "LiteralNode", ",", "error", ")", "{", "goType", ":=", "reflect", ".", "TypeOf", "(", "value", ")", "\n", "var", "hilType", "Type", "\n\n", "switch", "goType", ".", "Kind", "(", ")", "{", "case", "reflect", ".", "Bool", ":", "hilType", "=", "TypeBool", "\n", "case", "reflect", ".", "Int", ":", "hilType", "=", "TypeInt", "\n", "case", "reflect", ".", "Float64", ":", "hilType", "=", "TypeFloat", "\n", "case", "reflect", ".", "String", ":", "hilType", "=", "TypeString", "\n", "default", ":", "return", "nil", ",", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "value", ")", "\n", "}", "\n\n", "return", "&", "LiteralNode", "{", "Value", ":", "value", ",", "Typex", ":", "hilType", ",", "Posx", ":", "pos", ",", "}", ",", "nil", "\n", "}" ]
// NewLiteralNode returns a new literal node representing the given // literal Go value, which must correspond to one of the primitive types // supported by HIL. Lists and maps cannot currently be constructed via // this function. // // If an inappropriately-typed value is provided, this function will // return an error. The main intended use of this function is to produce // "synthetic" literals from constants in code, where the value type is // well known at compile time. To easily store these in global variables, // see also MustNewLiteralNode.
[ "NewLiteralNode", "returns", "a", "new", "literal", "node", "representing", "the", "given", "literal", "Go", "value", "which", "must", "correspond", "to", "one", "of", "the", "primitive", "types", "supported", "by", "HIL", ".", "Lists", "and", "maps", "cannot", "currently", "be", "constructed", "via", "this", "function", ".", "If", "an", "inappropriately", "-", "typed", "value", "is", "provided", "this", "function", "will", "return", "an", "error", ".", "The", "main", "intended", "use", "of", "this", "function", "is", "to", "produce", "synthetic", "literals", "from", "constants", "in", "code", "where", "the", "value", "type", "is", "well", "known", "at", "compile", "time", ".", "To", "easily", "store", "these", "in", "global", "variables", "see", "also", "MustNewLiteralNode", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L26-L48
5,578
hashicorp/hil
ast/literal.go
MustNewLiteralNode
func MustNewLiteralNode(value interface{}, pos Pos) *LiteralNode { node, err := NewLiteralNode(value, pos) if err != nil { panic(err) } return node }
go
func MustNewLiteralNode(value interface{}, pos Pos) *LiteralNode { node, err := NewLiteralNode(value, pos) if err != nil { panic(err) } return node }
[ "func", "MustNewLiteralNode", "(", "value", "interface", "{", "}", ",", "pos", "Pos", ")", "*", "LiteralNode", "{", "node", ",", "err", ":=", "NewLiteralNode", "(", "value", ",", "pos", ")", "\n", "if", "err", "!=", "nil", "{", "panic", "(", "err", ")", "\n", "}", "\n", "return", "node", "\n", "}" ]
// MustNewLiteralNode wraps NewLiteralNode and panics if an error is // returned, thus allowing valid literal nodes to be easily assigned to // global variables.
[ "MustNewLiteralNode", "wraps", "NewLiteralNode", "and", "panics", "if", "an", "error", "is", "returned", "thus", "allowing", "valid", "literal", "nodes", "to", "be", "easily", "assigned", "to", "global", "variables", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L53-L59
5,579
hashicorp/hil
ast/literal.go
IsUnknown
func (n *LiteralNode) IsUnknown() bool { return IsUnknown(Variable{ Type: n.Typex, Value: n.Value, }) }
go
func (n *LiteralNode) IsUnknown() bool { return IsUnknown(Variable{ Type: n.Typex, Value: n.Value, }) }
[ "func", "(", "n", "*", "LiteralNode", ")", "IsUnknown", "(", ")", "bool", "{", "return", "IsUnknown", "(", "Variable", "{", "Type", ":", "n", ".", "Typex", ",", "Value", ":", "n", ".", "Value", ",", "}", ")", "\n", "}" ]
// IsUnknown returns true either if the node's value is itself unknown // of if it is a collection containing any unknown elements, deeply.
[ "IsUnknown", "returns", "true", "either", "if", "the", "node", "s", "value", "is", "itself", "unknown", "of", "if", "it", "is", "a", "collection", "containing", "any", "unknown", "elements", "deeply", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L83-L88
5,580
hashicorp/hil
ast/unknown.go
IsUnknown
func IsUnknown(v Variable) bool { // If it is unknown itself, return true if v.Type == TypeUnknown { return true } // If it is a container type, check the values switch v.Type { case TypeList: for _, el := range v.Value.([]Variable) { if IsUnknown(el) { return true } } case TypeMap: for _, el := range v.Value.(map[string]Variable) { if IsUnknown(el) { return true } } default: } // Not a container type or survive the above checks return false }
go
func IsUnknown(v Variable) bool { // If it is unknown itself, return true if v.Type == TypeUnknown { return true } // If it is a container type, check the values switch v.Type { case TypeList: for _, el := range v.Value.([]Variable) { if IsUnknown(el) { return true } } case TypeMap: for _, el := range v.Value.(map[string]Variable) { if IsUnknown(el) { return true } } default: } // Not a container type or survive the above checks return false }
[ "func", "IsUnknown", "(", "v", "Variable", ")", "bool", "{", "// If it is unknown itself, return true", "if", "v", ".", "Type", "==", "TypeUnknown", "{", "return", "true", "\n", "}", "\n\n", "// If it is a container type, check the values", "switch", "v", ".", "Type", "{", "case", "TypeList", ":", "for", "_", ",", "el", ":=", "range", "v", ".", "Value", ".", "(", "[", "]", "Variable", ")", "{", "if", "IsUnknown", "(", "el", ")", "{", "return", "true", "\n", "}", "\n", "}", "\n", "case", "TypeMap", ":", "for", "_", ",", "el", ":=", "range", "v", ".", "Value", ".", "(", "map", "[", "string", "]", "Variable", ")", "{", "if", "IsUnknown", "(", "el", ")", "{", "return", "true", "\n", "}", "\n", "}", "\n", "default", ":", "}", "\n\n", "// Not a container type or survive the above checks", "return", "false", "\n", "}" ]
// IsUnknown reports whether a variable is unknown or contains any value // that is unknown. This will recurse into lists and maps and so on.
[ "IsUnknown", "reports", "whether", "a", "variable", "is", "unknown", "or", "contains", "any", "value", "that", "is", "unknown", ".", "This", "will", "recurse", "into", "lists", "and", "maps", "and", "so", "on", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/unknown.go#L5-L30
5,581
hashicorp/hil
walk.go
Walk
func Walk(v interface{}, cb WalkFn) error { walker := &interpolationWalker{F: cb} return reflectwalk.Walk(v, walker) }
go
func Walk(v interface{}, cb WalkFn) error { walker := &interpolationWalker{F: cb} return reflectwalk.Walk(v, walker) }
[ "func", "Walk", "(", "v", "interface", "{", "}", ",", "cb", "WalkFn", ")", "error", "{", "walker", ":=", "&", "interpolationWalker", "{", "F", ":", "cb", "}", "\n", "return", "reflectwalk", ".", "Walk", "(", "v", ",", "walker", ")", "\n", "}" ]
// Walk will walk an arbitrary Go structure and parse any string as an // HIL program and call the callback cb to determine what to replace it // with. // // This function is very useful for arbitrary HIL program interpolation // across a complex configuration structure. Due to the heavy use of // reflection in this function, it is recommend to write many unit tests // with your typical configuration structures to hilp mitigate the risk // of panics.
[ "Walk", "will", "walk", "an", "arbitrary", "Go", "structure", "and", "parse", "any", "string", "as", "an", "HIL", "program", "and", "call", "the", "callback", "cb", "to", "determine", "what", "to", "replace", "it", "with", ".", "This", "function", "is", "very", "useful", "for", "arbitrary", "HIL", "program", "interpolation", "across", "a", "complex", "configuration", "structure", ".", "Due", "to", "the", "heavy", "use", "of", "reflection", "in", "this", "function", "it", "is", "recommend", "to", "write", "many", "unit", "tests", "with", "your", "typical", "configuration", "structures", "to", "hilp", "mitigate", "the", "risk", "of", "panics", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/walk.go#L47-L50
5,582
hashicorp/hil
scanner/scanner.go
Scan
func Scan(s string, startPos ast.Pos) <-chan *Token { ch := make(chan *Token) go scan(s, ch, startPos) return ch }
go
func Scan(s string, startPos ast.Pos) <-chan *Token { ch := make(chan *Token) go scan(s, ch, startPos) return ch }
[ "func", "Scan", "(", "s", "string", ",", "startPos", "ast", ".", "Pos", ")", "<-", "chan", "*", "Token", "{", "ch", ":=", "make", "(", "chan", "*", "Token", ")", "\n", "go", "scan", "(", "s", ",", "ch", ",", "startPos", ")", "\n", "return", "ch", "\n", "}" ]
// Scan returns a channel that recieves Tokens from the given input string. // // The scanner's job is just to partition the string into meaningful parts. // It doesn't do any transformation of the raw input string, so the caller // must deal with any further interpretation required, such as parsing INTEGER // tokens into real ints, or dealing with escape sequences in LITERAL or // STRING tokens. // // Strings in the returned tokens are slices from the original string. // // startPos should be set to ast.InitPos unless the caller knows that // this interpolation string is part of a larger file and knows the position // of the first character in that larger file.
[ "Scan", "returns", "a", "channel", "that", "recieves", "Tokens", "from", "the", "given", "input", "string", ".", "The", "scanner", "s", "job", "is", "just", "to", "partition", "the", "string", "into", "meaningful", "parts", ".", "It", "doesn", "t", "do", "any", "transformation", "of", "the", "raw", "input", "string", "so", "the", "caller", "must", "deal", "with", "any", "further", "interpretation", "required", "such", "as", "parsing", "INTEGER", "tokens", "into", "real", "ints", "or", "dealing", "with", "escape", "sequences", "in", "LITERAL", "or", "STRING", "tokens", ".", "Strings", "in", "the", "returned", "tokens", "are", "slices", "from", "the", "original", "string", ".", "startPos", "should", "be", "set", "to", "ast", ".", "InitPos", "unless", "the", "caller", "knows", "that", "this", "interpolation", "string", "is", "part", "of", "a", "larger", "file", "and", "knows", "the", "position", "of", "the", "first", "character", "in", "that", "larger", "file", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/scanner.go#L23-L27
5,583
hashicorp/hil
scanner/scanner.go
scanIdentifier
func scanIdentifier(s string) (string, int) { byteLen := 0 runeLen := 0 for { if byteLen >= len(s) { break } nextRune, size := utf8.DecodeRuneInString(s[byteLen:]) if !(nextRune == '_' || nextRune == '-' || nextRune == '.' || nextRune == '*' || unicode.IsNumber(nextRune) || unicode.IsLetter(nextRune) || unicode.IsMark(nextRune)) { break } // If we reach a star, it must be between periods to be part // of the same identifier. if nextRune == '*' && s[byteLen-1] != '.' { break } // If our previous character was a star, then the current must // be period. Otherwise, undo that and exit. if byteLen > 0 && s[byteLen-1] == '*' && nextRune != '.' { byteLen-- if s[byteLen-1] == '.' { byteLen-- } break } byteLen = byteLen + size runeLen = runeLen + 1 } return s[:byteLen], runeLen }
go
func scanIdentifier(s string) (string, int) { byteLen := 0 runeLen := 0 for { if byteLen >= len(s) { break } nextRune, size := utf8.DecodeRuneInString(s[byteLen:]) if !(nextRune == '_' || nextRune == '-' || nextRune == '.' || nextRune == '*' || unicode.IsNumber(nextRune) || unicode.IsLetter(nextRune) || unicode.IsMark(nextRune)) { break } // If we reach a star, it must be between periods to be part // of the same identifier. if nextRune == '*' && s[byteLen-1] != '.' { break } // If our previous character was a star, then the current must // be period. Otherwise, undo that and exit. if byteLen > 0 && s[byteLen-1] == '*' && nextRune != '.' { byteLen-- if s[byteLen-1] == '.' { byteLen-- } break } byteLen = byteLen + size runeLen = runeLen + 1 } return s[:byteLen], runeLen }
[ "func", "scanIdentifier", "(", "s", "string", ")", "(", "string", ",", "int", ")", "{", "byteLen", ":=", "0", "\n", "runeLen", ":=", "0", "\n", "for", "{", "if", "byteLen", ">=", "len", "(", "s", ")", "{", "break", "\n", "}", "\n\n", "nextRune", ",", "size", ":=", "utf8", ".", "DecodeRuneInString", "(", "s", "[", "byteLen", ":", "]", ")", "\n", "if", "!", "(", "nextRune", "==", "'_'", "||", "nextRune", "==", "'-'", "||", "nextRune", "==", "'.'", "||", "nextRune", "==", "'*'", "||", "unicode", ".", "IsNumber", "(", "nextRune", ")", "||", "unicode", ".", "IsLetter", "(", "nextRune", ")", "||", "unicode", ".", "IsMark", "(", "nextRune", ")", ")", "{", "break", "\n", "}", "\n\n", "// If we reach a star, it must be between periods to be part", "// of the same identifier.", "if", "nextRune", "==", "'*'", "&&", "s", "[", "byteLen", "-", "1", "]", "!=", "'.'", "{", "break", "\n", "}", "\n\n", "// If our previous character was a star, then the current must", "// be period. Otherwise, undo that and exit.", "if", "byteLen", ">", "0", "&&", "s", "[", "byteLen", "-", "1", "]", "==", "'*'", "&&", "nextRune", "!=", "'.'", "{", "byteLen", "--", "\n", "if", "s", "[", "byteLen", "-", "1", "]", "==", "'.'", "{", "byteLen", "--", "\n", "}", "\n\n", "break", "\n", "}", "\n\n", "byteLen", "=", "byteLen", "+", "size", "\n", "runeLen", "=", "runeLen", "+", "1", "\n", "}", "\n\n", "return", "s", "[", ":", "byteLen", "]", ",", "runeLen", "\n", "}" ]
// scanIdentifier returns the extent of the prefix of the string that // represents a valid identifier, along with the length of that prefix // in runes. // // Identifiers may contain utf8-encoded non-Latin letters, which will // cause the returned "rune length" to be shorter than the byte length // of the returned string.
[ "scanIdentifier", "returns", "the", "extent", "of", "the", "prefix", "of", "the", "string", "that", "represents", "a", "valid", "identifier", "along", "with", "the", "length", "of", "that", "prefix", "in", "runes", ".", "Identifiers", "may", "contain", "utf8", "-", "encoded", "non", "-", "Latin", "letters", "which", "will", "cause", "the", "returned", "rune", "length", "to", "be", "shorter", "than", "the", "byte", "length", "of", "the", "returned", "string", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/scanner.go#L477-L518
5,584
hashicorp/hil
ast/scope.go
NewVariable
func NewVariable(v interface{}) (result Variable, err error) { switch v := reflect.ValueOf(v); v.Kind() { case reflect.String: result.Type = TypeString default: err = fmt.Errorf("Unknown type: %s", v.Kind()) } result.Value = v return }
go
func NewVariable(v interface{}) (result Variable, err error) { switch v := reflect.ValueOf(v); v.Kind() { case reflect.String: result.Type = TypeString default: err = fmt.Errorf("Unknown type: %s", v.Kind()) } result.Value = v return }
[ "func", "NewVariable", "(", "v", "interface", "{", "}", ")", "(", "result", "Variable", ",", "err", "error", ")", "{", "switch", "v", ":=", "reflect", ".", "ValueOf", "(", "v", ")", ";", "v", ".", "Kind", "(", ")", "{", "case", "reflect", ".", "String", ":", "result", ".", "Type", "=", "TypeString", "\n", "default", ":", "err", "=", "fmt", ".", "Errorf", "(", "\"", "\"", ",", "v", ".", "Kind", "(", ")", ")", "\n", "}", "\n\n", "result", ".", "Value", "=", "v", "\n", "return", "\n", "}" ]
// NewVariable creates a new Variable for the given value. This will // attempt to infer the correct type. If it can't, an error will be returned.
[ "NewVariable", "creates", "a", "new", "Variable", "for", "the", "given", "value", ".", "This", "will", "attempt", "to", "infer", "the", "correct", "type", ".", "If", "it", "can", "t", "an", "error", "will", "be", "returned", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/scope.go#L24-L34
5,585
hashicorp/hil
ast/scope.go
String
func (v Variable) String() string { return fmt.Sprintf("{Variable (%s): %+v}", v.Type, v.Value) }
go
func (v Variable) String() string { return fmt.Sprintf("{Variable (%s): %+v}", v.Type, v.Value) }
[ "func", "(", "v", "Variable", ")", "String", "(", ")", "string", "{", "return", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "v", ".", "Type", ",", "v", ".", "Value", ")", "\n", "}" ]
// String implements Stringer on Variable, displaying the type and value // of the Variable.
[ "String", "implements", "Stringer", "on", "Variable", "displaying", "the", "type", "and", "value", "of", "the", "Variable", "." ]
97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63
https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/scope.go#L38-L40
5,586
minio/sha256-simd
sha256blockAvx512_amd64.go
NewAvx512
func NewAvx512(a512srv *Avx512Server) hash.Hash { uid := atomic.AddUint64(&uidCounter, 1) return &Avx512Digest{uid: uid, a512srv: a512srv} }
go
func NewAvx512(a512srv *Avx512Server) hash.Hash { uid := atomic.AddUint64(&uidCounter, 1) return &Avx512Digest{uid: uid, a512srv: a512srv} }
[ "func", "NewAvx512", "(", "a512srv", "*", "Avx512Server", ")", "hash", ".", "Hash", "{", "uid", ":=", "atomic", ".", "AddUint64", "(", "&", "uidCounter", ",", "1", ")", "\n", "return", "&", "Avx512Digest", "{", "uid", ":", "uid", ",", "a512srv", ":", "a512srv", "}", "\n", "}" ]
// NewAvx512 - initialize sha256 Avx512 implementation.
[ "NewAvx512", "-", "initialize", "sha256", "Avx512", "implementation", "." ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L40-L43
5,587
minio/sha256-simd
sha256blockAvx512_amd64.go
Reset
func (d *Avx512Digest) Reset() { d.a512srv.blocksCh <- blockInput{uid: d.uid, reset: true} d.nx = 0 d.len = 0 d.final = false }
go
func (d *Avx512Digest) Reset() { d.a512srv.blocksCh <- blockInput{uid: d.uid, reset: true} d.nx = 0 d.len = 0 d.final = false }
[ "func", "(", "d", "*", "Avx512Digest", ")", "Reset", "(", ")", "{", "d", ".", "a512srv", ".", "blocksCh", "<-", "blockInput", "{", "uid", ":", "d", ".", "uid", ",", "reset", ":", "true", "}", "\n", "d", ".", "nx", "=", "0", "\n", "d", ".", "len", "=", "0", "\n", "d", ".", "final", "=", "false", "\n", "}" ]
// Reset - reset sha digest to its initial values
[ "Reset", "-", "reset", "sha", "digest", "to", "its", "initial", "values" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L63-L68
5,588
minio/sha256-simd
sha256blockAvx512_amd64.go
Write
func (d *Avx512Digest) Write(p []byte) (nn int, err error) { if d.final { return 0, errors.New("Avx512Digest already finalized. Reset first before writing again") } nn = len(p) d.len += uint64(nn) if d.nx > 0 { n := copy(d.x[d.nx:], p) d.nx += n if d.nx == chunk { d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: d.x[:]} d.nx = 0 } p = p[n:] } if len(p) >= chunk { n := len(p) &^ (chunk - 1) d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: p[:n]} p = p[n:] } if len(p) > 0 { d.nx = copy(d.x[:], p) } return }
go
func (d *Avx512Digest) Write(p []byte) (nn int, err error) { if d.final { return 0, errors.New("Avx512Digest already finalized. Reset first before writing again") } nn = len(p) d.len += uint64(nn) if d.nx > 0 { n := copy(d.x[d.nx:], p) d.nx += n if d.nx == chunk { d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: d.x[:]} d.nx = 0 } p = p[n:] } if len(p) >= chunk { n := len(p) &^ (chunk - 1) d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: p[:n]} p = p[n:] } if len(p) > 0 { d.nx = copy(d.x[:], p) } return }
[ "func", "(", "d", "*", "Avx512Digest", ")", "Write", "(", "p", "[", "]", "byte", ")", "(", "nn", "int", ",", "err", "error", ")", "{", "if", "d", ".", "final", "{", "return", "0", ",", "errors", ".", "New", "(", "\"", "\"", ")", "\n", "}", "\n\n", "nn", "=", "len", "(", "p", ")", "\n", "d", ".", "len", "+=", "uint64", "(", "nn", ")", "\n", "if", "d", ".", "nx", ">", "0", "{", "n", ":=", "copy", "(", "d", ".", "x", "[", "d", ".", "nx", ":", "]", ",", "p", ")", "\n", "d", ".", "nx", "+=", "n", "\n", "if", "d", ".", "nx", "==", "chunk", "{", "d", ".", "a512srv", ".", "blocksCh", "<-", "blockInput", "{", "uid", ":", "d", ".", "uid", ",", "msg", ":", "d", ".", "x", "[", ":", "]", "}", "\n", "d", ".", "nx", "=", "0", "\n", "}", "\n", "p", "=", "p", "[", "n", ":", "]", "\n", "}", "\n", "if", "len", "(", "p", ")", ">=", "chunk", "{", "n", ":=", "len", "(", "p", ")", "&^", "(", "chunk", "-", "1", ")", "\n", "d", ".", "a512srv", ".", "blocksCh", "<-", "blockInput", "{", "uid", ":", "d", ".", "uid", ",", "msg", ":", "p", "[", ":", "n", "]", "}", "\n", "p", "=", "p", "[", "n", ":", "]", "\n", "}", "\n", "if", "len", "(", "p", ")", ">", "0", "{", "d", ".", "nx", "=", "copy", "(", "d", ".", "x", "[", ":", "]", ",", "p", ")", "\n", "}", "\n", "return", "\n", "}" ]
// Write to digest
[ "Write", "to", "digest" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L71-L97
5,589
minio/sha256-simd
sha256blockAvx512_amd64.go
Sum
func (d *Avx512Digest) Sum(in []byte) (result []byte) { if d.final { return append(in, d.result[:]...) } trail := make([]byte, 0, 128) len := d.len // Padding. Add a 1 bit and 0 bits until 56 bytes mod 64. var tmp [64]byte tmp[0] = 0x80 if len%64 < 56 { trail = append(d.x[:d.nx], tmp[0:56-len%64]...) } else { trail = append(d.x[:d.nx], tmp[0:64+56-len%64]...) } d.nx = 0 // Length in bits. len <<= 3 for i := uint(0); i < 8; i++ { tmp[i] = byte(len >> (56 - 8*i)) } trail = append(trail, tmp[0:8]...) sumCh := make(chan [Size]byte) d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: trail, final: true, sumCh: sumCh} d.result = <-sumCh d.final = true return append(in, d.result[:]...) }
go
func (d *Avx512Digest) Sum(in []byte) (result []byte) { if d.final { return append(in, d.result[:]...) } trail := make([]byte, 0, 128) len := d.len // Padding. Add a 1 bit and 0 bits until 56 bytes mod 64. var tmp [64]byte tmp[0] = 0x80 if len%64 < 56 { trail = append(d.x[:d.nx], tmp[0:56-len%64]...) } else { trail = append(d.x[:d.nx], tmp[0:64+56-len%64]...) } d.nx = 0 // Length in bits. len <<= 3 for i := uint(0); i < 8; i++ { tmp[i] = byte(len >> (56 - 8*i)) } trail = append(trail, tmp[0:8]...) sumCh := make(chan [Size]byte) d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: trail, final: true, sumCh: sumCh} d.result = <-sumCh d.final = true return append(in, d.result[:]...) }
[ "func", "(", "d", "*", "Avx512Digest", ")", "Sum", "(", "in", "[", "]", "byte", ")", "(", "result", "[", "]", "byte", ")", "{", "if", "d", ".", "final", "{", "return", "append", "(", "in", ",", "d", ".", "result", "[", ":", "]", "...", ")", "\n", "}", "\n\n", "trail", ":=", "make", "(", "[", "]", "byte", ",", "0", ",", "128", ")", "\n\n", "len", ":=", "d", ".", "len", "\n", "// Padding. Add a 1 bit and 0 bits until 56 bytes mod 64.", "var", "tmp", "[", "64", "]", "byte", "\n", "tmp", "[", "0", "]", "=", "0x80", "\n", "if", "len", "%", "64", "<", "56", "{", "trail", "=", "append", "(", "d", ".", "x", "[", ":", "d", ".", "nx", "]", ",", "tmp", "[", "0", ":", "56", "-", "len", "%", "64", "]", "...", ")", "\n", "}", "else", "{", "trail", "=", "append", "(", "d", ".", "x", "[", ":", "d", ".", "nx", "]", ",", "tmp", "[", "0", ":", "64", "+", "56", "-", "len", "%", "64", "]", "...", ")", "\n", "}", "\n", "d", ".", "nx", "=", "0", "\n\n", "// Length in bits.", "len", "<<=", "3", "\n", "for", "i", ":=", "uint", "(", "0", ")", ";", "i", "<", "8", ";", "i", "++", "{", "tmp", "[", "i", "]", "=", "byte", "(", "len", ">>", "(", "56", "-", "8", "*", "i", ")", ")", "\n", "}", "\n", "trail", "=", "append", "(", "trail", ",", "tmp", "[", "0", ":", "8", "]", "...", ")", "\n\n", "sumCh", ":=", "make", "(", "chan", "[", "Size", "]", "byte", ")", "\n", "d", ".", "a512srv", ".", "blocksCh", "<-", "blockInput", "{", "uid", ":", "d", ".", "uid", ",", "msg", ":", "trail", ",", "final", ":", "true", ",", "sumCh", ":", "sumCh", "}", "\n", "d", ".", "result", "=", "<-", "sumCh", "\n", "d", ".", "final", "=", "true", "\n", "return", "append", "(", "in", ",", "d", ".", "result", "[", ":", "]", "...", ")", "\n", "}" ]
// Sum - Return sha256 sum in bytes
[ "Sum", "-", "Return", "sha256", "sum", "in", "bytes" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L100-L131
5,590
minio/sha256-simd
sha256blockAvx512_amd64.go
blockAvx512
func blockAvx512(digests *[512]byte, input [16][]byte, mask []uint64) [16][Size]byte { scratch := [512]byte{} sha256X16Avx512(digests, &scratch, &table, mask, input) output := [16][Size]byte{} for i := 0; i < 16; i++ { output[i] = getDigest(i, digests[:]) } return output }
go
func blockAvx512(digests *[512]byte, input [16][]byte, mask []uint64) [16][Size]byte { scratch := [512]byte{} sha256X16Avx512(digests, &scratch, &table, mask, input) output := [16][Size]byte{} for i := 0; i < 16; i++ { output[i] = getDigest(i, digests[:]) } return output }
[ "func", "blockAvx512", "(", "digests", "*", "[", "512", "]", "byte", ",", "input", "[", "16", "]", "[", "]", "byte", ",", "mask", "[", "]", "uint64", ")", "[", "16", "]", "[", "Size", "]", "byte", "{", "scratch", ":=", "[", "512", "]", "byte", "{", "}", "\n", "sha256X16Avx512", "(", "digests", ",", "&", "scratch", ",", "&", "table", ",", "mask", ",", "input", ")", "\n\n", "output", ":=", "[", "16", "]", "[", "Size", "]", "byte", "{", "}", "\n", "for", "i", ":=", "0", ";", "i", "<", "16", ";", "i", "++", "{", "output", "[", "i", "]", "=", "getDigest", "(", "i", ",", "digests", "[", ":", "]", ")", "\n", "}", "\n\n", "return", "output", "\n", "}" ]
// Interface function to assembly ode
[ "Interface", "function", "to", "assembly", "ode" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L264-L275
5,591
minio/sha256-simd
sha256blockAvx512_amd64.go
NewAvx512Server
func NewAvx512Server() *Avx512Server { a512srv := &Avx512Server{} a512srv.digests = make(map[uint64][Size]byte) a512srv.blocksCh = make(chan blockInput) // Start a single thread for reading from the input channel go a512srv.Process() return a512srv }
go
func NewAvx512Server() *Avx512Server { a512srv := &Avx512Server{} a512srv.digests = make(map[uint64][Size]byte) a512srv.blocksCh = make(chan blockInput) // Start a single thread for reading from the input channel go a512srv.Process() return a512srv }
[ "func", "NewAvx512Server", "(", ")", "*", "Avx512Server", "{", "a512srv", ":=", "&", "Avx512Server", "{", "}", "\n", "a512srv", ".", "digests", "=", "make", "(", "map", "[", "uint64", "]", "[", "Size", "]", "byte", ")", "\n", "a512srv", ".", "blocksCh", "=", "make", "(", "chan", "blockInput", ")", "\n\n", "// Start a single thread for reading from the input channel", "go", "a512srv", ".", "Process", "(", ")", "\n", "return", "a512srv", "\n", "}" ]
// NewAvx512Server - Create new object for parallel processing handling
[ "NewAvx512Server", "-", "Create", "new", "object", "for", "parallel", "processing", "handling" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L311-L319
5,592
minio/sha256-simd
sha256blockAvx512_amd64.go
Process
func (a512srv *Avx512Server) Process() { for { select { case block := <-a512srv.blocksCh: if block.reset { a512srv.reset(block.uid) continue } index := block.uid & 0xf // fmt.Println("Adding message:", block.uid, index) if a512srv.lanes[index].block != nil { // If slot is already filled, process all inputs //fmt.Println("Invoking Blocks()") a512srv.blocks() } a512srv.totalIn++ a512srv.lanes[index] = Avx512LaneInfo{uid: block.uid, block: block.msg} if block.final { a512srv.lanes[index].outputCh = block.sumCh } if a512srv.totalIn == len(a512srv.lanes) { // fmt.Println("Invoking Blocks() while FULL: ") a512srv.blocks() } // TODO: test with larger timeout case <-time.After(1 * time.Microsecond): for _, lane := range a512srv.lanes { if lane.block != nil { // check if there is any input to process // fmt.Println("Invoking Blocks() on TIMEOUT: ") a512srv.blocks() break // we are done } } } } }
go
func (a512srv *Avx512Server) Process() { for { select { case block := <-a512srv.blocksCh: if block.reset { a512srv.reset(block.uid) continue } index := block.uid & 0xf // fmt.Println("Adding message:", block.uid, index) if a512srv.lanes[index].block != nil { // If slot is already filled, process all inputs //fmt.Println("Invoking Blocks()") a512srv.blocks() } a512srv.totalIn++ a512srv.lanes[index] = Avx512LaneInfo{uid: block.uid, block: block.msg} if block.final { a512srv.lanes[index].outputCh = block.sumCh } if a512srv.totalIn == len(a512srv.lanes) { // fmt.Println("Invoking Blocks() while FULL: ") a512srv.blocks() } // TODO: test with larger timeout case <-time.After(1 * time.Microsecond): for _, lane := range a512srv.lanes { if lane.block != nil { // check if there is any input to process // fmt.Println("Invoking Blocks() on TIMEOUT: ") a512srv.blocks() break // we are done } } } } }
[ "func", "(", "a512srv", "*", "Avx512Server", ")", "Process", "(", ")", "{", "for", "{", "select", "{", "case", "block", ":=", "<-", "a512srv", ".", "blocksCh", ":", "if", "block", ".", "reset", "{", "a512srv", ".", "reset", "(", "block", ".", "uid", ")", "\n", "continue", "\n", "}", "\n", "index", ":=", "block", ".", "uid", "&", "0xf", "\n", "// fmt.Println(\"Adding message:\", block.uid, index)", "if", "a512srv", ".", "lanes", "[", "index", "]", ".", "block", "!=", "nil", "{", "// If slot is already filled, process all inputs", "//fmt.Println(\"Invoking Blocks()\")", "a512srv", ".", "blocks", "(", ")", "\n", "}", "\n", "a512srv", ".", "totalIn", "++", "\n", "a512srv", ".", "lanes", "[", "index", "]", "=", "Avx512LaneInfo", "{", "uid", ":", "block", ".", "uid", ",", "block", ":", "block", ".", "msg", "}", "\n", "if", "block", ".", "final", "{", "a512srv", ".", "lanes", "[", "index", "]", ".", "outputCh", "=", "block", ".", "sumCh", "\n", "}", "\n", "if", "a512srv", ".", "totalIn", "==", "len", "(", "a512srv", ".", "lanes", ")", "{", "// fmt.Println(\"Invoking Blocks() while FULL: \")", "a512srv", ".", "blocks", "(", ")", "\n", "}", "\n\n", "// TODO: test with larger timeout", "case", "<-", "time", ".", "After", "(", "1", "*", "time", ".", "Microsecond", ")", ":", "for", "_", ",", "lane", ":=", "range", "a512srv", ".", "lanes", "{", "if", "lane", ".", "block", "!=", "nil", "{", "// check if there is any input to process", "// fmt.Println(\"Invoking Blocks() on TIMEOUT: \")", "a512srv", ".", "blocks", "(", ")", "\n", "break", "// we are done", "\n", "}", "\n", "}", "\n", "}", "\n", "}", "\n", "}" ]
// Process - Sole handler for reading from the input channel
[ "Process", "-", "Sole", "handler", "for", "reading", "from", "the", "input", "channel" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L322-L358
5,593
minio/sha256-simd
sha256blockAvx512_amd64.go
reset
func (a512srv *Avx512Server) reset(uid uint64) { // Check if there is a message still waiting to be processed (and remove if so) for i, lane := range a512srv.lanes { if lane.uid == uid { if lane.block != nil { a512srv.lanes[i] = Avx512LaneInfo{} // clear message a512srv.totalIn-- } } } // Delete entry from hash map delete(a512srv.digests, uid) }
go
func (a512srv *Avx512Server) reset(uid uint64) { // Check if there is a message still waiting to be processed (and remove if so) for i, lane := range a512srv.lanes { if lane.uid == uid { if lane.block != nil { a512srv.lanes[i] = Avx512LaneInfo{} // clear message a512srv.totalIn-- } } } // Delete entry from hash map delete(a512srv.digests, uid) }
[ "func", "(", "a512srv", "*", "Avx512Server", ")", "reset", "(", "uid", "uint64", ")", "{", "// Check if there is a message still waiting to be processed (and remove if so)", "for", "i", ",", "lane", ":=", "range", "a512srv", ".", "lanes", "{", "if", "lane", ".", "uid", "==", "uid", "{", "if", "lane", ".", "block", "!=", "nil", "{", "a512srv", ".", "lanes", "[", "i", "]", "=", "Avx512LaneInfo", "{", "}", "// clear message", "\n", "a512srv", ".", "totalIn", "--", "\n", "}", "\n", "}", "\n", "}", "\n\n", "// Delete entry from hash map", "delete", "(", "a512srv", ".", "digests", ",", "uid", ")", "\n", "}" ]
// Do a reset for this calculation
[ "Do", "a", "reset", "for", "this", "calculation" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L361-L375
5,594
minio/sha256-simd
sha256blockAvx512_amd64.go
blocks
func (a512srv *Avx512Server) blocks() (err error) { inputs := [16][]byte{} for i := range inputs { inputs[i] = a512srv.lanes[i].block } mask := expandMask(genMask(inputs)) outputs := blockAvx512(a512srv.getDigests(), inputs, mask) a512srv.totalIn = 0 for i := 0; i < len(outputs); i++ { uid, outputCh := a512srv.lanes[i].uid, a512srv.lanes[i].outputCh a512srv.digests[uid] = outputs[i] a512srv.lanes[i] = Avx512LaneInfo{} if outputCh != nil { // Send back result outputCh <- outputs[i] delete(a512srv.digests, uid) // Delete entry from hashmap } } return }
go
func (a512srv *Avx512Server) blocks() (err error) { inputs := [16][]byte{} for i := range inputs { inputs[i] = a512srv.lanes[i].block } mask := expandMask(genMask(inputs)) outputs := blockAvx512(a512srv.getDigests(), inputs, mask) a512srv.totalIn = 0 for i := 0; i < len(outputs); i++ { uid, outputCh := a512srv.lanes[i].uid, a512srv.lanes[i].outputCh a512srv.digests[uid] = outputs[i] a512srv.lanes[i] = Avx512LaneInfo{} if outputCh != nil { // Send back result outputCh <- outputs[i] delete(a512srv.digests, uid) // Delete entry from hashmap } } return }
[ "func", "(", "a512srv", "*", "Avx512Server", ")", "blocks", "(", ")", "(", "err", "error", ")", "{", "inputs", ":=", "[", "16", "]", "[", "]", "byte", "{", "}", "\n", "for", "i", ":=", "range", "inputs", "{", "inputs", "[", "i", "]", "=", "a512srv", ".", "lanes", "[", "i", "]", ".", "block", "\n", "}", "\n\n", "mask", ":=", "expandMask", "(", "genMask", "(", "inputs", ")", ")", "\n", "outputs", ":=", "blockAvx512", "(", "a512srv", ".", "getDigests", "(", ")", ",", "inputs", ",", "mask", ")", "\n\n", "a512srv", ".", "totalIn", "=", "0", "\n", "for", "i", ":=", "0", ";", "i", "<", "len", "(", "outputs", ")", ";", "i", "++", "{", "uid", ",", "outputCh", ":=", "a512srv", ".", "lanes", "[", "i", "]", ".", "uid", ",", "a512srv", ".", "lanes", "[", "i", "]", ".", "outputCh", "\n", "a512srv", ".", "digests", "[", "uid", "]", "=", "outputs", "[", "i", "]", "\n", "a512srv", ".", "lanes", "[", "i", "]", "=", "Avx512LaneInfo", "{", "}", "\n\n", "if", "outputCh", "!=", "nil", "{", "// Send back result", "outputCh", "<-", "outputs", "[", "i", "]", "\n", "delete", "(", "a512srv", ".", "digests", ",", "uid", ")", "// Delete entry from hashmap", "\n", "}", "\n", "}", "\n", "return", "\n", "}" ]
// Invoke assembly and send results back
[ "Invoke", "assembly", "and", "send", "results", "back" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L378-L401
5,595
minio/sha256-simd
sha256blockAvx512_amd64.go
Sum
func (a512srv *Avx512Server) Sum(uid uint64, p []byte) [32]byte { sumCh := make(chan [32]byte) a512srv.blocksCh <- blockInput{uid: uid, msg: p, final: true, sumCh: sumCh} return <-sumCh }
go
func (a512srv *Avx512Server) Sum(uid uint64, p []byte) [32]byte { sumCh := make(chan [32]byte) a512srv.blocksCh <- blockInput{uid: uid, msg: p, final: true, sumCh: sumCh} return <-sumCh }
[ "func", "(", "a512srv", "*", "Avx512Server", ")", "Sum", "(", "uid", "uint64", ",", "p", "[", "]", "byte", ")", "[", "32", "]", "byte", "{", "sumCh", ":=", "make", "(", "chan", "[", "32", "]", "byte", ")", "\n", "a512srv", ".", "blocksCh", "<-", "blockInput", "{", "uid", ":", "uid", ",", "msg", ":", "p", ",", "final", ":", "true", ",", "sumCh", ":", "sumCh", "}", "\n", "return", "<-", "sumCh", "\n", "}" ]
// Sum - return sha256 sum in bytes for a given sum id.
[ "Sum", "-", "return", "sha256", "sum", "in", "bytes", "for", "a", "given", "sum", "id", "." ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L409-L413
5,596
minio/sha256-simd
sha256.go
Reset
func (d *digest) Reset() { d.h[0] = init0 d.h[1] = init1 d.h[2] = init2 d.h[3] = init3 d.h[4] = init4 d.h[5] = init5 d.h[6] = init6 d.h[7] = init7 d.nx = 0 d.len = 0 }
go
func (d *digest) Reset() { d.h[0] = init0 d.h[1] = init1 d.h[2] = init2 d.h[3] = init3 d.h[4] = init4 d.h[5] = init5 d.h[6] = init6 d.h[7] = init7 d.nx = 0 d.len = 0 }
[ "func", "(", "d", "*", "digest", ")", "Reset", "(", ")", "{", "d", ".", "h", "[", "0", "]", "=", "init0", "\n", "d", ".", "h", "[", "1", "]", "=", "init1", "\n", "d", ".", "h", "[", "2", "]", "=", "init2", "\n", "d", ".", "h", "[", "3", "]", "=", "init3", "\n", "d", ".", "h", "[", "4", "]", "=", "init4", "\n", "d", ".", "h", "[", "5", "]", "=", "init5", "\n", "d", ".", "h", "[", "6", "]", "=", "init6", "\n", "d", ".", "h", "[", "7", "]", "=", "init7", "\n", "d", ".", "nx", "=", "0", "\n", "d", ".", "len", "=", "0", "\n", "}" ]
// Reset digest back to default
[ "Reset", "digest", "back", "to", "default" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L53-L64
5,597
minio/sha256-simd
sha256.go
New
func New() hash.Hash { if blockfunc != blockfuncGeneric { d := new(digest) d.Reset() return d } // Fallback to the standard golang implementation // if no features were found. return sha256.New() }
go
func New() hash.Hash { if blockfunc != blockfuncGeneric { d := new(digest) d.Reset() return d } // Fallback to the standard golang implementation // if no features were found. return sha256.New() }
[ "func", "New", "(", ")", "hash", ".", "Hash", "{", "if", "blockfunc", "!=", "blockfuncGeneric", "{", "d", ":=", "new", "(", "digest", ")", "\n", "d", ".", "Reset", "(", ")", "\n", "return", "d", "\n", "}", "\n", "// Fallback to the standard golang implementation", "// if no features were found.", "return", "sha256", ".", "New", "(", ")", "\n", "}" ]
// New returns a new hash.Hash computing the SHA256 checksum.
[ "New", "returns", "a", "new", "hash", ".", "Hash", "computing", "the", "SHA256", "checksum", "." ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L102-L111
5,598
minio/sha256-simd
sha256.go
Sum256
func Sum256(data []byte) (result [Size]byte) { var d digest d.Reset() d.Write(data) result = d.checkSum() return }
go
func Sum256(data []byte) (result [Size]byte) { var d digest d.Reset() d.Write(data) result = d.checkSum() return }
[ "func", "Sum256", "(", "data", "[", "]", "byte", ")", "(", "result", "[", "Size", "]", "byte", ")", "{", "var", "d", "digest", "\n", "d", ".", "Reset", "(", ")", "\n", "d", ".", "Write", "(", "data", ")", "\n", "result", "=", "d", ".", "checkSum", "(", ")", "\n", "return", "\n", "}" ]
// Sum256 - single caller sha256 helper
[ "Sum256", "-", "single", "caller", "sha256", "helper" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L114-L120
5,599
minio/sha256-simd
sha256.go
Sum
func (d *digest) Sum(in []byte) []byte { // Make a copy of d0 so that caller can keep writing and summing. d0 := *d hash := d0.checkSum() return append(in, hash[:]...) }
go
func (d *digest) Sum(in []byte) []byte { // Make a copy of d0 so that caller can keep writing and summing. d0 := *d hash := d0.checkSum() return append(in, hash[:]...) }
[ "func", "(", "d", "*", "digest", ")", "Sum", "(", "in", "[", "]", "byte", ")", "[", "]", "byte", "{", "// Make a copy of d0 so that caller can keep writing and summing.", "d0", ":=", "*", "d", "\n", "hash", ":=", "d0", ".", "checkSum", "(", ")", "\n", "return", "append", "(", "in", ",", "hash", "[", ":", "]", "...", ")", "\n", "}" ]
// Return sha256 sum in bytes
[ "Return", "sha256", "sum", "in", "bytes" ]
05b4dd3047e5d6e86cb4e0477164b850cd896261
https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L153-L158