id
int32 0
167k
| repo
stringlengths 5
54
| path
stringlengths 4
155
| func_name
stringlengths 1
118
| original_string
stringlengths 52
85.5k
| language
stringclasses 1
value | code
stringlengths 52
85.5k
| code_tokens
sequencelengths 21
1.41k
| docstring
stringlengths 6
2.61k
| docstring_tokens
sequencelengths 3
215
| sha
stringlengths 40
40
| url
stringlengths 85
252
|
---|---|---|---|---|---|---|---|---|---|---|---|
5,500 | kotakanbe/go-cve-dictionary | log/log.go | SetLogger | func SetLogger(logDir string, quiet, debug, logJSON bool) {
stderrHundler := logger.StderrHandler
logFormat := logger.LogfmtFormat()
if logJSON {
logFormat = logger.JsonFormatEx(false, true)
stderrHundler = logger.StreamHandler(os.Stderr, logFormat)
}
lvlHundler := logger.LvlFilterHandler(logger.LvlInfo, stderrHundler)
if debug {
lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, stderrHundler)
}
if quiet {
lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, logger.DiscardHandler())
pp.SetDefaultOutput(ioutil.Discard)
}
if _, err := os.Stat(logDir); os.IsNotExist(err) {
if err := os.Mkdir(logDir, 0700); err != nil {
logger.Error("Failed to create log directory", "err", err)
}
}
var hundler logger.Handler
if _, err := os.Stat(logDir); err == nil {
logPath := filepath.Join(logDir, "cve-dictionary.log")
hundler = logger.MultiHandler(
logger.Must.FileHandler(logPath, logFormat),
lvlHundler,
)
} else {
hundler = lvlHundler
}
logger.Root().SetHandler(hundler)
} | go | func SetLogger(logDir string, quiet, debug, logJSON bool) {
stderrHundler := logger.StderrHandler
logFormat := logger.LogfmtFormat()
if logJSON {
logFormat = logger.JsonFormatEx(false, true)
stderrHundler = logger.StreamHandler(os.Stderr, logFormat)
}
lvlHundler := logger.LvlFilterHandler(logger.LvlInfo, stderrHundler)
if debug {
lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, stderrHundler)
}
if quiet {
lvlHundler = logger.LvlFilterHandler(logger.LvlDebug, logger.DiscardHandler())
pp.SetDefaultOutput(ioutil.Discard)
}
if _, err := os.Stat(logDir); os.IsNotExist(err) {
if err := os.Mkdir(logDir, 0700); err != nil {
logger.Error("Failed to create log directory", "err", err)
}
}
var hundler logger.Handler
if _, err := os.Stat(logDir); err == nil {
logPath := filepath.Join(logDir, "cve-dictionary.log")
hundler = logger.MultiHandler(
logger.Must.FileHandler(logPath, logFormat),
lvlHundler,
)
} else {
hundler = lvlHundler
}
logger.Root().SetHandler(hundler)
} | [
"func",
"SetLogger",
"(",
"logDir",
"string",
",",
"quiet",
",",
"debug",
",",
"logJSON",
"bool",
")",
"{",
"stderrHundler",
":=",
"logger",
".",
"StderrHandler",
"\n",
"logFormat",
":=",
"logger",
".",
"LogfmtFormat",
"(",
")",
"\n",
"if",
"logJSON",
"{",
"logFormat",
"=",
"logger",
".",
"JsonFormatEx",
"(",
"false",
",",
"true",
")",
"\n",
"stderrHundler",
"=",
"logger",
".",
"StreamHandler",
"(",
"os",
".",
"Stderr",
",",
"logFormat",
")",
"\n",
"}",
"\n\n",
"lvlHundler",
":=",
"logger",
".",
"LvlFilterHandler",
"(",
"logger",
".",
"LvlInfo",
",",
"stderrHundler",
")",
"\n",
"if",
"debug",
"{",
"lvlHundler",
"=",
"logger",
".",
"LvlFilterHandler",
"(",
"logger",
".",
"LvlDebug",
",",
"stderrHundler",
")",
"\n",
"}",
"\n",
"if",
"quiet",
"{",
"lvlHundler",
"=",
"logger",
".",
"LvlFilterHandler",
"(",
"logger",
".",
"LvlDebug",
",",
"logger",
".",
"DiscardHandler",
"(",
")",
")",
"\n",
"pp",
".",
"SetDefaultOutput",
"(",
"ioutil",
".",
"Discard",
")",
"\n",
"}",
"\n\n",
"if",
"_",
",",
"err",
":=",
"os",
".",
"Stat",
"(",
"logDir",
")",
";",
"os",
".",
"IsNotExist",
"(",
"err",
")",
"{",
"if",
"err",
":=",
"os",
".",
"Mkdir",
"(",
"logDir",
",",
"0700",
")",
";",
"err",
"!=",
"nil",
"{",
"logger",
".",
"Error",
"(",
"\"",
"\"",
",",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"\n",
"var",
"hundler",
"logger",
".",
"Handler",
"\n",
"if",
"_",
",",
"err",
":=",
"os",
".",
"Stat",
"(",
"logDir",
")",
";",
"err",
"==",
"nil",
"{",
"logPath",
":=",
"filepath",
".",
"Join",
"(",
"logDir",
",",
"\"",
"\"",
")",
"\n",
"hundler",
"=",
"logger",
".",
"MultiHandler",
"(",
"logger",
".",
"Must",
".",
"FileHandler",
"(",
"logPath",
",",
"logFormat",
")",
",",
"lvlHundler",
",",
")",
"\n",
"}",
"else",
"{",
"hundler",
"=",
"lvlHundler",
"\n",
"}",
"\n",
"logger",
".",
"Root",
"(",
")",
".",
"SetHandler",
"(",
"hundler",
")",
"\n",
"}"
] | // SetLogger set logger | [
"SetLogger",
"set",
"logger"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/log/log.go#L14-L47 |
5,501 | kotakanbe/go-cve-dictionary | log/log.go | Fatalf | func Fatalf(format string, args ...interface{}) {
logger.Crit(fmt.Sprintf(format, args...))
} | go | func Fatalf(format string, args ...interface{}) {
logger.Crit(fmt.Sprintf(format, args...))
} | [
"func",
"Fatalf",
"(",
"format",
"string",
",",
"args",
"...",
"interface",
"{",
"}",
")",
"{",
"logger",
".",
"Crit",
"(",
"fmt",
".",
"Sprintf",
"(",
"format",
",",
"args",
"...",
")",
")",
"\n",
"}"
] | // Fatalf is wrapper function | [
"Fatalf",
"is",
"wrapper",
"function"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/log/log.go#L70-L72 |
5,502 | kotakanbe/go-cve-dictionary | server/server.go | Start | func Start(logDir string, driver db.DB) error {
e := echo.New()
e.Debug = config.Conf.Debug
// Middleware
e.Use(middleware.Logger())
e.Use(middleware.Recover())
// setup access logger
logPath := filepath.Join(logDir, "access.log")
if _, err := os.Stat(logPath); os.IsNotExist(err) {
if _, err := os.Create(logPath); err != nil {
return err
}
}
f, err := os.OpenFile(logPath, os.O_APPEND|os.O_WRONLY, 0600)
if err != nil {
return err
}
defer f.Close()
e.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{
Output: f,
}))
// Routes
e.GET("/health", health())
e.GET("/cves/:id", getCve(driver))
e.POST("/cpes", getCveByCpeName(driver))
bindURL := fmt.Sprintf("%s:%s", config.Conf.Bind, config.Conf.Port)
log.Infof("Listening on %s", bindURL)
e.Start(bindURL)
return nil
} | go | func Start(logDir string, driver db.DB) error {
e := echo.New()
e.Debug = config.Conf.Debug
// Middleware
e.Use(middleware.Logger())
e.Use(middleware.Recover())
// setup access logger
logPath := filepath.Join(logDir, "access.log")
if _, err := os.Stat(logPath); os.IsNotExist(err) {
if _, err := os.Create(logPath); err != nil {
return err
}
}
f, err := os.OpenFile(logPath, os.O_APPEND|os.O_WRONLY, 0600)
if err != nil {
return err
}
defer f.Close()
e.Use(middleware.LoggerWithConfig(middleware.LoggerConfig{
Output: f,
}))
// Routes
e.GET("/health", health())
e.GET("/cves/:id", getCve(driver))
e.POST("/cpes", getCveByCpeName(driver))
bindURL := fmt.Sprintf("%s:%s", config.Conf.Bind, config.Conf.Port)
log.Infof("Listening on %s", bindURL)
e.Start(bindURL)
return nil
} | [
"func",
"Start",
"(",
"logDir",
"string",
",",
"driver",
"db",
".",
"DB",
")",
"error",
"{",
"e",
":=",
"echo",
".",
"New",
"(",
")",
"\n",
"e",
".",
"Debug",
"=",
"config",
".",
"Conf",
".",
"Debug",
"\n\n",
"// Middleware",
"e",
".",
"Use",
"(",
"middleware",
".",
"Logger",
"(",
")",
")",
"\n",
"e",
".",
"Use",
"(",
"middleware",
".",
"Recover",
"(",
")",
")",
"\n\n",
"// setup access logger",
"logPath",
":=",
"filepath",
".",
"Join",
"(",
"logDir",
",",
"\"",
"\"",
")",
"\n",
"if",
"_",
",",
"err",
":=",
"os",
".",
"Stat",
"(",
"logPath",
")",
";",
"os",
".",
"IsNotExist",
"(",
"err",
")",
"{",
"if",
"_",
",",
"err",
":=",
"os",
".",
"Create",
"(",
"logPath",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"f",
",",
"err",
":=",
"os",
".",
"OpenFile",
"(",
"logPath",
",",
"os",
".",
"O_APPEND",
"|",
"os",
".",
"O_WRONLY",
",",
"0600",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"defer",
"f",
".",
"Close",
"(",
")",
"\n",
"e",
".",
"Use",
"(",
"middleware",
".",
"LoggerWithConfig",
"(",
"middleware",
".",
"LoggerConfig",
"{",
"Output",
":",
"f",
",",
"}",
")",
")",
"\n\n",
"// Routes",
"e",
".",
"GET",
"(",
"\"",
"\"",
",",
"health",
"(",
")",
")",
"\n",
"e",
".",
"GET",
"(",
"\"",
"\"",
",",
"getCve",
"(",
"driver",
")",
")",
"\n",
"e",
".",
"POST",
"(",
"\"",
"\"",
",",
"getCveByCpeName",
"(",
"driver",
")",
")",
"\n\n",
"bindURL",
":=",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"config",
".",
"Conf",
".",
"Bind",
",",
"config",
".",
"Conf",
".",
"Port",
")",
"\n",
"log",
".",
"Infof",
"(",
"\"",
"\"",
",",
"bindURL",
")",
"\n\n",
"e",
".",
"Start",
"(",
"bindURL",
")",
"\n",
"return",
"nil",
"\n",
"}"
] | // Start starts CVE dictionary HTTP Server. | [
"Start",
"starts",
"CVE",
"dictionary",
"HTTP",
"Server",
"."
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/server/server.go#L17-L51 |
5,503 | kotakanbe/go-cve-dictionary | db/redis.go | NewRedis | func NewRedis(dbType, dbpath string, debugSQL bool) (driver *RedisDriver, locked bool, err error) {
driver = &RedisDriver{
name: dbType,
}
log.Debugf("Opening DB (%s).", driver.Name())
if err = driver.OpenDB(dbType, dbpath, debugSQL); err != nil {
return
}
return
} | go | func NewRedis(dbType, dbpath string, debugSQL bool) (driver *RedisDriver, locked bool, err error) {
driver = &RedisDriver{
name: dbType,
}
log.Debugf("Opening DB (%s).", driver.Name())
if err = driver.OpenDB(dbType, dbpath, debugSQL); err != nil {
return
}
return
} | [
"func",
"NewRedis",
"(",
"dbType",
",",
"dbpath",
"string",
",",
"debugSQL",
"bool",
")",
"(",
"driver",
"*",
"RedisDriver",
",",
"locked",
"bool",
",",
"err",
"error",
")",
"{",
"driver",
"=",
"&",
"RedisDriver",
"{",
"name",
":",
"dbType",
",",
"}",
"\n",
"log",
".",
"Debugf",
"(",
"\"",
"\"",
",",
"driver",
".",
"Name",
"(",
")",
")",
"\n",
"if",
"err",
"=",
"driver",
".",
"OpenDB",
"(",
"dbType",
",",
"dbpath",
",",
"debugSQL",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n\n",
"return",
"\n",
"}"
] | // NewRedis return Redis driver | [
"NewRedis",
"return",
"Redis",
"driver"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L61-L71 |
5,504 | kotakanbe/go-cve-dictionary | db/redis.go | InsertJvn | func (r *RedisDriver) InsertJvn(cves []models.CveDetail) error {
log.Infof("Inserting fetched CVEs...")
var err error
var refreshedJvns []string
bar := pb.New(len(cves))
if c.Conf.Quiet {
bar.SetWriter(ioutil.Discard)
} else {
bar.SetWriter(os.Stderr)
}
bar.Start()
for chunked := range chunkSlice(cves, 10) {
var pipe redis.Pipeliner
pipe = r.conn.Pipeline()
for _, c := range chunked {
bar.Increment()
cpes := make([]models.Cpe, len(c.Jvn.Cpes))
copy(cpes, c.Jvn.Cpes)
c.Jvn.Cpes = nil
var jj []byte
if jj, err = json.Marshal(c.Jvn); err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
refreshedJvns = append(refreshedJvns, c.CveID)
if result := pipe.HSet(hashKeyPrefix+c.CveID, "Jvn", string(jj)); result.Err() != nil {
return fmt.Errorf("Failed to HSet CVE. err: %s", result.Err())
}
for _, cpe := range cpes {
if result := pipe.ZAdd(
fmt.Sprintf("%s%s::%s", hashKeyPrefix, cpe.Vendor, cpe.Product),
redis.Z{Score: 0, Member: c.CveID},
); result.Err() != nil {
return fmt.Errorf("Failed to ZAdd cpe. err: %s", result.Err())
}
}
var jc []byte
if jc, err = json.Marshal(cpes); err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
if result := pipe.HSet(cpeHashKeyPrefix+c.CveID, "Jvn", string(jc)); result.Err() != nil {
return fmt.Errorf("Failed to HSet CPE. err: %s", result.Err())
}
}
if _, err = pipe.Exec(); err != nil {
return fmt.Errorf("Failed to exec pipeline. err: %s", err)
}
}
bar.Finish()
log.Infof("Refreshed %d Jvns.", len(refreshedJvns))
// log.Debugf("%v", refreshedJvns)
return nil
} | go | func (r *RedisDriver) InsertJvn(cves []models.CveDetail) error {
log.Infof("Inserting fetched CVEs...")
var err error
var refreshedJvns []string
bar := pb.New(len(cves))
if c.Conf.Quiet {
bar.SetWriter(ioutil.Discard)
} else {
bar.SetWriter(os.Stderr)
}
bar.Start()
for chunked := range chunkSlice(cves, 10) {
var pipe redis.Pipeliner
pipe = r.conn.Pipeline()
for _, c := range chunked {
bar.Increment()
cpes := make([]models.Cpe, len(c.Jvn.Cpes))
copy(cpes, c.Jvn.Cpes)
c.Jvn.Cpes = nil
var jj []byte
if jj, err = json.Marshal(c.Jvn); err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
refreshedJvns = append(refreshedJvns, c.CveID)
if result := pipe.HSet(hashKeyPrefix+c.CveID, "Jvn", string(jj)); result.Err() != nil {
return fmt.Errorf("Failed to HSet CVE. err: %s", result.Err())
}
for _, cpe := range cpes {
if result := pipe.ZAdd(
fmt.Sprintf("%s%s::%s", hashKeyPrefix, cpe.Vendor, cpe.Product),
redis.Z{Score: 0, Member: c.CveID},
); result.Err() != nil {
return fmt.Errorf("Failed to ZAdd cpe. err: %s", result.Err())
}
}
var jc []byte
if jc, err = json.Marshal(cpes); err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
if result := pipe.HSet(cpeHashKeyPrefix+c.CveID, "Jvn", string(jc)); result.Err() != nil {
return fmt.Errorf("Failed to HSet CPE. err: %s", result.Err())
}
}
if _, err = pipe.Exec(); err != nil {
return fmt.Errorf("Failed to exec pipeline. err: %s", err)
}
}
bar.Finish()
log.Infof("Refreshed %d Jvns.", len(refreshedJvns))
// log.Debugf("%v", refreshedJvns)
return nil
} | [
"func",
"(",
"r",
"*",
"RedisDriver",
")",
"InsertJvn",
"(",
"cves",
"[",
"]",
"models",
".",
"CveDetail",
")",
"error",
"{",
"log",
".",
"Infof",
"(",
"\"",
"\"",
")",
"\n",
"var",
"err",
"error",
"\n",
"var",
"refreshedJvns",
"[",
"]",
"string",
"\n",
"bar",
":=",
"pb",
".",
"New",
"(",
"len",
"(",
"cves",
")",
")",
"\n",
"if",
"c",
".",
"Conf",
".",
"Quiet",
"{",
"bar",
".",
"SetWriter",
"(",
"ioutil",
".",
"Discard",
")",
"\n",
"}",
"else",
"{",
"bar",
".",
"SetWriter",
"(",
"os",
".",
"Stderr",
")",
"\n",
"}",
"\n",
"bar",
".",
"Start",
"(",
")",
"\n\n",
"for",
"chunked",
":=",
"range",
"chunkSlice",
"(",
"cves",
",",
"10",
")",
"{",
"var",
"pipe",
"redis",
".",
"Pipeliner",
"\n",
"pipe",
"=",
"r",
".",
"conn",
".",
"Pipeline",
"(",
")",
"\n",
"for",
"_",
",",
"c",
":=",
"range",
"chunked",
"{",
"bar",
".",
"Increment",
"(",
")",
"\n\n",
"cpes",
":=",
"make",
"(",
"[",
"]",
"models",
".",
"Cpe",
",",
"len",
"(",
"c",
".",
"Jvn",
".",
"Cpes",
")",
")",
"\n",
"copy",
"(",
"cpes",
",",
"c",
".",
"Jvn",
".",
"Cpes",
")",
"\n",
"c",
".",
"Jvn",
".",
"Cpes",
"=",
"nil",
"\n\n",
"var",
"jj",
"[",
"]",
"byte",
"\n",
"if",
"jj",
",",
"err",
"=",
"json",
".",
"Marshal",
"(",
"c",
".",
"Jvn",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"refreshedJvns",
"=",
"append",
"(",
"refreshedJvns",
",",
"c",
".",
"CveID",
")",
"\n",
"if",
"result",
":=",
"pipe",
".",
"HSet",
"(",
"hashKeyPrefix",
"+",
"c",
".",
"CveID",
",",
"\"",
"\"",
",",
"string",
"(",
"jj",
")",
")",
";",
"result",
".",
"Err",
"(",
")",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"result",
".",
"Err",
"(",
")",
")",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"cpe",
":=",
"range",
"cpes",
"{",
"if",
"result",
":=",
"pipe",
".",
"ZAdd",
"(",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"hashKeyPrefix",
",",
"cpe",
".",
"Vendor",
",",
"cpe",
".",
"Product",
")",
",",
"redis",
".",
"Z",
"{",
"Score",
":",
"0",
",",
"Member",
":",
"c",
".",
"CveID",
"}",
",",
")",
";",
"result",
".",
"Err",
"(",
")",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"result",
".",
"Err",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"\n",
"var",
"jc",
"[",
"]",
"byte",
"\n",
"if",
"jc",
",",
"err",
"=",
"json",
".",
"Marshal",
"(",
"cpes",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"if",
"result",
":=",
"pipe",
".",
"HSet",
"(",
"cpeHashKeyPrefix",
"+",
"c",
".",
"CveID",
",",
"\"",
"\"",
",",
"string",
"(",
"jc",
")",
")",
";",
"result",
".",
"Err",
"(",
")",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"result",
".",
"Err",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"_",
",",
"err",
"=",
"pipe",
".",
"Exec",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"\n",
"bar",
".",
"Finish",
"(",
")",
"\n",
"log",
".",
"Infof",
"(",
"\"",
"\"",
",",
"len",
"(",
"refreshedJvns",
")",
")",
"\n",
"// log.Debugf(\"%v\", refreshedJvns)",
"return",
"nil",
"\n",
"}"
] | // InsertJvn insert items fetched from JVN. | [
"InsertJvn",
"insert",
"items",
"fetched",
"from",
"JVN",
"."
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L274-L329 |
5,505 | kotakanbe/go-cve-dictionary | db/redis.go | GetFetchedFeedMeta | func (r *RedisDriver) GetFetchedFeedMeta(url string) (*models.FeedMeta, error) {
var result *redis.StringStringMapCmd
if result = r.conn.HGetAll(hashKeyPrefix + "Meta"); result.Err() != nil {
return nil, result.Err()
}
meta := &models.FeedMeta{}
if s, ok := result.Val()[url]; ok {
if err := json.Unmarshal([]byte(s), meta); err != nil {
return nil, err
}
return meta, nil
}
return meta, nil
} | go | func (r *RedisDriver) GetFetchedFeedMeta(url string) (*models.FeedMeta, error) {
var result *redis.StringStringMapCmd
if result = r.conn.HGetAll(hashKeyPrefix + "Meta"); result.Err() != nil {
return nil, result.Err()
}
meta := &models.FeedMeta{}
if s, ok := result.Val()[url]; ok {
if err := json.Unmarshal([]byte(s), meta); err != nil {
return nil, err
}
return meta, nil
}
return meta, nil
} | [
"func",
"(",
"r",
"*",
"RedisDriver",
")",
"GetFetchedFeedMeta",
"(",
"url",
"string",
")",
"(",
"*",
"models",
".",
"FeedMeta",
",",
"error",
")",
"{",
"var",
"result",
"*",
"redis",
".",
"StringStringMapCmd",
"\n",
"if",
"result",
"=",
"r",
".",
"conn",
".",
"HGetAll",
"(",
"hashKeyPrefix",
"+",
"\"",
"\"",
")",
";",
"result",
".",
"Err",
"(",
")",
"!=",
"nil",
"{",
"return",
"nil",
",",
"result",
".",
"Err",
"(",
")",
"\n",
"}",
"\n",
"meta",
":=",
"&",
"models",
".",
"FeedMeta",
"{",
"}",
"\n",
"if",
"s",
",",
"ok",
":=",
"result",
".",
"Val",
"(",
")",
"[",
"url",
"]",
";",
"ok",
"{",
"if",
"err",
":=",
"json",
".",
"Unmarshal",
"(",
"[",
"]",
"byte",
"(",
"s",
")",
",",
"meta",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"return",
"meta",
",",
"nil",
"\n",
"}",
"\n",
"return",
"meta",
",",
"nil",
"\n",
"}"
] | // GetFetchedFeedMeta selects hash in metafile of the year | [
"GetFetchedFeedMeta",
"selects",
"hash",
"in",
"metafile",
"of",
"the",
"year"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L458-L471 |
5,506 | kotakanbe/go-cve-dictionary | db/redis.go | UpsertFeedHash | func (r *RedisDriver) UpsertFeedHash(m models.FeedMeta) error {
jn, err := json.Marshal(m)
if err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
var pipe redis.Pipeliner
pipe = r.conn.Pipeline()
if result := pipe.HSet(hashKeyPrefix+"Meta", m.URL, jn); result.Err() != nil {
return fmt.Errorf("Failed to HSet META. err: %s", result.Err())
}
if _, err := pipe.Exec(); err != nil {
return fmt.Errorf("Failed to exec pipeline. err: %s", err)
}
return nil
} | go | func (r *RedisDriver) UpsertFeedHash(m models.FeedMeta) error {
jn, err := json.Marshal(m)
if err != nil {
return fmt.Errorf("Failed to marshal json. err: %s", err)
}
var pipe redis.Pipeliner
pipe = r.conn.Pipeline()
if result := pipe.HSet(hashKeyPrefix+"Meta", m.URL, jn); result.Err() != nil {
return fmt.Errorf("Failed to HSet META. err: %s", result.Err())
}
if _, err := pipe.Exec(); err != nil {
return fmt.Errorf("Failed to exec pipeline. err: %s", err)
}
return nil
} | [
"func",
"(",
"r",
"*",
"RedisDriver",
")",
"UpsertFeedHash",
"(",
"m",
"models",
".",
"FeedMeta",
")",
"error",
"{",
"jn",
",",
"err",
":=",
"json",
".",
"Marshal",
"(",
"m",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n\n",
"var",
"pipe",
"redis",
".",
"Pipeliner",
"\n",
"pipe",
"=",
"r",
".",
"conn",
".",
"Pipeline",
"(",
")",
"\n",
"if",
"result",
":=",
"pipe",
".",
"HSet",
"(",
"hashKeyPrefix",
"+",
"\"",
"\"",
",",
"m",
".",
"URL",
",",
"jn",
")",
";",
"result",
".",
"Err",
"(",
")",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"result",
".",
"Err",
"(",
")",
")",
"\n",
"}",
"\n",
"if",
"_",
",",
"err",
":=",
"pipe",
".",
"Exec",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
] | // UpsertFeedHash selects hash in metafile of the year | [
"UpsertFeedHash",
"selects",
"hash",
"in",
"metafile",
"of",
"the",
"year"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/db/redis.go#L474-L489 |
5,507 | kotakanbe/go-cve-dictionary | models/models.go | UpToDate | func (f FeedMeta) UpToDate() bool {
return !f.Newly() && f.Hash == f.LatestHash
} | go | func (f FeedMeta) UpToDate() bool {
return !f.Newly() && f.Hash == f.LatestHash
} | [
"func",
"(",
"f",
"FeedMeta",
")",
"UpToDate",
"(",
")",
"bool",
"{",
"return",
"!",
"f",
".",
"Newly",
"(",
")",
"&&",
"f",
".",
"Hash",
"==",
"f",
".",
"LatestHash",
"\n",
"}"
] | // UpToDate checks whether last fetched feed is up to date | [
"UpToDate",
"checks",
"whether",
"last",
"fetched",
"feed",
"is",
"up",
"to",
"date"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L25-L27 |
5,508 | kotakanbe/go-cve-dictionary | models/models.go | OutDated | func (f FeedMeta) OutDated() bool {
return !f.Newly() && f.Hash != f.LatestHash
} | go | func (f FeedMeta) OutDated() bool {
return !f.Newly() && f.Hash != f.LatestHash
} | [
"func",
"(",
"f",
"FeedMeta",
")",
"OutDated",
"(",
")",
"bool",
"{",
"return",
"!",
"f",
".",
"Newly",
"(",
")",
"&&",
"f",
".",
"Hash",
"!=",
"f",
".",
"LatestHash",
"\n",
"}"
] | // OutDated checks whether last fetched feed is out dated | [
"OutDated",
"checks",
"whether",
"last",
"fetched",
"feed",
"is",
"out",
"dated"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L30-L32 |
5,509 | kotakanbe/go-cve-dictionary | models/models.go | StatusForStdout | func (f FeedMeta) StatusForStdout() string {
if f.Newly() {
return "Newly"
} else if f.OutDated() {
red := color.New(color.FgRed, color.Bold).SprintFunc()
return red("Out-Dated")
} else if f.UpToDate() {
return color.GreenString("Up-to-Date")
}
return "Unknown"
} | go | func (f FeedMeta) StatusForStdout() string {
if f.Newly() {
return "Newly"
} else if f.OutDated() {
red := color.New(color.FgRed, color.Bold).SprintFunc()
return red("Out-Dated")
} else if f.UpToDate() {
return color.GreenString("Up-to-Date")
}
return "Unknown"
} | [
"func",
"(",
"f",
"FeedMeta",
")",
"StatusForStdout",
"(",
")",
"string",
"{",
"if",
"f",
".",
"Newly",
"(",
")",
"{",
"return",
"\"",
"\"",
"\n",
"}",
"else",
"if",
"f",
".",
"OutDated",
"(",
")",
"{",
"red",
":=",
"color",
".",
"New",
"(",
"color",
".",
"FgRed",
",",
"color",
".",
"Bold",
")",
".",
"SprintFunc",
"(",
")",
"\n",
"return",
"red",
"(",
"\"",
"\"",
")",
"\n",
"}",
"else",
"if",
"f",
".",
"UpToDate",
"(",
")",
"{",
"return",
"color",
".",
"GreenString",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"return",
"\"",
"\"",
"\n",
"}"
] | // StatusForStdout returns a status of fetched feed | [
"StatusForStdout",
"returns",
"a",
"status",
"of",
"fetched",
"feed"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L40-L50 |
5,510 | kotakanbe/go-cve-dictionary | models/models.go | Year | func (f FeedMeta) Year() (year string, xml bool, err error) {
switch f.source() {
case nvdxml:
return strings.TrimSuffix(
strings.Split(f.URL, "nvdcve-2.0-")[1], ".xml.gz"), true, nil
case nvdjson:
return strings.TrimSuffix(
strings.Split(f.URL, "nvdcve-1.0-")[1], ".json.gz"), false, nil
case jvn:
if strings.HasSuffix(f.URL, "jvndb.rdf") {
return "modified", true, nil
} else if strings.HasSuffix(f.URL, "jvndb_new.rdf") {
return "recent", true, nil
} else {
return strings.TrimSuffix(
strings.Split(f.URL, "jvndb_")[1], ".rdf"), true, nil
}
default:
return "", false, fmt.Errorf("Failed to parse URL: %s", f.URL)
}
} | go | func (f FeedMeta) Year() (year string, xml bool, err error) {
switch f.source() {
case nvdxml:
return strings.TrimSuffix(
strings.Split(f.URL, "nvdcve-2.0-")[1], ".xml.gz"), true, nil
case nvdjson:
return strings.TrimSuffix(
strings.Split(f.URL, "nvdcve-1.0-")[1], ".json.gz"), false, nil
case jvn:
if strings.HasSuffix(f.URL, "jvndb.rdf") {
return "modified", true, nil
} else if strings.HasSuffix(f.URL, "jvndb_new.rdf") {
return "recent", true, nil
} else {
return strings.TrimSuffix(
strings.Split(f.URL, "jvndb_")[1], ".rdf"), true, nil
}
default:
return "", false, fmt.Errorf("Failed to parse URL: %s", f.URL)
}
} | [
"func",
"(",
"f",
"FeedMeta",
")",
"Year",
"(",
")",
"(",
"year",
"string",
",",
"xml",
"bool",
",",
"err",
"error",
")",
"{",
"switch",
"f",
".",
"source",
"(",
")",
"{",
"case",
"nvdxml",
":",
"return",
"strings",
".",
"TrimSuffix",
"(",
"strings",
".",
"Split",
"(",
"f",
".",
"URL",
",",
"\"",
"\"",
")",
"[",
"1",
"]",
",",
"\"",
"\"",
")",
",",
"true",
",",
"nil",
"\n",
"case",
"nvdjson",
":",
"return",
"strings",
".",
"TrimSuffix",
"(",
"strings",
".",
"Split",
"(",
"f",
".",
"URL",
",",
"\"",
"\"",
")",
"[",
"1",
"]",
",",
"\"",
"\"",
")",
",",
"false",
",",
"nil",
"\n",
"case",
"jvn",
":",
"if",
"strings",
".",
"HasSuffix",
"(",
"f",
".",
"URL",
",",
"\"",
"\"",
")",
"{",
"return",
"\"",
"\"",
",",
"true",
",",
"nil",
"\n",
"}",
"else",
"if",
"strings",
".",
"HasSuffix",
"(",
"f",
".",
"URL",
",",
"\"",
"\"",
")",
"{",
"return",
"\"",
"\"",
",",
"true",
",",
"nil",
"\n",
"}",
"else",
"{",
"return",
"strings",
".",
"TrimSuffix",
"(",
"strings",
".",
"Split",
"(",
"f",
".",
"URL",
",",
"\"",
"\"",
")",
"[",
"1",
"]",
",",
"\"",
"\"",
")",
",",
"true",
",",
"nil",
"\n",
"}",
"\n",
"default",
":",
"return",
"\"",
"\"",
",",
"false",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"f",
".",
"URL",
")",
"\n",
"}",
"\n",
"}"
] | // Year returns year, whether xml or not of the feed | [
"Year",
"returns",
"year",
"whether",
"xml",
"or",
"not",
"of",
"the",
"feed"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L93-L113 |
5,511 | kotakanbe/go-cve-dictionary | models/models.go | ToTableWriterRow | func (f FeedMeta) ToTableWriterRow() []string {
y, _, _ := f.Year()
fetched, latest := f.modifiedTimesToStrs()
return []string{
f.color(f.source()),
f.color(y),
f.StatusForStdout(),
f.color(fetched),
f.color(latest),
}
} | go | func (f FeedMeta) ToTableWriterRow() []string {
y, _, _ := f.Year()
fetched, latest := f.modifiedTimesToStrs()
return []string{
f.color(f.source()),
f.color(y),
f.StatusForStdout(),
f.color(fetched),
f.color(latest),
}
} | [
"func",
"(",
"f",
"FeedMeta",
")",
"ToTableWriterRow",
"(",
")",
"[",
"]",
"string",
"{",
"y",
",",
"_",
",",
"_",
":=",
"f",
".",
"Year",
"(",
")",
"\n",
"fetched",
",",
"latest",
":=",
"f",
".",
"modifiedTimesToStrs",
"(",
")",
"\n",
"return",
"[",
"]",
"string",
"{",
"f",
".",
"color",
"(",
"f",
".",
"source",
"(",
")",
")",
",",
"f",
".",
"color",
"(",
"y",
")",
",",
"f",
".",
"StatusForStdout",
"(",
")",
",",
"f",
".",
"color",
"(",
"fetched",
")",
",",
"f",
".",
"color",
"(",
"latest",
")",
",",
"}",
"\n",
"}"
] | // ToTableWriterRow generate data for table writer | [
"ToTableWriterRow",
"generate",
"data",
"for",
"table",
"writer"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/models/models.go#L135-L145 |
5,512 | kotakanbe/go-cve-dictionary | fetcher/nvd/json/nvd.go | FetchConvert | func FetchConvert(metas []models.FeedMeta) (cves []models.CveDetail, err error) {
reqs := []fetcher.FetchRequest{}
for _, meta := range metas {
reqs = append(reqs, fetcher.FetchRequest{
URL: meta.URL,
GZIP: true,
})
}
results, err := fetcher.FetchFeedFiles(reqs)
if err != nil {
return nil,
fmt.Errorf("Failed to fetch. err: %s", err)
}
for _, res := range results {
nvd := NvdJSON{}
if err = json.Unmarshal(res.Body, &nvd); err != nil {
return nil, fmt.Errorf(
"Failed to unmarshal. url: %s, err: %s",
res.URL, err)
}
for _, item := range nvd.CveItems {
cve, err := convertToModel(&item)
if err != nil {
return nil, fmt.Errorf("Failed to convert to model. cve: %s, err: %s",
item.Cve.CveDataMeta.ID, err)
}
cves = append(cves, *cve)
}
}
return
} | go | func FetchConvert(metas []models.FeedMeta) (cves []models.CveDetail, err error) {
reqs := []fetcher.FetchRequest{}
for _, meta := range metas {
reqs = append(reqs, fetcher.FetchRequest{
URL: meta.URL,
GZIP: true,
})
}
results, err := fetcher.FetchFeedFiles(reqs)
if err != nil {
return nil,
fmt.Errorf("Failed to fetch. err: %s", err)
}
for _, res := range results {
nvd := NvdJSON{}
if err = json.Unmarshal(res.Body, &nvd); err != nil {
return nil, fmt.Errorf(
"Failed to unmarshal. url: %s, err: %s",
res.URL, err)
}
for _, item := range nvd.CveItems {
cve, err := convertToModel(&item)
if err != nil {
return nil, fmt.Errorf("Failed to convert to model. cve: %s, err: %s",
item.Cve.CveDataMeta.ID, err)
}
cves = append(cves, *cve)
}
}
return
} | [
"func",
"FetchConvert",
"(",
"metas",
"[",
"]",
"models",
".",
"FeedMeta",
")",
"(",
"cves",
"[",
"]",
"models",
".",
"CveDetail",
",",
"err",
"error",
")",
"{",
"reqs",
":=",
"[",
"]",
"fetcher",
".",
"FetchRequest",
"{",
"}",
"\n",
"for",
"_",
",",
"meta",
":=",
"range",
"metas",
"{",
"reqs",
"=",
"append",
"(",
"reqs",
",",
"fetcher",
".",
"FetchRequest",
"{",
"URL",
":",
"meta",
".",
"URL",
",",
"GZIP",
":",
"true",
",",
"}",
")",
"\n",
"}",
"\n\n",
"results",
",",
"err",
":=",
"fetcher",
".",
"FetchFeedFiles",
"(",
"reqs",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"res",
":=",
"range",
"results",
"{",
"nvd",
":=",
"NvdJSON",
"{",
"}",
"\n",
"if",
"err",
"=",
"json",
".",
"Unmarshal",
"(",
"res",
".",
"Body",
",",
"&",
"nvd",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"res",
".",
"URL",
",",
"err",
")",
"\n",
"}",
"\n",
"for",
"_",
",",
"item",
":=",
"range",
"nvd",
".",
"CveItems",
"{",
"cve",
",",
"err",
":=",
"convertToModel",
"(",
"&",
"item",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"item",
".",
"Cve",
".",
"CveDataMeta",
".",
"ID",
",",
"err",
")",
"\n",
"}",
"\n",
"cves",
"=",
"append",
"(",
"cves",
",",
"*",
"cve",
")",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // FetchConvert Fetch CVE vulnerability information from NVD | [
"FetchConvert",
"Fetch",
"CVE",
"vulnerability",
"information",
"from",
"NVD"
] | 5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a | https://github.com/kotakanbe/go-cve-dictionary/blob/5fe52611f0b8dff9f95374d9cd7bdb23cc5fc67a/fetcher/nvd/json/nvd.go#L17-L49 |
5,513 | go-pg/migrations | collection.go | Register | func (c *Collection) Register(fns ...func(DB) error) error {
return c.register(false, fns...)
} | go | func (c *Collection) Register(fns ...func(DB) error) error {
return c.register(false, fns...)
} | [
"func",
"(",
"c",
"*",
"Collection",
")",
"Register",
"(",
"fns",
"...",
"func",
"(",
"DB",
")",
"error",
")",
"error",
"{",
"return",
"c",
".",
"register",
"(",
"false",
",",
"fns",
"...",
")",
"\n",
"}"
] | // Register registers new database migration. Must be called
// from a file with name like "1_initialize_db.go". | [
"Register",
"registers",
"new",
"database",
"migration",
".",
"Must",
"be",
"called",
"from",
"a",
"file",
"with",
"name",
"like",
"1_initialize_db",
".",
"go",
"."
] | 42c5771cbefbd8c12aa3e935e082b388186764bb | https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L74-L76 |
5,514 | go-pg/migrations | collection.go | RegisterTx | func (c *Collection) RegisterTx(fns ...func(DB) error) error {
return c.register(true, fns...)
} | go | func (c *Collection) RegisterTx(fns ...func(DB) error) error {
return c.register(true, fns...)
} | [
"func",
"(",
"c",
"*",
"Collection",
")",
"RegisterTx",
"(",
"fns",
"...",
"func",
"(",
"DB",
")",
"error",
")",
"error",
"{",
"return",
"c",
".",
"register",
"(",
"true",
",",
"fns",
"...",
")",
"\n",
"}"
] | // RegisterTx is like Register, but migration will be run in a transaction. | [
"RegisterTx",
"is",
"like",
"Register",
"but",
"migration",
"will",
"be",
"run",
"in",
"a",
"transaction",
"."
] | 42c5771cbefbd8c12aa3e935e082b388186764bb | https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L79-L81 |
5,515 | go-pg/migrations | collection.go | DiscoverSQLMigrations | func (c *Collection) DiscoverSQLMigrations(dir string) error {
dir, err := filepath.Abs(dir)
if err != nil {
return err
}
if c.isVisitedDir(dir) {
return nil
}
if _, err := os.Stat(dir); os.IsNotExist(err) {
return nil
}
var ms []*Migration
newMigration := func(version int64) *Migration {
for i := range ms {
m := ms[i]
if m.Version == version {
return m
}
}
ms = append(ms, &Migration{
Version: version,
})
return ms[len(ms)-1]
}
files, err := ioutil.ReadDir(dir)
if err != nil {
return err
}
for _, f := range files {
if f.IsDir() {
continue
}
fileName := f.Name()
if !strings.HasSuffix(fileName, ".sql") {
continue
}
idx := strings.IndexByte(fileName, '_')
if idx == -1 {
err := fmt.Errorf(
"file=%q must have name in format version_comment, e.g. 1_initial",
fileName)
return err
}
version, err := strconv.ParseInt(fileName[:idx], 10, 64)
if err != nil {
return err
}
m := newMigration(version)
filePath := filepath.Join(dir, fileName)
if strings.HasSuffix(fileName, ".up.sql") {
if m.Up != nil {
return fmt.Errorf("migration=%d already has Up func", version)
}
m.UpTx = strings.HasSuffix(fileName, ".tx.up.sql")
m.Up = newSQLMigration(filePath)
continue
}
if strings.HasSuffix(fileName, ".down.sql") {
if m.Down != nil {
return fmt.Errorf("migration=%d already has Down func", version)
}
m.DownTx = strings.HasSuffix(fileName, ".tx.down.sql")
m.Down = newSQLMigration(filePath)
continue
}
return fmt.Errorf(
"file=%q must have extension .up.sql or .down.sql", fileName)
}
for _, m := range ms {
c.addMigration(m)
}
return nil
} | go | func (c *Collection) DiscoverSQLMigrations(dir string) error {
dir, err := filepath.Abs(dir)
if err != nil {
return err
}
if c.isVisitedDir(dir) {
return nil
}
if _, err := os.Stat(dir); os.IsNotExist(err) {
return nil
}
var ms []*Migration
newMigration := func(version int64) *Migration {
for i := range ms {
m := ms[i]
if m.Version == version {
return m
}
}
ms = append(ms, &Migration{
Version: version,
})
return ms[len(ms)-1]
}
files, err := ioutil.ReadDir(dir)
if err != nil {
return err
}
for _, f := range files {
if f.IsDir() {
continue
}
fileName := f.Name()
if !strings.HasSuffix(fileName, ".sql") {
continue
}
idx := strings.IndexByte(fileName, '_')
if idx == -1 {
err := fmt.Errorf(
"file=%q must have name in format version_comment, e.g. 1_initial",
fileName)
return err
}
version, err := strconv.ParseInt(fileName[:idx], 10, 64)
if err != nil {
return err
}
m := newMigration(version)
filePath := filepath.Join(dir, fileName)
if strings.HasSuffix(fileName, ".up.sql") {
if m.Up != nil {
return fmt.Errorf("migration=%d already has Up func", version)
}
m.UpTx = strings.HasSuffix(fileName, ".tx.up.sql")
m.Up = newSQLMigration(filePath)
continue
}
if strings.HasSuffix(fileName, ".down.sql") {
if m.Down != nil {
return fmt.Errorf("migration=%d already has Down func", version)
}
m.DownTx = strings.HasSuffix(fileName, ".tx.down.sql")
m.Down = newSQLMigration(filePath)
continue
}
return fmt.Errorf(
"file=%q must have extension .up.sql or .down.sql", fileName)
}
for _, m := range ms {
c.addMigration(m)
}
return nil
} | [
"func",
"(",
"c",
"*",
"Collection",
")",
"DiscoverSQLMigrations",
"(",
"dir",
"string",
")",
"error",
"{",
"dir",
",",
"err",
":=",
"filepath",
".",
"Abs",
"(",
"dir",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"if",
"c",
".",
"isVisitedDir",
"(",
"dir",
")",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"if",
"_",
",",
"err",
":=",
"os",
".",
"Stat",
"(",
"dir",
")",
";",
"os",
".",
"IsNotExist",
"(",
"err",
")",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"var",
"ms",
"[",
"]",
"*",
"Migration",
"\n",
"newMigration",
":=",
"func",
"(",
"version",
"int64",
")",
"*",
"Migration",
"{",
"for",
"i",
":=",
"range",
"ms",
"{",
"m",
":=",
"ms",
"[",
"i",
"]",
"\n",
"if",
"m",
".",
"Version",
"==",
"version",
"{",
"return",
"m",
"\n",
"}",
"\n",
"}",
"\n\n",
"ms",
"=",
"append",
"(",
"ms",
",",
"&",
"Migration",
"{",
"Version",
":",
"version",
",",
"}",
")",
"\n",
"return",
"ms",
"[",
"len",
"(",
"ms",
")",
"-",
"1",
"]",
"\n",
"}",
"\n\n",
"files",
",",
"err",
":=",
"ioutil",
".",
"ReadDir",
"(",
"dir",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"f",
":=",
"range",
"files",
"{",
"if",
"f",
".",
"IsDir",
"(",
")",
"{",
"continue",
"\n",
"}",
"\n\n",
"fileName",
":=",
"f",
".",
"Name",
"(",
")",
"\n",
"if",
"!",
"strings",
".",
"HasSuffix",
"(",
"fileName",
",",
"\"",
"\"",
")",
"{",
"continue",
"\n",
"}",
"\n\n",
"idx",
":=",
"strings",
".",
"IndexByte",
"(",
"fileName",
",",
"'_'",
")",
"\n",
"if",
"idx",
"==",
"-",
"1",
"{",
"err",
":=",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"fileName",
")",
"\n",
"return",
"err",
"\n",
"}",
"\n\n",
"version",
",",
"err",
":=",
"strconv",
".",
"ParseInt",
"(",
"fileName",
"[",
":",
"idx",
"]",
",",
"10",
",",
"64",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"m",
":=",
"newMigration",
"(",
"version",
")",
"\n",
"filePath",
":=",
"filepath",
".",
"Join",
"(",
"dir",
",",
"fileName",
")",
"\n\n",
"if",
"strings",
".",
"HasSuffix",
"(",
"fileName",
",",
"\"",
"\"",
")",
"{",
"if",
"m",
".",
"Up",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"version",
")",
"\n",
"}",
"\n",
"m",
".",
"UpTx",
"=",
"strings",
".",
"HasSuffix",
"(",
"fileName",
",",
"\"",
"\"",
")",
"\n",
"m",
".",
"Up",
"=",
"newSQLMigration",
"(",
"filePath",
")",
"\n",
"continue",
"\n",
"}",
"\n\n",
"if",
"strings",
".",
"HasSuffix",
"(",
"fileName",
",",
"\"",
"\"",
")",
"{",
"if",
"m",
".",
"Down",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"version",
")",
"\n",
"}",
"\n",
"m",
".",
"DownTx",
"=",
"strings",
".",
"HasSuffix",
"(",
"fileName",
",",
"\"",
"\"",
")",
"\n",
"m",
".",
"Down",
"=",
"newSQLMigration",
"(",
"filePath",
")",
"\n",
"continue",
"\n",
"}",
"\n\n",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"fileName",
")",
"\n",
"}",
"\n\n",
"for",
"_",
",",
"m",
":=",
"range",
"ms",
"{",
"c",
".",
"addMigration",
"(",
"m",
")",
"\n",
"}",
"\n\n",
"return",
"nil",
"\n",
"}"
] | // DiscoverSQLMigrations scan the dir for files with .sql extension
// and adds discovered SQL migrations to the collection. | [
"DiscoverSQLMigrations",
"scan",
"the",
"dir",
"for",
"files",
"with",
".",
"sql",
"extension",
"and",
"adds",
"discovered",
"SQL",
"migrations",
"to",
"the",
"collection",
"."
] | 42c5771cbefbd8c12aa3e935e082b388186764bb | https://github.com/go-pg/migrations/blob/42c5771cbefbd8c12aa3e935e082b388186764bb/collection.go#L144-L231 |
5,516 | tcnksm/go-httpstat | httpstat.go | Format | func (r Result) Format(s fmt.State, verb rune) {
switch verb {
case 'v':
if s.Flag('+') {
var buf bytes.Buffer
fmt.Fprintf(&buf, "DNS lookup: %4d ms\n",
int(r.DNSLookup/time.Millisecond))
fmt.Fprintf(&buf, "TCP connection: %4d ms\n",
int(r.TCPConnection/time.Millisecond))
fmt.Fprintf(&buf, "TLS handshake: %4d ms\n",
int(r.TLSHandshake/time.Millisecond))
fmt.Fprintf(&buf, "Server processing: %4d ms\n",
int(r.ServerProcessing/time.Millisecond))
if r.total > 0 {
fmt.Fprintf(&buf, "Content transfer: %4d ms\n\n",
int(r.contentTransfer/time.Millisecond))
} else {
fmt.Fprintf(&buf, "Content transfer: %4s ms\n\n", "-")
}
fmt.Fprintf(&buf, "Name Lookup: %4d ms\n",
int(r.NameLookup/time.Millisecond))
fmt.Fprintf(&buf, "Connect: %4d ms\n",
int(r.Connect/time.Millisecond))
fmt.Fprintf(&buf, "Pre Transfer: %4d ms\n",
int(r.Pretransfer/time.Millisecond))
fmt.Fprintf(&buf, "Start Transfer: %4d ms\n",
int(r.StartTransfer/time.Millisecond))
if r.total > 0 {
fmt.Fprintf(&buf, "Total: %4d ms\n",
int(r.total/time.Millisecond))
} else {
fmt.Fprintf(&buf, "Total: %4s ms\n", "-")
}
io.WriteString(s, buf.String())
return
}
fallthrough
case 's', 'q':
d := r.durations()
list := make([]string, 0, len(d))
for k, v := range d {
// Handle when End function is not called
if (k == "ContentTransfer" || k == "Total") && r.t5.IsZero() {
list = append(list, fmt.Sprintf("%s: - ms", k))
continue
}
list = append(list, fmt.Sprintf("%s: %d ms", k, v/time.Millisecond))
}
io.WriteString(s, strings.Join(list, ", "))
}
} | go | func (r Result) Format(s fmt.State, verb rune) {
switch verb {
case 'v':
if s.Flag('+') {
var buf bytes.Buffer
fmt.Fprintf(&buf, "DNS lookup: %4d ms\n",
int(r.DNSLookup/time.Millisecond))
fmt.Fprintf(&buf, "TCP connection: %4d ms\n",
int(r.TCPConnection/time.Millisecond))
fmt.Fprintf(&buf, "TLS handshake: %4d ms\n",
int(r.TLSHandshake/time.Millisecond))
fmt.Fprintf(&buf, "Server processing: %4d ms\n",
int(r.ServerProcessing/time.Millisecond))
if r.total > 0 {
fmt.Fprintf(&buf, "Content transfer: %4d ms\n\n",
int(r.contentTransfer/time.Millisecond))
} else {
fmt.Fprintf(&buf, "Content transfer: %4s ms\n\n", "-")
}
fmt.Fprintf(&buf, "Name Lookup: %4d ms\n",
int(r.NameLookup/time.Millisecond))
fmt.Fprintf(&buf, "Connect: %4d ms\n",
int(r.Connect/time.Millisecond))
fmt.Fprintf(&buf, "Pre Transfer: %4d ms\n",
int(r.Pretransfer/time.Millisecond))
fmt.Fprintf(&buf, "Start Transfer: %4d ms\n",
int(r.StartTransfer/time.Millisecond))
if r.total > 0 {
fmt.Fprintf(&buf, "Total: %4d ms\n",
int(r.total/time.Millisecond))
} else {
fmt.Fprintf(&buf, "Total: %4s ms\n", "-")
}
io.WriteString(s, buf.String())
return
}
fallthrough
case 's', 'q':
d := r.durations()
list := make([]string, 0, len(d))
for k, v := range d {
// Handle when End function is not called
if (k == "ContentTransfer" || k == "Total") && r.t5.IsZero() {
list = append(list, fmt.Sprintf("%s: - ms", k))
continue
}
list = append(list, fmt.Sprintf("%s: %d ms", k, v/time.Millisecond))
}
io.WriteString(s, strings.Join(list, ", "))
}
} | [
"func",
"(",
"r",
"Result",
")",
"Format",
"(",
"s",
"fmt",
".",
"State",
",",
"verb",
"rune",
")",
"{",
"switch",
"verb",
"{",
"case",
"'v'",
":",
"if",
"s",
".",
"Flag",
"(",
"'+'",
")",
"{",
"var",
"buf",
"bytes",
".",
"Buffer",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"DNSLookup",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"TCPConnection",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"TLSHandshake",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"ServerProcessing",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n\n",
"if",
"r",
".",
"total",
">",
"0",
"{",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"contentTransfer",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"}",
"else",
"{",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\\n",
"\"",
",",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"NameLookup",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"Connect",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"Pretransfer",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"StartTransfer",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n\n",
"if",
"r",
".",
"total",
">",
"0",
"{",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"int",
"(",
"r",
".",
"total",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"}",
"else",
"{",
"fmt",
".",
"Fprintf",
"(",
"&",
"buf",
",",
"\"",
"\\n",
"\"",
",",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"io",
".",
"WriteString",
"(",
"s",
",",
"buf",
".",
"String",
"(",
")",
")",
"\n",
"return",
"\n",
"}",
"\n\n",
"fallthrough",
"\n",
"case",
"'s'",
",",
"'q'",
":",
"d",
":=",
"r",
".",
"durations",
"(",
")",
"\n",
"list",
":=",
"make",
"(",
"[",
"]",
"string",
",",
"0",
",",
"len",
"(",
"d",
")",
")",
"\n",
"for",
"k",
",",
"v",
":=",
"range",
"d",
"{",
"// Handle when End function is not called",
"if",
"(",
"k",
"==",
"\"",
"\"",
"||",
"k",
"==",
"\"",
"\"",
")",
"&&",
"r",
".",
"t5",
".",
"IsZero",
"(",
")",
"{",
"list",
"=",
"append",
"(",
"list",
",",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"k",
")",
")",
"\n",
"continue",
"\n",
"}",
"\n",
"list",
"=",
"append",
"(",
"list",
",",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"k",
",",
"v",
"/",
"time",
".",
"Millisecond",
")",
")",
"\n",
"}",
"\n",
"io",
".",
"WriteString",
"(",
"s",
",",
"strings",
".",
"Join",
"(",
"list",
",",
"\"",
"\"",
")",
")",
"\n",
"}",
"\n\n",
"}"
] | // Format formats stats result. | [
"Format",
"formats",
"stats",
"result",
"."
] | 6962beb23919c3de82c788a87b36f4f901ab514d | https://github.com/tcnksm/go-httpstat/blob/6962beb23919c3de82c788a87b36f4f901ab514d/httpstat.go#L72-L127 |
5,517 | tcnksm/go-httpstat | httpstat.go | WithHTTPStat | func WithHTTPStat(ctx context.Context, r *Result) context.Context {
return withClientTrace(ctx, r)
} | go | func WithHTTPStat(ctx context.Context, r *Result) context.Context {
return withClientTrace(ctx, r)
} | [
"func",
"WithHTTPStat",
"(",
"ctx",
"context",
".",
"Context",
",",
"r",
"*",
"Result",
")",
"context",
".",
"Context",
"{",
"return",
"withClientTrace",
"(",
"ctx",
",",
"r",
")",
"\n",
"}"
] | // WithHTTPStat is a wrapper of httptrace.WithClientTrace. It records the
// time of each httptrace hooks. | [
"WithHTTPStat",
"is",
"a",
"wrapper",
"of",
"httptrace",
".",
"WithClientTrace",
".",
"It",
"records",
"the",
"time",
"of",
"each",
"httptrace",
"hooks",
"."
] | 6962beb23919c3de82c788a87b36f4f901ab514d | https://github.com/tcnksm/go-httpstat/blob/6962beb23919c3de82c788a87b36f4f901ab514d/httpstat.go#L131-L133 |
5,518 | PuerkitoBio/gocrawl | worker.go | run | func (w *worker) run() {
defer func() {
w.logFunc(LogInfo, "worker done.")
w.wg.Done()
}()
// Enter loop to process URLs until stop signal is received
for {
var idleChan <-chan time.Time
w.logFunc(LogInfo, "waiting for pop...")
// Initialize the idle timeout channel, if required
if w.opts.WorkerIdleTTL > 0 {
idleChan = time.After(w.opts.WorkerIdleTTL)
}
select {
case <-w.stop:
w.logFunc(LogInfo, "stop signal received.")
return
case <-idleChan:
w.logFunc(LogInfo, "idle timeout received.")
w.sendResponse(nil, false, nil, true)
return
case batch := <-w.pop:
// Got a batch of urls to crawl, loop and check at each iteration if a stop
// is received.
for _, ctx := range batch {
w.logFunc(LogInfo, "popped: %s", ctx.url)
if ctx.IsRobotsURL() {
w.requestRobotsTxt(ctx)
} else if w.isAllowedPerRobotsPolicies(ctx.url) {
w.requestURL(ctx, ctx.HeadBeforeGet)
} else {
// Must still notify Crawler that this URL was processed, although not visited
w.opts.Extender.Disallowed(ctx)
w.sendResponse(ctx, false, nil, false)
}
// No need to check for idle timeout here, no idling while looping through
// a batch of URLs.
select {
case <-w.stop:
w.logFunc(LogInfo, "stop signal received.")
return
default:
// Nothing, just continue...
}
}
}
}
} | go | func (w *worker) run() {
defer func() {
w.logFunc(LogInfo, "worker done.")
w.wg.Done()
}()
// Enter loop to process URLs until stop signal is received
for {
var idleChan <-chan time.Time
w.logFunc(LogInfo, "waiting for pop...")
// Initialize the idle timeout channel, if required
if w.opts.WorkerIdleTTL > 0 {
idleChan = time.After(w.opts.WorkerIdleTTL)
}
select {
case <-w.stop:
w.logFunc(LogInfo, "stop signal received.")
return
case <-idleChan:
w.logFunc(LogInfo, "idle timeout received.")
w.sendResponse(nil, false, nil, true)
return
case batch := <-w.pop:
// Got a batch of urls to crawl, loop and check at each iteration if a stop
// is received.
for _, ctx := range batch {
w.logFunc(LogInfo, "popped: %s", ctx.url)
if ctx.IsRobotsURL() {
w.requestRobotsTxt(ctx)
} else if w.isAllowedPerRobotsPolicies(ctx.url) {
w.requestURL(ctx, ctx.HeadBeforeGet)
} else {
// Must still notify Crawler that this URL was processed, although not visited
w.opts.Extender.Disallowed(ctx)
w.sendResponse(ctx, false, nil, false)
}
// No need to check for idle timeout here, no idling while looping through
// a batch of URLs.
select {
case <-w.stop:
w.logFunc(LogInfo, "stop signal received.")
return
default:
// Nothing, just continue...
}
}
}
}
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"run",
"(",
")",
"{",
"defer",
"func",
"(",
")",
"{",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"w",
".",
"wg",
".",
"Done",
"(",
")",
"\n",
"}",
"(",
")",
"\n\n",
"// Enter loop to process URLs until stop signal is received",
"for",
"{",
"var",
"idleChan",
"<-",
"chan",
"time",
".",
"Time",
"\n\n",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n\n",
"// Initialize the idle timeout channel, if required",
"if",
"w",
".",
"opts",
".",
"WorkerIdleTTL",
">",
"0",
"{",
"idleChan",
"=",
"time",
".",
"After",
"(",
"w",
".",
"opts",
".",
"WorkerIdleTTL",
")",
"\n",
"}",
"\n\n",
"select",
"{",
"case",
"<-",
"w",
".",
"stop",
":",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n\n",
"case",
"<-",
"idleChan",
":",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"w",
".",
"sendResponse",
"(",
"nil",
",",
"false",
",",
"nil",
",",
"true",
")",
"\n",
"return",
"\n\n",
"case",
"batch",
":=",
"<-",
"w",
".",
"pop",
":",
"// Got a batch of urls to crawl, loop and check at each iteration if a stop",
"// is received.",
"for",
"_",
",",
"ctx",
":=",
"range",
"batch",
"{",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
")",
"\n\n",
"if",
"ctx",
".",
"IsRobotsURL",
"(",
")",
"{",
"w",
".",
"requestRobotsTxt",
"(",
"ctx",
")",
"\n",
"}",
"else",
"if",
"w",
".",
"isAllowedPerRobotsPolicies",
"(",
"ctx",
".",
"url",
")",
"{",
"w",
".",
"requestURL",
"(",
"ctx",
",",
"ctx",
".",
"HeadBeforeGet",
")",
"\n",
"}",
"else",
"{",
"// Must still notify Crawler that this URL was processed, although not visited",
"w",
".",
"opts",
".",
"Extender",
".",
"Disallowed",
"(",
"ctx",
")",
"\n",
"w",
".",
"sendResponse",
"(",
"ctx",
",",
"false",
",",
"nil",
",",
"false",
")",
"\n",
"}",
"\n\n",
"// No need to check for idle timeout here, no idling while looping through",
"// a batch of URLs.",
"select",
"{",
"case",
"<-",
"w",
".",
"stop",
":",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"default",
":",
"// Nothing, just continue...",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // Start crawling the host. | [
"Start",
"crawling",
"the",
"host",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L48-L104 |
5,519 | PuerkitoBio/gocrawl | worker.go | isAllowedPerRobotsPolicies | func (w *worker) isAllowedPerRobotsPolicies(u *url.URL) bool {
if w.robotsGroup != nil {
// Is this URL allowed per robots.txt policy?
ok := w.robotsGroup.Test(u.Path)
if !ok {
w.logFunc(LogIgnored, "ignored on robots.txt policy: %s", u.String())
}
return ok
}
// No robots.txt = everything is allowed
return true
} | go | func (w *worker) isAllowedPerRobotsPolicies(u *url.URL) bool {
if w.robotsGroup != nil {
// Is this URL allowed per robots.txt policy?
ok := w.robotsGroup.Test(u.Path)
if !ok {
w.logFunc(LogIgnored, "ignored on robots.txt policy: %s", u.String())
}
return ok
}
// No robots.txt = everything is allowed
return true
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"isAllowedPerRobotsPolicies",
"(",
"u",
"*",
"url",
".",
"URL",
")",
"bool",
"{",
"if",
"w",
".",
"robotsGroup",
"!=",
"nil",
"{",
"// Is this URL allowed per robots.txt policy?",
"ok",
":=",
"w",
".",
"robotsGroup",
".",
"Test",
"(",
"u",
".",
"Path",
")",
"\n",
"if",
"!",
"ok",
"{",
"w",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"u",
".",
"String",
"(",
")",
")",
"\n",
"}",
"\n",
"return",
"ok",
"\n",
"}",
"\n\n",
"// No robots.txt = everything is allowed",
"return",
"true",
"\n",
"}"
] | // Checks if the given URL can be fetched based on robots.txt policies. | [
"Checks",
"if",
"the",
"given",
"URL",
"can",
"be",
"fetched",
"based",
"on",
"robots",
".",
"txt",
"policies",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L107-L119 |
5,520 | PuerkitoBio/gocrawl | worker.go | requestURL | func (w *worker) requestURL(ctx *URLContext, headRequest bool) {
if res, ok := w.fetchURL(ctx, w.opts.UserAgent, headRequest); ok {
var harvested interface{}
var visited bool
// Close the body on function end
defer res.Body.Close()
// Any 2xx status code is good to go
if res.StatusCode >= 200 && res.StatusCode < 300 {
// Success, visit the URL
harvested = w.visitURL(ctx, res)
visited = true
} else {
// Error based on status code received
w.opts.Extender.Error(newCrawlErrorMessage(ctx, res.Status, CekHttpStatusCode))
w.logFunc(LogError, "ERROR status code for %s: %s", ctx.url, res.Status)
}
w.sendResponse(ctx, visited, harvested, false)
}
} | go | func (w *worker) requestURL(ctx *URLContext, headRequest bool) {
if res, ok := w.fetchURL(ctx, w.opts.UserAgent, headRequest); ok {
var harvested interface{}
var visited bool
// Close the body on function end
defer res.Body.Close()
// Any 2xx status code is good to go
if res.StatusCode >= 200 && res.StatusCode < 300 {
// Success, visit the URL
harvested = w.visitURL(ctx, res)
visited = true
} else {
// Error based on status code received
w.opts.Extender.Error(newCrawlErrorMessage(ctx, res.Status, CekHttpStatusCode))
w.logFunc(LogError, "ERROR status code for %s: %s", ctx.url, res.Status)
}
w.sendResponse(ctx, visited, harvested, false)
}
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"requestURL",
"(",
"ctx",
"*",
"URLContext",
",",
"headRequest",
"bool",
")",
"{",
"if",
"res",
",",
"ok",
":=",
"w",
".",
"fetchURL",
"(",
"ctx",
",",
"w",
".",
"opts",
".",
"UserAgent",
",",
"headRequest",
")",
";",
"ok",
"{",
"var",
"harvested",
"interface",
"{",
"}",
"\n",
"var",
"visited",
"bool",
"\n\n",
"// Close the body on function end",
"defer",
"res",
".",
"Body",
".",
"Close",
"(",
")",
"\n\n",
"// Any 2xx status code is good to go",
"if",
"res",
".",
"StatusCode",
">=",
"200",
"&&",
"res",
".",
"StatusCode",
"<",
"300",
"{",
"// Success, visit the URL",
"harvested",
"=",
"w",
".",
"visitURL",
"(",
"ctx",
",",
"res",
")",
"\n",
"visited",
"=",
"true",
"\n",
"}",
"else",
"{",
"// Error based on status code received",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlErrorMessage",
"(",
"ctx",
",",
"res",
".",
"Status",
",",
"CekHttpStatusCode",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
",",
"res",
".",
"Status",
")",
"\n",
"}",
"\n",
"w",
".",
"sendResponse",
"(",
"ctx",
",",
"visited",
",",
"harvested",
",",
"false",
")",
"\n",
"}",
"\n",
"}"
] | // Process the specified URL. | [
"Process",
"the",
"specified",
"URL",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L122-L142 |
5,521 | PuerkitoBio/gocrawl | worker.go | requestRobotsTxt | func (w *worker) requestRobotsTxt(ctx *URLContext) {
// Ask if it should be fetched
if robData, reqRob := w.opts.Extender.RequestRobots(ctx, w.opts.RobotUserAgent); !reqRob {
w.logFunc(LogInfo, "using robots.txt from cache")
w.robotsGroup = w.getRobotsTxtGroup(ctx, robData, nil)
} else if res, ok := w.fetchURL(ctx, w.opts.UserAgent, false); ok {
// Close the body on function end
defer res.Body.Close()
w.robotsGroup = w.getRobotsTxtGroup(ctx, nil, res)
}
} | go | func (w *worker) requestRobotsTxt(ctx *URLContext) {
// Ask if it should be fetched
if robData, reqRob := w.opts.Extender.RequestRobots(ctx, w.opts.RobotUserAgent); !reqRob {
w.logFunc(LogInfo, "using robots.txt from cache")
w.robotsGroup = w.getRobotsTxtGroup(ctx, robData, nil)
} else if res, ok := w.fetchURL(ctx, w.opts.UserAgent, false); ok {
// Close the body on function end
defer res.Body.Close()
w.robotsGroup = w.getRobotsTxtGroup(ctx, nil, res)
}
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"requestRobotsTxt",
"(",
"ctx",
"*",
"URLContext",
")",
"{",
"// Ask if it should be fetched",
"if",
"robData",
",",
"reqRob",
":=",
"w",
".",
"opts",
".",
"Extender",
".",
"RequestRobots",
"(",
"ctx",
",",
"w",
".",
"opts",
".",
"RobotUserAgent",
")",
";",
"!",
"reqRob",
"{",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"w",
".",
"robotsGroup",
"=",
"w",
".",
"getRobotsTxtGroup",
"(",
"ctx",
",",
"robData",
",",
"nil",
")",
"\n\n",
"}",
"else",
"if",
"res",
",",
"ok",
":=",
"w",
".",
"fetchURL",
"(",
"ctx",
",",
"w",
".",
"opts",
".",
"UserAgent",
",",
"false",
")",
";",
"ok",
"{",
"// Close the body on function end",
"defer",
"res",
".",
"Body",
".",
"Close",
"(",
")",
"\n",
"w",
".",
"robotsGroup",
"=",
"w",
".",
"getRobotsTxtGroup",
"(",
"ctx",
",",
"nil",
",",
"res",
")",
"\n",
"}",
"\n",
"}"
] | // Process the robots.txt URL. | [
"Process",
"the",
"robots",
".",
"txt",
"URL",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L145-L156 |
5,522 | PuerkitoBio/gocrawl | worker.go | getRobotsTxtGroup | func (w *worker) getRobotsTxtGroup(ctx *URLContext, b []byte, res *http.Response) (g *robotstxt.Group) {
var data *robotstxt.RobotsData
var e error
if res != nil {
var buf bytes.Buffer
io.Copy(&buf, res.Body)
res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes()))
data, e = robotstxt.FromResponse(res)
// Rewind the res.Body (by re-creating it from the bytes)
res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes()))
// Error or not, the robots.txt has been fetched, so notify
w.opts.Extender.FetchedRobots(ctx, res)
} else {
data, e = robotstxt.FromBytes(b)
}
// If robots data cannot be parsed, will return nil, which will allow access by default.
// Reasonable, since by default no robots.txt means full access, so invalid
// robots.txt is similar behavior.
if e != nil {
w.opts.Extender.Error(newCrawlError(nil, e, CekParseRobots))
w.logFunc(LogError, "ERROR parsing robots.txt for host %s: %s", w.host, e)
} else {
g = data.FindGroup(w.opts.RobotUserAgent)
}
return g
} | go | func (w *worker) getRobotsTxtGroup(ctx *URLContext, b []byte, res *http.Response) (g *robotstxt.Group) {
var data *robotstxt.RobotsData
var e error
if res != nil {
var buf bytes.Buffer
io.Copy(&buf, res.Body)
res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes()))
data, e = robotstxt.FromResponse(res)
// Rewind the res.Body (by re-creating it from the bytes)
res.Body = ioutil.NopCloser(bytes.NewReader(buf.Bytes()))
// Error or not, the robots.txt has been fetched, so notify
w.opts.Extender.FetchedRobots(ctx, res)
} else {
data, e = robotstxt.FromBytes(b)
}
// If robots data cannot be parsed, will return nil, which will allow access by default.
// Reasonable, since by default no robots.txt means full access, so invalid
// robots.txt is similar behavior.
if e != nil {
w.opts.Extender.Error(newCrawlError(nil, e, CekParseRobots))
w.logFunc(LogError, "ERROR parsing robots.txt for host %s: %s", w.host, e)
} else {
g = data.FindGroup(w.opts.RobotUserAgent)
}
return g
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"getRobotsTxtGroup",
"(",
"ctx",
"*",
"URLContext",
",",
"b",
"[",
"]",
"byte",
",",
"res",
"*",
"http",
".",
"Response",
")",
"(",
"g",
"*",
"robotstxt",
".",
"Group",
")",
"{",
"var",
"data",
"*",
"robotstxt",
".",
"RobotsData",
"\n",
"var",
"e",
"error",
"\n\n",
"if",
"res",
"!=",
"nil",
"{",
"var",
"buf",
"bytes",
".",
"Buffer",
"\n",
"io",
".",
"Copy",
"(",
"&",
"buf",
",",
"res",
".",
"Body",
")",
"\n",
"res",
".",
"Body",
"=",
"ioutil",
".",
"NopCloser",
"(",
"bytes",
".",
"NewReader",
"(",
"buf",
".",
"Bytes",
"(",
")",
")",
")",
"\n",
"data",
",",
"e",
"=",
"robotstxt",
".",
"FromResponse",
"(",
"res",
")",
"\n",
"// Rewind the res.Body (by re-creating it from the bytes)",
"res",
".",
"Body",
"=",
"ioutil",
".",
"NopCloser",
"(",
"bytes",
".",
"NewReader",
"(",
"buf",
".",
"Bytes",
"(",
")",
")",
")",
"\n",
"// Error or not, the robots.txt has been fetched, so notify",
"w",
".",
"opts",
".",
"Extender",
".",
"FetchedRobots",
"(",
"ctx",
",",
"res",
")",
"\n",
"}",
"else",
"{",
"data",
",",
"e",
"=",
"robotstxt",
".",
"FromBytes",
"(",
"b",
")",
"\n",
"}",
"\n\n",
"// If robots data cannot be parsed, will return nil, which will allow access by default.",
"// Reasonable, since by default no robots.txt means full access, so invalid",
"// robots.txt is similar behavior.",
"if",
"e",
"!=",
"nil",
"{",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"nil",
",",
"e",
",",
"CekParseRobots",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"w",
".",
"host",
",",
"e",
")",
"\n",
"}",
"else",
"{",
"g",
"=",
"data",
".",
"FindGroup",
"(",
"w",
".",
"opts",
".",
"RobotUserAgent",
")",
"\n",
"}",
"\n",
"return",
"g",
"\n",
"}"
] | // Get the robots.txt group for this crawler. | [
"Get",
"the",
"robots",
".",
"txt",
"group",
"for",
"this",
"crawler",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L159-L186 |
5,523 | PuerkitoBio/gocrawl | worker.go | setCrawlDelay | func (w *worker) setCrawlDelay() {
var robDelay time.Duration
if w.robotsGroup != nil {
robDelay = w.robotsGroup.CrawlDelay
}
w.lastCrawlDelay = w.opts.Extender.ComputeDelay(w.host,
&DelayInfo{
w.opts.CrawlDelay,
robDelay,
w.lastCrawlDelay,
},
w.lastFetch)
w.logFunc(LogInfo, "using crawl-delay: %v", w.lastCrawlDelay)
} | go | func (w *worker) setCrawlDelay() {
var robDelay time.Duration
if w.robotsGroup != nil {
robDelay = w.robotsGroup.CrawlDelay
}
w.lastCrawlDelay = w.opts.Extender.ComputeDelay(w.host,
&DelayInfo{
w.opts.CrawlDelay,
robDelay,
w.lastCrawlDelay,
},
w.lastFetch)
w.logFunc(LogInfo, "using crawl-delay: %v", w.lastCrawlDelay)
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"setCrawlDelay",
"(",
")",
"{",
"var",
"robDelay",
"time",
".",
"Duration",
"\n\n",
"if",
"w",
".",
"robotsGroup",
"!=",
"nil",
"{",
"robDelay",
"=",
"w",
".",
"robotsGroup",
".",
"CrawlDelay",
"\n",
"}",
"\n",
"w",
".",
"lastCrawlDelay",
"=",
"w",
".",
"opts",
".",
"Extender",
".",
"ComputeDelay",
"(",
"w",
".",
"host",
",",
"&",
"DelayInfo",
"{",
"w",
".",
"opts",
".",
"CrawlDelay",
",",
"robDelay",
",",
"w",
".",
"lastCrawlDelay",
",",
"}",
",",
"w",
".",
"lastFetch",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
",",
"w",
".",
"lastCrawlDelay",
")",
"\n",
"}"
] | // Set the crawl delay between this request and the next. | [
"Set",
"the",
"crawl",
"delay",
"between",
"this",
"request",
"and",
"the",
"next",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L189-L203 |
5,524 | PuerkitoBio/gocrawl | worker.go | fetchURL | func (w *worker) fetchURL(ctx *URLContext, agent string, headRequest bool) (res *http.Response, ok bool) {
var e error
var silent bool
for {
// Wait for crawl delay, if one is pending.
w.logFunc(LogTrace, "waiting for crawl delay")
if w.wait != nil {
<-w.wait
w.wait = nil
}
// Compute the next delay
w.setCrawlDelay()
// Compute the fetch duration
now := time.Now()
// Request the URL
if res, e = w.opts.Extender.Fetch(ctx, agent, headRequest); e != nil {
// Check if this is an ErrEnqueueRedirect, in which case we will enqueue
// the redirect-to URL.
if ue, ok := e.(*url.Error); ok {
// We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect
if ue.Err == ErrEnqueueRedirect {
// Do not notify this error outside of this if block, this is not a
// "real" error. We either enqueue the new URL, or fail to parse it,
// and then stop processing the current URL.
silent = true
// Parse the URL in the context of the original URL (so that relative URLs are ok).
// Absolute URLs that point to another host are ok too.
if ur, e := ctx.url.Parse(ue.URL); e != nil {
// Notify error
w.opts.Extender.Error(newCrawlError(nil, e, CekParseRedirectURL))
w.logFunc(LogError, "ERROR parsing redirect URL %s: %s", ue.URL, e)
} else {
w.logFunc(LogTrace, "redirect to %s from %s, linked from %s", ur, ctx.URL(), ctx.SourceURL())
// Enqueue the redirect-to URL with the original source
rCtx := ctx.cloneForRedirect(ur, w.opts.URLNormalizationFlags)
w.enqueue <- rCtx
}
}
}
// No fetch, so set to nil
w.lastFetch = nil
if !silent {
// Notify error
w.opts.Extender.Error(newCrawlError(ctx, e, CekFetch))
w.logFunc(LogError, "ERROR fetching %s: %s", ctx.url, e)
}
// Return from this URL crawl
w.sendResponse(ctx, false, nil, false)
return nil, false
}
// Get the fetch duration
fetchDuration := time.Now().Sub(now)
// Crawl delay starts now.
w.wait = time.After(w.lastCrawlDelay)
// Keep trace of this last fetch info
w.lastFetch = &FetchInfo{
ctx,
fetchDuration,
res.StatusCode,
headRequest,
}
if headRequest {
// Close the HEAD request's body
defer res.Body.Close()
// Next up is GET request, maybe
headRequest = false
// Ask caller if we should proceed with a GET
if !w.opts.Extender.RequestGet(ctx, res) {
w.logFunc(LogIgnored, "ignored on HEAD filter policy: %s", ctx.url)
w.sendResponse(ctx, false, nil, false)
ok = false
break
}
} else {
ok = true
break
}
}
return
} | go | func (w *worker) fetchURL(ctx *URLContext, agent string, headRequest bool) (res *http.Response, ok bool) {
var e error
var silent bool
for {
// Wait for crawl delay, if one is pending.
w.logFunc(LogTrace, "waiting for crawl delay")
if w.wait != nil {
<-w.wait
w.wait = nil
}
// Compute the next delay
w.setCrawlDelay()
// Compute the fetch duration
now := time.Now()
// Request the URL
if res, e = w.opts.Extender.Fetch(ctx, agent, headRequest); e != nil {
// Check if this is an ErrEnqueueRedirect, in which case we will enqueue
// the redirect-to URL.
if ue, ok := e.(*url.Error); ok {
// We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect
if ue.Err == ErrEnqueueRedirect {
// Do not notify this error outside of this if block, this is not a
// "real" error. We either enqueue the new URL, or fail to parse it,
// and then stop processing the current URL.
silent = true
// Parse the URL in the context of the original URL (so that relative URLs are ok).
// Absolute URLs that point to another host are ok too.
if ur, e := ctx.url.Parse(ue.URL); e != nil {
// Notify error
w.opts.Extender.Error(newCrawlError(nil, e, CekParseRedirectURL))
w.logFunc(LogError, "ERROR parsing redirect URL %s: %s", ue.URL, e)
} else {
w.logFunc(LogTrace, "redirect to %s from %s, linked from %s", ur, ctx.URL(), ctx.SourceURL())
// Enqueue the redirect-to URL with the original source
rCtx := ctx.cloneForRedirect(ur, w.opts.URLNormalizationFlags)
w.enqueue <- rCtx
}
}
}
// No fetch, so set to nil
w.lastFetch = nil
if !silent {
// Notify error
w.opts.Extender.Error(newCrawlError(ctx, e, CekFetch))
w.logFunc(LogError, "ERROR fetching %s: %s", ctx.url, e)
}
// Return from this URL crawl
w.sendResponse(ctx, false, nil, false)
return nil, false
}
// Get the fetch duration
fetchDuration := time.Now().Sub(now)
// Crawl delay starts now.
w.wait = time.After(w.lastCrawlDelay)
// Keep trace of this last fetch info
w.lastFetch = &FetchInfo{
ctx,
fetchDuration,
res.StatusCode,
headRequest,
}
if headRequest {
// Close the HEAD request's body
defer res.Body.Close()
// Next up is GET request, maybe
headRequest = false
// Ask caller if we should proceed with a GET
if !w.opts.Extender.RequestGet(ctx, res) {
w.logFunc(LogIgnored, "ignored on HEAD filter policy: %s", ctx.url)
w.sendResponse(ctx, false, nil, false)
ok = false
break
}
} else {
ok = true
break
}
}
return
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"fetchURL",
"(",
"ctx",
"*",
"URLContext",
",",
"agent",
"string",
",",
"headRequest",
"bool",
")",
"(",
"res",
"*",
"http",
".",
"Response",
",",
"ok",
"bool",
")",
"{",
"var",
"e",
"error",
"\n",
"var",
"silent",
"bool",
"\n\n",
"for",
"{",
"// Wait for crawl delay, if one is pending.",
"w",
".",
"logFunc",
"(",
"LogTrace",
",",
"\"",
"\"",
")",
"\n",
"if",
"w",
".",
"wait",
"!=",
"nil",
"{",
"<-",
"w",
".",
"wait",
"\n",
"w",
".",
"wait",
"=",
"nil",
"\n",
"}",
"\n\n",
"// Compute the next delay",
"w",
".",
"setCrawlDelay",
"(",
")",
"\n\n",
"// Compute the fetch duration",
"now",
":=",
"time",
".",
"Now",
"(",
")",
"\n\n",
"// Request the URL",
"if",
"res",
",",
"e",
"=",
"w",
".",
"opts",
".",
"Extender",
".",
"Fetch",
"(",
"ctx",
",",
"agent",
",",
"headRequest",
")",
";",
"e",
"!=",
"nil",
"{",
"// Check if this is an ErrEnqueueRedirect, in which case we will enqueue",
"// the redirect-to URL.",
"if",
"ue",
",",
"ok",
":=",
"e",
".",
"(",
"*",
"url",
".",
"Error",
")",
";",
"ok",
"{",
"// We have a *url.Error, check if it was returned because of an ErrEnqueueRedirect",
"if",
"ue",
".",
"Err",
"==",
"ErrEnqueueRedirect",
"{",
"// Do not notify this error outside of this if block, this is not a",
"// \"real\" error. We either enqueue the new URL, or fail to parse it,",
"// and then stop processing the current URL.",
"silent",
"=",
"true",
"\n",
"// Parse the URL in the context of the original URL (so that relative URLs are ok).",
"// Absolute URLs that point to another host are ok too.",
"if",
"ur",
",",
"e",
":=",
"ctx",
".",
"url",
".",
"Parse",
"(",
"ue",
".",
"URL",
")",
";",
"e",
"!=",
"nil",
"{",
"// Notify error",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"nil",
",",
"e",
",",
"CekParseRedirectURL",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ue",
".",
"URL",
",",
"e",
")",
"\n",
"}",
"else",
"{",
"w",
".",
"logFunc",
"(",
"LogTrace",
",",
"\"",
"\"",
",",
"ur",
",",
"ctx",
".",
"URL",
"(",
")",
",",
"ctx",
".",
"SourceURL",
"(",
")",
")",
"\n",
"// Enqueue the redirect-to URL with the original source",
"rCtx",
":=",
"ctx",
".",
"cloneForRedirect",
"(",
"ur",
",",
"w",
".",
"opts",
".",
"URLNormalizationFlags",
")",
"\n",
"w",
".",
"enqueue",
"<-",
"rCtx",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n\n",
"// No fetch, so set to nil",
"w",
".",
"lastFetch",
"=",
"nil",
"\n\n",
"if",
"!",
"silent",
"{",
"// Notify error",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"ctx",
",",
"e",
",",
"CekFetch",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
",",
"e",
")",
"\n",
"}",
"\n\n",
"// Return from this URL crawl",
"w",
".",
"sendResponse",
"(",
"ctx",
",",
"false",
",",
"nil",
",",
"false",
")",
"\n",
"return",
"nil",
",",
"false",
"\n\n",
"}",
"\n",
"// Get the fetch duration",
"fetchDuration",
":=",
"time",
".",
"Now",
"(",
")",
".",
"Sub",
"(",
"now",
")",
"\n",
"// Crawl delay starts now.",
"w",
".",
"wait",
"=",
"time",
".",
"After",
"(",
"w",
".",
"lastCrawlDelay",
")",
"\n\n",
"// Keep trace of this last fetch info",
"w",
".",
"lastFetch",
"=",
"&",
"FetchInfo",
"{",
"ctx",
",",
"fetchDuration",
",",
"res",
".",
"StatusCode",
",",
"headRequest",
",",
"}",
"\n\n",
"if",
"headRequest",
"{",
"// Close the HEAD request's body",
"defer",
"res",
".",
"Body",
".",
"Close",
"(",
")",
"\n",
"// Next up is GET request, maybe",
"headRequest",
"=",
"false",
"\n",
"// Ask caller if we should proceed with a GET",
"if",
"!",
"w",
".",
"opts",
".",
"Extender",
".",
"RequestGet",
"(",
"ctx",
",",
"res",
")",
"{",
"w",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
")",
"\n",
"w",
".",
"sendResponse",
"(",
"ctx",
",",
"false",
",",
"nil",
",",
"false",
")",
"\n",
"ok",
"=",
"false",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"else",
"{",
"ok",
"=",
"true",
"\n",
"break",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Request the specified URL and return the response. | [
"Request",
"the",
"specified",
"URL",
"and",
"return",
"the",
"response",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L206-L295 |
5,525 | PuerkitoBio/gocrawl | worker.go | sendResponse | func (w *worker) sendResponse(ctx *URLContext, visited bool, harvested interface{}, idleDeath bool) {
// Push harvested urls back to crawler, even if empty (uses the channel communication
// to decrement reference count of pending URLs)
if ctx == nil || !isRobotsURL(ctx.url) {
// If a stop signal has been received, ignore the response, since the push
// channel may be full and could block indefinitely.
select {
case <-w.stop:
w.logFunc(LogInfo, "ignoring send response, will stop.")
return
default:
// Nothing, just continue...
}
// No stop signal, send the response
res := &workerResponse{
ctx,
visited,
harvested,
w.host,
idleDeath,
}
w.push <- res
}
} | go | func (w *worker) sendResponse(ctx *URLContext, visited bool, harvested interface{}, idleDeath bool) {
// Push harvested urls back to crawler, even if empty (uses the channel communication
// to decrement reference count of pending URLs)
if ctx == nil || !isRobotsURL(ctx.url) {
// If a stop signal has been received, ignore the response, since the push
// channel may be full and could block indefinitely.
select {
case <-w.stop:
w.logFunc(LogInfo, "ignoring send response, will stop.")
return
default:
// Nothing, just continue...
}
// No stop signal, send the response
res := &workerResponse{
ctx,
visited,
harvested,
w.host,
idleDeath,
}
w.push <- res
}
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"sendResponse",
"(",
"ctx",
"*",
"URLContext",
",",
"visited",
"bool",
",",
"harvested",
"interface",
"{",
"}",
",",
"idleDeath",
"bool",
")",
"{",
"// Push harvested urls back to crawler, even if empty (uses the channel communication",
"// to decrement reference count of pending URLs)",
"if",
"ctx",
"==",
"nil",
"||",
"!",
"isRobotsURL",
"(",
"ctx",
".",
"url",
")",
"{",
"// If a stop signal has been received, ignore the response, since the push",
"// channel may be full and could block indefinitely.",
"select",
"{",
"case",
"<-",
"w",
".",
"stop",
":",
"w",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"default",
":",
"// Nothing, just continue...",
"}",
"\n\n",
"// No stop signal, send the response",
"res",
":=",
"&",
"workerResponse",
"{",
"ctx",
",",
"visited",
",",
"harvested",
",",
"w",
".",
"host",
",",
"idleDeath",
",",
"}",
"\n",
"w",
".",
"push",
"<-",
"res",
"\n",
"}",
"\n",
"}"
] | // Send a response to the crawler. | [
"Send",
"a",
"response",
"to",
"the",
"crawler",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L298-L322 |
5,526 | PuerkitoBio/gocrawl | worker.go | visitURL | func (w *worker) visitURL(ctx *URLContext, res *http.Response) interface{} {
var doc *goquery.Document
var harvested interface{}
var doLinks bool
// Load a goquery document and call the visitor function
if bd, e := ioutil.ReadAll(res.Body); e != nil {
w.opts.Extender.Error(newCrawlError(ctx, e, CekReadBody))
w.logFunc(LogError, "ERROR reading body %s: %s", ctx.url, e)
} else {
if node, e := html.Parse(bytes.NewBuffer(bd)); e != nil {
w.opts.Extender.Error(newCrawlError(ctx, e, CekParseBody))
w.logFunc(LogError, "ERROR parsing %s: %s", ctx.url, e)
} else {
doc = goquery.NewDocumentFromNode(node)
doc.Url = res.Request.URL
}
// Re-assign the body so it can be consumed by the visitor function
res.Body = ioutil.NopCloser(bytes.NewBuffer(bd))
}
// Visit the document (with nil goquery doc if failed to load)
if harvested, doLinks = w.opts.Extender.Visit(ctx, res, doc); doLinks {
// Links were not processed by the visitor, so process links
if doc != nil {
harvested = w.processLinks(doc)
} else {
w.opts.Extender.Error(newCrawlErrorMessage(ctx, "No goquery document to process links.", CekProcessLinks))
w.logFunc(LogError, "ERROR processing links %s", ctx.url)
}
}
// Notify that this URL has been visited
w.opts.Extender.Visited(ctx, harvested)
return harvested
} | go | func (w *worker) visitURL(ctx *URLContext, res *http.Response) interface{} {
var doc *goquery.Document
var harvested interface{}
var doLinks bool
// Load a goquery document and call the visitor function
if bd, e := ioutil.ReadAll(res.Body); e != nil {
w.opts.Extender.Error(newCrawlError(ctx, e, CekReadBody))
w.logFunc(LogError, "ERROR reading body %s: %s", ctx.url, e)
} else {
if node, e := html.Parse(bytes.NewBuffer(bd)); e != nil {
w.opts.Extender.Error(newCrawlError(ctx, e, CekParseBody))
w.logFunc(LogError, "ERROR parsing %s: %s", ctx.url, e)
} else {
doc = goquery.NewDocumentFromNode(node)
doc.Url = res.Request.URL
}
// Re-assign the body so it can be consumed by the visitor function
res.Body = ioutil.NopCloser(bytes.NewBuffer(bd))
}
// Visit the document (with nil goquery doc if failed to load)
if harvested, doLinks = w.opts.Extender.Visit(ctx, res, doc); doLinks {
// Links were not processed by the visitor, so process links
if doc != nil {
harvested = w.processLinks(doc)
} else {
w.opts.Extender.Error(newCrawlErrorMessage(ctx, "No goquery document to process links.", CekProcessLinks))
w.logFunc(LogError, "ERROR processing links %s", ctx.url)
}
}
// Notify that this URL has been visited
w.opts.Extender.Visited(ctx, harvested)
return harvested
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"visitURL",
"(",
"ctx",
"*",
"URLContext",
",",
"res",
"*",
"http",
".",
"Response",
")",
"interface",
"{",
"}",
"{",
"var",
"doc",
"*",
"goquery",
".",
"Document",
"\n",
"var",
"harvested",
"interface",
"{",
"}",
"\n",
"var",
"doLinks",
"bool",
"\n\n",
"// Load a goquery document and call the visitor function",
"if",
"bd",
",",
"e",
":=",
"ioutil",
".",
"ReadAll",
"(",
"res",
".",
"Body",
")",
";",
"e",
"!=",
"nil",
"{",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"ctx",
",",
"e",
",",
"CekReadBody",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
",",
"e",
")",
"\n",
"}",
"else",
"{",
"if",
"node",
",",
"e",
":=",
"html",
".",
"Parse",
"(",
"bytes",
".",
"NewBuffer",
"(",
"bd",
")",
")",
";",
"e",
"!=",
"nil",
"{",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"ctx",
",",
"e",
",",
"CekParseBody",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
",",
"e",
")",
"\n",
"}",
"else",
"{",
"doc",
"=",
"goquery",
".",
"NewDocumentFromNode",
"(",
"node",
")",
"\n",
"doc",
".",
"Url",
"=",
"res",
".",
"Request",
".",
"URL",
"\n",
"}",
"\n",
"// Re-assign the body so it can be consumed by the visitor function",
"res",
".",
"Body",
"=",
"ioutil",
".",
"NopCloser",
"(",
"bytes",
".",
"NewBuffer",
"(",
"bd",
")",
")",
"\n",
"}",
"\n\n",
"// Visit the document (with nil goquery doc if failed to load)",
"if",
"harvested",
",",
"doLinks",
"=",
"w",
".",
"opts",
".",
"Extender",
".",
"Visit",
"(",
"ctx",
",",
"res",
",",
"doc",
")",
";",
"doLinks",
"{",
"// Links were not processed by the visitor, so process links",
"if",
"doc",
"!=",
"nil",
"{",
"harvested",
"=",
"w",
".",
"processLinks",
"(",
"doc",
")",
"\n",
"}",
"else",
"{",
"w",
".",
"opts",
".",
"Extender",
".",
"Error",
"(",
"newCrawlErrorMessage",
"(",
"ctx",
",",
"\"",
"\"",
",",
"CekProcessLinks",
")",
")",
"\n",
"w",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
")",
"\n",
"}",
"\n",
"}",
"\n",
"// Notify that this URL has been visited",
"w",
".",
"opts",
".",
"Extender",
".",
"Visited",
"(",
"ctx",
",",
"harvested",
")",
"\n\n",
"return",
"harvested",
"\n",
"}"
] | // Process the response for a URL. | [
"Process",
"the",
"response",
"for",
"a",
"URL",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L325-L360 |
5,527 | PuerkitoBio/gocrawl | worker.go | processLinks | func (w *worker) processLinks(doc *goquery.Document) (result []*url.URL) {
baseURL, _ := doc.Find("base[href]").Attr("href")
urls := doc.Find("a[href]").Map(func(_ int, s *goquery.Selection) string {
val, _ := s.Attr("href")
if baseURL != "" {
val = handleBaseTag(doc.Url, baseURL, val)
}
return val
})
for _, s := range urls {
// If href starts with "#", then it points to this same exact URL, ignore (will fail to parse anyway)
if len(s) > 0 && !strings.HasPrefix(s, "#") {
if parsed, e := url.Parse(s); e == nil {
parsed = doc.Url.ResolveReference(parsed)
result = append(result, parsed)
} else {
w.logFunc(LogIgnored, "ignore on unparsable policy %s: %s", s, e.Error())
}
}
}
return
} | go | func (w *worker) processLinks(doc *goquery.Document) (result []*url.URL) {
baseURL, _ := doc.Find("base[href]").Attr("href")
urls := doc.Find("a[href]").Map(func(_ int, s *goquery.Selection) string {
val, _ := s.Attr("href")
if baseURL != "" {
val = handleBaseTag(doc.Url, baseURL, val)
}
return val
})
for _, s := range urls {
// If href starts with "#", then it points to this same exact URL, ignore (will fail to parse anyway)
if len(s) > 0 && !strings.HasPrefix(s, "#") {
if parsed, e := url.Parse(s); e == nil {
parsed = doc.Url.ResolveReference(parsed)
result = append(result, parsed)
} else {
w.logFunc(LogIgnored, "ignore on unparsable policy %s: %s", s, e.Error())
}
}
}
return
} | [
"func",
"(",
"w",
"*",
"worker",
")",
"processLinks",
"(",
"doc",
"*",
"goquery",
".",
"Document",
")",
"(",
"result",
"[",
"]",
"*",
"url",
".",
"URL",
")",
"{",
"baseURL",
",",
"_",
":=",
"doc",
".",
"Find",
"(",
"\"",
"\"",
")",
".",
"Attr",
"(",
"\"",
"\"",
")",
"\n",
"urls",
":=",
"doc",
".",
"Find",
"(",
"\"",
"\"",
")",
".",
"Map",
"(",
"func",
"(",
"_",
"int",
",",
"s",
"*",
"goquery",
".",
"Selection",
")",
"string",
"{",
"val",
",",
"_",
":=",
"s",
".",
"Attr",
"(",
"\"",
"\"",
")",
"\n",
"if",
"baseURL",
"!=",
"\"",
"\"",
"{",
"val",
"=",
"handleBaseTag",
"(",
"doc",
".",
"Url",
",",
"baseURL",
",",
"val",
")",
"\n",
"}",
"\n",
"return",
"val",
"\n",
"}",
")",
"\n",
"for",
"_",
",",
"s",
":=",
"range",
"urls",
"{",
"// If href starts with \"#\", then it points to this same exact URL, ignore (will fail to parse anyway)",
"if",
"len",
"(",
"s",
")",
">",
"0",
"&&",
"!",
"strings",
".",
"HasPrefix",
"(",
"s",
",",
"\"",
"\"",
")",
"{",
"if",
"parsed",
",",
"e",
":=",
"url",
".",
"Parse",
"(",
"s",
")",
";",
"e",
"==",
"nil",
"{",
"parsed",
"=",
"doc",
".",
"Url",
".",
"ResolveReference",
"(",
"parsed",
")",
"\n",
"result",
"=",
"append",
"(",
"result",
",",
"parsed",
")",
"\n",
"}",
"else",
"{",
"w",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"s",
",",
"e",
".",
"Error",
"(",
")",
")",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Scrape the document's content to gather all links | [
"Scrape",
"the",
"document",
"s",
"content",
"to",
"gather",
"all",
"links"
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/worker.go#L385-L406 |
5,528 | PuerkitoBio/gocrawl | ext.go | Log | func (de *DefaultExtender) Log(logFlags LogFlags, msgLevel LogFlags, msg string) {
if logFlags&msgLevel == msgLevel {
log.Println(msg)
}
} | go | func (de *DefaultExtender) Log(logFlags LogFlags, msgLevel LogFlags, msg string) {
if logFlags&msgLevel == msgLevel {
log.Println(msg)
}
} | [
"func",
"(",
"de",
"*",
"DefaultExtender",
")",
"Log",
"(",
"logFlags",
"LogFlags",
",",
"msgLevel",
"LogFlags",
",",
"msg",
"string",
")",
"{",
"if",
"logFlags",
"&",
"msgLevel",
"==",
"msgLevel",
"{",
"log",
".",
"Println",
"(",
"msg",
")",
"\n",
"}",
"\n",
"}"
] | // Log prints to the standard error by default, based on the requested log verbosity. | [
"Log",
"prints",
"to",
"the",
"standard",
"error",
"by",
"default",
"based",
"on",
"the",
"requested",
"log",
"verbosity",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L101-L105 |
5,529 | PuerkitoBio/gocrawl | ext.go | ComputeDelay | func (de *DefaultExtender) ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration {
if di.RobotsDelay > 0 {
return di.RobotsDelay
}
return di.OptsDelay
} | go | func (de *DefaultExtender) ComputeDelay(host string, di *DelayInfo, lastFetch *FetchInfo) time.Duration {
if di.RobotsDelay > 0 {
return di.RobotsDelay
}
return di.OptsDelay
} | [
"func",
"(",
"de",
"*",
"DefaultExtender",
")",
"ComputeDelay",
"(",
"host",
"string",
",",
"di",
"*",
"DelayInfo",
",",
"lastFetch",
"*",
"FetchInfo",
")",
"time",
".",
"Duration",
"{",
"if",
"di",
".",
"RobotsDelay",
">",
"0",
"{",
"return",
"di",
".",
"RobotsDelay",
"\n",
"}",
"\n",
"return",
"di",
".",
"OptsDelay",
"\n",
"}"
] | // ComputeDelay returns the delay specified in the Crawler's Options, unless a
// crawl-delay is specified in the robots.txt file, which has precedence. | [
"ComputeDelay",
"returns",
"the",
"delay",
"specified",
"in",
"the",
"Crawler",
"s",
"Options",
"unless",
"a",
"crawl",
"-",
"delay",
"is",
"specified",
"in",
"the",
"robots",
".",
"txt",
"file",
"which",
"has",
"precedence",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L109-L114 |
5,530 | PuerkitoBio/gocrawl | ext.go | Visit | func (de *DefaultExtender) Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool) {
return nil, true
} | go | func (de *DefaultExtender) Visit(ctx *URLContext, res *http.Response, doc *goquery.Document) (harvested interface{}, findLinks bool) {
return nil, true
} | [
"func",
"(",
"de",
"*",
"DefaultExtender",
")",
"Visit",
"(",
"ctx",
"*",
"URLContext",
",",
"res",
"*",
"http",
".",
"Response",
",",
"doc",
"*",
"goquery",
".",
"Document",
")",
"(",
"harvested",
"interface",
"{",
"}",
",",
"findLinks",
"bool",
")",
"{",
"return",
"nil",
",",
"true",
"\n",
"}"
] | // Visit asks the worker to harvest the links in this page. | [
"Visit",
"asks",
"the",
"worker",
"to",
"harvest",
"the",
"links",
"in",
"this",
"page",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/ext.go#L180-L182 |
5,531 | PuerkitoBio/gocrawl | crawler.go | NewCrawlerWithOptions | func NewCrawlerWithOptions(opts *Options) *Crawler {
ret := new(Crawler)
ret.Options = opts
return ret
} | go | func NewCrawlerWithOptions(opts *Options) *Crawler {
ret := new(Crawler)
ret.Options = opts
return ret
} | [
"func",
"NewCrawlerWithOptions",
"(",
"opts",
"*",
"Options",
")",
"*",
"Crawler",
"{",
"ret",
":=",
"new",
"(",
"Crawler",
")",
"\n",
"ret",
".",
"Options",
"=",
"opts",
"\n",
"return",
"ret",
"\n",
"}"
] | // NewCrawlerWithOptions returns a Crawler initialized with the
// provided Options. | [
"NewCrawlerWithOptions",
"returns",
"a",
"Crawler",
"initialized",
"with",
"the",
"provided",
"Options",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L42-L46 |
5,532 | PuerkitoBio/gocrawl | crawler.go | init | func (c *Crawler) init(ctxs []*URLContext) {
// Initialize the internal hosts map
c.hosts = make(map[string]struct{}, len(ctxs))
for _, ctx := range ctxs {
// Add this normalized URL's host if it is not already there.
if _, ok := c.hosts[ctx.normalizedURL.Host]; !ok {
c.hosts[ctx.normalizedURL.Host] = struct{}{}
}
}
hostCount := len(c.hosts)
l := len(ctxs)
c.logFunc(LogTrace, "init() - seeds length: %d", l)
c.logFunc(LogTrace, "init() - host count: %d", hostCount)
c.logFunc(LogInfo, "robot user-agent: %s", c.Options.RobotUserAgent)
// Create a shiny new WaitGroup
c.wg = new(sync.WaitGroup)
// Initialize the visits fields
c.visited = make(map[string]struct{}, l)
c.pushPopRefCount, c.visits = 0, 0
// Create the workers map and the push channel (the channel used by workers
// to communicate back to the crawler)
c.stop = make(chan struct{})
if c.Options.SameHostOnly {
c.workers, c.push = make(map[string]*worker, hostCount),
make(chan *workerResponse, hostCount)
} else {
c.workers, c.push = make(map[string]*worker, c.Options.HostBufferFactor*hostCount),
make(chan *workerResponse, c.Options.HostBufferFactor*hostCount)
}
// Create and pass the enqueue channel
c.enqueue = make(chan interface{}, c.Options.EnqueueChanBuffer)
c.setExtenderEnqueueChan()
} | go | func (c *Crawler) init(ctxs []*URLContext) {
// Initialize the internal hosts map
c.hosts = make(map[string]struct{}, len(ctxs))
for _, ctx := range ctxs {
// Add this normalized URL's host if it is not already there.
if _, ok := c.hosts[ctx.normalizedURL.Host]; !ok {
c.hosts[ctx.normalizedURL.Host] = struct{}{}
}
}
hostCount := len(c.hosts)
l := len(ctxs)
c.logFunc(LogTrace, "init() - seeds length: %d", l)
c.logFunc(LogTrace, "init() - host count: %d", hostCount)
c.logFunc(LogInfo, "robot user-agent: %s", c.Options.RobotUserAgent)
// Create a shiny new WaitGroup
c.wg = new(sync.WaitGroup)
// Initialize the visits fields
c.visited = make(map[string]struct{}, l)
c.pushPopRefCount, c.visits = 0, 0
// Create the workers map and the push channel (the channel used by workers
// to communicate back to the crawler)
c.stop = make(chan struct{})
if c.Options.SameHostOnly {
c.workers, c.push = make(map[string]*worker, hostCount),
make(chan *workerResponse, hostCount)
} else {
c.workers, c.push = make(map[string]*worker, c.Options.HostBufferFactor*hostCount),
make(chan *workerResponse, c.Options.HostBufferFactor*hostCount)
}
// Create and pass the enqueue channel
c.enqueue = make(chan interface{}, c.Options.EnqueueChanBuffer)
c.setExtenderEnqueueChan()
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"init",
"(",
"ctxs",
"[",
"]",
"*",
"URLContext",
")",
"{",
"// Initialize the internal hosts map",
"c",
".",
"hosts",
"=",
"make",
"(",
"map",
"[",
"string",
"]",
"struct",
"{",
"}",
",",
"len",
"(",
"ctxs",
")",
")",
"\n",
"for",
"_",
",",
"ctx",
":=",
"range",
"ctxs",
"{",
"// Add this normalized URL's host if it is not already there.",
"if",
"_",
",",
"ok",
":=",
"c",
".",
"hosts",
"[",
"ctx",
".",
"normalizedURL",
".",
"Host",
"]",
";",
"!",
"ok",
"{",
"c",
".",
"hosts",
"[",
"ctx",
".",
"normalizedURL",
".",
"Host",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"}",
"\n",
"}",
"\n\n",
"hostCount",
":=",
"len",
"(",
"c",
".",
"hosts",
")",
"\n",
"l",
":=",
"len",
"(",
"ctxs",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogTrace",
",",
"\"",
"\"",
",",
"l",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogTrace",
",",
"\"",
"\"",
",",
"hostCount",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
",",
"c",
".",
"Options",
".",
"RobotUserAgent",
")",
"\n\n",
"// Create a shiny new WaitGroup",
"c",
".",
"wg",
"=",
"new",
"(",
"sync",
".",
"WaitGroup",
")",
"\n\n",
"// Initialize the visits fields",
"c",
".",
"visited",
"=",
"make",
"(",
"map",
"[",
"string",
"]",
"struct",
"{",
"}",
",",
"l",
")",
"\n",
"c",
".",
"pushPopRefCount",
",",
"c",
".",
"visits",
"=",
"0",
",",
"0",
"\n\n",
"// Create the workers map and the push channel (the channel used by workers",
"// to communicate back to the crawler)",
"c",
".",
"stop",
"=",
"make",
"(",
"chan",
"struct",
"{",
"}",
")",
"\n",
"if",
"c",
".",
"Options",
".",
"SameHostOnly",
"{",
"c",
".",
"workers",
",",
"c",
".",
"push",
"=",
"make",
"(",
"map",
"[",
"string",
"]",
"*",
"worker",
",",
"hostCount",
")",
",",
"make",
"(",
"chan",
"*",
"workerResponse",
",",
"hostCount",
")",
"\n",
"}",
"else",
"{",
"c",
".",
"workers",
",",
"c",
".",
"push",
"=",
"make",
"(",
"map",
"[",
"string",
"]",
"*",
"worker",
",",
"c",
".",
"Options",
".",
"HostBufferFactor",
"*",
"hostCount",
")",
",",
"make",
"(",
"chan",
"*",
"workerResponse",
",",
"c",
".",
"Options",
".",
"HostBufferFactor",
"*",
"hostCount",
")",
"\n",
"}",
"\n",
"// Create and pass the enqueue channel",
"c",
".",
"enqueue",
"=",
"make",
"(",
"chan",
"interface",
"{",
"}",
",",
"c",
".",
"Options",
".",
"EnqueueChanBuffer",
")",
"\n",
"c",
".",
"setExtenderEnqueueChan",
"(",
")",
"\n",
"}"
] | // Initialize the Crawler's internal fields before a crawling execution. | [
"Initialize",
"the",
"Crawler",
"s",
"internal",
"fields",
"before",
"a",
"crawling",
"execution",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L78-L114 |
5,533 | PuerkitoBio/gocrawl | crawler.go | setExtenderEnqueueChan | func (c *Crawler) setExtenderEnqueueChan() {
defer func() {
if err := recover(); err != nil {
// Panic can happen if the field exists on a pointer struct, but that
// pointer is nil.
c.logFunc(LogError, "cannot set the enqueue channel: %s", err)
}
}()
// Using reflection, check if the extender has a `EnqueueChan` field
// of type `chan<- interface{}`. If it does, set it to the crawler's
// enqueue channel.
v := reflect.ValueOf(c.Options.Extender)
el := v.Elem()
if el.Kind() != reflect.Struct {
c.logFunc(LogInfo, "extender is not a struct, cannot set the enqueue channel")
return
}
ec := el.FieldByName("EnqueueChan")
if !ec.IsValid() {
c.logFunc(LogInfo, "extender.EnqueueChan does not exist, cannot set the enqueue channel")
return
}
t := ec.Type()
if t.Kind() != reflect.Chan || t.ChanDir() != reflect.SendDir {
c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel")
return
}
tt := t.Elem()
if tt.Kind() != reflect.Interface || tt.NumMethod() != 0 {
c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel")
return
}
src := reflect.ValueOf(c.enqueue)
ec.Set(src)
} | go | func (c *Crawler) setExtenderEnqueueChan() {
defer func() {
if err := recover(); err != nil {
// Panic can happen if the field exists on a pointer struct, but that
// pointer is nil.
c.logFunc(LogError, "cannot set the enqueue channel: %s", err)
}
}()
// Using reflection, check if the extender has a `EnqueueChan` field
// of type `chan<- interface{}`. If it does, set it to the crawler's
// enqueue channel.
v := reflect.ValueOf(c.Options.Extender)
el := v.Elem()
if el.Kind() != reflect.Struct {
c.logFunc(LogInfo, "extender is not a struct, cannot set the enqueue channel")
return
}
ec := el.FieldByName("EnqueueChan")
if !ec.IsValid() {
c.logFunc(LogInfo, "extender.EnqueueChan does not exist, cannot set the enqueue channel")
return
}
t := ec.Type()
if t.Kind() != reflect.Chan || t.ChanDir() != reflect.SendDir {
c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel")
return
}
tt := t.Elem()
if tt.Kind() != reflect.Interface || tt.NumMethod() != 0 {
c.logFunc(LogInfo, "extender.EnqueueChan is not of type chan<-interface{}, cannot set the enqueue channel")
return
}
src := reflect.ValueOf(c.enqueue)
ec.Set(src)
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"setExtenderEnqueueChan",
"(",
")",
"{",
"defer",
"func",
"(",
")",
"{",
"if",
"err",
":=",
"recover",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"// Panic can happen if the field exists on a pointer struct, but that",
"// pointer is nil.",
"c",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"(",
")",
"\n\n",
"// Using reflection, check if the extender has a `EnqueueChan` field",
"// of type `chan<- interface{}`. If it does, set it to the crawler's",
"// enqueue channel.",
"v",
":=",
"reflect",
".",
"ValueOf",
"(",
"c",
".",
"Options",
".",
"Extender",
")",
"\n",
"el",
":=",
"v",
".",
"Elem",
"(",
")",
"\n",
"if",
"el",
".",
"Kind",
"(",
")",
"!=",
"reflect",
".",
"Struct",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"}",
"\n",
"ec",
":=",
"el",
".",
"FieldByName",
"(",
"\"",
"\"",
")",
"\n",
"if",
"!",
"ec",
".",
"IsValid",
"(",
")",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"}",
"\n",
"t",
":=",
"ec",
".",
"Type",
"(",
")",
"\n",
"if",
"t",
".",
"Kind",
"(",
")",
"!=",
"reflect",
".",
"Chan",
"||",
"t",
".",
"ChanDir",
"(",
")",
"!=",
"reflect",
".",
"SendDir",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"}",
"\n",
"tt",
":=",
"t",
".",
"Elem",
"(",
")",
"\n",
"if",
"tt",
".",
"Kind",
"(",
")",
"!=",
"reflect",
".",
"Interface",
"||",
"tt",
".",
"NumMethod",
"(",
")",
"!=",
"0",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"return",
"\n",
"}",
"\n",
"src",
":=",
"reflect",
".",
"ValueOf",
"(",
"c",
".",
"enqueue",
")",
"\n",
"ec",
".",
"Set",
"(",
"src",
")",
"\n",
"}"
] | // Set the Enqueue channel on the extender, based on the naming convention. | [
"Set",
"the",
"Enqueue",
"channel",
"on",
"the",
"extender",
"based",
"on",
"the",
"naming",
"convention",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L117-L152 |
5,534 | PuerkitoBio/gocrawl | crawler.go | launchWorker | func (c *Crawler) launchWorker(ctx *URLContext) *worker {
// Initialize index and channels
i := len(c.workers) + 1
pop := newPopChannel()
// Create the worker
w := &worker{
host: ctx.normalizedURL.Host,
index: i,
push: c.push,
pop: pop,
stop: c.stop,
enqueue: c.enqueue,
wg: c.wg,
logFunc: getLogFunc(c.Options.Extender, c.Options.LogFlags, i),
opts: c.Options,
}
// Increment wait group count
c.wg.Add(1)
// Launch worker
go w.run()
c.logFunc(LogInfo, "worker %d launched for host %s", i, w.host)
c.workers[w.host] = w
return w
} | go | func (c *Crawler) launchWorker(ctx *URLContext) *worker {
// Initialize index and channels
i := len(c.workers) + 1
pop := newPopChannel()
// Create the worker
w := &worker{
host: ctx.normalizedURL.Host,
index: i,
push: c.push,
pop: pop,
stop: c.stop,
enqueue: c.enqueue,
wg: c.wg,
logFunc: getLogFunc(c.Options.Extender, c.Options.LogFlags, i),
opts: c.Options,
}
// Increment wait group count
c.wg.Add(1)
// Launch worker
go w.run()
c.logFunc(LogInfo, "worker %d launched for host %s", i, w.host)
c.workers[w.host] = w
return w
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"launchWorker",
"(",
"ctx",
"*",
"URLContext",
")",
"*",
"worker",
"{",
"// Initialize index and channels",
"i",
":=",
"len",
"(",
"c",
".",
"workers",
")",
"+",
"1",
"\n",
"pop",
":=",
"newPopChannel",
"(",
")",
"\n\n",
"// Create the worker",
"w",
":=",
"&",
"worker",
"{",
"host",
":",
"ctx",
".",
"normalizedURL",
".",
"Host",
",",
"index",
":",
"i",
",",
"push",
":",
"c",
".",
"push",
",",
"pop",
":",
"pop",
",",
"stop",
":",
"c",
".",
"stop",
",",
"enqueue",
":",
"c",
".",
"enqueue",
",",
"wg",
":",
"c",
".",
"wg",
",",
"logFunc",
":",
"getLogFunc",
"(",
"c",
".",
"Options",
".",
"Extender",
",",
"c",
".",
"Options",
".",
"LogFlags",
",",
"i",
")",
",",
"opts",
":",
"c",
".",
"Options",
",",
"}",
"\n\n",
"// Increment wait group count",
"c",
".",
"wg",
".",
"Add",
"(",
"1",
")",
"\n\n",
"// Launch worker",
"go",
"w",
".",
"run",
"(",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
",",
"i",
",",
"w",
".",
"host",
")",
"\n",
"c",
".",
"workers",
"[",
"w",
".",
"host",
"]",
"=",
"w",
"\n\n",
"return",
"w",
"\n",
"}"
] | // Launch a new worker goroutine for a given host. | [
"Launch",
"a",
"new",
"worker",
"goroutine",
"for",
"a",
"given",
"host",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L155-L182 |
5,535 | PuerkitoBio/gocrawl | crawler.go | isSameHost | func (c *Crawler) isSameHost(ctx *URLContext) bool {
// If there is a source URL, then just check if the new URL is from the same host
if ctx.normalizedSourceURL != nil {
return ctx.normalizedURL.Host == ctx.normalizedSourceURL.Host
}
// Otherwise, check if the URL is from one of the seed hosts
_, ok := c.hosts[ctx.normalizedURL.Host]
return ok
} | go | func (c *Crawler) isSameHost(ctx *URLContext) bool {
// If there is a source URL, then just check if the new URL is from the same host
if ctx.normalizedSourceURL != nil {
return ctx.normalizedURL.Host == ctx.normalizedSourceURL.Host
}
// Otherwise, check if the URL is from one of the seed hosts
_, ok := c.hosts[ctx.normalizedURL.Host]
return ok
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"isSameHost",
"(",
"ctx",
"*",
"URLContext",
")",
"bool",
"{",
"// If there is a source URL, then just check if the new URL is from the same host",
"if",
"ctx",
".",
"normalizedSourceURL",
"!=",
"nil",
"{",
"return",
"ctx",
".",
"normalizedURL",
".",
"Host",
"==",
"ctx",
".",
"normalizedSourceURL",
".",
"Host",
"\n",
"}",
"\n\n",
"// Otherwise, check if the URL is from one of the seed hosts",
"_",
",",
"ok",
":=",
"c",
".",
"hosts",
"[",
"ctx",
".",
"normalizedURL",
".",
"Host",
"]",
"\n",
"return",
"ok",
"\n",
"}"
] | // Check if the specified URL is from the same host as its source URL, or if
// nil, from the same host as one of the seed URLs. | [
"Check",
"if",
"the",
"specified",
"URL",
"is",
"from",
"the",
"same",
"host",
"as",
"its",
"source",
"URL",
"or",
"if",
"nil",
"from",
"the",
"same",
"host",
"as",
"one",
"of",
"the",
"seed",
"URLs",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L186-L195 |
5,536 | PuerkitoBio/gocrawl | crawler.go | enqueueUrls | func (c *Crawler) enqueueUrls(ctxs []*URLContext) (cnt int) {
for _, ctx := range ctxs {
var isVisited, enqueue bool
// Cannot directly enqueue a robots.txt URL, since it is managed as a special case
// in the worker (doesn't return a response to crawler).
if ctx.IsRobotsURL() {
continue
}
// Check if it has been visited before, using the normalized URL
_, isVisited = c.visited[ctx.normalizedURL.String()]
// Filter the URL
if enqueue = c.Options.Extender.Filter(ctx, isVisited); !enqueue {
// Filter said NOT to use this url, so continue with next
c.logFunc(LogIgnored, "ignore on filter policy: %s", ctx.normalizedURL)
continue
}
// Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed,
// and comply with the same host policy if requested.
if !ctx.normalizedURL.IsAbs() {
// Only absolute URLs are processed, so ignore
c.logFunc(LogIgnored, "ignore on absolute policy: %s", ctx.normalizedURL)
} else if !strings.HasPrefix(ctx.normalizedURL.Scheme, "http") {
c.logFunc(LogIgnored, "ignore on scheme policy: %s", ctx.normalizedURL)
} else if c.Options.SameHostOnly && !c.isSameHost(ctx) {
// Only allow URLs coming from the same host
c.logFunc(LogIgnored, "ignore on same host policy: %s", ctx.normalizedURL)
} else {
// All is good, visit this URL (robots.txt verification is done by worker)
// Possible caveat: if the normalization changes the host, it is possible
// that the robots.txt fetched for this host would differ from the one for
// the unnormalized host. However, this should be rare, and is a weird
// behaviour from the host (i.e. why would site.com differ in its rules
// from www.site.com) and can be fixed by using a different normalization
// flag. So this is an acceptable behaviour for gocrawl.
// Launch worker if required, based on the host of the normalized URL
w, ok := c.workers[ctx.normalizedURL.Host]
if !ok {
// No worker exists for this host, launch a new one
w = c.launchWorker(ctx)
// Automatically enqueue the robots.txt URL as first in line
if robCtx, e := ctx.getRobotsURLCtx(); e != nil {
c.Options.Extender.Error(newCrawlError(ctx, e, CekParseRobots))
c.logFunc(LogError, "ERROR parsing robots.txt from %s: %s", ctx.normalizedURL, e)
} else {
c.logFunc(LogEnqueued, "enqueue: %s", robCtx.url)
c.Options.Extender.Enqueued(robCtx)
w.pop.stack(robCtx)
}
}
cnt++
c.logFunc(LogEnqueued, "enqueue: %s", ctx.url)
c.Options.Extender.Enqueued(ctx)
w.pop.stack(ctx)
c.pushPopRefCount++
// Once it is stacked, it WILL be visited eventually, so add it to the visited slice
// (unless denied by robots.txt, but this is out of our hands, for all we
// care, it is visited).
if !isVisited {
// The visited map works with the normalized URL
c.visited[ctx.normalizedURL.String()] = struct{}{}
}
}
}
return
} | go | func (c *Crawler) enqueueUrls(ctxs []*URLContext) (cnt int) {
for _, ctx := range ctxs {
var isVisited, enqueue bool
// Cannot directly enqueue a robots.txt URL, since it is managed as a special case
// in the worker (doesn't return a response to crawler).
if ctx.IsRobotsURL() {
continue
}
// Check if it has been visited before, using the normalized URL
_, isVisited = c.visited[ctx.normalizedURL.String()]
// Filter the URL
if enqueue = c.Options.Extender.Filter(ctx, isVisited); !enqueue {
// Filter said NOT to use this url, so continue with next
c.logFunc(LogIgnored, "ignore on filter policy: %s", ctx.normalizedURL)
continue
}
// Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed,
// and comply with the same host policy if requested.
if !ctx.normalizedURL.IsAbs() {
// Only absolute URLs are processed, so ignore
c.logFunc(LogIgnored, "ignore on absolute policy: %s", ctx.normalizedURL)
} else if !strings.HasPrefix(ctx.normalizedURL.Scheme, "http") {
c.logFunc(LogIgnored, "ignore on scheme policy: %s", ctx.normalizedURL)
} else if c.Options.SameHostOnly && !c.isSameHost(ctx) {
// Only allow URLs coming from the same host
c.logFunc(LogIgnored, "ignore on same host policy: %s", ctx.normalizedURL)
} else {
// All is good, visit this URL (robots.txt verification is done by worker)
// Possible caveat: if the normalization changes the host, it is possible
// that the robots.txt fetched for this host would differ from the one for
// the unnormalized host. However, this should be rare, and is a weird
// behaviour from the host (i.e. why would site.com differ in its rules
// from www.site.com) and can be fixed by using a different normalization
// flag. So this is an acceptable behaviour for gocrawl.
// Launch worker if required, based on the host of the normalized URL
w, ok := c.workers[ctx.normalizedURL.Host]
if !ok {
// No worker exists for this host, launch a new one
w = c.launchWorker(ctx)
// Automatically enqueue the robots.txt URL as first in line
if robCtx, e := ctx.getRobotsURLCtx(); e != nil {
c.Options.Extender.Error(newCrawlError(ctx, e, CekParseRobots))
c.logFunc(LogError, "ERROR parsing robots.txt from %s: %s", ctx.normalizedURL, e)
} else {
c.logFunc(LogEnqueued, "enqueue: %s", robCtx.url)
c.Options.Extender.Enqueued(robCtx)
w.pop.stack(robCtx)
}
}
cnt++
c.logFunc(LogEnqueued, "enqueue: %s", ctx.url)
c.Options.Extender.Enqueued(ctx)
w.pop.stack(ctx)
c.pushPopRefCount++
// Once it is stacked, it WILL be visited eventually, so add it to the visited slice
// (unless denied by robots.txt, but this is out of our hands, for all we
// care, it is visited).
if !isVisited {
// The visited map works with the normalized URL
c.visited[ctx.normalizedURL.String()] = struct{}{}
}
}
}
return
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"enqueueUrls",
"(",
"ctxs",
"[",
"]",
"*",
"URLContext",
")",
"(",
"cnt",
"int",
")",
"{",
"for",
"_",
",",
"ctx",
":=",
"range",
"ctxs",
"{",
"var",
"isVisited",
",",
"enqueue",
"bool",
"\n\n",
"// Cannot directly enqueue a robots.txt URL, since it is managed as a special case",
"// in the worker (doesn't return a response to crawler).",
"if",
"ctx",
".",
"IsRobotsURL",
"(",
")",
"{",
"continue",
"\n",
"}",
"\n",
"// Check if it has been visited before, using the normalized URL",
"_",
",",
"isVisited",
"=",
"c",
".",
"visited",
"[",
"ctx",
".",
"normalizedURL",
".",
"String",
"(",
")",
"]",
"\n\n",
"// Filter the URL",
"if",
"enqueue",
"=",
"c",
".",
"Options",
".",
"Extender",
".",
"Filter",
"(",
"ctx",
",",
"isVisited",
")",
";",
"!",
"enqueue",
"{",
"// Filter said NOT to use this url, so continue with next",
"c",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"ctx",
".",
"normalizedURL",
")",
"\n",
"continue",
"\n",
"}",
"\n\n",
"// Even if filter said to use the URL, it still MUST be absolute, http(s)-prefixed,",
"// and comply with the same host policy if requested.",
"if",
"!",
"ctx",
".",
"normalizedURL",
".",
"IsAbs",
"(",
")",
"{",
"// Only absolute URLs are processed, so ignore",
"c",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"ctx",
".",
"normalizedURL",
")",
"\n\n",
"}",
"else",
"if",
"!",
"strings",
".",
"HasPrefix",
"(",
"ctx",
".",
"normalizedURL",
".",
"Scheme",
",",
"\"",
"\"",
")",
"{",
"c",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"ctx",
".",
"normalizedURL",
")",
"\n\n",
"}",
"else",
"if",
"c",
".",
"Options",
".",
"SameHostOnly",
"&&",
"!",
"c",
".",
"isSameHost",
"(",
"ctx",
")",
"{",
"// Only allow URLs coming from the same host",
"c",
".",
"logFunc",
"(",
"LogIgnored",
",",
"\"",
"\"",
",",
"ctx",
".",
"normalizedURL",
")",
"\n\n",
"}",
"else",
"{",
"// All is good, visit this URL (robots.txt verification is done by worker)",
"// Possible caveat: if the normalization changes the host, it is possible",
"// that the robots.txt fetched for this host would differ from the one for",
"// the unnormalized host. However, this should be rare, and is a weird",
"// behaviour from the host (i.e. why would site.com differ in its rules",
"// from www.site.com) and can be fixed by using a different normalization",
"// flag. So this is an acceptable behaviour for gocrawl.",
"// Launch worker if required, based on the host of the normalized URL",
"w",
",",
"ok",
":=",
"c",
".",
"workers",
"[",
"ctx",
".",
"normalizedURL",
".",
"Host",
"]",
"\n",
"if",
"!",
"ok",
"{",
"// No worker exists for this host, launch a new one",
"w",
"=",
"c",
".",
"launchWorker",
"(",
"ctx",
")",
"\n",
"// Automatically enqueue the robots.txt URL as first in line",
"if",
"robCtx",
",",
"e",
":=",
"ctx",
".",
"getRobotsURLCtx",
"(",
")",
";",
"e",
"!=",
"nil",
"{",
"c",
".",
"Options",
".",
"Extender",
".",
"Error",
"(",
"newCrawlError",
"(",
"ctx",
",",
"e",
",",
"CekParseRobots",
")",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"ctx",
".",
"normalizedURL",
",",
"e",
")",
"\n",
"}",
"else",
"{",
"c",
".",
"logFunc",
"(",
"LogEnqueued",
",",
"\"",
"\"",
",",
"robCtx",
".",
"url",
")",
"\n",
"c",
".",
"Options",
".",
"Extender",
".",
"Enqueued",
"(",
"robCtx",
")",
"\n",
"w",
".",
"pop",
".",
"stack",
"(",
"robCtx",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"cnt",
"++",
"\n",
"c",
".",
"logFunc",
"(",
"LogEnqueued",
",",
"\"",
"\"",
",",
"ctx",
".",
"url",
")",
"\n",
"c",
".",
"Options",
".",
"Extender",
".",
"Enqueued",
"(",
"ctx",
")",
"\n",
"w",
".",
"pop",
".",
"stack",
"(",
"ctx",
")",
"\n",
"c",
".",
"pushPopRefCount",
"++",
"\n\n",
"// Once it is stacked, it WILL be visited eventually, so add it to the visited slice",
"// (unless denied by robots.txt, but this is out of our hands, for all we",
"// care, it is visited).",
"if",
"!",
"isVisited",
"{",
"// The visited map works with the normalized URL",
"c",
".",
"visited",
"[",
"ctx",
".",
"normalizedURL",
".",
"String",
"(",
")",
"]",
"=",
"struct",
"{",
"}",
"{",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Enqueue the URLs returned from the worker, as long as it complies with the
// selection policies. | [
"Enqueue",
"the",
"URLs",
"returned",
"from",
"the",
"worker",
"as",
"long",
"as",
"it",
"complies",
"with",
"the",
"selection",
"policies",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L199-L273 |
5,537 | PuerkitoBio/gocrawl | crawler.go | collectUrls | func (c *Crawler) collectUrls() error {
defer func() {
c.logFunc(LogInfo, "waiting for goroutines to complete...")
c.wg.Wait()
c.logFunc(LogInfo, "crawler done.")
}()
for {
// By checking this after each channel reception, there is a bug if the worker
// wants to reenqueue following an error or a redirection. The pushPopRefCount
// temporarily gets to zero before the new URL is enqueued. Check the length
// of the enqueue channel to see if this is really over, or just this temporary
// state.
//
// Check if refcount is zero - MUST be before the select statement, so that if
// no valid seeds are enqueued, the crawler stops.
if c.pushPopRefCount == 0 && len(c.enqueue) == 0 {
c.logFunc(LogInfo, "sending STOP signals...")
close(c.stop)
return nil
}
select {
case res := <-c.push:
// Received a response, check if it contains URLs to enqueue
if res.visited {
c.visits++
if c.Options.MaxVisits > 0 && c.visits >= c.Options.MaxVisits {
// Limit reached, request workers to stop
c.logFunc(LogInfo, "sending STOP signals...")
close(c.stop)
return ErrMaxVisits
}
}
if res.idleDeath {
// The worker timed out from its Idle TTL delay, remove from active workers
delete(c.workers, res.host)
c.logFunc(LogInfo, "worker for host %s cleared on idle policy", res.host)
} else {
c.enqueueUrls(c.toURLContexts(res.harvestedURLs, res.ctx.url))
c.pushPopRefCount--
}
case enq := <-c.enqueue:
// Received a command to enqueue a URL, proceed
ctxs := c.toURLContexts(enq, nil)
c.logFunc(LogTrace, "receive url(s) to enqueue %v", toStringArrayContextURL(ctxs))
c.enqueueUrls(ctxs)
case <-c.stop:
return ErrInterrupted
}
}
} | go | func (c *Crawler) collectUrls() error {
defer func() {
c.logFunc(LogInfo, "waiting for goroutines to complete...")
c.wg.Wait()
c.logFunc(LogInfo, "crawler done.")
}()
for {
// By checking this after each channel reception, there is a bug if the worker
// wants to reenqueue following an error or a redirection. The pushPopRefCount
// temporarily gets to zero before the new URL is enqueued. Check the length
// of the enqueue channel to see if this is really over, or just this temporary
// state.
//
// Check if refcount is zero - MUST be before the select statement, so that if
// no valid seeds are enqueued, the crawler stops.
if c.pushPopRefCount == 0 && len(c.enqueue) == 0 {
c.logFunc(LogInfo, "sending STOP signals...")
close(c.stop)
return nil
}
select {
case res := <-c.push:
// Received a response, check if it contains URLs to enqueue
if res.visited {
c.visits++
if c.Options.MaxVisits > 0 && c.visits >= c.Options.MaxVisits {
// Limit reached, request workers to stop
c.logFunc(LogInfo, "sending STOP signals...")
close(c.stop)
return ErrMaxVisits
}
}
if res.idleDeath {
// The worker timed out from its Idle TTL delay, remove from active workers
delete(c.workers, res.host)
c.logFunc(LogInfo, "worker for host %s cleared on idle policy", res.host)
} else {
c.enqueueUrls(c.toURLContexts(res.harvestedURLs, res.ctx.url))
c.pushPopRefCount--
}
case enq := <-c.enqueue:
// Received a command to enqueue a URL, proceed
ctxs := c.toURLContexts(enq, nil)
c.logFunc(LogTrace, "receive url(s) to enqueue %v", toStringArrayContextURL(ctxs))
c.enqueueUrls(ctxs)
case <-c.stop:
return ErrInterrupted
}
}
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"collectUrls",
"(",
")",
"error",
"{",
"defer",
"func",
"(",
")",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"c",
".",
"wg",
".",
"Wait",
"(",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"}",
"(",
")",
"\n\n",
"for",
"{",
"// By checking this after each channel reception, there is a bug if the worker",
"// wants to reenqueue following an error or a redirection. The pushPopRefCount",
"// temporarily gets to zero before the new URL is enqueued. Check the length",
"// of the enqueue channel to see if this is really over, or just this temporary",
"// state.",
"//",
"// Check if refcount is zero - MUST be before the select statement, so that if",
"// no valid seeds are enqueued, the crawler stops.",
"if",
"c",
".",
"pushPopRefCount",
"==",
"0",
"&&",
"len",
"(",
"c",
".",
"enqueue",
")",
"==",
"0",
"{",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"close",
"(",
"c",
".",
"stop",
")",
"\n",
"return",
"nil",
"\n",
"}",
"\n\n",
"select",
"{",
"case",
"res",
":=",
"<-",
"c",
".",
"push",
":",
"// Received a response, check if it contains URLs to enqueue",
"if",
"res",
".",
"visited",
"{",
"c",
".",
"visits",
"++",
"\n",
"if",
"c",
".",
"Options",
".",
"MaxVisits",
">",
"0",
"&&",
"c",
".",
"visits",
">=",
"c",
".",
"Options",
".",
"MaxVisits",
"{",
"// Limit reached, request workers to stop",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
")",
"\n",
"close",
"(",
"c",
".",
"stop",
")",
"\n",
"return",
"ErrMaxVisits",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"res",
".",
"idleDeath",
"{",
"// The worker timed out from its Idle TTL delay, remove from active workers",
"delete",
"(",
"c",
".",
"workers",
",",
"res",
".",
"host",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogInfo",
",",
"\"",
"\"",
",",
"res",
".",
"host",
")",
"\n",
"}",
"else",
"{",
"c",
".",
"enqueueUrls",
"(",
"c",
".",
"toURLContexts",
"(",
"res",
".",
"harvestedURLs",
",",
"res",
".",
"ctx",
".",
"url",
")",
")",
"\n",
"c",
".",
"pushPopRefCount",
"--",
"\n",
"}",
"\n\n",
"case",
"enq",
":=",
"<-",
"c",
".",
"enqueue",
":",
"// Received a command to enqueue a URL, proceed",
"ctxs",
":=",
"c",
".",
"toURLContexts",
"(",
"enq",
",",
"nil",
")",
"\n",
"c",
".",
"logFunc",
"(",
"LogTrace",
",",
"\"",
"\"",
",",
"toStringArrayContextURL",
"(",
"ctxs",
")",
")",
"\n",
"c",
".",
"enqueueUrls",
"(",
"ctxs",
")",
"\n",
"case",
"<-",
"c",
".",
"stop",
":",
"return",
"ErrInterrupted",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // This is the main loop of the crawler, waiting for responses from the workers
// and processing these responses. | [
"This",
"is",
"the",
"main",
"loop",
"of",
"the",
"crawler",
"waiting",
"for",
"responses",
"from",
"the",
"workers",
"and",
"processing",
"these",
"responses",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L277-L329 |
5,538 | PuerkitoBio/gocrawl | crawler.go | Stop | func (c *Crawler) Stop() {
defer func() {
if err := recover(); err != nil {
c.logFunc(LogError, "error when manually stopping crawler: %s", err)
}
}()
// this channel may be closed already
close(c.stop)
} | go | func (c *Crawler) Stop() {
defer func() {
if err := recover(); err != nil {
c.logFunc(LogError, "error when manually stopping crawler: %s", err)
}
}()
// this channel may be closed already
close(c.stop)
} | [
"func",
"(",
"c",
"*",
"Crawler",
")",
"Stop",
"(",
")",
"{",
"defer",
"func",
"(",
")",
"{",
"if",
"err",
":=",
"recover",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"c",
".",
"logFunc",
"(",
"LogError",
",",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"(",
")",
"\n\n",
"// this channel may be closed already",
"close",
"(",
"c",
".",
"stop",
")",
"\n",
"}"
] | // Stop terminates the crawler. | [
"Stop",
"terminates",
"the",
"crawler",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/crawler.go#L332-L341 |
5,539 | PuerkitoBio/gocrawl | errors.go | Error | func (ce CrawlError) Error() string {
if ce.Err != nil {
return ce.Err.Error()
}
return ce.msg
} | go | func (ce CrawlError) Error() string {
if ce.Err != nil {
return ce.Err.Error()
}
return ce.msg
} | [
"func",
"(",
"ce",
"CrawlError",
")",
"Error",
"(",
")",
"string",
"{",
"if",
"ce",
".",
"Err",
"!=",
"nil",
"{",
"return",
"ce",
".",
"Err",
".",
"Error",
"(",
")",
"\n",
"}",
"\n",
"return",
"ce",
".",
"msg",
"\n",
"}"
] | // Error implements of the error interface for CrawlError. | [
"Error",
"implements",
"of",
"the",
"error",
"interface",
"for",
"CrawlError",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L70-L75 |
5,540 | PuerkitoBio/gocrawl | errors.go | newCrawlError | func newCrawlError(ctx *URLContext, e error, kind CrawlErrorKind) *CrawlError {
return &CrawlError{ctx, e, kind, ""}
} | go | func newCrawlError(ctx *URLContext, e error, kind CrawlErrorKind) *CrawlError {
return &CrawlError{ctx, e, kind, ""}
} | [
"func",
"newCrawlError",
"(",
"ctx",
"*",
"URLContext",
",",
"e",
"error",
",",
"kind",
"CrawlErrorKind",
")",
"*",
"CrawlError",
"{",
"return",
"&",
"CrawlError",
"{",
"ctx",
",",
"e",
",",
"kind",
",",
"\"",
"\"",
"}",
"\n",
"}"
] | // Create a new CrawlError based on a source error. | [
"Create",
"a",
"new",
"CrawlError",
"based",
"on",
"a",
"source",
"error",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L78-L80 |
5,541 | PuerkitoBio/gocrawl | errors.go | newCrawlErrorMessage | func newCrawlErrorMessage(ctx *URLContext, msg string, kind CrawlErrorKind) *CrawlError {
return &CrawlError{ctx, nil, kind, msg}
} | go | func newCrawlErrorMessage(ctx *URLContext, msg string, kind CrawlErrorKind) *CrawlError {
return &CrawlError{ctx, nil, kind, msg}
} | [
"func",
"newCrawlErrorMessage",
"(",
"ctx",
"*",
"URLContext",
",",
"msg",
"string",
",",
"kind",
"CrawlErrorKind",
")",
"*",
"CrawlError",
"{",
"return",
"&",
"CrawlError",
"{",
"ctx",
",",
"nil",
",",
"kind",
",",
"msg",
"}",
"\n",
"}"
] | // Create a new CrawlError with the specified message. | [
"Create",
"a",
"new",
"CrawlError",
"with",
"the",
"specified",
"message",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/errors.go#L83-L85 |
5,542 | PuerkitoBio/gocrawl | urlcontext.go | cloneForRedirect | func (uc *URLContext) cloneForRedirect(dst *url.URL, normFlags purell.NormalizationFlags) *URLContext {
var src, normalizedSrc *url.URL
if uc.sourceURL != nil {
src = &url.URL{}
*src = *uc.sourceURL
}
if src == nil && uc.url != nil {
// if the current context doesn't have a source URL, use its URL as
// source (e.g. for a seed URL that triggers a redirect)
src = &url.URL{}
*src = *uc.url
}
if uc.normalizedSourceURL != nil {
normalizedSrc = &url.URL{}
*normalizedSrc = *uc.normalizedSourceURL
}
if normalizedSrc == nil {
normalizedSrc = &url.URL{}
*normalizedSrc = *uc.normalizedURL
}
rawDst := &url.URL{}
*rawDst = *dst
purell.NormalizeURL(dst, normFlags)
return &URLContext{
HeadBeforeGet: uc.HeadBeforeGet,
State: uc.State,
url: rawDst,
normalizedURL: dst,
sourceURL: src,
normalizedSourceURL: normalizedSrc,
}
} | go | func (uc *URLContext) cloneForRedirect(dst *url.URL, normFlags purell.NormalizationFlags) *URLContext {
var src, normalizedSrc *url.URL
if uc.sourceURL != nil {
src = &url.URL{}
*src = *uc.sourceURL
}
if src == nil && uc.url != nil {
// if the current context doesn't have a source URL, use its URL as
// source (e.g. for a seed URL that triggers a redirect)
src = &url.URL{}
*src = *uc.url
}
if uc.normalizedSourceURL != nil {
normalizedSrc = &url.URL{}
*normalizedSrc = *uc.normalizedSourceURL
}
if normalizedSrc == nil {
normalizedSrc = &url.URL{}
*normalizedSrc = *uc.normalizedURL
}
rawDst := &url.URL{}
*rawDst = *dst
purell.NormalizeURL(dst, normFlags)
return &URLContext{
HeadBeforeGet: uc.HeadBeforeGet,
State: uc.State,
url: rawDst,
normalizedURL: dst,
sourceURL: src,
normalizedSourceURL: normalizedSrc,
}
} | [
"func",
"(",
"uc",
"*",
"URLContext",
")",
"cloneForRedirect",
"(",
"dst",
"*",
"url",
".",
"URL",
",",
"normFlags",
"purell",
".",
"NormalizationFlags",
")",
"*",
"URLContext",
"{",
"var",
"src",
",",
"normalizedSrc",
"*",
"url",
".",
"URL",
"\n",
"if",
"uc",
".",
"sourceURL",
"!=",
"nil",
"{",
"src",
"=",
"&",
"url",
".",
"URL",
"{",
"}",
"\n",
"*",
"src",
"=",
"*",
"uc",
".",
"sourceURL",
"\n",
"}",
"\n",
"if",
"src",
"==",
"nil",
"&&",
"uc",
".",
"url",
"!=",
"nil",
"{",
"// if the current context doesn't have a source URL, use its URL as",
"// source (e.g. for a seed URL that triggers a redirect)",
"src",
"=",
"&",
"url",
".",
"URL",
"{",
"}",
"\n",
"*",
"src",
"=",
"*",
"uc",
".",
"url",
"\n",
"}",
"\n\n",
"if",
"uc",
".",
"normalizedSourceURL",
"!=",
"nil",
"{",
"normalizedSrc",
"=",
"&",
"url",
".",
"URL",
"{",
"}",
"\n",
"*",
"normalizedSrc",
"=",
"*",
"uc",
".",
"normalizedSourceURL",
"\n",
"}",
"\n",
"if",
"normalizedSrc",
"==",
"nil",
"{",
"normalizedSrc",
"=",
"&",
"url",
".",
"URL",
"{",
"}",
"\n",
"*",
"normalizedSrc",
"=",
"*",
"uc",
".",
"normalizedURL",
"\n",
"}",
"\n\n",
"rawDst",
":=",
"&",
"url",
".",
"URL",
"{",
"}",
"\n",
"*",
"rawDst",
"=",
"*",
"dst",
"\n",
"purell",
".",
"NormalizeURL",
"(",
"dst",
",",
"normFlags",
")",
"\n",
"return",
"&",
"URLContext",
"{",
"HeadBeforeGet",
":",
"uc",
".",
"HeadBeforeGet",
",",
"State",
":",
"uc",
".",
"State",
",",
"url",
":",
"rawDst",
",",
"normalizedURL",
":",
"dst",
",",
"sourceURL",
":",
"src",
",",
"normalizedSourceURL",
":",
"normalizedSrc",
",",
"}",
"\n",
"}"
] | // cloneForRedirect returns a new URLContext with the given
// destination URL with the same sourceURL and normalizedSourceURL. | [
"cloneForRedirect",
"returns",
"a",
"new",
"URLContext",
"with",
"the",
"given",
"destination",
"URL",
"with",
"the",
"same",
"sourceURL",
"and",
"normalizedSourceURL",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/urlcontext.go#L65-L98 |
5,543 | PuerkitoBio/gocrawl | options.go | NewOptions | func NewOptions(ext Extender) *Options {
// Use defaults except for Extender
return &Options{
DefaultUserAgent,
DefaultRobotUserAgent,
0,
DefaultEnqueueChanBuffer,
DefaultHostBufferFactor,
DefaultCrawlDelay,
DefaultIdleTTL,
true,
false,
DefaultNormalizationFlags,
LogError,
ext,
}
} | go | func NewOptions(ext Extender) *Options {
// Use defaults except for Extender
return &Options{
DefaultUserAgent,
DefaultRobotUserAgent,
0,
DefaultEnqueueChanBuffer,
DefaultHostBufferFactor,
DefaultCrawlDelay,
DefaultIdleTTL,
true,
false,
DefaultNormalizationFlags,
LogError,
ext,
}
} | [
"func",
"NewOptions",
"(",
"ext",
"Extender",
")",
"*",
"Options",
"{",
"// Use defaults except for Extender",
"return",
"&",
"Options",
"{",
"DefaultUserAgent",
",",
"DefaultRobotUserAgent",
",",
"0",
",",
"DefaultEnqueueChanBuffer",
",",
"DefaultHostBufferFactor",
",",
"DefaultCrawlDelay",
",",
"DefaultIdleTTL",
",",
"true",
",",
"false",
",",
"DefaultNormalizationFlags",
",",
"LogError",
",",
"ext",
",",
"}",
"\n",
"}"
] | // NewOptions creates a new set of Options with default values
// using the provided Extender. The RobotUserAgent option should
// be set to the name of your crawler, it is used to find the matching
// entry in the robots.txt file. | [
"NewOptions",
"creates",
"a",
"new",
"set",
"of",
"Options",
"with",
"default",
"values",
"using",
"the",
"provided",
"Extender",
".",
"The",
"RobotUserAgent",
"option",
"should",
"be",
"set",
"to",
"the",
"name",
"of",
"your",
"crawler",
"it",
"is",
"used",
"to",
"find",
"the",
"matching",
"entry",
"in",
"the",
"robots",
".",
"txt",
"file",
"."
] | 0aa0507f0cf9eee3726aeabf9848501a5ca4d960 | https://github.com/PuerkitoBio/gocrawl/blob/0aa0507f0cf9eee3726aeabf9848501a5ca4d960/options.go#L84-L100 |
5,544 | iancoleman/strcase | camel.go | toCamelInitCase | func toCamelInitCase(s string, initCase bool) string {
s = addWordBoundariesToNumbers(s)
s = strings.Trim(s, " ")
n := ""
capNext := initCase
for _, v := range s {
if v >= 'A' && v <= 'Z' {
n += string(v)
}
if v >= '0' && v <= '9' {
n += string(v)
}
if v >= 'a' && v <= 'z' {
if capNext {
n += strings.ToUpper(string(v))
} else {
n += string(v)
}
}
if v == '_' || v == ' ' || v == '-' {
capNext = true
} else {
capNext = false
}
}
return n
} | go | func toCamelInitCase(s string, initCase bool) string {
s = addWordBoundariesToNumbers(s)
s = strings.Trim(s, " ")
n := ""
capNext := initCase
for _, v := range s {
if v >= 'A' && v <= 'Z' {
n += string(v)
}
if v >= '0' && v <= '9' {
n += string(v)
}
if v >= 'a' && v <= 'z' {
if capNext {
n += strings.ToUpper(string(v))
} else {
n += string(v)
}
}
if v == '_' || v == ' ' || v == '-' {
capNext = true
} else {
capNext = false
}
}
return n
} | [
"func",
"toCamelInitCase",
"(",
"s",
"string",
",",
"initCase",
"bool",
")",
"string",
"{",
"s",
"=",
"addWordBoundariesToNumbers",
"(",
"s",
")",
"\n",
"s",
"=",
"strings",
".",
"Trim",
"(",
"s",
",",
"\"",
"\"",
")",
"\n",
"n",
":=",
"\"",
"\"",
"\n",
"capNext",
":=",
"initCase",
"\n",
"for",
"_",
",",
"v",
":=",
"range",
"s",
"{",
"if",
"v",
">=",
"'A'",
"&&",
"v",
"<=",
"'Z'",
"{",
"n",
"+=",
"string",
"(",
"v",
")",
"\n",
"}",
"\n",
"if",
"v",
">=",
"'0'",
"&&",
"v",
"<=",
"'9'",
"{",
"n",
"+=",
"string",
"(",
"v",
")",
"\n",
"}",
"\n",
"if",
"v",
">=",
"'a'",
"&&",
"v",
"<=",
"'z'",
"{",
"if",
"capNext",
"{",
"n",
"+=",
"strings",
".",
"ToUpper",
"(",
"string",
"(",
"v",
")",
")",
"\n",
"}",
"else",
"{",
"n",
"+=",
"string",
"(",
"v",
")",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"v",
"==",
"'_'",
"||",
"v",
"==",
"' '",
"||",
"v",
"==",
"'-'",
"{",
"capNext",
"=",
"true",
"\n",
"}",
"else",
"{",
"capNext",
"=",
"false",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"n",
"\n",
"}"
] | // Converts a string to CamelCase | [
"Converts",
"a",
"string",
"to",
"CamelCase"
] | e506e3ef73653e84c592ba44aab577a46678f68c | https://github.com/iancoleman/strcase/blob/e506e3ef73653e84c592ba44aab577a46678f68c/camel.go#L33-L59 |
5,545 | iancoleman/strcase | camel.go | ToLowerCamel | func ToLowerCamel(s string) string {
if s == "" {
return s
}
if r := rune(s[0]); r >= 'A' && r <= 'Z' {
s = strings.ToLower(string(r)) + s[1:]
}
return toCamelInitCase(s, false)
} | go | func ToLowerCamel(s string) string {
if s == "" {
return s
}
if r := rune(s[0]); r >= 'A' && r <= 'Z' {
s = strings.ToLower(string(r)) + s[1:]
}
return toCamelInitCase(s, false)
} | [
"func",
"ToLowerCamel",
"(",
"s",
"string",
")",
"string",
"{",
"if",
"s",
"==",
"\"",
"\"",
"{",
"return",
"s",
"\n",
"}",
"\n",
"if",
"r",
":=",
"rune",
"(",
"s",
"[",
"0",
"]",
")",
";",
"r",
">=",
"'A'",
"&&",
"r",
"<=",
"'Z'",
"{",
"s",
"=",
"strings",
".",
"ToLower",
"(",
"string",
"(",
"r",
")",
")",
"+",
"s",
"[",
"1",
":",
"]",
"\n",
"}",
"\n",
"return",
"toCamelInitCase",
"(",
"s",
",",
"false",
")",
"\n",
"}"
] | // ToLowerCamel converts a string to lowerCamelCase | [
"ToLowerCamel",
"converts",
"a",
"string",
"to",
"lowerCamelCase"
] | e506e3ef73653e84c592ba44aab577a46678f68c | https://github.com/iancoleman/strcase/blob/e506e3ef73653e84c592ba44aab577a46678f68c/camel.go#L67-L75 |
5,546 | codeskyblue/go-sh | sh.go | Call | func (s *Session) Call(name string, a ...interface{}) error {
return s.Command(name, a...).Run()
} | go | func (s *Session) Call(name string, a ...interface{}) error {
return s.Command(name, a...).Run()
} | [
"func",
"(",
"s",
"*",
"Session",
")",
"Call",
"(",
"name",
"string",
",",
"a",
"...",
"interface",
"{",
"}",
")",
"error",
"{",
"return",
"s",
".",
"Command",
"(",
"name",
",",
"a",
"...",
")",
".",
"Run",
"(",
")",
"\n",
"}"
] | // combine Command and Run | [
"combine",
"Command",
"and",
"Run"
] | 76bd3d59ff27c0518096e0af6c165bf626b809ec | https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/sh.go#L130-L132 |
5,547 | codeskyblue/go-sh | pipe.go | UnmarshalJSON | func (s *Session) UnmarshalJSON(data interface{}) (err error) {
bufrw := bytes.NewBuffer(nil)
s.Stdout = bufrw
if err = s.Run(); err != nil {
return
}
return json.NewDecoder(bufrw).Decode(data)
} | go | func (s *Session) UnmarshalJSON(data interface{}) (err error) {
bufrw := bytes.NewBuffer(nil)
s.Stdout = bufrw
if err = s.Run(); err != nil {
return
}
return json.NewDecoder(bufrw).Decode(data)
} | [
"func",
"(",
"s",
"*",
"Session",
")",
"UnmarshalJSON",
"(",
"data",
"interface",
"{",
"}",
")",
"(",
"err",
"error",
")",
"{",
"bufrw",
":=",
"bytes",
".",
"NewBuffer",
"(",
"nil",
")",
"\n",
"s",
".",
"Stdout",
"=",
"bufrw",
"\n",
"if",
"err",
"=",
"s",
".",
"Run",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"return",
"json",
".",
"NewDecoder",
"(",
"bufrw",
")",
".",
"Decode",
"(",
"data",
")",
"\n",
"}"
] | // unmarshal shell output to decode json | [
"unmarshal",
"shell",
"output",
"to",
"decode",
"json"
] | 76bd3d59ff27c0518096e0af6c165bf626b809ec | https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/pipe.go#L18-L25 |
5,548 | codeskyblue/go-sh | pipe.go | UnmarshalXML | func (s *Session) UnmarshalXML(data interface{}) (err error) {
bufrw := bytes.NewBuffer(nil)
s.Stdout = bufrw
if err = s.Run(); err != nil {
return
}
return xml.NewDecoder(bufrw).Decode(data)
} | go | func (s *Session) UnmarshalXML(data interface{}) (err error) {
bufrw := bytes.NewBuffer(nil)
s.Stdout = bufrw
if err = s.Run(); err != nil {
return
}
return xml.NewDecoder(bufrw).Decode(data)
} | [
"func",
"(",
"s",
"*",
"Session",
")",
"UnmarshalXML",
"(",
"data",
"interface",
"{",
"}",
")",
"(",
"err",
"error",
")",
"{",
"bufrw",
":=",
"bytes",
".",
"NewBuffer",
"(",
"nil",
")",
"\n",
"s",
".",
"Stdout",
"=",
"bufrw",
"\n",
"if",
"err",
"=",
"s",
".",
"Run",
"(",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"return",
"xml",
".",
"NewDecoder",
"(",
"bufrw",
")",
".",
"Decode",
"(",
"data",
")",
"\n",
"}"
] | // unmarshal command output into xml | [
"unmarshal",
"command",
"output",
"into",
"xml"
] | 76bd3d59ff27c0518096e0af6c165bf626b809ec | https://github.com/codeskyblue/go-sh/blob/76bd3d59ff27c0518096e0af6c165bf626b809ec/pipe.go#L28-L35 |
5,549 | go-audio/wav | encoder.go | NewEncoder | func NewEncoder(w io.WriteSeeker, sampleRate, bitDepth, numChans, audioFormat int) *Encoder {
return &Encoder{
w: w,
SampleRate: sampleRate,
BitDepth: bitDepth,
NumChans: numChans,
WavAudioFormat: audioFormat,
}
} | go | func NewEncoder(w io.WriteSeeker, sampleRate, bitDepth, numChans, audioFormat int) *Encoder {
return &Encoder{
w: w,
SampleRate: sampleRate,
BitDepth: bitDepth,
NumChans: numChans,
WavAudioFormat: audioFormat,
}
} | [
"func",
"NewEncoder",
"(",
"w",
"io",
".",
"WriteSeeker",
",",
"sampleRate",
",",
"bitDepth",
",",
"numChans",
",",
"audioFormat",
"int",
")",
"*",
"Encoder",
"{",
"return",
"&",
"Encoder",
"{",
"w",
":",
"w",
",",
"SampleRate",
":",
"sampleRate",
",",
"BitDepth",
":",
"bitDepth",
",",
"NumChans",
":",
"numChans",
",",
"WavAudioFormat",
":",
"audioFormat",
",",
"}",
"\n",
"}"
] | // NewEncoder creates a new encoder to create a new wav file.
// Don't forget to add Frames to the encoder before writing. | [
"NewEncoder",
"creates",
"a",
"new",
"encoder",
"to",
"create",
"a",
"new",
"wav",
"file",
".",
"Don",
"t",
"forget",
"to",
"add",
"Frames",
"to",
"the",
"encoder",
"before",
"writing",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L41-L49 |
5,550 | go-audio/wav | encoder.go | AddLE | func (e *Encoder) AddLE(src interface{}) error {
e.WrittenBytes += binary.Size(src)
return binary.Write(e.w, binary.LittleEndian, src)
} | go | func (e *Encoder) AddLE(src interface{}) error {
e.WrittenBytes += binary.Size(src)
return binary.Write(e.w, binary.LittleEndian, src)
} | [
"func",
"(",
"e",
"*",
"Encoder",
")",
"AddLE",
"(",
"src",
"interface",
"{",
"}",
")",
"error",
"{",
"e",
".",
"WrittenBytes",
"+=",
"binary",
".",
"Size",
"(",
"src",
")",
"\n",
"return",
"binary",
".",
"Write",
"(",
"e",
".",
"w",
",",
"binary",
".",
"LittleEndian",
",",
"src",
")",
"\n",
"}"
] | // AddLE serializes and adds the passed value using little endian | [
"AddLE",
"serializes",
"and",
"adds",
"the",
"passed",
"value",
"using",
"little",
"endian"
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L52-L55 |
5,551 | go-audio/wav | encoder.go | AddBE | func (e *Encoder) AddBE(src interface{}) error {
e.WrittenBytes += binary.Size(src)
return binary.Write(e.w, binary.BigEndian, src)
} | go | func (e *Encoder) AddBE(src interface{}) error {
e.WrittenBytes += binary.Size(src)
return binary.Write(e.w, binary.BigEndian, src)
} | [
"func",
"(",
"e",
"*",
"Encoder",
")",
"AddBE",
"(",
"src",
"interface",
"{",
"}",
")",
"error",
"{",
"e",
".",
"WrittenBytes",
"+=",
"binary",
".",
"Size",
"(",
"src",
")",
"\n",
"return",
"binary",
".",
"Write",
"(",
"e",
".",
"w",
",",
"binary",
".",
"BigEndian",
",",
"src",
")",
"\n",
"}"
] | // AddBE serializes and adds the passed value using big endian | [
"AddBE",
"serializes",
"and",
"adds",
"the",
"passed",
"value",
"using",
"big",
"endian"
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L58-L61 |
5,552 | go-audio/wav | encoder.go | WriteFrame | func (e *Encoder) WriteFrame(value interface{}) error {
if !e.wroteHeader {
e.writeHeader()
}
if !e.pcmChunkStarted {
// sound header
if err := e.AddLE(riff.DataFormatID); err != nil {
return fmt.Errorf("error encoding sound header %v", err)
}
e.pcmChunkStarted = true
// write a temporary chunksize
e.pcmChunkSizePos = e.WrittenBytes
if err := e.AddLE(uint32(42)); err != nil {
return fmt.Errorf("%v when writing wav data chunk size header", err)
}
}
e.frames++
return e.AddLE(value)
} | go | func (e *Encoder) WriteFrame(value interface{}) error {
if !e.wroteHeader {
e.writeHeader()
}
if !e.pcmChunkStarted {
// sound header
if err := e.AddLE(riff.DataFormatID); err != nil {
return fmt.Errorf("error encoding sound header %v", err)
}
e.pcmChunkStarted = true
// write a temporary chunksize
e.pcmChunkSizePos = e.WrittenBytes
if err := e.AddLE(uint32(42)); err != nil {
return fmt.Errorf("%v when writing wav data chunk size header", err)
}
}
e.frames++
return e.AddLE(value)
} | [
"func",
"(",
"e",
"*",
"Encoder",
")",
"WriteFrame",
"(",
"value",
"interface",
"{",
"}",
")",
"error",
"{",
"if",
"!",
"e",
".",
"wroteHeader",
"{",
"e",
".",
"writeHeader",
"(",
")",
"\n",
"}",
"\n",
"if",
"!",
"e",
".",
"pcmChunkStarted",
"{",
"// sound header",
"if",
"err",
":=",
"e",
".",
"AddLE",
"(",
"riff",
".",
"DataFormatID",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"e",
".",
"pcmChunkStarted",
"=",
"true",
"\n\n",
"// write a temporary chunksize",
"e",
".",
"pcmChunkSizePos",
"=",
"e",
".",
"WrittenBytes",
"\n",
"if",
"err",
":=",
"e",
".",
"AddLE",
"(",
"uint32",
"(",
"42",
")",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"e",
".",
"frames",
"++",
"\n",
"return",
"e",
".",
"AddLE",
"(",
"value",
")",
"\n",
"}"
] | // WriteFrame writes a single frame of data to the underlying writer. | [
"WriteFrame",
"writes",
"a",
"single",
"frame",
"of",
"data",
"to",
"the",
"underlying",
"writer",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/encoder.go#L199-L219 |
5,553 | go-audio/wav | cue_chunk.go | DecodeCueChunk | func DecodeCueChunk(d *Decoder, ch *riff.Chunk) error {
if ch == nil {
return fmt.Errorf("can't decode a nil chunk")
}
if d == nil {
return fmt.Errorf("nil decoder")
}
if ch.ID == CIDCue {
// read the entire chunk in memory
buf := make([]byte, ch.Size)
var err error
if _, err = ch.Read(buf); err != nil {
return fmt.Errorf("failed to read the CUE chunk - %v", err)
}
r := bytes.NewReader(buf)
var nbrCues uint32
if err := binary.Read(r, binary.LittleEndian, &nbrCues); err != nil {
return fmt.Errorf("failed to read the number of cues - %v", err)
}
if nbrCues > 0 {
if d.Metadata == nil {
d.Metadata = &Metadata{}
}
d.Metadata.CuePoints = []*CuePoint{}
scratch := make([]byte, 4)
for i := uint32(0); i < nbrCues; i++ {
c := &CuePoint{}
if _, err = r.Read(scratch); err != nil {
return fmt.Errorf("failed to read the cue point ID")
}
copy(c.ID[:], scratch[:4])
if err := binary.Read(r, binary.LittleEndian, &c.Position); err != nil {
return err
}
if _, err = r.Read(scratch); err != nil {
return fmt.Errorf("failed to read the data chunk id")
}
copy(c.DataChunkID[:], scratch[:4])
if err := binary.Read(r, binary.LittleEndian, &c.ChunkStart); err != nil {
return err
}
if err := binary.Read(r, binary.LittleEndian, &c.BlockStart); err != nil {
return err
}
if err := binary.Read(r, binary.LittleEndian, &c.SampleOffset); err != nil {
return err
}
d.Metadata.CuePoints = append(d.Metadata.CuePoints, c)
}
}
}
return nil
} | go | func DecodeCueChunk(d *Decoder, ch *riff.Chunk) error {
if ch == nil {
return fmt.Errorf("can't decode a nil chunk")
}
if d == nil {
return fmt.Errorf("nil decoder")
}
if ch.ID == CIDCue {
// read the entire chunk in memory
buf := make([]byte, ch.Size)
var err error
if _, err = ch.Read(buf); err != nil {
return fmt.Errorf("failed to read the CUE chunk - %v", err)
}
r := bytes.NewReader(buf)
var nbrCues uint32
if err := binary.Read(r, binary.LittleEndian, &nbrCues); err != nil {
return fmt.Errorf("failed to read the number of cues - %v", err)
}
if nbrCues > 0 {
if d.Metadata == nil {
d.Metadata = &Metadata{}
}
d.Metadata.CuePoints = []*CuePoint{}
scratch := make([]byte, 4)
for i := uint32(0); i < nbrCues; i++ {
c := &CuePoint{}
if _, err = r.Read(scratch); err != nil {
return fmt.Errorf("failed to read the cue point ID")
}
copy(c.ID[:], scratch[:4])
if err := binary.Read(r, binary.LittleEndian, &c.Position); err != nil {
return err
}
if _, err = r.Read(scratch); err != nil {
return fmt.Errorf("failed to read the data chunk id")
}
copy(c.DataChunkID[:], scratch[:4])
if err := binary.Read(r, binary.LittleEndian, &c.ChunkStart); err != nil {
return err
}
if err := binary.Read(r, binary.LittleEndian, &c.BlockStart); err != nil {
return err
}
if err := binary.Read(r, binary.LittleEndian, &c.SampleOffset); err != nil {
return err
}
d.Metadata.CuePoints = append(d.Metadata.CuePoints, c)
}
}
}
return nil
} | [
"func",
"DecodeCueChunk",
"(",
"d",
"*",
"Decoder",
",",
"ch",
"*",
"riff",
".",
"Chunk",
")",
"error",
"{",
"if",
"ch",
"==",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"if",
"d",
"==",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"if",
"ch",
".",
"ID",
"==",
"CIDCue",
"{",
"// read the entire chunk in memory",
"buf",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"ch",
".",
"Size",
")",
"\n",
"var",
"err",
"error",
"\n",
"if",
"_",
",",
"err",
"=",
"ch",
".",
"Read",
"(",
"buf",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"r",
":=",
"bytes",
".",
"NewReader",
"(",
"buf",
")",
"\n",
"var",
"nbrCues",
"uint32",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"r",
",",
"binary",
".",
"LittleEndian",
",",
"&",
"nbrCues",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"err",
")",
"\n",
"}",
"\n",
"if",
"nbrCues",
">",
"0",
"{",
"if",
"d",
".",
"Metadata",
"==",
"nil",
"{",
"d",
".",
"Metadata",
"=",
"&",
"Metadata",
"{",
"}",
"\n",
"}",
"\n",
"d",
".",
"Metadata",
".",
"CuePoints",
"=",
"[",
"]",
"*",
"CuePoint",
"{",
"}",
"\n",
"scratch",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"4",
")",
"\n",
"for",
"i",
":=",
"uint32",
"(",
"0",
")",
";",
"i",
"<",
"nbrCues",
";",
"i",
"++",
"{",
"c",
":=",
"&",
"CuePoint",
"{",
"}",
"\n",
"if",
"_",
",",
"err",
"=",
"r",
".",
"Read",
"(",
"scratch",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"copy",
"(",
"c",
".",
"ID",
"[",
":",
"]",
",",
"scratch",
"[",
":",
"4",
"]",
")",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"r",
",",
"binary",
".",
"LittleEndian",
",",
"&",
"c",
".",
"Position",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"if",
"_",
",",
"err",
"=",
"r",
".",
"Read",
"(",
"scratch",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"copy",
"(",
"c",
".",
"DataChunkID",
"[",
":",
"]",
",",
"scratch",
"[",
":",
"4",
"]",
")",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"r",
",",
"binary",
".",
"LittleEndian",
",",
"&",
"c",
".",
"ChunkStart",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"r",
",",
"binary",
".",
"LittleEndian",
",",
"&",
"c",
".",
"BlockStart",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"r",
",",
"binary",
".",
"LittleEndian",
",",
"&",
"c",
".",
"SampleOffset",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"d",
".",
"Metadata",
".",
"CuePoints",
"=",
"append",
"(",
"d",
".",
"Metadata",
".",
"CuePoints",
",",
"c",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"}",
"\n",
"return",
"nil",
"\n",
"}"
] | // DecodeCueChunk decodes the optional cue chunk and extracts cue points. | [
"DecodeCueChunk",
"decodes",
"the",
"optional",
"cue",
"chunk",
"and",
"extracts",
"cue",
"points",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/cue_chunk.go#L53-L106 |
5,554 | go-audio/wav | decoder.go | NewDecoder | func NewDecoder(r io.ReadSeeker) *Decoder {
return &Decoder{
r: r,
parser: riff.New(r),
}
} | go | func NewDecoder(r io.ReadSeeker) *Decoder {
return &Decoder{
r: r,
parser: riff.New(r),
}
} | [
"func",
"NewDecoder",
"(",
"r",
"io",
".",
"ReadSeeker",
")",
"*",
"Decoder",
"{",
"return",
"&",
"Decoder",
"{",
"r",
":",
"r",
",",
"parser",
":",
"riff",
".",
"New",
"(",
"r",
")",
",",
"}",
"\n",
"}"
] | // NewDecoder creates a decoder for the passed wav reader.
// Note that the reader doesn't get rewinded as the container is processed. | [
"NewDecoder",
"creates",
"a",
"decoder",
"for",
"the",
"passed",
"wav",
"reader",
".",
"Note",
"that",
"the",
"reader",
"doesn",
"t",
"get",
"rewinded",
"as",
"the",
"container",
"is",
"processed",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L50-L55 |
5,555 | go-audio/wav | decoder.go | Seek | func (d *Decoder) Seek(offset int64, whence int) (int64, error) {
return d.r.Seek(offset, whence)
} | go | func (d *Decoder) Seek(offset int64, whence int) (int64, error) {
return d.r.Seek(offset, whence)
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Seek",
"(",
"offset",
"int64",
",",
"whence",
"int",
")",
"(",
"int64",
",",
"error",
")",
"{",
"return",
"d",
".",
"r",
".",
"Seek",
"(",
"offset",
",",
"whence",
")",
"\n",
"}"
] | // Seek provides access to the cursor position in the PCM data | [
"Seek",
"provides",
"access",
"to",
"the",
"cursor",
"position",
"in",
"the",
"PCM",
"data"
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L58-L60 |
5,556 | go-audio/wav | decoder.go | Err | func (d *Decoder) Err() error {
if d.err == io.EOF {
return nil
}
return d.err
} | go | func (d *Decoder) Err() error {
if d.err == io.EOF {
return nil
}
return d.err
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Err",
"(",
")",
"error",
"{",
"if",
"d",
".",
"err",
"==",
"io",
".",
"EOF",
"{",
"return",
"nil",
"\n",
"}",
"\n",
"return",
"d",
".",
"err",
"\n",
"}"
] | // Err returns the first non-EOF error that was encountered by the Decoder. | [
"Err",
"returns",
"the",
"first",
"non",
"-",
"EOF",
"error",
"that",
"was",
"encountered",
"by",
"the",
"Decoder",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L79-L84 |
5,557 | go-audio/wav | decoder.go | EOF | func (d *Decoder) EOF() bool {
if d == nil || d.err == io.EOF {
return true
}
return false
} | go | func (d *Decoder) EOF() bool {
if d == nil || d.err == io.EOF {
return true
}
return false
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"EOF",
"(",
")",
"bool",
"{",
"if",
"d",
"==",
"nil",
"||",
"d",
".",
"err",
"==",
"io",
".",
"EOF",
"{",
"return",
"true",
"\n",
"}",
"\n",
"return",
"false",
"\n",
"}"
] | // EOF returns positively if the underlying reader reached the end of file. | [
"EOF",
"returns",
"positively",
"if",
"the",
"underlying",
"reader",
"reached",
"the",
"end",
"of",
"file",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L87-L92 |
5,558 | go-audio/wav | decoder.go | ReadMetadata | func (d *Decoder) ReadMetadata() {
if d.Metadata != nil {
return
}
d.ReadInfo()
if d.Err() != nil || d.Metadata != nil {
return
}
var (
chunk *riff.Chunk
err error
)
for err == nil {
chunk, err = d.parser.NextChunk()
if err != nil {
break
}
switch chunk.ID {
case CIDList:
if err = DecodeListChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
if d.Metadata != nil && d.Metadata.SamplerInfo != nil {
// we got everything we were looking for
break
}
case CIDSmpl:
if err = DecodeSamplerChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
case CIDCue:
if err = DecodeCueChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
default:
// fmt.Println(string(chunk.ID[:]))
chunk.Drain()
}
}
} | go | func (d *Decoder) ReadMetadata() {
if d.Metadata != nil {
return
}
d.ReadInfo()
if d.Err() != nil || d.Metadata != nil {
return
}
var (
chunk *riff.Chunk
err error
)
for err == nil {
chunk, err = d.parser.NextChunk()
if err != nil {
break
}
switch chunk.ID {
case CIDList:
if err = DecodeListChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
if d.Metadata != nil && d.Metadata.SamplerInfo != nil {
// we got everything we were looking for
break
}
case CIDSmpl:
if err = DecodeSamplerChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
case CIDCue:
if err = DecodeCueChunk(d, chunk); err != nil {
if err != io.EOF {
d.err = err
}
}
default:
// fmt.Println(string(chunk.ID[:]))
chunk.Drain()
}
}
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"ReadMetadata",
"(",
")",
"{",
"if",
"d",
".",
"Metadata",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"d",
".",
"ReadInfo",
"(",
")",
"\n",
"if",
"d",
".",
"Err",
"(",
")",
"!=",
"nil",
"||",
"d",
".",
"Metadata",
"!=",
"nil",
"{",
"return",
"\n",
"}",
"\n",
"var",
"(",
"chunk",
"*",
"riff",
".",
"Chunk",
"\n",
"err",
"error",
"\n",
")",
"\n",
"for",
"err",
"==",
"nil",
"{",
"chunk",
",",
"err",
"=",
"d",
".",
"parser",
".",
"NextChunk",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"break",
"\n",
"}",
"\n\n",
"switch",
"chunk",
".",
"ID",
"{",
"case",
"CIDList",
":",
"if",
"err",
"=",
"DecodeListChunk",
"(",
"d",
",",
"chunk",
")",
";",
"err",
"!=",
"nil",
"{",
"if",
"err",
"!=",
"io",
".",
"EOF",
"{",
"d",
".",
"err",
"=",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"if",
"d",
".",
"Metadata",
"!=",
"nil",
"&&",
"d",
".",
"Metadata",
".",
"SamplerInfo",
"!=",
"nil",
"{",
"// we got everything we were looking for",
"break",
"\n",
"}",
"\n",
"case",
"CIDSmpl",
":",
"if",
"err",
"=",
"DecodeSamplerChunk",
"(",
"d",
",",
"chunk",
")",
";",
"err",
"!=",
"nil",
"{",
"if",
"err",
"!=",
"io",
".",
"EOF",
"{",
"d",
".",
"err",
"=",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"case",
"CIDCue",
":",
"if",
"err",
"=",
"DecodeCueChunk",
"(",
"d",
",",
"chunk",
")",
";",
"err",
"!=",
"nil",
"{",
"if",
"err",
"!=",
"io",
".",
"EOF",
"{",
"d",
".",
"err",
"=",
"err",
"\n",
"}",
"\n",
"}",
"\n",
"default",
":",
"// fmt.Println(string(chunk.ID[:]))",
"chunk",
".",
"Drain",
"(",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"}"
] | // ReadMetadata parses the file for extra metadata such as the INFO list chunk.
// The entire file will be read and should be rewinded if more data must be
// accessed. | [
"ReadMetadata",
"parses",
"the",
"file",
"for",
"extra",
"metadata",
"such",
"as",
"the",
"INFO",
"list",
"chunk",
".",
"The",
"entire",
"file",
"will",
"be",
"read",
"and",
"should",
"be",
"rewinded",
"if",
"more",
"data",
"must",
"be",
"accessed",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L122-L169 |
5,559 | go-audio/wav | decoder.go | Duration | func (d *Decoder) Duration() (time.Duration, error) {
if d == nil || d.parser == nil {
return 0, errors.New("can't calculate the duration of a nil pointer")
}
return d.parser.Duration()
} | go | func (d *Decoder) Duration() (time.Duration, error) {
if d == nil || d.parser == nil {
return 0, errors.New("can't calculate the duration of a nil pointer")
}
return d.parser.Duration()
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"Duration",
"(",
")",
"(",
"time",
".",
"Duration",
",",
"error",
")",
"{",
"if",
"d",
"==",
"nil",
"||",
"d",
".",
"parser",
"==",
"nil",
"{",
"return",
"0",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n",
"return",
"d",
".",
"parser",
".",
"Duration",
"(",
")",
"\n",
"}"
] | // Duration returns the time duration for the current audio container | [
"Duration",
"returns",
"the",
"time",
"duration",
"for",
"the",
"current",
"audio",
"container"
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L380-L385 |
5,560 | go-audio/wav | decoder.go | readHeaders | func (d *Decoder) readHeaders() error {
if d == nil || d.NumChans > 0 {
return nil
}
id, size, err := d.parser.IDnSize()
if err != nil {
return err
}
d.parser.ID = id
if d.parser.ID != riff.RiffID {
return fmt.Errorf("%s - %s", d.parser.ID, riff.ErrFmtNotSupported)
}
d.parser.Size = size
if err := binary.Read(d.r, binary.BigEndian, &d.parser.Format); err != nil {
return err
}
var chunk *riff.Chunk
var rewindBytes int64
for err == nil {
chunk, err = d.parser.NextChunk()
if err != nil {
break
}
if chunk.ID == riff.FmtID {
chunk.DecodeWavHeader(d.parser)
d.NumChans = d.parser.NumChannels
d.BitDepth = d.parser.BitsPerSample
d.SampleRate = d.parser.SampleRate
d.WavAudioFormat = d.parser.WavAudioFormat
d.AvgBytesPerSec = d.parser.AvgBytesPerSec
if rewindBytes > 0 {
d.r.Seek(-(rewindBytes + int64(chunk.Size) + 8), 1)
}
break
} else if chunk.ID == CIDList {
// The list chunk can be in the header or footer
// because so many players don't support that chunk properly
// it is recommended to have it at the end of the file.
DecodeListChunk(d, chunk)
// unexpected chunk order, might be a bext chunk
rewindBytes += int64(chunk.Size) + 8
} else if chunk.ID == CIDSmpl {
DecodeSamplerChunk(d, chunk)
rewindBytes += int64(chunk.Size) + 8
} else {
// unexpected chunk order, might be a bext chunk
rewindBytes += int64(chunk.Size) + 8
// drain the chunk
io.CopyN(ioutil.Discard, d.r, int64(chunk.Size))
}
}
return d.err
} | go | func (d *Decoder) readHeaders() error {
if d == nil || d.NumChans > 0 {
return nil
}
id, size, err := d.parser.IDnSize()
if err != nil {
return err
}
d.parser.ID = id
if d.parser.ID != riff.RiffID {
return fmt.Errorf("%s - %s", d.parser.ID, riff.ErrFmtNotSupported)
}
d.parser.Size = size
if err := binary.Read(d.r, binary.BigEndian, &d.parser.Format); err != nil {
return err
}
var chunk *riff.Chunk
var rewindBytes int64
for err == nil {
chunk, err = d.parser.NextChunk()
if err != nil {
break
}
if chunk.ID == riff.FmtID {
chunk.DecodeWavHeader(d.parser)
d.NumChans = d.parser.NumChannels
d.BitDepth = d.parser.BitsPerSample
d.SampleRate = d.parser.SampleRate
d.WavAudioFormat = d.parser.WavAudioFormat
d.AvgBytesPerSec = d.parser.AvgBytesPerSec
if rewindBytes > 0 {
d.r.Seek(-(rewindBytes + int64(chunk.Size) + 8), 1)
}
break
} else if chunk.ID == CIDList {
// The list chunk can be in the header or footer
// because so many players don't support that chunk properly
// it is recommended to have it at the end of the file.
DecodeListChunk(d, chunk)
// unexpected chunk order, might be a bext chunk
rewindBytes += int64(chunk.Size) + 8
} else if chunk.ID == CIDSmpl {
DecodeSamplerChunk(d, chunk)
rewindBytes += int64(chunk.Size) + 8
} else {
// unexpected chunk order, might be a bext chunk
rewindBytes += int64(chunk.Size) + 8
// drain the chunk
io.CopyN(ioutil.Discard, d.r, int64(chunk.Size))
}
}
return d.err
} | [
"func",
"(",
"d",
"*",
"Decoder",
")",
"readHeaders",
"(",
")",
"error",
"{",
"if",
"d",
"==",
"nil",
"||",
"d",
".",
"NumChans",
">",
"0",
"{",
"return",
"nil",
"\n",
"}",
"\n\n",
"id",
",",
"size",
",",
"err",
":=",
"d",
".",
"parser",
".",
"IDnSize",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n",
"d",
".",
"parser",
".",
"ID",
"=",
"id",
"\n",
"if",
"d",
".",
"parser",
".",
"ID",
"!=",
"riff",
".",
"RiffID",
"{",
"return",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"d",
".",
"parser",
".",
"ID",
",",
"riff",
".",
"ErrFmtNotSupported",
")",
"\n",
"}",
"\n",
"d",
".",
"parser",
".",
"Size",
"=",
"size",
"\n",
"if",
"err",
":=",
"binary",
".",
"Read",
"(",
"d",
".",
"r",
",",
"binary",
".",
"BigEndian",
",",
"&",
"d",
".",
"parser",
".",
"Format",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"err",
"\n",
"}",
"\n\n",
"var",
"chunk",
"*",
"riff",
".",
"Chunk",
"\n",
"var",
"rewindBytes",
"int64",
"\n\n",
"for",
"err",
"==",
"nil",
"{",
"chunk",
",",
"err",
"=",
"d",
".",
"parser",
".",
"NextChunk",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"break",
"\n",
"}",
"\n\n",
"if",
"chunk",
".",
"ID",
"==",
"riff",
".",
"FmtID",
"{",
"chunk",
".",
"DecodeWavHeader",
"(",
"d",
".",
"parser",
")",
"\n",
"d",
".",
"NumChans",
"=",
"d",
".",
"parser",
".",
"NumChannels",
"\n",
"d",
".",
"BitDepth",
"=",
"d",
".",
"parser",
".",
"BitsPerSample",
"\n",
"d",
".",
"SampleRate",
"=",
"d",
".",
"parser",
".",
"SampleRate",
"\n",
"d",
".",
"WavAudioFormat",
"=",
"d",
".",
"parser",
".",
"WavAudioFormat",
"\n",
"d",
".",
"AvgBytesPerSec",
"=",
"d",
".",
"parser",
".",
"AvgBytesPerSec",
"\n\n",
"if",
"rewindBytes",
">",
"0",
"{",
"d",
".",
"r",
".",
"Seek",
"(",
"-",
"(",
"rewindBytes",
"+",
"int64",
"(",
"chunk",
".",
"Size",
")",
"+",
"8",
")",
",",
"1",
")",
"\n",
"}",
"\n",
"break",
"\n",
"}",
"else",
"if",
"chunk",
".",
"ID",
"==",
"CIDList",
"{",
"// The list chunk can be in the header or footer",
"// because so many players don't support that chunk properly",
"// it is recommended to have it at the end of the file.",
"DecodeListChunk",
"(",
"d",
",",
"chunk",
")",
"\n",
"// unexpected chunk order, might be a bext chunk",
"rewindBytes",
"+=",
"int64",
"(",
"chunk",
".",
"Size",
")",
"+",
"8",
"\n",
"}",
"else",
"if",
"chunk",
".",
"ID",
"==",
"CIDSmpl",
"{",
"DecodeSamplerChunk",
"(",
"d",
",",
"chunk",
")",
"\n",
"rewindBytes",
"+=",
"int64",
"(",
"chunk",
".",
"Size",
")",
"+",
"8",
"\n",
"}",
"else",
"{",
"// unexpected chunk order, might be a bext chunk",
"rewindBytes",
"+=",
"int64",
"(",
"chunk",
".",
"Size",
")",
"+",
"8",
"\n",
"// drain the chunk",
"io",
".",
"CopyN",
"(",
"ioutil",
".",
"Discard",
",",
"d",
".",
"r",
",",
"int64",
"(",
"chunk",
".",
"Size",
")",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"d",
".",
"err",
"\n",
"}"
] | // readHeaders is safe to call multiple times | [
"readHeaders",
"is",
"safe",
"to",
"call",
"multiple",
"times"
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L393-L451 |
5,561 | go-audio/wav | decoder.go | sampleFloat64DecodeFunc | func sampleFloat64DecodeFunc(bitsPerSample int) (func([]byte) float64, error) {
bytesPerSample := bitsPerSample / 8
switch bytesPerSample {
case 1:
// 8bit values are unsigned
return func(s []byte) float64 {
return float64(uint8(s[0]))
}, nil
case 2:
return func(s []byte) float64 {
return float64(int(s[0]) + int(s[1])<<8)
}, nil
case 3:
return func(s []byte) float64 {
var output int32
output |= int32(s[2]) << 0
output |= int32(s[1]) << 8
output |= int32(s[0]) << 16
return float64(output)
}, nil
case 4:
// TODO: fix the float64 conversion (current int implementation)
return func(s []byte) float64 {
return float64(int(s[0]) + int(s[1])<<8 + int(s[2])<<16 + int(s[3])<<24)
}, nil
default:
return nil, fmt.Errorf("unhandled byte depth:%d", bitsPerSample)
}
} | go | func sampleFloat64DecodeFunc(bitsPerSample int) (func([]byte) float64, error) {
bytesPerSample := bitsPerSample / 8
switch bytesPerSample {
case 1:
// 8bit values are unsigned
return func(s []byte) float64 {
return float64(uint8(s[0]))
}, nil
case 2:
return func(s []byte) float64 {
return float64(int(s[0]) + int(s[1])<<8)
}, nil
case 3:
return func(s []byte) float64 {
var output int32
output |= int32(s[2]) << 0
output |= int32(s[1]) << 8
output |= int32(s[0]) << 16
return float64(output)
}, nil
case 4:
// TODO: fix the float64 conversion (current int implementation)
return func(s []byte) float64 {
return float64(int(s[0]) + int(s[1])<<8 + int(s[2])<<16 + int(s[3])<<24)
}, nil
default:
return nil, fmt.Errorf("unhandled byte depth:%d", bitsPerSample)
}
} | [
"func",
"sampleFloat64DecodeFunc",
"(",
"bitsPerSample",
"int",
")",
"(",
"func",
"(",
"[",
"]",
"byte",
")",
"float64",
",",
"error",
")",
"{",
"bytesPerSample",
":=",
"bitsPerSample",
"/",
"8",
"\n",
"switch",
"bytesPerSample",
"{",
"case",
"1",
":",
"// 8bit values are unsigned",
"return",
"func",
"(",
"s",
"[",
"]",
"byte",
")",
"float64",
"{",
"return",
"float64",
"(",
"uint8",
"(",
"s",
"[",
"0",
"]",
")",
")",
"\n",
"}",
",",
"nil",
"\n",
"case",
"2",
":",
"return",
"func",
"(",
"s",
"[",
"]",
"byte",
")",
"float64",
"{",
"return",
"float64",
"(",
"int",
"(",
"s",
"[",
"0",
"]",
")",
"+",
"int",
"(",
"s",
"[",
"1",
"]",
")",
"<<",
"8",
")",
"\n",
"}",
",",
"nil",
"\n",
"case",
"3",
":",
"return",
"func",
"(",
"s",
"[",
"]",
"byte",
")",
"float64",
"{",
"var",
"output",
"int32",
"\n",
"output",
"|=",
"int32",
"(",
"s",
"[",
"2",
"]",
")",
"<<",
"0",
"\n",
"output",
"|=",
"int32",
"(",
"s",
"[",
"1",
"]",
")",
"<<",
"8",
"\n",
"output",
"|=",
"int32",
"(",
"s",
"[",
"0",
"]",
")",
"<<",
"16",
"\n",
"return",
"float64",
"(",
"output",
")",
"\n",
"}",
",",
"nil",
"\n",
"case",
"4",
":",
"// TODO: fix the float64 conversion (current int implementation)",
"return",
"func",
"(",
"s",
"[",
"]",
"byte",
")",
"float64",
"{",
"return",
"float64",
"(",
"int",
"(",
"s",
"[",
"0",
"]",
")",
"+",
"int",
"(",
"s",
"[",
"1",
"]",
")",
"<<",
"8",
"+",
"int",
"(",
"s",
"[",
"2",
"]",
")",
"<<",
"16",
"+",
"int",
"(",
"s",
"[",
"3",
"]",
")",
"<<",
"24",
")",
"\n",
"}",
",",
"nil",
"\n",
"default",
":",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"bitsPerSample",
")",
"\n",
"}",
"\n",
"}"
] | // sampleDecodeFloat64Func returns a function that can be used to convert
// a byte range into a float64 value based on the amount of bits used per sample. | [
"sampleDecodeFloat64Func",
"returns",
"a",
"function",
"that",
"can",
"be",
"used",
"to",
"convert",
"a",
"byte",
"range",
"into",
"a",
"float64",
"value",
"based",
"on",
"the",
"amount",
"of",
"bits",
"used",
"per",
"sample",
"."
] | cc323235893be1cbdb031cb52acfc0184b366870 | https://github.com/go-audio/wav/blob/cc323235893be1cbdb031cb52acfc0184b366870/decoder.go#L495-L523 |
5,562 | kr/pretty | diff.go | Diff | func Diff(a, b interface{}) (desc []string) {
Pdiff((*sbuf)(&desc), a, b)
return desc
} | go | func Diff(a, b interface{}) (desc []string) {
Pdiff((*sbuf)(&desc), a, b)
return desc
} | [
"func",
"Diff",
"(",
"a",
",",
"b",
"interface",
"{",
"}",
")",
"(",
"desc",
"[",
"]",
"string",
")",
"{",
"Pdiff",
"(",
"(",
"*",
"sbuf",
")",
"(",
"&",
"desc",
")",
",",
"a",
",",
"b",
")",
"\n",
"return",
"desc",
"\n",
"}"
] | // Diff returns a slice where each element describes
// a difference between a and b. | [
"Diff",
"returns",
"a",
"slice",
"where",
"each",
"element",
"describes",
"a",
"difference",
"between",
"a",
"and",
"b",
"."
] | 73f6ac0b30a98e433b289500d779f50c1a6f0712 | https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L18-L21 |
5,563 | kr/pretty | diff.go | Fdiff | func Fdiff(w io.Writer, a, b interface{}) {
Pdiff(&wprintfer{w}, a, b)
} | go | func Fdiff(w io.Writer, a, b interface{}) {
Pdiff(&wprintfer{w}, a, b)
} | [
"func",
"Fdiff",
"(",
"w",
"io",
".",
"Writer",
",",
"a",
",",
"b",
"interface",
"{",
"}",
")",
"{",
"Pdiff",
"(",
"&",
"wprintfer",
"{",
"w",
"}",
",",
"a",
",",
"b",
")",
"\n",
"}"
] | // Fdiff writes to w a description of the differences between a and b. | [
"Fdiff",
"writes",
"to",
"w",
"a",
"description",
"of",
"the",
"differences",
"between",
"a",
"and",
"b",
"."
] | 73f6ac0b30a98e433b289500d779f50c1a6f0712 | https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L32-L34 |
5,564 | kr/pretty | diff.go | Pdiff | func Pdiff(p Printfer, a, b interface{}) {
diffPrinter{w: p}.diff(reflect.ValueOf(a), reflect.ValueOf(b))
} | go | func Pdiff(p Printfer, a, b interface{}) {
diffPrinter{w: p}.diff(reflect.ValueOf(a), reflect.ValueOf(b))
} | [
"func",
"Pdiff",
"(",
"p",
"Printfer",
",",
"a",
",",
"b",
"interface",
"{",
"}",
")",
"{",
"diffPrinter",
"{",
"w",
":",
"p",
"}",
".",
"diff",
"(",
"reflect",
".",
"ValueOf",
"(",
"a",
")",
",",
"reflect",
".",
"ValueOf",
"(",
"b",
")",
")",
"\n",
"}"
] | // Pdiff prints to p a description of the differences between a and b.
// It calls Printf once for each difference, with no trailing newline.
// The standard library log.Logger is a Printfer. | [
"Pdiff",
"prints",
"to",
"p",
"a",
"description",
"of",
"the",
"differences",
"between",
"a",
"and",
"b",
".",
"It",
"calls",
"Printf",
"once",
"for",
"each",
"difference",
"with",
"no",
"trailing",
"newline",
".",
"The",
"standard",
"library",
"log",
".",
"Logger",
"is",
"a",
"Printfer",
"."
] | 73f6ac0b30a98e433b289500d779f50c1a6f0712 | https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L43-L45 |
5,565 | kr/pretty | diff.go | Ldiff | func Ldiff(l Logfer, a, b interface{}) {
Pdiff(&logprintfer{l}, a, b)
} | go | func Ldiff(l Logfer, a, b interface{}) {
Pdiff(&logprintfer{l}, a, b)
} | [
"func",
"Ldiff",
"(",
"l",
"Logfer",
",",
"a",
",",
"b",
"interface",
"{",
"}",
")",
"{",
"Pdiff",
"(",
"&",
"logprintfer",
"{",
"l",
"}",
",",
"a",
",",
"b",
")",
"\n",
"}"
] | // Ldiff prints to l a description of the differences between a and b.
// It calls Logf once for each difference, with no trailing newline.
// The standard library testing.T and testing.B are Logfers. | [
"Ldiff",
"prints",
"to",
"l",
"a",
"description",
"of",
"the",
"differences",
"between",
"a",
"and",
"b",
".",
"It",
"calls",
"Logf",
"once",
"for",
"each",
"difference",
"with",
"no",
"trailing",
"newline",
".",
"The",
"standard",
"library",
"testing",
".",
"T",
"and",
"testing",
".",
"B",
"are",
"Logfers",
"."
] | 73f6ac0b30a98e433b289500d779f50c1a6f0712 | https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L62-L64 |
5,566 | kr/pretty | diff.go | keyEqual | func keyEqual(av, bv reflect.Value) bool {
if !av.IsValid() && !bv.IsValid() {
return true
}
if !av.IsValid() || !bv.IsValid() || av.Type() != bv.Type() {
return false
}
switch kind := av.Kind(); kind {
case reflect.Bool:
a, b := av.Bool(), bv.Bool()
return a == b
case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:
a, b := av.Int(), bv.Int()
return a == b
case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:
a, b := av.Uint(), bv.Uint()
return a == b
case reflect.Float32, reflect.Float64:
a, b := av.Float(), bv.Float()
return a == b
case reflect.Complex64, reflect.Complex128:
a, b := av.Complex(), bv.Complex()
return a == b
case reflect.Array:
for i := 0; i < av.Len(); i++ {
if !keyEqual(av.Index(i), bv.Index(i)) {
return false
}
}
return true
case reflect.Chan, reflect.UnsafePointer, reflect.Ptr:
a, b := av.Pointer(), bv.Pointer()
return a == b
case reflect.Interface:
return keyEqual(av.Elem(), bv.Elem())
case reflect.String:
a, b := av.String(), bv.String()
return a == b
case reflect.Struct:
for i := 0; i < av.NumField(); i++ {
if !keyEqual(av.Field(i), bv.Field(i)) {
return false
}
}
return true
default:
panic("invalid map key type " + av.Type().String())
}
} | go | func keyEqual(av, bv reflect.Value) bool {
if !av.IsValid() && !bv.IsValid() {
return true
}
if !av.IsValid() || !bv.IsValid() || av.Type() != bv.Type() {
return false
}
switch kind := av.Kind(); kind {
case reflect.Bool:
a, b := av.Bool(), bv.Bool()
return a == b
case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:
a, b := av.Int(), bv.Int()
return a == b
case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:
a, b := av.Uint(), bv.Uint()
return a == b
case reflect.Float32, reflect.Float64:
a, b := av.Float(), bv.Float()
return a == b
case reflect.Complex64, reflect.Complex128:
a, b := av.Complex(), bv.Complex()
return a == b
case reflect.Array:
for i := 0; i < av.Len(); i++ {
if !keyEqual(av.Index(i), bv.Index(i)) {
return false
}
}
return true
case reflect.Chan, reflect.UnsafePointer, reflect.Ptr:
a, b := av.Pointer(), bv.Pointer()
return a == b
case reflect.Interface:
return keyEqual(av.Elem(), bv.Elem())
case reflect.String:
a, b := av.String(), bv.String()
return a == b
case reflect.Struct:
for i := 0; i < av.NumField(); i++ {
if !keyEqual(av.Field(i), bv.Field(i)) {
return false
}
}
return true
default:
panic("invalid map key type " + av.Type().String())
}
} | [
"func",
"keyEqual",
"(",
"av",
",",
"bv",
"reflect",
".",
"Value",
")",
"bool",
"{",
"if",
"!",
"av",
".",
"IsValid",
"(",
")",
"&&",
"!",
"bv",
".",
"IsValid",
"(",
")",
"{",
"return",
"true",
"\n",
"}",
"\n",
"if",
"!",
"av",
".",
"IsValid",
"(",
")",
"||",
"!",
"bv",
".",
"IsValid",
"(",
")",
"||",
"av",
".",
"Type",
"(",
")",
"!=",
"bv",
".",
"Type",
"(",
")",
"{",
"return",
"false",
"\n",
"}",
"\n",
"switch",
"kind",
":=",
"av",
".",
"Kind",
"(",
")",
";",
"kind",
"{",
"case",
"reflect",
".",
"Bool",
":",
"a",
",",
"b",
":=",
"av",
".",
"Bool",
"(",
")",
",",
"bv",
".",
"Bool",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Int",
",",
"reflect",
".",
"Int8",
",",
"reflect",
".",
"Int16",
",",
"reflect",
".",
"Int32",
",",
"reflect",
".",
"Int64",
":",
"a",
",",
"b",
":=",
"av",
".",
"Int",
"(",
")",
",",
"bv",
".",
"Int",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Uint",
",",
"reflect",
".",
"Uint8",
",",
"reflect",
".",
"Uint16",
",",
"reflect",
".",
"Uint32",
",",
"reflect",
".",
"Uint64",
",",
"reflect",
".",
"Uintptr",
":",
"a",
",",
"b",
":=",
"av",
".",
"Uint",
"(",
")",
",",
"bv",
".",
"Uint",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Float32",
",",
"reflect",
".",
"Float64",
":",
"a",
",",
"b",
":=",
"av",
".",
"Float",
"(",
")",
",",
"bv",
".",
"Float",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Complex64",
",",
"reflect",
".",
"Complex128",
":",
"a",
",",
"b",
":=",
"av",
".",
"Complex",
"(",
")",
",",
"bv",
".",
"Complex",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Array",
":",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"av",
".",
"Len",
"(",
")",
";",
"i",
"++",
"{",
"if",
"!",
"keyEqual",
"(",
"av",
".",
"Index",
"(",
"i",
")",
",",
"bv",
".",
"Index",
"(",
"i",
")",
")",
"{",
"return",
"false",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"true",
"\n",
"case",
"reflect",
".",
"Chan",
",",
"reflect",
".",
"UnsafePointer",
",",
"reflect",
".",
"Ptr",
":",
"a",
",",
"b",
":=",
"av",
".",
"Pointer",
"(",
")",
",",
"bv",
".",
"Pointer",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Interface",
":",
"return",
"keyEqual",
"(",
"av",
".",
"Elem",
"(",
")",
",",
"bv",
".",
"Elem",
"(",
")",
")",
"\n",
"case",
"reflect",
".",
"String",
":",
"a",
",",
"b",
":=",
"av",
".",
"String",
"(",
")",
",",
"bv",
".",
"String",
"(",
")",
"\n",
"return",
"a",
"==",
"b",
"\n",
"case",
"reflect",
".",
"Struct",
":",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"av",
".",
"NumField",
"(",
")",
";",
"i",
"++",
"{",
"if",
"!",
"keyEqual",
"(",
"av",
".",
"Field",
"(",
"i",
")",
",",
"bv",
".",
"Field",
"(",
"i",
")",
")",
"{",
"return",
"false",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"true",
"\n",
"default",
":",
"panic",
"(",
"\"",
"\"",
"+",
"av",
".",
"Type",
"(",
")",
".",
"String",
"(",
")",
")",
"\n",
"}",
"\n",
"}"
] | // keyEqual compares a and b for equality.
// Both a and b must be valid map keys. | [
"keyEqual",
"compares",
"a",
"and",
"b",
"for",
"equality",
".",
"Both",
"a",
"and",
"b",
"must",
"be",
"valid",
"map",
"keys",
"."
] | 73f6ac0b30a98e433b289500d779f50c1a6f0712 | https://github.com/kr/pretty/blob/73f6ac0b30a98e433b289500d779f50c1a6f0712/diff.go#L188-L236 |
5,567 | hashicorp/hil | parser/error.go | TokenErrorf | func TokenErrorf(token *scanner.Token, format string, args ...interface{}) error {
return Errorf(token.Pos, format, args...)
} | go | func TokenErrorf(token *scanner.Token, format string, args ...interface{}) error {
return Errorf(token.Pos, format, args...)
} | [
"func",
"TokenErrorf",
"(",
"token",
"*",
"scanner",
".",
"Token",
",",
"format",
"string",
",",
"args",
"...",
"interface",
"{",
"}",
")",
"error",
"{",
"return",
"Errorf",
"(",
"token",
".",
"Pos",
",",
"format",
",",
"args",
"...",
")",
"\n",
"}"
] | // TokenErrorf is a convenient wrapper around Errorf that uses the
// position of the given token. | [
"TokenErrorf",
"is",
"a",
"convenient",
"wrapper",
"around",
"Errorf",
"that",
"uses",
"the",
"position",
"of",
"the",
"given",
"token",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/error.go#L24-L26 |
5,568 | hashicorp/hil | parse.go | ParseWithPosition | func ParseWithPosition(v string, pos ast.Pos) (ast.Node, error) {
ch := scanner.Scan(v, pos)
return parser.Parse(ch)
} | go | func ParseWithPosition(v string, pos ast.Pos) (ast.Node, error) {
ch := scanner.Scan(v, pos)
return parser.Parse(ch)
} | [
"func",
"ParseWithPosition",
"(",
"v",
"string",
",",
"pos",
"ast",
".",
"Pos",
")",
"(",
"ast",
".",
"Node",
",",
"error",
")",
"{",
"ch",
":=",
"scanner",
".",
"Scan",
"(",
"v",
",",
"pos",
")",
"\n",
"return",
"parser",
".",
"Parse",
"(",
"ch",
")",
"\n",
"}"
] | // ParseWithPosition is like Parse except that it overrides the source
// row and column position of the first character in the string, which should
// be 1-based.
//
// This can be used when HIL is embedded in another language and the outer
// parser knows the row and column where the HIL expression started within
// the overall source file. | [
"ParseWithPosition",
"is",
"like",
"Parse",
"except",
"that",
"it",
"overrides",
"the",
"source",
"row",
"and",
"column",
"position",
"of",
"the",
"first",
"character",
"in",
"the",
"string",
"which",
"should",
"be",
"1",
"-",
"based",
".",
"This",
"can",
"be",
"used",
"when",
"HIL",
"is",
"embedded",
"in",
"another",
"language",
"and",
"the",
"outer",
"parser",
"knows",
"the",
"row",
"and",
"column",
"where",
"the",
"HIL",
"expression",
"started",
"within",
"the",
"overall",
"source",
"file",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parse.go#L26-L29 |
5,569 | hashicorp/hil | scanner/peeker.go | Peek | func (p *Peeker) Peek() *Token {
if p.peeked == nil {
p.peeked = <-p.ch
}
return p.peeked
} | go | func (p *Peeker) Peek() *Token {
if p.peeked == nil {
p.peeked = <-p.ch
}
return p.peeked
} | [
"func",
"(",
"p",
"*",
"Peeker",
")",
"Peek",
"(",
")",
"*",
"Token",
"{",
"if",
"p",
".",
"peeked",
"==",
"nil",
"{",
"p",
".",
"peeked",
"=",
"<-",
"p",
".",
"ch",
"\n",
"}",
"\n",
"return",
"p",
".",
"peeked",
"\n",
"}"
] | // Peek returns the next token in the stream without consuming it. A
// subsequent call to Read will return the same token. | [
"Peek",
"returns",
"the",
"next",
"token",
"in",
"the",
"stream",
"without",
"consuming",
"it",
".",
"A",
"subsequent",
"call",
"to",
"Read",
"will",
"return",
"the",
"same",
"token",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L20-L25 |
5,570 | hashicorp/hil | scanner/peeker.go | Read | func (p *Peeker) Read() *Token {
token := p.Peek()
// As a special case, we will produce the EOF token forever once
// it is reached.
if token.Type != EOF {
p.peeked = nil
}
return token
} | go | func (p *Peeker) Read() *Token {
token := p.Peek()
// As a special case, we will produce the EOF token forever once
// it is reached.
if token.Type != EOF {
p.peeked = nil
}
return token
} | [
"func",
"(",
"p",
"*",
"Peeker",
")",
"Read",
"(",
")",
"*",
"Token",
"{",
"token",
":=",
"p",
".",
"Peek",
"(",
")",
"\n\n",
"// As a special case, we will produce the EOF token forever once",
"// it is reached.",
"if",
"token",
".",
"Type",
"!=",
"EOF",
"{",
"p",
".",
"peeked",
"=",
"nil",
"\n",
"}",
"\n\n",
"return",
"token",
"\n",
"}"
] | // Read consumes the next token in the stream and returns it. | [
"Read",
"consumes",
"the",
"next",
"token",
"in",
"the",
"stream",
"and",
"returns",
"it",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L28-L38 |
5,571 | hashicorp/hil | scanner/peeker.go | Close | func (p *Peeker) Close() {
for _ = range p.ch {
// discard
}
// Install a synthetic EOF token in 'peeked' in case someone
// erroneously calls Peek() or Read() after we've closed.
p.peeked = &Token{
Type: EOF,
Content: "",
}
} | go | func (p *Peeker) Close() {
for _ = range p.ch {
// discard
}
// Install a synthetic EOF token in 'peeked' in case someone
// erroneously calls Peek() or Read() after we've closed.
p.peeked = &Token{
Type: EOF,
Content: "",
}
} | [
"func",
"(",
"p",
"*",
"Peeker",
")",
"Close",
"(",
")",
"{",
"for",
"_",
"=",
"range",
"p",
".",
"ch",
"{",
"// discard",
"}",
"\n",
"// Install a synthetic EOF token in 'peeked' in case someone",
"// erroneously calls Peek() or Read() after we've closed.",
"p",
".",
"peeked",
"=",
"&",
"Token",
"{",
"Type",
":",
"EOF",
",",
"Content",
":",
"\"",
"\"",
",",
"}",
"\n",
"}"
] | // Close ensures that the token stream has been exhausted, to prevent
// the goroutine in the underlying scanner from leaking.
//
// It's not necessary to call this if the caller reads the token stream
// to EOF, since that implicitly closes the scanner. | [
"Close",
"ensures",
"that",
"the",
"token",
"stream",
"has",
"been",
"exhausted",
"to",
"prevent",
"the",
"goroutine",
"in",
"the",
"underlying",
"scanner",
"from",
"leaking",
".",
"It",
"s",
"not",
"necessary",
"to",
"call",
"this",
"if",
"the",
"caller",
"reads",
"the",
"token",
"stream",
"to",
"EOF",
"since",
"that",
"implicitly",
"closes",
"the",
"scanner",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/peeker.go#L45-L55 |
5,572 | hashicorp/hil | parser/parser.go | parseInterpolationSeq | func (p *parser) parseInterpolationSeq(quoted bool) (ast.Node, error) {
literalType := scanner.LITERAL
endType := scanner.EOF
if quoted {
// exceptions for quoted sequences
literalType = scanner.STRING
endType = scanner.CQUOTE
}
startPos := p.peeker.Peek().Pos
if quoted {
tok := p.peeker.Read()
if tok.Type != scanner.OQUOTE {
return nil, ExpectationError("open quote", tok)
}
}
var exprs []ast.Node
for {
tok := p.peeker.Read()
if tok.Type == endType {
break
}
switch tok.Type {
case literalType:
val, err := p.parseStringToken(tok)
if err != nil {
return nil, err
}
exprs = append(exprs, &ast.LiteralNode{
Value: val,
Typex: ast.TypeString,
Posx: tok.Pos,
})
case scanner.BEGIN:
expr, err := p.ParseInterpolation()
if err != nil {
return nil, err
}
exprs = append(exprs, expr)
default:
return nil, ExpectationError(`"${"`, tok)
}
}
if len(exprs) == 0 {
// If we have no parts at all then the input must've
// been an empty string.
exprs = append(exprs, &ast.LiteralNode{
Value: "",
Typex: ast.TypeString,
Posx: startPos,
})
}
// As a special case, if our "Output" contains only one expression
// and it's a literal string then we'll hoist it up to be our
// direct return value, so callers can easily recognize a string
// that has no interpolations at all.
if len(exprs) == 1 {
if lit, ok := exprs[0].(*ast.LiteralNode); ok {
if lit.Typex == ast.TypeString {
return lit, nil
}
}
}
return &ast.Output{
Exprs: exprs,
Posx: startPos,
}, nil
} | go | func (p *parser) parseInterpolationSeq(quoted bool) (ast.Node, error) {
literalType := scanner.LITERAL
endType := scanner.EOF
if quoted {
// exceptions for quoted sequences
literalType = scanner.STRING
endType = scanner.CQUOTE
}
startPos := p.peeker.Peek().Pos
if quoted {
tok := p.peeker.Read()
if tok.Type != scanner.OQUOTE {
return nil, ExpectationError("open quote", tok)
}
}
var exprs []ast.Node
for {
tok := p.peeker.Read()
if tok.Type == endType {
break
}
switch tok.Type {
case literalType:
val, err := p.parseStringToken(tok)
if err != nil {
return nil, err
}
exprs = append(exprs, &ast.LiteralNode{
Value: val,
Typex: ast.TypeString,
Posx: tok.Pos,
})
case scanner.BEGIN:
expr, err := p.ParseInterpolation()
if err != nil {
return nil, err
}
exprs = append(exprs, expr)
default:
return nil, ExpectationError(`"${"`, tok)
}
}
if len(exprs) == 0 {
// If we have no parts at all then the input must've
// been an empty string.
exprs = append(exprs, &ast.LiteralNode{
Value: "",
Typex: ast.TypeString,
Posx: startPos,
})
}
// As a special case, if our "Output" contains only one expression
// and it's a literal string then we'll hoist it up to be our
// direct return value, so callers can easily recognize a string
// that has no interpolations at all.
if len(exprs) == 1 {
if lit, ok := exprs[0].(*ast.LiteralNode); ok {
if lit.Typex == ast.TypeString {
return lit, nil
}
}
}
return &ast.Output{
Exprs: exprs,
Posx: startPos,
}, nil
} | [
"func",
"(",
"p",
"*",
"parser",
")",
"parseInterpolationSeq",
"(",
"quoted",
"bool",
")",
"(",
"ast",
".",
"Node",
",",
"error",
")",
"{",
"literalType",
":=",
"scanner",
".",
"LITERAL",
"\n",
"endType",
":=",
"scanner",
".",
"EOF",
"\n",
"if",
"quoted",
"{",
"// exceptions for quoted sequences",
"literalType",
"=",
"scanner",
".",
"STRING",
"\n",
"endType",
"=",
"scanner",
".",
"CQUOTE",
"\n",
"}",
"\n\n",
"startPos",
":=",
"p",
".",
"peeker",
".",
"Peek",
"(",
")",
".",
"Pos",
"\n\n",
"if",
"quoted",
"{",
"tok",
":=",
"p",
".",
"peeker",
".",
"Read",
"(",
")",
"\n",
"if",
"tok",
".",
"Type",
"!=",
"scanner",
".",
"OQUOTE",
"{",
"return",
"nil",
",",
"ExpectationError",
"(",
"\"",
"\"",
",",
"tok",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"var",
"exprs",
"[",
"]",
"ast",
".",
"Node",
"\n",
"for",
"{",
"tok",
":=",
"p",
".",
"peeker",
".",
"Read",
"(",
")",
"\n\n",
"if",
"tok",
".",
"Type",
"==",
"endType",
"{",
"break",
"\n",
"}",
"\n\n",
"switch",
"tok",
".",
"Type",
"{",
"case",
"literalType",
":",
"val",
",",
"err",
":=",
"p",
".",
"parseStringToken",
"(",
"tok",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"exprs",
"=",
"append",
"(",
"exprs",
",",
"&",
"ast",
".",
"LiteralNode",
"{",
"Value",
":",
"val",
",",
"Typex",
":",
"ast",
".",
"TypeString",
",",
"Posx",
":",
"tok",
".",
"Pos",
",",
"}",
")",
"\n",
"case",
"scanner",
".",
"BEGIN",
":",
"expr",
",",
"err",
":=",
"p",
".",
"ParseInterpolation",
"(",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"exprs",
"=",
"append",
"(",
"exprs",
",",
"expr",
")",
"\n",
"default",
":",
"return",
"nil",
",",
"ExpectationError",
"(",
"`\"${\"`",
",",
"tok",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"len",
"(",
"exprs",
")",
"==",
"0",
"{",
"// If we have no parts at all then the input must've",
"// been an empty string.",
"exprs",
"=",
"append",
"(",
"exprs",
",",
"&",
"ast",
".",
"LiteralNode",
"{",
"Value",
":",
"\"",
"\"",
",",
"Typex",
":",
"ast",
".",
"TypeString",
",",
"Posx",
":",
"startPos",
",",
"}",
")",
"\n",
"}",
"\n\n",
"// As a special case, if our \"Output\" contains only one expression",
"// and it's a literal string then we'll hoist it up to be our",
"// direct return value, so callers can easily recognize a string",
"// that has no interpolations at all.",
"if",
"len",
"(",
"exprs",
")",
"==",
"1",
"{",
"if",
"lit",
",",
"ok",
":=",
"exprs",
"[",
"0",
"]",
".",
"(",
"*",
"ast",
".",
"LiteralNode",
")",
";",
"ok",
"{",
"if",
"lit",
".",
"Typex",
"==",
"ast",
".",
"TypeString",
"{",
"return",
"lit",
",",
"nil",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n\n",
"return",
"&",
"ast",
".",
"Output",
"{",
"Exprs",
":",
"exprs",
",",
"Posx",
":",
"startPos",
",",
"}",
",",
"nil",
"\n",
"}"
] | // parseInterpolationSeq parses either the top-level sequence of literals
// and interpolation expressions or a similar sequence within a quoted
// string inside an interpolation expression. The latter case is requested
// by setting 'quoted' to true. | [
"parseInterpolationSeq",
"parses",
"either",
"the",
"top",
"-",
"level",
"sequence",
"of",
"literals",
"and",
"interpolation",
"expressions",
"or",
"a",
"similar",
"sequence",
"within",
"a",
"quoted",
"string",
"inside",
"an",
"interpolation",
"expression",
".",
"The",
"latter",
"case",
"is",
"requested",
"by",
"setting",
"quoted",
"to",
"true",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L35-L109 |
5,573 | hashicorp/hil | parser/parser.go | parseStringToken | func (p *parser) parseStringToken(tok *scanner.Token) (string, error) {
var backslashes bool
switch tok.Type {
case scanner.LITERAL:
backslashes = false
case scanner.STRING:
backslashes = true
default:
panic("unsupported string token type")
}
raw := []byte(tok.Content)
buf := make([]byte, 0, len(raw))
for i := 0; i < len(raw); i++ {
b := raw[i]
more := len(raw) > (i + 1)
if b == '$' {
if more && raw[i+1] == '$' {
// skip over the second dollar sign
i++
}
} else if backslashes && b == '\\' {
if !more {
return "", Errorf(
ast.Pos{
Column: tok.Pos.Column + utf8.RuneCount(raw[:i]),
Line: tok.Pos.Line,
},
`unfinished backslash escape sequence`,
)
}
escapeType := raw[i+1]
switch escapeType {
case '\\':
// skip over the second slash
i++
case 'n':
b = '\n'
i++
case '"':
b = '"'
i++
default:
return "", Errorf(
ast.Pos{
Column: tok.Pos.Column + utf8.RuneCount(raw[:i]),
Line: tok.Pos.Line,
},
`invalid backslash escape sequence`,
)
}
}
buf = append(buf, b)
}
return string(buf), nil
} | go | func (p *parser) parseStringToken(tok *scanner.Token) (string, error) {
var backslashes bool
switch tok.Type {
case scanner.LITERAL:
backslashes = false
case scanner.STRING:
backslashes = true
default:
panic("unsupported string token type")
}
raw := []byte(tok.Content)
buf := make([]byte, 0, len(raw))
for i := 0; i < len(raw); i++ {
b := raw[i]
more := len(raw) > (i + 1)
if b == '$' {
if more && raw[i+1] == '$' {
// skip over the second dollar sign
i++
}
} else if backslashes && b == '\\' {
if !more {
return "", Errorf(
ast.Pos{
Column: tok.Pos.Column + utf8.RuneCount(raw[:i]),
Line: tok.Pos.Line,
},
`unfinished backslash escape sequence`,
)
}
escapeType := raw[i+1]
switch escapeType {
case '\\':
// skip over the second slash
i++
case 'n':
b = '\n'
i++
case '"':
b = '"'
i++
default:
return "", Errorf(
ast.Pos{
Column: tok.Pos.Column + utf8.RuneCount(raw[:i]),
Line: tok.Pos.Line,
},
`invalid backslash escape sequence`,
)
}
}
buf = append(buf, b)
}
return string(buf), nil
} | [
"func",
"(",
"p",
"*",
"parser",
")",
"parseStringToken",
"(",
"tok",
"*",
"scanner",
".",
"Token",
")",
"(",
"string",
",",
"error",
")",
"{",
"var",
"backslashes",
"bool",
"\n",
"switch",
"tok",
".",
"Type",
"{",
"case",
"scanner",
".",
"LITERAL",
":",
"backslashes",
"=",
"false",
"\n",
"case",
"scanner",
".",
"STRING",
":",
"backslashes",
"=",
"true",
"\n",
"default",
":",
"panic",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"raw",
":=",
"[",
"]",
"byte",
"(",
"tok",
".",
"Content",
")",
"\n",
"buf",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"0",
",",
"len",
"(",
"raw",
")",
")",
"\n\n",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"len",
"(",
"raw",
")",
";",
"i",
"++",
"{",
"b",
":=",
"raw",
"[",
"i",
"]",
"\n",
"more",
":=",
"len",
"(",
"raw",
")",
">",
"(",
"i",
"+",
"1",
")",
"\n\n",
"if",
"b",
"==",
"'$'",
"{",
"if",
"more",
"&&",
"raw",
"[",
"i",
"+",
"1",
"]",
"==",
"'$'",
"{",
"// skip over the second dollar sign",
"i",
"++",
"\n",
"}",
"\n",
"}",
"else",
"if",
"backslashes",
"&&",
"b",
"==",
"'\\\\'",
"{",
"if",
"!",
"more",
"{",
"return",
"\"",
"\"",
",",
"Errorf",
"(",
"ast",
".",
"Pos",
"{",
"Column",
":",
"tok",
".",
"Pos",
".",
"Column",
"+",
"utf8",
".",
"RuneCount",
"(",
"raw",
"[",
":",
"i",
"]",
")",
",",
"Line",
":",
"tok",
".",
"Pos",
".",
"Line",
",",
"}",
",",
"`unfinished backslash escape sequence`",
",",
")",
"\n",
"}",
"\n",
"escapeType",
":=",
"raw",
"[",
"i",
"+",
"1",
"]",
"\n",
"switch",
"escapeType",
"{",
"case",
"'\\\\'",
":",
"// skip over the second slash",
"i",
"++",
"\n",
"case",
"'n'",
":",
"b",
"=",
"'\\n'",
"\n",
"i",
"++",
"\n",
"case",
"'\"'",
":",
"b",
"=",
"'\"'",
"\n",
"i",
"++",
"\n",
"default",
":",
"return",
"\"",
"\"",
",",
"Errorf",
"(",
"ast",
".",
"Pos",
"{",
"Column",
":",
"tok",
".",
"Pos",
".",
"Column",
"+",
"utf8",
".",
"RuneCount",
"(",
"raw",
"[",
":",
"i",
"]",
")",
",",
"Line",
":",
"tok",
".",
"Pos",
".",
"Line",
",",
"}",
",",
"`invalid backslash escape sequence`",
",",
")",
"\n",
"}",
"\n",
"}",
"\n\n",
"buf",
"=",
"append",
"(",
"buf",
",",
"b",
")",
"\n",
"}",
"\n\n",
"return",
"string",
"(",
"buf",
")",
",",
"nil",
"\n",
"}"
] | // parseStringToken takes a token of either LITERAL or STRING type and
// returns the interpreted string, after processing any relevant
// escape sequences. | [
"parseStringToken",
"takes",
"a",
"token",
"of",
"either",
"LITERAL",
"or",
"STRING",
"type",
"and",
"returns",
"the",
"interpreted",
"string",
"after",
"processing",
"any",
"relevant",
"escape",
"sequences",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L114-L173 |
5,574 | hashicorp/hil | parser/parser.go | parseBinaryOps | func (p *parser) parseBinaryOps(ops []map[scanner.TokenType]ast.ArithmeticOp) (ast.Node, error) {
if len(ops) == 0 {
// We've run out of operators, so now we'll just try to parse a term.
return p.ParseExpressionTerm()
}
thisLevel := ops[0]
remaining := ops[1:]
startPos := p.peeker.Peek().Pos
var lhs, rhs ast.Node
operator := ast.ArithmeticOpInvalid
var err error
// parse a term that might be the first operand of a binary
// expression or it might just be a standalone term, but
// we won't know until we've parsed it and can look ahead
// to see if there's an operator token.
lhs, err = p.parseBinaryOps(remaining)
if err != nil {
return nil, err
}
// We'll keep eating up arithmetic operators until we run
// out, so that operators with the same precedence will combine in a
// left-associative manner:
// a+b+c => (a+b)+c, not a+(b+c)
//
// Should we later want to have right-associative operators, a way
// to achieve that would be to call back up to ParseExpression here
// instead of iteratively parsing only the remaining operators.
for {
next := p.peeker.Peek()
var newOperator ast.ArithmeticOp
var ok bool
if newOperator, ok = thisLevel[next.Type]; !ok {
break
}
// Are we extending an expression started on
// the previous iteration?
if operator != ast.ArithmeticOpInvalid {
lhs = &ast.Arithmetic{
Op: operator,
Exprs: []ast.Node{lhs, rhs},
Posx: startPos,
}
}
operator = newOperator
p.peeker.Read() // eat operator token
rhs, err = p.parseBinaryOps(remaining)
if err != nil {
return nil, err
}
}
if operator != ast.ArithmeticOpInvalid {
return &ast.Arithmetic{
Op: operator,
Exprs: []ast.Node{lhs, rhs},
Posx: startPos,
}, nil
} else {
return lhs, nil
}
} | go | func (p *parser) parseBinaryOps(ops []map[scanner.TokenType]ast.ArithmeticOp) (ast.Node, error) {
if len(ops) == 0 {
// We've run out of operators, so now we'll just try to parse a term.
return p.ParseExpressionTerm()
}
thisLevel := ops[0]
remaining := ops[1:]
startPos := p.peeker.Peek().Pos
var lhs, rhs ast.Node
operator := ast.ArithmeticOpInvalid
var err error
// parse a term that might be the first operand of a binary
// expression or it might just be a standalone term, but
// we won't know until we've parsed it and can look ahead
// to see if there's an operator token.
lhs, err = p.parseBinaryOps(remaining)
if err != nil {
return nil, err
}
// We'll keep eating up arithmetic operators until we run
// out, so that operators with the same precedence will combine in a
// left-associative manner:
// a+b+c => (a+b)+c, not a+(b+c)
//
// Should we later want to have right-associative operators, a way
// to achieve that would be to call back up to ParseExpression here
// instead of iteratively parsing only the remaining operators.
for {
next := p.peeker.Peek()
var newOperator ast.ArithmeticOp
var ok bool
if newOperator, ok = thisLevel[next.Type]; !ok {
break
}
// Are we extending an expression started on
// the previous iteration?
if operator != ast.ArithmeticOpInvalid {
lhs = &ast.Arithmetic{
Op: operator,
Exprs: []ast.Node{lhs, rhs},
Posx: startPos,
}
}
operator = newOperator
p.peeker.Read() // eat operator token
rhs, err = p.parseBinaryOps(remaining)
if err != nil {
return nil, err
}
}
if operator != ast.ArithmeticOpInvalid {
return &ast.Arithmetic{
Op: operator,
Exprs: []ast.Node{lhs, rhs},
Posx: startPos,
}, nil
} else {
return lhs, nil
}
} | [
"func",
"(",
"p",
"*",
"parser",
")",
"parseBinaryOps",
"(",
"ops",
"[",
"]",
"map",
"[",
"scanner",
".",
"TokenType",
"]",
"ast",
".",
"ArithmeticOp",
")",
"(",
"ast",
".",
"Node",
",",
"error",
")",
"{",
"if",
"len",
"(",
"ops",
")",
"==",
"0",
"{",
"// We've run out of operators, so now we'll just try to parse a term.",
"return",
"p",
".",
"ParseExpressionTerm",
"(",
")",
"\n",
"}",
"\n\n",
"thisLevel",
":=",
"ops",
"[",
"0",
"]",
"\n",
"remaining",
":=",
"ops",
"[",
"1",
":",
"]",
"\n\n",
"startPos",
":=",
"p",
".",
"peeker",
".",
"Peek",
"(",
")",
".",
"Pos",
"\n\n",
"var",
"lhs",
",",
"rhs",
"ast",
".",
"Node",
"\n",
"operator",
":=",
"ast",
".",
"ArithmeticOpInvalid",
"\n",
"var",
"err",
"error",
"\n\n",
"// parse a term that might be the first operand of a binary",
"// expression or it might just be a standalone term, but",
"// we won't know until we've parsed it and can look ahead",
"// to see if there's an operator token.",
"lhs",
",",
"err",
"=",
"p",
".",
"parseBinaryOps",
"(",
"remaining",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n\n",
"// We'll keep eating up arithmetic operators until we run",
"// out, so that operators with the same precedence will combine in a",
"// left-associative manner:",
"// a+b+c => (a+b)+c, not a+(b+c)",
"//",
"// Should we later want to have right-associative operators, a way",
"// to achieve that would be to call back up to ParseExpression here",
"// instead of iteratively parsing only the remaining operators.",
"for",
"{",
"next",
":=",
"p",
".",
"peeker",
".",
"Peek",
"(",
")",
"\n",
"var",
"newOperator",
"ast",
".",
"ArithmeticOp",
"\n",
"var",
"ok",
"bool",
"\n",
"if",
"newOperator",
",",
"ok",
"=",
"thisLevel",
"[",
"next",
".",
"Type",
"]",
";",
"!",
"ok",
"{",
"break",
"\n",
"}",
"\n\n",
"// Are we extending an expression started on",
"// the previous iteration?",
"if",
"operator",
"!=",
"ast",
".",
"ArithmeticOpInvalid",
"{",
"lhs",
"=",
"&",
"ast",
".",
"Arithmetic",
"{",
"Op",
":",
"operator",
",",
"Exprs",
":",
"[",
"]",
"ast",
".",
"Node",
"{",
"lhs",
",",
"rhs",
"}",
",",
"Posx",
":",
"startPos",
",",
"}",
"\n",
"}",
"\n\n",
"operator",
"=",
"newOperator",
"\n",
"p",
".",
"peeker",
".",
"Read",
"(",
")",
"// eat operator token",
"\n",
"rhs",
",",
"err",
"=",
"p",
".",
"parseBinaryOps",
"(",
"remaining",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"err",
"\n",
"}",
"\n",
"}",
"\n\n",
"if",
"operator",
"!=",
"ast",
".",
"ArithmeticOpInvalid",
"{",
"return",
"&",
"ast",
".",
"Arithmetic",
"{",
"Op",
":",
"operator",
",",
"Exprs",
":",
"[",
"]",
"ast",
".",
"Node",
"{",
"lhs",
",",
"rhs",
"}",
",",
"Posx",
":",
"startPos",
",",
"}",
",",
"nil",
"\n",
"}",
"else",
"{",
"return",
"lhs",
",",
"nil",
"\n",
"}",
"\n",
"}"
] | // parseBinaryOps calls itself recursively to work through all of the
// operator precedence groups, and then eventually calls ParseExpressionTerm
// for each operand. | [
"parseBinaryOps",
"calls",
"itself",
"recursively",
"to",
"work",
"through",
"all",
"of",
"the",
"operator",
"precedence",
"groups",
"and",
"then",
"eventually",
"calls",
"ParseExpressionTerm",
"for",
"each",
"operand",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/parser/parser.go#L248-L315 |
5,575 | hashicorp/hil | eval.go | internalEval | func internalEval(root ast.Node, config *EvalConfig) (interface{}, ast.Type, error) {
// Copy the scope so we can add our builtins
if config == nil {
config = new(EvalConfig)
}
scope := registerBuiltins(config.GlobalScope)
implicitMap := map[ast.Type]map[ast.Type]string{
ast.TypeFloat: {
ast.TypeInt: "__builtin_FloatToInt",
ast.TypeString: "__builtin_FloatToString",
},
ast.TypeInt: {
ast.TypeFloat: "__builtin_IntToFloat",
ast.TypeString: "__builtin_IntToString",
},
ast.TypeString: {
ast.TypeInt: "__builtin_StringToInt",
ast.TypeFloat: "__builtin_StringToFloat",
ast.TypeBool: "__builtin_StringToBool",
},
ast.TypeBool: {
ast.TypeString: "__builtin_BoolToString",
},
}
// Build our own semantic checks that we always run
tv := &TypeCheck{Scope: scope, Implicit: implicitMap}
ic := &IdentifierCheck{Scope: scope}
// Build up the semantic checks for execution
checks := make(
[]SemanticChecker,
len(config.SemanticChecks),
len(config.SemanticChecks)+2)
copy(checks, config.SemanticChecks)
checks = append(checks, ic.Visit)
checks = append(checks, tv.Visit)
// Run the semantic checks
for _, check := range checks {
if err := check(root); err != nil {
return nil, ast.TypeInvalid, err
}
}
// Execute
v := &evalVisitor{Scope: scope}
return v.Visit(root)
} | go | func internalEval(root ast.Node, config *EvalConfig) (interface{}, ast.Type, error) {
// Copy the scope so we can add our builtins
if config == nil {
config = new(EvalConfig)
}
scope := registerBuiltins(config.GlobalScope)
implicitMap := map[ast.Type]map[ast.Type]string{
ast.TypeFloat: {
ast.TypeInt: "__builtin_FloatToInt",
ast.TypeString: "__builtin_FloatToString",
},
ast.TypeInt: {
ast.TypeFloat: "__builtin_IntToFloat",
ast.TypeString: "__builtin_IntToString",
},
ast.TypeString: {
ast.TypeInt: "__builtin_StringToInt",
ast.TypeFloat: "__builtin_StringToFloat",
ast.TypeBool: "__builtin_StringToBool",
},
ast.TypeBool: {
ast.TypeString: "__builtin_BoolToString",
},
}
// Build our own semantic checks that we always run
tv := &TypeCheck{Scope: scope, Implicit: implicitMap}
ic := &IdentifierCheck{Scope: scope}
// Build up the semantic checks for execution
checks := make(
[]SemanticChecker,
len(config.SemanticChecks),
len(config.SemanticChecks)+2)
copy(checks, config.SemanticChecks)
checks = append(checks, ic.Visit)
checks = append(checks, tv.Visit)
// Run the semantic checks
for _, check := range checks {
if err := check(root); err != nil {
return nil, ast.TypeInvalid, err
}
}
// Execute
v := &evalVisitor{Scope: scope}
return v.Visit(root)
} | [
"func",
"internalEval",
"(",
"root",
"ast",
".",
"Node",
",",
"config",
"*",
"EvalConfig",
")",
"(",
"interface",
"{",
"}",
",",
"ast",
".",
"Type",
",",
"error",
")",
"{",
"// Copy the scope so we can add our builtins",
"if",
"config",
"==",
"nil",
"{",
"config",
"=",
"new",
"(",
"EvalConfig",
")",
"\n",
"}",
"\n",
"scope",
":=",
"registerBuiltins",
"(",
"config",
".",
"GlobalScope",
")",
"\n",
"implicitMap",
":=",
"map",
"[",
"ast",
".",
"Type",
"]",
"map",
"[",
"ast",
".",
"Type",
"]",
"string",
"{",
"ast",
".",
"TypeFloat",
":",
"{",
"ast",
".",
"TypeInt",
":",
"\"",
"\"",
",",
"ast",
".",
"TypeString",
":",
"\"",
"\"",
",",
"}",
",",
"ast",
".",
"TypeInt",
":",
"{",
"ast",
".",
"TypeFloat",
":",
"\"",
"\"",
",",
"ast",
".",
"TypeString",
":",
"\"",
"\"",
",",
"}",
",",
"ast",
".",
"TypeString",
":",
"{",
"ast",
".",
"TypeInt",
":",
"\"",
"\"",
",",
"ast",
".",
"TypeFloat",
":",
"\"",
"\"",
",",
"ast",
".",
"TypeBool",
":",
"\"",
"\"",
",",
"}",
",",
"ast",
".",
"TypeBool",
":",
"{",
"ast",
".",
"TypeString",
":",
"\"",
"\"",
",",
"}",
",",
"}",
"\n\n",
"// Build our own semantic checks that we always run",
"tv",
":=",
"&",
"TypeCheck",
"{",
"Scope",
":",
"scope",
",",
"Implicit",
":",
"implicitMap",
"}",
"\n",
"ic",
":=",
"&",
"IdentifierCheck",
"{",
"Scope",
":",
"scope",
"}",
"\n\n",
"// Build up the semantic checks for execution",
"checks",
":=",
"make",
"(",
"[",
"]",
"SemanticChecker",
",",
"len",
"(",
"config",
".",
"SemanticChecks",
")",
",",
"len",
"(",
"config",
".",
"SemanticChecks",
")",
"+",
"2",
")",
"\n",
"copy",
"(",
"checks",
",",
"config",
".",
"SemanticChecks",
")",
"\n",
"checks",
"=",
"append",
"(",
"checks",
",",
"ic",
".",
"Visit",
")",
"\n",
"checks",
"=",
"append",
"(",
"checks",
",",
"tv",
".",
"Visit",
")",
"\n\n",
"// Run the semantic checks",
"for",
"_",
",",
"check",
":=",
"range",
"checks",
"{",
"if",
"err",
":=",
"check",
"(",
"root",
")",
";",
"err",
"!=",
"nil",
"{",
"return",
"nil",
",",
"ast",
".",
"TypeInvalid",
",",
"err",
"\n",
"}",
"\n",
"}",
"\n\n",
"// Execute",
"v",
":=",
"&",
"evalVisitor",
"{",
"Scope",
":",
"scope",
"}",
"\n",
"return",
"v",
".",
"Visit",
"(",
"root",
")",
"\n",
"}"
] | // Eval evaluates the given AST tree and returns its output value, the type
// of the output, and any error that occurred. | [
"Eval",
"evaluates",
"the",
"given",
"AST",
"tree",
"and",
"returns",
"its",
"output",
"value",
"the",
"type",
"of",
"the",
"output",
"and",
"any",
"error",
"that",
"occurred",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/eval.go#L106-L154 |
5,576 | hashicorp/hil | eval.go | evalNode | func evalNode(raw ast.Node) (EvalNode, error) {
switch n := raw.(type) {
case *ast.Index:
return &evalIndex{n}, nil
case *ast.Call:
return &evalCall{n}, nil
case *ast.Conditional:
return &evalConditional{n}, nil
case *ast.Output:
return &evalOutput{n}, nil
case *ast.LiteralNode:
return &evalLiteralNode{n}, nil
case *ast.VariableAccess:
return &evalVariableAccess{n}, nil
default:
en, ok := n.(EvalNode)
if !ok {
return nil, fmt.Errorf("node doesn't support evaluation: %#v", raw)
}
return en, nil
}
} | go | func evalNode(raw ast.Node) (EvalNode, error) {
switch n := raw.(type) {
case *ast.Index:
return &evalIndex{n}, nil
case *ast.Call:
return &evalCall{n}, nil
case *ast.Conditional:
return &evalConditional{n}, nil
case *ast.Output:
return &evalOutput{n}, nil
case *ast.LiteralNode:
return &evalLiteralNode{n}, nil
case *ast.VariableAccess:
return &evalVariableAccess{n}, nil
default:
en, ok := n.(EvalNode)
if !ok {
return nil, fmt.Errorf("node doesn't support evaluation: %#v", raw)
}
return en, nil
}
} | [
"func",
"evalNode",
"(",
"raw",
"ast",
".",
"Node",
")",
"(",
"EvalNode",
",",
"error",
")",
"{",
"switch",
"n",
":=",
"raw",
".",
"(",
"type",
")",
"{",
"case",
"*",
"ast",
".",
"Index",
":",
"return",
"&",
"evalIndex",
"{",
"n",
"}",
",",
"nil",
"\n",
"case",
"*",
"ast",
".",
"Call",
":",
"return",
"&",
"evalCall",
"{",
"n",
"}",
",",
"nil",
"\n",
"case",
"*",
"ast",
".",
"Conditional",
":",
"return",
"&",
"evalConditional",
"{",
"n",
"}",
",",
"nil",
"\n",
"case",
"*",
"ast",
".",
"Output",
":",
"return",
"&",
"evalOutput",
"{",
"n",
"}",
",",
"nil",
"\n",
"case",
"*",
"ast",
".",
"LiteralNode",
":",
"return",
"&",
"evalLiteralNode",
"{",
"n",
"}",
",",
"nil",
"\n",
"case",
"*",
"ast",
".",
"VariableAccess",
":",
"return",
"&",
"evalVariableAccess",
"{",
"n",
"}",
",",
"nil",
"\n",
"default",
":",
"en",
",",
"ok",
":=",
"n",
".",
"(",
"EvalNode",
")",
"\n",
"if",
"!",
"ok",
"{",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"raw",
")",
"\n",
"}",
"\n\n",
"return",
"en",
",",
"nil",
"\n",
"}",
"\n",
"}"
] | // evalNode is a private function that returns an EvalNode for built-in
// types as well as any other EvalNode implementations. | [
"evalNode",
"is",
"a",
"private",
"function",
"that",
"returns",
"an",
"EvalNode",
"for",
"built",
"-",
"in",
"types",
"as",
"well",
"as",
"any",
"other",
"EvalNode",
"implementations",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/eval.go#L237-L259 |
5,577 | hashicorp/hil | ast/literal.go | NewLiteralNode | func NewLiteralNode(value interface{}, pos Pos) (*LiteralNode, error) {
goType := reflect.TypeOf(value)
var hilType Type
switch goType.Kind() {
case reflect.Bool:
hilType = TypeBool
case reflect.Int:
hilType = TypeInt
case reflect.Float64:
hilType = TypeFloat
case reflect.String:
hilType = TypeString
default:
return nil, fmt.Errorf("unsupported literal node type: %T", value)
}
return &LiteralNode{
Value: value,
Typex: hilType,
Posx: pos,
}, nil
} | go | func NewLiteralNode(value interface{}, pos Pos) (*LiteralNode, error) {
goType := reflect.TypeOf(value)
var hilType Type
switch goType.Kind() {
case reflect.Bool:
hilType = TypeBool
case reflect.Int:
hilType = TypeInt
case reflect.Float64:
hilType = TypeFloat
case reflect.String:
hilType = TypeString
default:
return nil, fmt.Errorf("unsupported literal node type: %T", value)
}
return &LiteralNode{
Value: value,
Typex: hilType,
Posx: pos,
}, nil
} | [
"func",
"NewLiteralNode",
"(",
"value",
"interface",
"{",
"}",
",",
"pos",
"Pos",
")",
"(",
"*",
"LiteralNode",
",",
"error",
")",
"{",
"goType",
":=",
"reflect",
".",
"TypeOf",
"(",
"value",
")",
"\n",
"var",
"hilType",
"Type",
"\n\n",
"switch",
"goType",
".",
"Kind",
"(",
")",
"{",
"case",
"reflect",
".",
"Bool",
":",
"hilType",
"=",
"TypeBool",
"\n",
"case",
"reflect",
".",
"Int",
":",
"hilType",
"=",
"TypeInt",
"\n",
"case",
"reflect",
".",
"Float64",
":",
"hilType",
"=",
"TypeFloat",
"\n",
"case",
"reflect",
".",
"String",
":",
"hilType",
"=",
"TypeString",
"\n",
"default",
":",
"return",
"nil",
",",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"value",
")",
"\n",
"}",
"\n\n",
"return",
"&",
"LiteralNode",
"{",
"Value",
":",
"value",
",",
"Typex",
":",
"hilType",
",",
"Posx",
":",
"pos",
",",
"}",
",",
"nil",
"\n",
"}"
] | // NewLiteralNode returns a new literal node representing the given
// literal Go value, which must correspond to one of the primitive types
// supported by HIL. Lists and maps cannot currently be constructed via
// this function.
//
// If an inappropriately-typed value is provided, this function will
// return an error. The main intended use of this function is to produce
// "synthetic" literals from constants in code, where the value type is
// well known at compile time. To easily store these in global variables,
// see also MustNewLiteralNode. | [
"NewLiteralNode",
"returns",
"a",
"new",
"literal",
"node",
"representing",
"the",
"given",
"literal",
"Go",
"value",
"which",
"must",
"correspond",
"to",
"one",
"of",
"the",
"primitive",
"types",
"supported",
"by",
"HIL",
".",
"Lists",
"and",
"maps",
"cannot",
"currently",
"be",
"constructed",
"via",
"this",
"function",
".",
"If",
"an",
"inappropriately",
"-",
"typed",
"value",
"is",
"provided",
"this",
"function",
"will",
"return",
"an",
"error",
".",
"The",
"main",
"intended",
"use",
"of",
"this",
"function",
"is",
"to",
"produce",
"synthetic",
"literals",
"from",
"constants",
"in",
"code",
"where",
"the",
"value",
"type",
"is",
"well",
"known",
"at",
"compile",
"time",
".",
"To",
"easily",
"store",
"these",
"in",
"global",
"variables",
"see",
"also",
"MustNewLiteralNode",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L26-L48 |
5,578 | hashicorp/hil | ast/literal.go | MustNewLiteralNode | func MustNewLiteralNode(value interface{}, pos Pos) *LiteralNode {
node, err := NewLiteralNode(value, pos)
if err != nil {
panic(err)
}
return node
} | go | func MustNewLiteralNode(value interface{}, pos Pos) *LiteralNode {
node, err := NewLiteralNode(value, pos)
if err != nil {
panic(err)
}
return node
} | [
"func",
"MustNewLiteralNode",
"(",
"value",
"interface",
"{",
"}",
",",
"pos",
"Pos",
")",
"*",
"LiteralNode",
"{",
"node",
",",
"err",
":=",
"NewLiteralNode",
"(",
"value",
",",
"pos",
")",
"\n",
"if",
"err",
"!=",
"nil",
"{",
"panic",
"(",
"err",
")",
"\n",
"}",
"\n",
"return",
"node",
"\n",
"}"
] | // MustNewLiteralNode wraps NewLiteralNode and panics if an error is
// returned, thus allowing valid literal nodes to be easily assigned to
// global variables. | [
"MustNewLiteralNode",
"wraps",
"NewLiteralNode",
"and",
"panics",
"if",
"an",
"error",
"is",
"returned",
"thus",
"allowing",
"valid",
"literal",
"nodes",
"to",
"be",
"easily",
"assigned",
"to",
"global",
"variables",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L53-L59 |
5,579 | hashicorp/hil | ast/literal.go | IsUnknown | func (n *LiteralNode) IsUnknown() bool {
return IsUnknown(Variable{
Type: n.Typex,
Value: n.Value,
})
} | go | func (n *LiteralNode) IsUnknown() bool {
return IsUnknown(Variable{
Type: n.Typex,
Value: n.Value,
})
} | [
"func",
"(",
"n",
"*",
"LiteralNode",
")",
"IsUnknown",
"(",
")",
"bool",
"{",
"return",
"IsUnknown",
"(",
"Variable",
"{",
"Type",
":",
"n",
".",
"Typex",
",",
"Value",
":",
"n",
".",
"Value",
",",
"}",
")",
"\n",
"}"
] | // IsUnknown returns true either if the node's value is itself unknown
// of if it is a collection containing any unknown elements, deeply. | [
"IsUnknown",
"returns",
"true",
"either",
"if",
"the",
"node",
"s",
"value",
"is",
"itself",
"unknown",
"of",
"if",
"it",
"is",
"a",
"collection",
"containing",
"any",
"unknown",
"elements",
"deeply",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/literal.go#L83-L88 |
5,580 | hashicorp/hil | ast/unknown.go | IsUnknown | func IsUnknown(v Variable) bool {
// If it is unknown itself, return true
if v.Type == TypeUnknown {
return true
}
// If it is a container type, check the values
switch v.Type {
case TypeList:
for _, el := range v.Value.([]Variable) {
if IsUnknown(el) {
return true
}
}
case TypeMap:
for _, el := range v.Value.(map[string]Variable) {
if IsUnknown(el) {
return true
}
}
default:
}
// Not a container type or survive the above checks
return false
} | go | func IsUnknown(v Variable) bool {
// If it is unknown itself, return true
if v.Type == TypeUnknown {
return true
}
// If it is a container type, check the values
switch v.Type {
case TypeList:
for _, el := range v.Value.([]Variable) {
if IsUnknown(el) {
return true
}
}
case TypeMap:
for _, el := range v.Value.(map[string]Variable) {
if IsUnknown(el) {
return true
}
}
default:
}
// Not a container type or survive the above checks
return false
} | [
"func",
"IsUnknown",
"(",
"v",
"Variable",
")",
"bool",
"{",
"// If it is unknown itself, return true",
"if",
"v",
".",
"Type",
"==",
"TypeUnknown",
"{",
"return",
"true",
"\n",
"}",
"\n\n",
"// If it is a container type, check the values",
"switch",
"v",
".",
"Type",
"{",
"case",
"TypeList",
":",
"for",
"_",
",",
"el",
":=",
"range",
"v",
".",
"Value",
".",
"(",
"[",
"]",
"Variable",
")",
"{",
"if",
"IsUnknown",
"(",
"el",
")",
"{",
"return",
"true",
"\n",
"}",
"\n",
"}",
"\n",
"case",
"TypeMap",
":",
"for",
"_",
",",
"el",
":=",
"range",
"v",
".",
"Value",
".",
"(",
"map",
"[",
"string",
"]",
"Variable",
")",
"{",
"if",
"IsUnknown",
"(",
"el",
")",
"{",
"return",
"true",
"\n",
"}",
"\n",
"}",
"\n",
"default",
":",
"}",
"\n\n",
"// Not a container type or survive the above checks",
"return",
"false",
"\n",
"}"
] | // IsUnknown reports whether a variable is unknown or contains any value
// that is unknown. This will recurse into lists and maps and so on. | [
"IsUnknown",
"reports",
"whether",
"a",
"variable",
"is",
"unknown",
"or",
"contains",
"any",
"value",
"that",
"is",
"unknown",
".",
"This",
"will",
"recurse",
"into",
"lists",
"and",
"maps",
"and",
"so",
"on",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/unknown.go#L5-L30 |
5,581 | hashicorp/hil | walk.go | Walk | func Walk(v interface{}, cb WalkFn) error {
walker := &interpolationWalker{F: cb}
return reflectwalk.Walk(v, walker)
} | go | func Walk(v interface{}, cb WalkFn) error {
walker := &interpolationWalker{F: cb}
return reflectwalk.Walk(v, walker)
} | [
"func",
"Walk",
"(",
"v",
"interface",
"{",
"}",
",",
"cb",
"WalkFn",
")",
"error",
"{",
"walker",
":=",
"&",
"interpolationWalker",
"{",
"F",
":",
"cb",
"}",
"\n",
"return",
"reflectwalk",
".",
"Walk",
"(",
"v",
",",
"walker",
")",
"\n",
"}"
] | // Walk will walk an arbitrary Go structure and parse any string as an
// HIL program and call the callback cb to determine what to replace it
// with.
//
// This function is very useful for arbitrary HIL program interpolation
// across a complex configuration structure. Due to the heavy use of
// reflection in this function, it is recommend to write many unit tests
// with your typical configuration structures to hilp mitigate the risk
// of panics. | [
"Walk",
"will",
"walk",
"an",
"arbitrary",
"Go",
"structure",
"and",
"parse",
"any",
"string",
"as",
"an",
"HIL",
"program",
"and",
"call",
"the",
"callback",
"cb",
"to",
"determine",
"what",
"to",
"replace",
"it",
"with",
".",
"This",
"function",
"is",
"very",
"useful",
"for",
"arbitrary",
"HIL",
"program",
"interpolation",
"across",
"a",
"complex",
"configuration",
"structure",
".",
"Due",
"to",
"the",
"heavy",
"use",
"of",
"reflection",
"in",
"this",
"function",
"it",
"is",
"recommend",
"to",
"write",
"many",
"unit",
"tests",
"with",
"your",
"typical",
"configuration",
"structures",
"to",
"hilp",
"mitigate",
"the",
"risk",
"of",
"panics",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/walk.go#L47-L50 |
5,582 | hashicorp/hil | scanner/scanner.go | Scan | func Scan(s string, startPos ast.Pos) <-chan *Token {
ch := make(chan *Token)
go scan(s, ch, startPos)
return ch
} | go | func Scan(s string, startPos ast.Pos) <-chan *Token {
ch := make(chan *Token)
go scan(s, ch, startPos)
return ch
} | [
"func",
"Scan",
"(",
"s",
"string",
",",
"startPos",
"ast",
".",
"Pos",
")",
"<-",
"chan",
"*",
"Token",
"{",
"ch",
":=",
"make",
"(",
"chan",
"*",
"Token",
")",
"\n",
"go",
"scan",
"(",
"s",
",",
"ch",
",",
"startPos",
")",
"\n",
"return",
"ch",
"\n",
"}"
] | // Scan returns a channel that recieves Tokens from the given input string.
//
// The scanner's job is just to partition the string into meaningful parts.
// It doesn't do any transformation of the raw input string, so the caller
// must deal with any further interpretation required, such as parsing INTEGER
// tokens into real ints, or dealing with escape sequences in LITERAL or
// STRING tokens.
//
// Strings in the returned tokens are slices from the original string.
//
// startPos should be set to ast.InitPos unless the caller knows that
// this interpolation string is part of a larger file and knows the position
// of the first character in that larger file. | [
"Scan",
"returns",
"a",
"channel",
"that",
"recieves",
"Tokens",
"from",
"the",
"given",
"input",
"string",
".",
"The",
"scanner",
"s",
"job",
"is",
"just",
"to",
"partition",
"the",
"string",
"into",
"meaningful",
"parts",
".",
"It",
"doesn",
"t",
"do",
"any",
"transformation",
"of",
"the",
"raw",
"input",
"string",
"so",
"the",
"caller",
"must",
"deal",
"with",
"any",
"further",
"interpretation",
"required",
"such",
"as",
"parsing",
"INTEGER",
"tokens",
"into",
"real",
"ints",
"or",
"dealing",
"with",
"escape",
"sequences",
"in",
"LITERAL",
"or",
"STRING",
"tokens",
".",
"Strings",
"in",
"the",
"returned",
"tokens",
"are",
"slices",
"from",
"the",
"original",
"string",
".",
"startPos",
"should",
"be",
"set",
"to",
"ast",
".",
"InitPos",
"unless",
"the",
"caller",
"knows",
"that",
"this",
"interpolation",
"string",
"is",
"part",
"of",
"a",
"larger",
"file",
"and",
"knows",
"the",
"position",
"of",
"the",
"first",
"character",
"in",
"that",
"larger",
"file",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/scanner.go#L23-L27 |
5,583 | hashicorp/hil | scanner/scanner.go | scanIdentifier | func scanIdentifier(s string) (string, int) {
byteLen := 0
runeLen := 0
for {
if byteLen >= len(s) {
break
}
nextRune, size := utf8.DecodeRuneInString(s[byteLen:])
if !(nextRune == '_' ||
nextRune == '-' ||
nextRune == '.' ||
nextRune == '*' ||
unicode.IsNumber(nextRune) ||
unicode.IsLetter(nextRune) ||
unicode.IsMark(nextRune)) {
break
}
// If we reach a star, it must be between periods to be part
// of the same identifier.
if nextRune == '*' && s[byteLen-1] != '.' {
break
}
// If our previous character was a star, then the current must
// be period. Otherwise, undo that and exit.
if byteLen > 0 && s[byteLen-1] == '*' && nextRune != '.' {
byteLen--
if s[byteLen-1] == '.' {
byteLen--
}
break
}
byteLen = byteLen + size
runeLen = runeLen + 1
}
return s[:byteLen], runeLen
} | go | func scanIdentifier(s string) (string, int) {
byteLen := 0
runeLen := 0
for {
if byteLen >= len(s) {
break
}
nextRune, size := utf8.DecodeRuneInString(s[byteLen:])
if !(nextRune == '_' ||
nextRune == '-' ||
nextRune == '.' ||
nextRune == '*' ||
unicode.IsNumber(nextRune) ||
unicode.IsLetter(nextRune) ||
unicode.IsMark(nextRune)) {
break
}
// If we reach a star, it must be between periods to be part
// of the same identifier.
if nextRune == '*' && s[byteLen-1] != '.' {
break
}
// If our previous character was a star, then the current must
// be period. Otherwise, undo that and exit.
if byteLen > 0 && s[byteLen-1] == '*' && nextRune != '.' {
byteLen--
if s[byteLen-1] == '.' {
byteLen--
}
break
}
byteLen = byteLen + size
runeLen = runeLen + 1
}
return s[:byteLen], runeLen
} | [
"func",
"scanIdentifier",
"(",
"s",
"string",
")",
"(",
"string",
",",
"int",
")",
"{",
"byteLen",
":=",
"0",
"\n",
"runeLen",
":=",
"0",
"\n",
"for",
"{",
"if",
"byteLen",
">=",
"len",
"(",
"s",
")",
"{",
"break",
"\n",
"}",
"\n\n",
"nextRune",
",",
"size",
":=",
"utf8",
".",
"DecodeRuneInString",
"(",
"s",
"[",
"byteLen",
":",
"]",
")",
"\n",
"if",
"!",
"(",
"nextRune",
"==",
"'_'",
"||",
"nextRune",
"==",
"'-'",
"||",
"nextRune",
"==",
"'.'",
"||",
"nextRune",
"==",
"'*'",
"||",
"unicode",
".",
"IsNumber",
"(",
"nextRune",
")",
"||",
"unicode",
".",
"IsLetter",
"(",
"nextRune",
")",
"||",
"unicode",
".",
"IsMark",
"(",
"nextRune",
")",
")",
"{",
"break",
"\n",
"}",
"\n\n",
"// If we reach a star, it must be between periods to be part",
"// of the same identifier.",
"if",
"nextRune",
"==",
"'*'",
"&&",
"s",
"[",
"byteLen",
"-",
"1",
"]",
"!=",
"'.'",
"{",
"break",
"\n",
"}",
"\n\n",
"// If our previous character was a star, then the current must",
"// be period. Otherwise, undo that and exit.",
"if",
"byteLen",
">",
"0",
"&&",
"s",
"[",
"byteLen",
"-",
"1",
"]",
"==",
"'*'",
"&&",
"nextRune",
"!=",
"'.'",
"{",
"byteLen",
"--",
"\n",
"if",
"s",
"[",
"byteLen",
"-",
"1",
"]",
"==",
"'.'",
"{",
"byteLen",
"--",
"\n",
"}",
"\n\n",
"break",
"\n",
"}",
"\n\n",
"byteLen",
"=",
"byteLen",
"+",
"size",
"\n",
"runeLen",
"=",
"runeLen",
"+",
"1",
"\n",
"}",
"\n\n",
"return",
"s",
"[",
":",
"byteLen",
"]",
",",
"runeLen",
"\n",
"}"
] | // scanIdentifier returns the extent of the prefix of the string that
// represents a valid identifier, along with the length of that prefix
// in runes.
//
// Identifiers may contain utf8-encoded non-Latin letters, which will
// cause the returned "rune length" to be shorter than the byte length
// of the returned string. | [
"scanIdentifier",
"returns",
"the",
"extent",
"of",
"the",
"prefix",
"of",
"the",
"string",
"that",
"represents",
"a",
"valid",
"identifier",
"along",
"with",
"the",
"length",
"of",
"that",
"prefix",
"in",
"runes",
".",
"Identifiers",
"may",
"contain",
"utf8",
"-",
"encoded",
"non",
"-",
"Latin",
"letters",
"which",
"will",
"cause",
"the",
"returned",
"rune",
"length",
"to",
"be",
"shorter",
"than",
"the",
"byte",
"length",
"of",
"the",
"returned",
"string",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/scanner/scanner.go#L477-L518 |
5,584 | hashicorp/hil | ast/scope.go | NewVariable | func NewVariable(v interface{}) (result Variable, err error) {
switch v := reflect.ValueOf(v); v.Kind() {
case reflect.String:
result.Type = TypeString
default:
err = fmt.Errorf("Unknown type: %s", v.Kind())
}
result.Value = v
return
} | go | func NewVariable(v interface{}) (result Variable, err error) {
switch v := reflect.ValueOf(v); v.Kind() {
case reflect.String:
result.Type = TypeString
default:
err = fmt.Errorf("Unknown type: %s", v.Kind())
}
result.Value = v
return
} | [
"func",
"NewVariable",
"(",
"v",
"interface",
"{",
"}",
")",
"(",
"result",
"Variable",
",",
"err",
"error",
")",
"{",
"switch",
"v",
":=",
"reflect",
".",
"ValueOf",
"(",
"v",
")",
";",
"v",
".",
"Kind",
"(",
")",
"{",
"case",
"reflect",
".",
"String",
":",
"result",
".",
"Type",
"=",
"TypeString",
"\n",
"default",
":",
"err",
"=",
"fmt",
".",
"Errorf",
"(",
"\"",
"\"",
",",
"v",
".",
"Kind",
"(",
")",
")",
"\n",
"}",
"\n\n",
"result",
".",
"Value",
"=",
"v",
"\n",
"return",
"\n",
"}"
] | // NewVariable creates a new Variable for the given value. This will
// attempt to infer the correct type. If it can't, an error will be returned. | [
"NewVariable",
"creates",
"a",
"new",
"Variable",
"for",
"the",
"given",
"value",
".",
"This",
"will",
"attempt",
"to",
"infer",
"the",
"correct",
"type",
".",
"If",
"it",
"can",
"t",
"an",
"error",
"will",
"be",
"returned",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/scope.go#L24-L34 |
5,585 | hashicorp/hil | ast/scope.go | String | func (v Variable) String() string {
return fmt.Sprintf("{Variable (%s): %+v}", v.Type, v.Value)
} | go | func (v Variable) String() string {
return fmt.Sprintf("{Variable (%s): %+v}", v.Type, v.Value)
} | [
"func",
"(",
"v",
"Variable",
")",
"String",
"(",
")",
"string",
"{",
"return",
"fmt",
".",
"Sprintf",
"(",
"\"",
"\"",
",",
"v",
".",
"Type",
",",
"v",
".",
"Value",
")",
"\n",
"}"
] | // String implements Stringer on Variable, displaying the type and value
// of the Variable. | [
"String",
"implements",
"Stringer",
"on",
"Variable",
"displaying",
"the",
"type",
"and",
"value",
"of",
"the",
"Variable",
"."
] | 97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63 | https://github.com/hashicorp/hil/blob/97b3a9cdfa9349086cfad7ea2fe3165bfe3cbf63/ast/scope.go#L38-L40 |
5,586 | minio/sha256-simd | sha256blockAvx512_amd64.go | NewAvx512 | func NewAvx512(a512srv *Avx512Server) hash.Hash {
uid := atomic.AddUint64(&uidCounter, 1)
return &Avx512Digest{uid: uid, a512srv: a512srv}
} | go | func NewAvx512(a512srv *Avx512Server) hash.Hash {
uid := atomic.AddUint64(&uidCounter, 1)
return &Avx512Digest{uid: uid, a512srv: a512srv}
} | [
"func",
"NewAvx512",
"(",
"a512srv",
"*",
"Avx512Server",
")",
"hash",
".",
"Hash",
"{",
"uid",
":=",
"atomic",
".",
"AddUint64",
"(",
"&",
"uidCounter",
",",
"1",
")",
"\n",
"return",
"&",
"Avx512Digest",
"{",
"uid",
":",
"uid",
",",
"a512srv",
":",
"a512srv",
"}",
"\n",
"}"
] | // NewAvx512 - initialize sha256 Avx512 implementation. | [
"NewAvx512",
"-",
"initialize",
"sha256",
"Avx512",
"implementation",
"."
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L40-L43 |
5,587 | minio/sha256-simd | sha256blockAvx512_amd64.go | Reset | func (d *Avx512Digest) Reset() {
d.a512srv.blocksCh <- blockInput{uid: d.uid, reset: true}
d.nx = 0
d.len = 0
d.final = false
} | go | func (d *Avx512Digest) Reset() {
d.a512srv.blocksCh <- blockInput{uid: d.uid, reset: true}
d.nx = 0
d.len = 0
d.final = false
} | [
"func",
"(",
"d",
"*",
"Avx512Digest",
")",
"Reset",
"(",
")",
"{",
"d",
".",
"a512srv",
".",
"blocksCh",
"<-",
"blockInput",
"{",
"uid",
":",
"d",
".",
"uid",
",",
"reset",
":",
"true",
"}",
"\n",
"d",
".",
"nx",
"=",
"0",
"\n",
"d",
".",
"len",
"=",
"0",
"\n",
"d",
".",
"final",
"=",
"false",
"\n",
"}"
] | // Reset - reset sha digest to its initial values | [
"Reset",
"-",
"reset",
"sha",
"digest",
"to",
"its",
"initial",
"values"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L63-L68 |
5,588 | minio/sha256-simd | sha256blockAvx512_amd64.go | Write | func (d *Avx512Digest) Write(p []byte) (nn int, err error) {
if d.final {
return 0, errors.New("Avx512Digest already finalized. Reset first before writing again")
}
nn = len(p)
d.len += uint64(nn)
if d.nx > 0 {
n := copy(d.x[d.nx:], p)
d.nx += n
if d.nx == chunk {
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: d.x[:]}
d.nx = 0
}
p = p[n:]
}
if len(p) >= chunk {
n := len(p) &^ (chunk - 1)
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: p[:n]}
p = p[n:]
}
if len(p) > 0 {
d.nx = copy(d.x[:], p)
}
return
} | go | func (d *Avx512Digest) Write(p []byte) (nn int, err error) {
if d.final {
return 0, errors.New("Avx512Digest already finalized. Reset first before writing again")
}
nn = len(p)
d.len += uint64(nn)
if d.nx > 0 {
n := copy(d.x[d.nx:], p)
d.nx += n
if d.nx == chunk {
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: d.x[:]}
d.nx = 0
}
p = p[n:]
}
if len(p) >= chunk {
n := len(p) &^ (chunk - 1)
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: p[:n]}
p = p[n:]
}
if len(p) > 0 {
d.nx = copy(d.x[:], p)
}
return
} | [
"func",
"(",
"d",
"*",
"Avx512Digest",
")",
"Write",
"(",
"p",
"[",
"]",
"byte",
")",
"(",
"nn",
"int",
",",
"err",
"error",
")",
"{",
"if",
"d",
".",
"final",
"{",
"return",
"0",
",",
"errors",
".",
"New",
"(",
"\"",
"\"",
")",
"\n",
"}",
"\n\n",
"nn",
"=",
"len",
"(",
"p",
")",
"\n",
"d",
".",
"len",
"+=",
"uint64",
"(",
"nn",
")",
"\n",
"if",
"d",
".",
"nx",
">",
"0",
"{",
"n",
":=",
"copy",
"(",
"d",
".",
"x",
"[",
"d",
".",
"nx",
":",
"]",
",",
"p",
")",
"\n",
"d",
".",
"nx",
"+=",
"n",
"\n",
"if",
"d",
".",
"nx",
"==",
"chunk",
"{",
"d",
".",
"a512srv",
".",
"blocksCh",
"<-",
"blockInput",
"{",
"uid",
":",
"d",
".",
"uid",
",",
"msg",
":",
"d",
".",
"x",
"[",
":",
"]",
"}",
"\n",
"d",
".",
"nx",
"=",
"0",
"\n",
"}",
"\n",
"p",
"=",
"p",
"[",
"n",
":",
"]",
"\n",
"}",
"\n",
"if",
"len",
"(",
"p",
")",
">=",
"chunk",
"{",
"n",
":=",
"len",
"(",
"p",
")",
"&^",
"(",
"chunk",
"-",
"1",
")",
"\n",
"d",
".",
"a512srv",
".",
"blocksCh",
"<-",
"blockInput",
"{",
"uid",
":",
"d",
".",
"uid",
",",
"msg",
":",
"p",
"[",
":",
"n",
"]",
"}",
"\n",
"p",
"=",
"p",
"[",
"n",
":",
"]",
"\n",
"}",
"\n",
"if",
"len",
"(",
"p",
")",
">",
"0",
"{",
"d",
".",
"nx",
"=",
"copy",
"(",
"d",
".",
"x",
"[",
":",
"]",
",",
"p",
")",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Write to digest | [
"Write",
"to",
"digest"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L71-L97 |
5,589 | minio/sha256-simd | sha256blockAvx512_amd64.go | Sum | func (d *Avx512Digest) Sum(in []byte) (result []byte) {
if d.final {
return append(in, d.result[:]...)
}
trail := make([]byte, 0, 128)
len := d.len
// Padding. Add a 1 bit and 0 bits until 56 bytes mod 64.
var tmp [64]byte
tmp[0] = 0x80
if len%64 < 56 {
trail = append(d.x[:d.nx], tmp[0:56-len%64]...)
} else {
trail = append(d.x[:d.nx], tmp[0:64+56-len%64]...)
}
d.nx = 0
// Length in bits.
len <<= 3
for i := uint(0); i < 8; i++ {
tmp[i] = byte(len >> (56 - 8*i))
}
trail = append(trail, tmp[0:8]...)
sumCh := make(chan [Size]byte)
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: trail, final: true, sumCh: sumCh}
d.result = <-sumCh
d.final = true
return append(in, d.result[:]...)
} | go | func (d *Avx512Digest) Sum(in []byte) (result []byte) {
if d.final {
return append(in, d.result[:]...)
}
trail := make([]byte, 0, 128)
len := d.len
// Padding. Add a 1 bit and 0 bits until 56 bytes mod 64.
var tmp [64]byte
tmp[0] = 0x80
if len%64 < 56 {
trail = append(d.x[:d.nx], tmp[0:56-len%64]...)
} else {
trail = append(d.x[:d.nx], tmp[0:64+56-len%64]...)
}
d.nx = 0
// Length in bits.
len <<= 3
for i := uint(0); i < 8; i++ {
tmp[i] = byte(len >> (56 - 8*i))
}
trail = append(trail, tmp[0:8]...)
sumCh := make(chan [Size]byte)
d.a512srv.blocksCh <- blockInput{uid: d.uid, msg: trail, final: true, sumCh: sumCh}
d.result = <-sumCh
d.final = true
return append(in, d.result[:]...)
} | [
"func",
"(",
"d",
"*",
"Avx512Digest",
")",
"Sum",
"(",
"in",
"[",
"]",
"byte",
")",
"(",
"result",
"[",
"]",
"byte",
")",
"{",
"if",
"d",
".",
"final",
"{",
"return",
"append",
"(",
"in",
",",
"d",
".",
"result",
"[",
":",
"]",
"...",
")",
"\n",
"}",
"\n\n",
"trail",
":=",
"make",
"(",
"[",
"]",
"byte",
",",
"0",
",",
"128",
")",
"\n\n",
"len",
":=",
"d",
".",
"len",
"\n",
"// Padding. Add a 1 bit and 0 bits until 56 bytes mod 64.",
"var",
"tmp",
"[",
"64",
"]",
"byte",
"\n",
"tmp",
"[",
"0",
"]",
"=",
"0x80",
"\n",
"if",
"len",
"%",
"64",
"<",
"56",
"{",
"trail",
"=",
"append",
"(",
"d",
".",
"x",
"[",
":",
"d",
".",
"nx",
"]",
",",
"tmp",
"[",
"0",
":",
"56",
"-",
"len",
"%",
"64",
"]",
"...",
")",
"\n",
"}",
"else",
"{",
"trail",
"=",
"append",
"(",
"d",
".",
"x",
"[",
":",
"d",
".",
"nx",
"]",
",",
"tmp",
"[",
"0",
":",
"64",
"+",
"56",
"-",
"len",
"%",
"64",
"]",
"...",
")",
"\n",
"}",
"\n",
"d",
".",
"nx",
"=",
"0",
"\n\n",
"// Length in bits.",
"len",
"<<=",
"3",
"\n",
"for",
"i",
":=",
"uint",
"(",
"0",
")",
";",
"i",
"<",
"8",
";",
"i",
"++",
"{",
"tmp",
"[",
"i",
"]",
"=",
"byte",
"(",
"len",
">>",
"(",
"56",
"-",
"8",
"*",
"i",
")",
")",
"\n",
"}",
"\n",
"trail",
"=",
"append",
"(",
"trail",
",",
"tmp",
"[",
"0",
":",
"8",
"]",
"...",
")",
"\n\n",
"sumCh",
":=",
"make",
"(",
"chan",
"[",
"Size",
"]",
"byte",
")",
"\n",
"d",
".",
"a512srv",
".",
"blocksCh",
"<-",
"blockInput",
"{",
"uid",
":",
"d",
".",
"uid",
",",
"msg",
":",
"trail",
",",
"final",
":",
"true",
",",
"sumCh",
":",
"sumCh",
"}",
"\n",
"d",
".",
"result",
"=",
"<-",
"sumCh",
"\n",
"d",
".",
"final",
"=",
"true",
"\n",
"return",
"append",
"(",
"in",
",",
"d",
".",
"result",
"[",
":",
"]",
"...",
")",
"\n",
"}"
] | // Sum - Return sha256 sum in bytes | [
"Sum",
"-",
"Return",
"sha256",
"sum",
"in",
"bytes"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L100-L131 |
5,590 | minio/sha256-simd | sha256blockAvx512_amd64.go | blockAvx512 | func blockAvx512(digests *[512]byte, input [16][]byte, mask []uint64) [16][Size]byte {
scratch := [512]byte{}
sha256X16Avx512(digests, &scratch, &table, mask, input)
output := [16][Size]byte{}
for i := 0; i < 16; i++ {
output[i] = getDigest(i, digests[:])
}
return output
} | go | func blockAvx512(digests *[512]byte, input [16][]byte, mask []uint64) [16][Size]byte {
scratch := [512]byte{}
sha256X16Avx512(digests, &scratch, &table, mask, input)
output := [16][Size]byte{}
for i := 0; i < 16; i++ {
output[i] = getDigest(i, digests[:])
}
return output
} | [
"func",
"blockAvx512",
"(",
"digests",
"*",
"[",
"512",
"]",
"byte",
",",
"input",
"[",
"16",
"]",
"[",
"]",
"byte",
",",
"mask",
"[",
"]",
"uint64",
")",
"[",
"16",
"]",
"[",
"Size",
"]",
"byte",
"{",
"scratch",
":=",
"[",
"512",
"]",
"byte",
"{",
"}",
"\n",
"sha256X16Avx512",
"(",
"digests",
",",
"&",
"scratch",
",",
"&",
"table",
",",
"mask",
",",
"input",
")",
"\n\n",
"output",
":=",
"[",
"16",
"]",
"[",
"Size",
"]",
"byte",
"{",
"}",
"\n",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"16",
";",
"i",
"++",
"{",
"output",
"[",
"i",
"]",
"=",
"getDigest",
"(",
"i",
",",
"digests",
"[",
":",
"]",
")",
"\n",
"}",
"\n\n",
"return",
"output",
"\n",
"}"
] | // Interface function to assembly ode | [
"Interface",
"function",
"to",
"assembly",
"ode"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L264-L275 |
5,591 | minio/sha256-simd | sha256blockAvx512_amd64.go | NewAvx512Server | func NewAvx512Server() *Avx512Server {
a512srv := &Avx512Server{}
a512srv.digests = make(map[uint64][Size]byte)
a512srv.blocksCh = make(chan blockInput)
// Start a single thread for reading from the input channel
go a512srv.Process()
return a512srv
} | go | func NewAvx512Server() *Avx512Server {
a512srv := &Avx512Server{}
a512srv.digests = make(map[uint64][Size]byte)
a512srv.blocksCh = make(chan blockInput)
// Start a single thread for reading from the input channel
go a512srv.Process()
return a512srv
} | [
"func",
"NewAvx512Server",
"(",
")",
"*",
"Avx512Server",
"{",
"a512srv",
":=",
"&",
"Avx512Server",
"{",
"}",
"\n",
"a512srv",
".",
"digests",
"=",
"make",
"(",
"map",
"[",
"uint64",
"]",
"[",
"Size",
"]",
"byte",
")",
"\n",
"a512srv",
".",
"blocksCh",
"=",
"make",
"(",
"chan",
"blockInput",
")",
"\n\n",
"// Start a single thread for reading from the input channel",
"go",
"a512srv",
".",
"Process",
"(",
")",
"\n",
"return",
"a512srv",
"\n",
"}"
] | // NewAvx512Server - Create new object for parallel processing handling | [
"NewAvx512Server",
"-",
"Create",
"new",
"object",
"for",
"parallel",
"processing",
"handling"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L311-L319 |
5,592 | minio/sha256-simd | sha256blockAvx512_amd64.go | Process | func (a512srv *Avx512Server) Process() {
for {
select {
case block := <-a512srv.blocksCh:
if block.reset {
a512srv.reset(block.uid)
continue
}
index := block.uid & 0xf
// fmt.Println("Adding message:", block.uid, index)
if a512srv.lanes[index].block != nil { // If slot is already filled, process all inputs
//fmt.Println("Invoking Blocks()")
a512srv.blocks()
}
a512srv.totalIn++
a512srv.lanes[index] = Avx512LaneInfo{uid: block.uid, block: block.msg}
if block.final {
a512srv.lanes[index].outputCh = block.sumCh
}
if a512srv.totalIn == len(a512srv.lanes) {
// fmt.Println("Invoking Blocks() while FULL: ")
a512srv.blocks()
}
// TODO: test with larger timeout
case <-time.After(1 * time.Microsecond):
for _, lane := range a512srv.lanes {
if lane.block != nil { // check if there is any input to process
// fmt.Println("Invoking Blocks() on TIMEOUT: ")
a512srv.blocks()
break // we are done
}
}
}
}
} | go | func (a512srv *Avx512Server) Process() {
for {
select {
case block := <-a512srv.blocksCh:
if block.reset {
a512srv.reset(block.uid)
continue
}
index := block.uid & 0xf
// fmt.Println("Adding message:", block.uid, index)
if a512srv.lanes[index].block != nil { // If slot is already filled, process all inputs
//fmt.Println("Invoking Blocks()")
a512srv.blocks()
}
a512srv.totalIn++
a512srv.lanes[index] = Avx512LaneInfo{uid: block.uid, block: block.msg}
if block.final {
a512srv.lanes[index].outputCh = block.sumCh
}
if a512srv.totalIn == len(a512srv.lanes) {
// fmt.Println("Invoking Blocks() while FULL: ")
a512srv.blocks()
}
// TODO: test with larger timeout
case <-time.After(1 * time.Microsecond):
for _, lane := range a512srv.lanes {
if lane.block != nil { // check if there is any input to process
// fmt.Println("Invoking Blocks() on TIMEOUT: ")
a512srv.blocks()
break // we are done
}
}
}
}
} | [
"func",
"(",
"a512srv",
"*",
"Avx512Server",
")",
"Process",
"(",
")",
"{",
"for",
"{",
"select",
"{",
"case",
"block",
":=",
"<-",
"a512srv",
".",
"blocksCh",
":",
"if",
"block",
".",
"reset",
"{",
"a512srv",
".",
"reset",
"(",
"block",
".",
"uid",
")",
"\n",
"continue",
"\n",
"}",
"\n",
"index",
":=",
"block",
".",
"uid",
"&",
"0xf",
"\n",
"// fmt.Println(\"Adding message:\", block.uid, index)",
"if",
"a512srv",
".",
"lanes",
"[",
"index",
"]",
".",
"block",
"!=",
"nil",
"{",
"// If slot is already filled, process all inputs",
"//fmt.Println(\"Invoking Blocks()\")",
"a512srv",
".",
"blocks",
"(",
")",
"\n",
"}",
"\n",
"a512srv",
".",
"totalIn",
"++",
"\n",
"a512srv",
".",
"lanes",
"[",
"index",
"]",
"=",
"Avx512LaneInfo",
"{",
"uid",
":",
"block",
".",
"uid",
",",
"block",
":",
"block",
".",
"msg",
"}",
"\n",
"if",
"block",
".",
"final",
"{",
"a512srv",
".",
"lanes",
"[",
"index",
"]",
".",
"outputCh",
"=",
"block",
".",
"sumCh",
"\n",
"}",
"\n",
"if",
"a512srv",
".",
"totalIn",
"==",
"len",
"(",
"a512srv",
".",
"lanes",
")",
"{",
"// fmt.Println(\"Invoking Blocks() while FULL: \")",
"a512srv",
".",
"blocks",
"(",
")",
"\n",
"}",
"\n\n",
"// TODO: test with larger timeout",
"case",
"<-",
"time",
".",
"After",
"(",
"1",
"*",
"time",
".",
"Microsecond",
")",
":",
"for",
"_",
",",
"lane",
":=",
"range",
"a512srv",
".",
"lanes",
"{",
"if",
"lane",
".",
"block",
"!=",
"nil",
"{",
"// check if there is any input to process",
"// fmt.Println(\"Invoking Blocks() on TIMEOUT: \")",
"a512srv",
".",
"blocks",
"(",
")",
"\n",
"break",
"// we are done",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n",
"}"
] | // Process - Sole handler for reading from the input channel | [
"Process",
"-",
"Sole",
"handler",
"for",
"reading",
"from",
"the",
"input",
"channel"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L322-L358 |
5,593 | minio/sha256-simd | sha256blockAvx512_amd64.go | reset | func (a512srv *Avx512Server) reset(uid uint64) {
// Check if there is a message still waiting to be processed (and remove if so)
for i, lane := range a512srv.lanes {
if lane.uid == uid {
if lane.block != nil {
a512srv.lanes[i] = Avx512LaneInfo{} // clear message
a512srv.totalIn--
}
}
}
// Delete entry from hash map
delete(a512srv.digests, uid)
} | go | func (a512srv *Avx512Server) reset(uid uint64) {
// Check if there is a message still waiting to be processed (and remove if so)
for i, lane := range a512srv.lanes {
if lane.uid == uid {
if lane.block != nil {
a512srv.lanes[i] = Avx512LaneInfo{} // clear message
a512srv.totalIn--
}
}
}
// Delete entry from hash map
delete(a512srv.digests, uid)
} | [
"func",
"(",
"a512srv",
"*",
"Avx512Server",
")",
"reset",
"(",
"uid",
"uint64",
")",
"{",
"// Check if there is a message still waiting to be processed (and remove if so)",
"for",
"i",
",",
"lane",
":=",
"range",
"a512srv",
".",
"lanes",
"{",
"if",
"lane",
".",
"uid",
"==",
"uid",
"{",
"if",
"lane",
".",
"block",
"!=",
"nil",
"{",
"a512srv",
".",
"lanes",
"[",
"i",
"]",
"=",
"Avx512LaneInfo",
"{",
"}",
"// clear message",
"\n",
"a512srv",
".",
"totalIn",
"--",
"\n",
"}",
"\n",
"}",
"\n",
"}",
"\n\n",
"// Delete entry from hash map",
"delete",
"(",
"a512srv",
".",
"digests",
",",
"uid",
")",
"\n",
"}"
] | // Do a reset for this calculation | [
"Do",
"a",
"reset",
"for",
"this",
"calculation"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L361-L375 |
5,594 | minio/sha256-simd | sha256blockAvx512_amd64.go | blocks | func (a512srv *Avx512Server) blocks() (err error) {
inputs := [16][]byte{}
for i := range inputs {
inputs[i] = a512srv.lanes[i].block
}
mask := expandMask(genMask(inputs))
outputs := blockAvx512(a512srv.getDigests(), inputs, mask)
a512srv.totalIn = 0
for i := 0; i < len(outputs); i++ {
uid, outputCh := a512srv.lanes[i].uid, a512srv.lanes[i].outputCh
a512srv.digests[uid] = outputs[i]
a512srv.lanes[i] = Avx512LaneInfo{}
if outputCh != nil {
// Send back result
outputCh <- outputs[i]
delete(a512srv.digests, uid) // Delete entry from hashmap
}
}
return
} | go | func (a512srv *Avx512Server) blocks() (err error) {
inputs := [16][]byte{}
for i := range inputs {
inputs[i] = a512srv.lanes[i].block
}
mask := expandMask(genMask(inputs))
outputs := blockAvx512(a512srv.getDigests(), inputs, mask)
a512srv.totalIn = 0
for i := 0; i < len(outputs); i++ {
uid, outputCh := a512srv.lanes[i].uid, a512srv.lanes[i].outputCh
a512srv.digests[uid] = outputs[i]
a512srv.lanes[i] = Avx512LaneInfo{}
if outputCh != nil {
// Send back result
outputCh <- outputs[i]
delete(a512srv.digests, uid) // Delete entry from hashmap
}
}
return
} | [
"func",
"(",
"a512srv",
"*",
"Avx512Server",
")",
"blocks",
"(",
")",
"(",
"err",
"error",
")",
"{",
"inputs",
":=",
"[",
"16",
"]",
"[",
"]",
"byte",
"{",
"}",
"\n",
"for",
"i",
":=",
"range",
"inputs",
"{",
"inputs",
"[",
"i",
"]",
"=",
"a512srv",
".",
"lanes",
"[",
"i",
"]",
".",
"block",
"\n",
"}",
"\n\n",
"mask",
":=",
"expandMask",
"(",
"genMask",
"(",
"inputs",
")",
")",
"\n",
"outputs",
":=",
"blockAvx512",
"(",
"a512srv",
".",
"getDigests",
"(",
")",
",",
"inputs",
",",
"mask",
")",
"\n\n",
"a512srv",
".",
"totalIn",
"=",
"0",
"\n",
"for",
"i",
":=",
"0",
";",
"i",
"<",
"len",
"(",
"outputs",
")",
";",
"i",
"++",
"{",
"uid",
",",
"outputCh",
":=",
"a512srv",
".",
"lanes",
"[",
"i",
"]",
".",
"uid",
",",
"a512srv",
".",
"lanes",
"[",
"i",
"]",
".",
"outputCh",
"\n",
"a512srv",
".",
"digests",
"[",
"uid",
"]",
"=",
"outputs",
"[",
"i",
"]",
"\n",
"a512srv",
".",
"lanes",
"[",
"i",
"]",
"=",
"Avx512LaneInfo",
"{",
"}",
"\n\n",
"if",
"outputCh",
"!=",
"nil",
"{",
"// Send back result",
"outputCh",
"<-",
"outputs",
"[",
"i",
"]",
"\n",
"delete",
"(",
"a512srv",
".",
"digests",
",",
"uid",
")",
"// Delete entry from hashmap",
"\n",
"}",
"\n",
"}",
"\n",
"return",
"\n",
"}"
] | // Invoke assembly and send results back | [
"Invoke",
"assembly",
"and",
"send",
"results",
"back"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L378-L401 |
5,595 | minio/sha256-simd | sha256blockAvx512_amd64.go | Sum | func (a512srv *Avx512Server) Sum(uid uint64, p []byte) [32]byte {
sumCh := make(chan [32]byte)
a512srv.blocksCh <- blockInput{uid: uid, msg: p, final: true, sumCh: sumCh}
return <-sumCh
} | go | func (a512srv *Avx512Server) Sum(uid uint64, p []byte) [32]byte {
sumCh := make(chan [32]byte)
a512srv.blocksCh <- blockInput{uid: uid, msg: p, final: true, sumCh: sumCh}
return <-sumCh
} | [
"func",
"(",
"a512srv",
"*",
"Avx512Server",
")",
"Sum",
"(",
"uid",
"uint64",
",",
"p",
"[",
"]",
"byte",
")",
"[",
"32",
"]",
"byte",
"{",
"sumCh",
":=",
"make",
"(",
"chan",
"[",
"32",
"]",
"byte",
")",
"\n",
"a512srv",
".",
"blocksCh",
"<-",
"blockInput",
"{",
"uid",
":",
"uid",
",",
"msg",
":",
"p",
",",
"final",
":",
"true",
",",
"sumCh",
":",
"sumCh",
"}",
"\n",
"return",
"<-",
"sumCh",
"\n",
"}"
] | // Sum - return sha256 sum in bytes for a given sum id. | [
"Sum",
"-",
"return",
"sha256",
"sum",
"in",
"bytes",
"for",
"a",
"given",
"sum",
"id",
"."
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256blockAvx512_amd64.go#L409-L413 |
5,596 | minio/sha256-simd | sha256.go | Reset | func (d *digest) Reset() {
d.h[0] = init0
d.h[1] = init1
d.h[2] = init2
d.h[3] = init3
d.h[4] = init4
d.h[5] = init5
d.h[6] = init6
d.h[7] = init7
d.nx = 0
d.len = 0
} | go | func (d *digest) Reset() {
d.h[0] = init0
d.h[1] = init1
d.h[2] = init2
d.h[3] = init3
d.h[4] = init4
d.h[5] = init5
d.h[6] = init6
d.h[7] = init7
d.nx = 0
d.len = 0
} | [
"func",
"(",
"d",
"*",
"digest",
")",
"Reset",
"(",
")",
"{",
"d",
".",
"h",
"[",
"0",
"]",
"=",
"init0",
"\n",
"d",
".",
"h",
"[",
"1",
"]",
"=",
"init1",
"\n",
"d",
".",
"h",
"[",
"2",
"]",
"=",
"init2",
"\n",
"d",
".",
"h",
"[",
"3",
"]",
"=",
"init3",
"\n",
"d",
".",
"h",
"[",
"4",
"]",
"=",
"init4",
"\n",
"d",
".",
"h",
"[",
"5",
"]",
"=",
"init5",
"\n",
"d",
".",
"h",
"[",
"6",
"]",
"=",
"init6",
"\n",
"d",
".",
"h",
"[",
"7",
"]",
"=",
"init7",
"\n",
"d",
".",
"nx",
"=",
"0",
"\n",
"d",
".",
"len",
"=",
"0",
"\n",
"}"
] | // Reset digest back to default | [
"Reset",
"digest",
"back",
"to",
"default"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L53-L64 |
5,597 | minio/sha256-simd | sha256.go | New | func New() hash.Hash {
if blockfunc != blockfuncGeneric {
d := new(digest)
d.Reset()
return d
}
// Fallback to the standard golang implementation
// if no features were found.
return sha256.New()
} | go | func New() hash.Hash {
if blockfunc != blockfuncGeneric {
d := new(digest)
d.Reset()
return d
}
// Fallback to the standard golang implementation
// if no features were found.
return sha256.New()
} | [
"func",
"New",
"(",
")",
"hash",
".",
"Hash",
"{",
"if",
"blockfunc",
"!=",
"blockfuncGeneric",
"{",
"d",
":=",
"new",
"(",
"digest",
")",
"\n",
"d",
".",
"Reset",
"(",
")",
"\n",
"return",
"d",
"\n",
"}",
"\n",
"// Fallback to the standard golang implementation",
"// if no features were found.",
"return",
"sha256",
".",
"New",
"(",
")",
"\n",
"}"
] | // New returns a new hash.Hash computing the SHA256 checksum. | [
"New",
"returns",
"a",
"new",
"hash",
".",
"Hash",
"computing",
"the",
"SHA256",
"checksum",
"."
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L102-L111 |
5,598 | minio/sha256-simd | sha256.go | Sum256 | func Sum256(data []byte) (result [Size]byte) {
var d digest
d.Reset()
d.Write(data)
result = d.checkSum()
return
} | go | func Sum256(data []byte) (result [Size]byte) {
var d digest
d.Reset()
d.Write(data)
result = d.checkSum()
return
} | [
"func",
"Sum256",
"(",
"data",
"[",
"]",
"byte",
")",
"(",
"result",
"[",
"Size",
"]",
"byte",
")",
"{",
"var",
"d",
"digest",
"\n",
"d",
".",
"Reset",
"(",
")",
"\n",
"d",
".",
"Write",
"(",
"data",
")",
"\n",
"result",
"=",
"d",
".",
"checkSum",
"(",
")",
"\n",
"return",
"\n",
"}"
] | // Sum256 - single caller sha256 helper | [
"Sum256",
"-",
"single",
"caller",
"sha256",
"helper"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L114-L120 |
5,599 | minio/sha256-simd | sha256.go | Sum | func (d *digest) Sum(in []byte) []byte {
// Make a copy of d0 so that caller can keep writing and summing.
d0 := *d
hash := d0.checkSum()
return append(in, hash[:]...)
} | go | func (d *digest) Sum(in []byte) []byte {
// Make a copy of d0 so that caller can keep writing and summing.
d0 := *d
hash := d0.checkSum()
return append(in, hash[:]...)
} | [
"func",
"(",
"d",
"*",
"digest",
")",
"Sum",
"(",
"in",
"[",
"]",
"byte",
")",
"[",
"]",
"byte",
"{",
"// Make a copy of d0 so that caller can keep writing and summing.",
"d0",
":=",
"*",
"d",
"\n",
"hash",
":=",
"d0",
".",
"checkSum",
"(",
")",
"\n",
"return",
"append",
"(",
"in",
",",
"hash",
"[",
":",
"]",
"...",
")",
"\n",
"}"
] | // Return sha256 sum in bytes | [
"Return",
"sha256",
"sum",
"in",
"bytes"
] | 05b4dd3047e5d6e86cb4e0477164b850cd896261 | https://github.com/minio/sha256-simd/blob/05b4dd3047e5d6e86cb4e0477164b850cd896261/sha256.go#L153-L158 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.