plugin/metrics: Switch to using promhttp instead of deprecated Handler (#1312)
prometheus.Handler is deprecated according to the godoc for the package so instead we're using promhttp. Additionally, we are exposing the Registry that metrics is using so other plugins that are not inside of coredns can read the registry. Otherwise, if we kept using the Default one, there's no way to access that from outside of the coredns repo since it is vendored.
This commit is contained in:
parent
1919913c98
commit
671d170619
6728 changed files with 1994787 additions and 16 deletions
4
Gopkg.lock
generated
4
Gopkg.lock
generated
|
@ -279,7 +279,7 @@
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
name = "github.com/prometheus/client_golang"
|
name = "github.com/prometheus/client_golang"
|
||||||
packages = ["prometheus"]
|
packages = ["prometheus","prometheus/promhttp"]
|
||||||
revision = "c5b7fccd204277076155f10851dad72b76a49317"
|
revision = "c5b7fccd204277076155f10851dad72b76a49317"
|
||||||
version = "v0.8.0"
|
version = "v0.8.0"
|
||||||
|
|
||||||
|
@ -380,6 +380,6 @@
|
||||||
[solve-meta]
|
[solve-meta]
|
||||||
analyzer-name = "dep"
|
analyzer-name = "dep"
|
||||||
analyzer-version = 1
|
analyzer-version = 1
|
||||||
inputs-digest = "be9300a30414c93aa44756868a7906a0a295b0910a662880741bcfac58b7b679"
|
inputs-digest = "802b3c477e5c45a1fc7283cfc3e9ed03c4dffb4bfe9fb406d252a8d8194975d1"
|
||||||
solver-name = "gps-cdcl"
|
solver-name = "gps-cdcl"
|
||||||
solver-version = 1
|
solver-version = 1
|
||||||
|
|
|
@ -5,29 +5,20 @@ import (
|
||||||
"log"
|
"log"
|
||||||
"net"
|
"net"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"os"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/coredns/coredns/plugin"
|
"github.com/coredns/coredns/plugin"
|
||||||
"github.com/coredns/coredns/plugin/metrics/vars"
|
"github.com/coredns/coredns/plugin/metrics/vars"
|
||||||
|
|
||||||
"github.com/prometheus/client_golang/prometheus"
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
|
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
|
||||||
prometheus.MustRegister(vars.RequestCount)
|
|
||||||
prometheus.MustRegister(vars.RequestDuration)
|
|
||||||
prometheus.MustRegister(vars.RequestSize)
|
|
||||||
prometheus.MustRegister(vars.RequestDo)
|
|
||||||
prometheus.MustRegister(vars.RequestType)
|
|
||||||
|
|
||||||
prometheus.MustRegister(vars.ResponseSize)
|
|
||||||
prometheus.MustRegister(vars.ResponseRcode)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Metrics holds the prometheus configuration. The metrics' path is fixed to be /metrics
|
// Metrics holds the prometheus configuration. The metrics' path is fixed to be /metrics
|
||||||
type Metrics struct {
|
type Metrics struct {
|
||||||
Next plugin.Handler
|
Next plugin.Handler
|
||||||
Addr string
|
Addr string
|
||||||
|
Reg *prometheus.Registry
|
||||||
ln net.Listener
|
ln net.Listener
|
||||||
mux *http.ServeMux
|
mux *http.ServeMux
|
||||||
|
|
||||||
|
@ -38,7 +29,24 @@ type Metrics struct {
|
||||||
|
|
||||||
// New returns a new instance of Metrics with the given address
|
// New returns a new instance of Metrics with the given address
|
||||||
func New(addr string) *Metrics {
|
func New(addr string) *Metrics {
|
||||||
return &Metrics{Addr: addr, zoneMap: make(map[string]bool)}
|
met := &Metrics{
|
||||||
|
Addr: addr,
|
||||||
|
Reg: prometheus.NewRegistry(),
|
||||||
|
zoneMap: make(map[string]bool),
|
||||||
|
}
|
||||||
|
// Add the default collectors
|
||||||
|
met.Reg.MustRegister(prometheus.NewGoCollector())
|
||||||
|
met.Reg.MustRegister(prometheus.NewProcessCollector(os.Getpid(), ""))
|
||||||
|
|
||||||
|
// Add all of our collectors
|
||||||
|
met.Reg.MustRegister(vars.RequestCount)
|
||||||
|
met.Reg.MustRegister(vars.RequestDuration)
|
||||||
|
met.Reg.MustRegister(vars.RequestSize)
|
||||||
|
met.Reg.MustRegister(vars.RequestDo)
|
||||||
|
met.Reg.MustRegister(vars.RequestType)
|
||||||
|
met.Reg.MustRegister(vars.ResponseSize)
|
||||||
|
met.Reg.MustRegister(vars.ResponseRcode)
|
||||||
|
return met
|
||||||
}
|
}
|
||||||
|
|
||||||
// AddZone adds zone z to m.
|
// AddZone adds zone z to m.
|
||||||
|
@ -77,7 +85,7 @@ func (m *Metrics) OnStartup() error {
|
||||||
ListenAddr = m.ln.Addr().String()
|
ListenAddr = m.ln.Addr().String()
|
||||||
|
|
||||||
m.mux = http.NewServeMux()
|
m.mux = http.NewServeMux()
|
||||||
m.mux.Handle("/metrics", prometheus.Handler())
|
m.mux.Handle("/metrics", promhttp.HandlerFor(m.Reg, promhttp.HandlerOpts{}))
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
http.Serve(m.ln, m.mux)
|
http.Serve(m.ln, m.mux)
|
||||||
|
|
9
vendor/github.com/PuerkitoBio/purell/benchmarks/v0.1.0
generated
vendored
Normal file
9
vendor/github.com/PuerkitoBio/purell/benchmarks/v0.1.0
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
PASS
|
||||||
|
BenchmarkSafe 500000 6131 ns/op
|
||||||
|
BenchmarkUsuallySafe 200000 7864 ns/op
|
||||||
|
BenchmarkUnsafe 100000 28560 ns/op
|
||||||
|
BenchmarkAllDWORD 50000 38722 ns/op
|
||||||
|
BenchmarkAllOctal 50000 40941 ns/op
|
||||||
|
BenchmarkAllHex 50000 44063 ns/op
|
||||||
|
BenchmarkAllCombined 50000 33613 ns/op
|
||||||
|
ok github.com/PuerkitoBio/purell 17.404s
|
31
vendor/github.com/Shopify/sarama/.github/CONTRIBUTING.md
generated
vendored
Normal file
31
vendor/github.com/Shopify/sarama/.github/CONTRIBUTING.md
generated
vendored
Normal file
|
@ -0,0 +1,31 @@
|
||||||
|
# Contributing
|
||||||
|
|
||||||
|
Contributions are always welcome, both reporting issues and submitting pull requests!
|
||||||
|
|
||||||
|
### Reporting issues
|
||||||
|
|
||||||
|
Please make sure to include any potentially useful information in the issue, so we can pinpoint the issue faster without going back and forth.
|
||||||
|
|
||||||
|
- What SHA of Sarama are you running? If this is not the latest SHA on the master branch, please try if the problem persists with the latest version.
|
||||||
|
- You can set `sarama.Logger` to a [log.Logger](http://golang.org/pkg/log/#Logger) instance to capture debug output. Please include it in your issue description.
|
||||||
|
- Also look at the logs of the Kafka broker you are connected to. If you see anything out of the ordinary, please include it.
|
||||||
|
|
||||||
|
Also, please include the following information about your environment, so we can help you faster:
|
||||||
|
|
||||||
|
- What version of Kafka are you using?
|
||||||
|
- What version of Go are you using?
|
||||||
|
- What are the values of your Producer/Consumer/Client configuration?
|
||||||
|
|
||||||
|
|
||||||
|
### Submitting pull requests
|
||||||
|
|
||||||
|
We will gladly accept bug fixes, or additions to this library. Please fork this library, commit & push your changes, and open a pull request. Because this library is in production use by many people and applications, we code review all additions. To make the review process go as smooth as possible, please consider the following.
|
||||||
|
|
||||||
|
- If you plan to work on something major, please open an issue to discuss the design first.
|
||||||
|
- Don't break backwards compatibility. If you really have to, open an issue to discuss this first.
|
||||||
|
- Make sure to use the `go fmt` command to format your code according to the standards. Even better, set up your editor to do this for you when saving.
|
||||||
|
- Run [go vet](https://godoc.org/golang.org/x/tools/cmd/vet) to detect any suspicious constructs in your code that could be bugs.
|
||||||
|
- Explicitly handle all error return values. If you really want to ignore an error value, you can assign it to `_`.You can use [errcheck](https://github.com/kisielk/errcheck) to verify whether you have handled all errors.
|
||||||
|
- You may also want to run [golint](https://github.com/golang/lint) as well to detect style problems.
|
||||||
|
- Add tests that cover the changes you made. Make sure to run `go test` with the `-race` argument to test for race conditions.
|
||||||
|
- Make sure your code is supported by all the Go versions we support. You can rely on [Travis CI](https://travis-ci.org/Shopify/sarama) for testing older Go versions
|
20
vendor/github.com/Shopify/sarama/.github/ISSUE_TEMPLATE.md
generated
vendored
Normal file
20
vendor/github.com/Shopify/sarama/.github/ISSUE_TEMPLATE.md
generated
vendored
Normal file
|
@ -0,0 +1,20 @@
|
||||||
|
##### Versions
|
||||||
|
|
||||||
|
*Please specify real version numbers or git SHAs, not just "Latest" since that changes fairly regularly.*
|
||||||
|
Sarama Version:
|
||||||
|
Kafka Version:
|
||||||
|
Go Version:
|
||||||
|
|
||||||
|
##### Configuration
|
||||||
|
|
||||||
|
What configuration values are you using for Sarama and Kafka?
|
||||||
|
|
||||||
|
##### Logs
|
||||||
|
|
||||||
|
When filing an issue please provide logs from Sarama and Kafka if at all
|
||||||
|
possible. You can set `sarama.Logger` to a `log.Logger` to capture Sarama debug
|
||||||
|
output.
|
||||||
|
|
||||||
|
##### Problem Description
|
||||||
|
|
||||||
|
|
9
vendor/github.com/Shopify/sarama/examples/README.md
generated
vendored
Normal file
9
vendor/github.com/Shopify/sarama/examples/README.md
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
# Sarama examples
|
||||||
|
|
||||||
|
This folder contains example applications to demonstrate the use of Sarama. For code snippet examples on how to use the different types in Sarama, see [Sarama's API documentation on godoc.org](https://godoc.org/github.com/Shopify/sarama)
|
||||||
|
|
||||||
|
In these examples, we use `github.com/Shopify/sarama` as import path. We do this to ensure all the examples are up to date with the latest changes in Sarama. For your own applications, you may want to use `gopkg.in/Shopify/sarama.v1` to lock into a stable API version.
|
||||||
|
|
||||||
|
#### HTTP server
|
||||||
|
|
||||||
|
[http_server](./http_server) is a simple HTTP server uses both the sync producer to produce data as part of the request handling cycle, as well as the async producer to maintain an access log. It also uses the [mocks subpackage](https://godoc.org/github.com/Shopify/sarama/mocks) to test both.
|
2
vendor/github.com/Shopify/sarama/examples/http_server/.gitignore
generated
vendored
Normal file
2
vendor/github.com/Shopify/sarama/examples/http_server/.gitignore
generated
vendored
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
http_server
|
||||||
|
http_server.test
|
7
vendor/github.com/Shopify/sarama/examples/http_server/README.md
generated
vendored
Normal file
7
vendor/github.com/Shopify/sarama/examples/http_server/README.md
generated
vendored
Normal file
|
@ -0,0 +1,7 @@
|
||||||
|
# HTTP server example
|
||||||
|
|
||||||
|
This HTTP server example shows you how to use the AsyncProducer and SyncProducer, and how to test them using mocks. The server simply sends the data of the HTTP request's query string to Kafka, and send a 200 result if that succeeds. For every request, it will send an access log entry to Kafka as well in the background.
|
||||||
|
|
||||||
|
If you need to know whether a message was successfully sent to the Kafka cluster before you can send your HTTP response, using the `SyncProducer` is probably the simplest way to achieve this. If you don't care, e.g. for the access log, using the `AsyncProducer` will let you fire and forget. You can send the HTTP response, while the message is being produced in the background.
|
||||||
|
|
||||||
|
One important thing to note is that both the `SyncProducer` and `AsyncProducer` are **thread-safe**. Go's `http.Server` handles requests concurrently in different goroutines, but you can use a single producer safely. This will actually achieve efficiency gains as the producer will be able to batch messages from concurrent requests together.
|
247
vendor/github.com/Shopify/sarama/examples/http_server/http_server.go
generated
vendored
Normal file
247
vendor/github.com/Shopify/sarama/examples/http_server/http_server.go
generated
vendored
Normal file
|
@ -0,0 +1,247 @@
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
|
||||||
|
"crypto/tls"
|
||||||
|
"crypto/x509"
|
||||||
|
"encoding/json"
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"log"
|
||||||
|
"net/http"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
addr = flag.String("addr", ":8080", "The address to bind to")
|
||||||
|
brokers = flag.String("brokers", os.Getenv("KAFKA_PEERS"), "The Kafka brokers to connect to, as a comma separated list")
|
||||||
|
verbose = flag.Bool("verbose", false, "Turn on Sarama logging")
|
||||||
|
certFile = flag.String("certificate", "", "The optional certificate file for client authentication")
|
||||||
|
keyFile = flag.String("key", "", "The optional key file for client authentication")
|
||||||
|
caFile = flag.String("ca", "", "The optional certificate authority file for TLS client authentication")
|
||||||
|
verifySsl = flag.Bool("verify", false, "Optional verify ssl certificates chain")
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
flag.Parse()
|
||||||
|
|
||||||
|
if *verbose {
|
||||||
|
sarama.Logger = log.New(os.Stdout, "[sarama] ", log.LstdFlags)
|
||||||
|
}
|
||||||
|
|
||||||
|
if *brokers == "" {
|
||||||
|
flag.PrintDefaults()
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
brokerList := strings.Split(*brokers, ",")
|
||||||
|
log.Printf("Kafka brokers: %s", strings.Join(brokerList, ", "))
|
||||||
|
|
||||||
|
server := &Server{
|
||||||
|
DataCollector: newDataCollector(brokerList),
|
||||||
|
AccessLogProducer: newAccessLogProducer(brokerList),
|
||||||
|
}
|
||||||
|
defer func() {
|
||||||
|
if err := server.Close(); err != nil {
|
||||||
|
log.Println("Failed to close server", err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
log.Fatal(server.Run(*addr))
|
||||||
|
}
|
||||||
|
|
||||||
|
func createTlsConfiguration() (t *tls.Config) {
|
||||||
|
if *certFile != "" && *keyFile != "" && *caFile != "" {
|
||||||
|
cert, err := tls.LoadX509KeyPair(*certFile, *keyFile)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
caCert, err := ioutil.ReadFile(*caFile)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
caCertPool := x509.NewCertPool()
|
||||||
|
caCertPool.AppendCertsFromPEM(caCert)
|
||||||
|
|
||||||
|
t = &tls.Config{
|
||||||
|
Certificates: []tls.Certificate{cert},
|
||||||
|
RootCAs: caCertPool,
|
||||||
|
InsecureSkipVerify: *verifySsl,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// will be nil by default if nothing is provided
|
||||||
|
return t
|
||||||
|
}
|
||||||
|
|
||||||
|
type Server struct {
|
||||||
|
DataCollector sarama.SyncProducer
|
||||||
|
AccessLogProducer sarama.AsyncProducer
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) Close() error {
|
||||||
|
if err := s.DataCollector.Close(); err != nil {
|
||||||
|
log.Println("Failed to shut down data collector cleanly", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := s.AccessLogProducer.Close(); err != nil {
|
||||||
|
log.Println("Failed to shut down access log producer cleanly", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) Handler() http.Handler {
|
||||||
|
return s.withAccessLog(s.collectQueryStringData())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) Run(addr string) error {
|
||||||
|
httpServer := &http.Server{
|
||||||
|
Addr: addr,
|
||||||
|
Handler: s.Handler(),
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Printf("Listening for requests on %s...\n", addr)
|
||||||
|
return httpServer.ListenAndServe()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) collectQueryStringData() http.Handler {
|
||||||
|
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
if r.URL.Path != "/" {
|
||||||
|
http.NotFound(w, r)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// We are not setting a message key, which means that all messages will
|
||||||
|
// be distributed randomly over the different partitions.
|
||||||
|
partition, offset, err := s.DataCollector.SendMessage(&sarama.ProducerMessage{
|
||||||
|
Topic: "important",
|
||||||
|
Value: sarama.StringEncoder(r.URL.RawQuery),
|
||||||
|
})
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
w.WriteHeader(http.StatusInternalServerError)
|
||||||
|
fmt.Fprintf(w, "Failed to store your data:, %s", err)
|
||||||
|
} else {
|
||||||
|
// The tuple (topic, partition, offset) can be used as a unique identifier
|
||||||
|
// for a message in a Kafka cluster.
|
||||||
|
fmt.Fprintf(w, "Your data is stored with unique identifier important/%d/%d", partition, offset)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
type accessLogEntry struct {
|
||||||
|
Method string `json:"method"`
|
||||||
|
Host string `json:"host"`
|
||||||
|
Path string `json:"path"`
|
||||||
|
IP string `json:"ip"`
|
||||||
|
ResponseTime float64 `json:"response_time"`
|
||||||
|
|
||||||
|
encoded []byte
|
||||||
|
err error
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ale *accessLogEntry) ensureEncoded() {
|
||||||
|
if ale.encoded == nil && ale.err == nil {
|
||||||
|
ale.encoded, ale.err = json.Marshal(ale)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ale *accessLogEntry) Length() int {
|
||||||
|
ale.ensureEncoded()
|
||||||
|
return len(ale.encoded)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ale *accessLogEntry) Encode() ([]byte, error) {
|
||||||
|
ale.ensureEncoded()
|
||||||
|
return ale.encoded, ale.err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) withAccessLog(next http.Handler) http.Handler {
|
||||||
|
|
||||||
|
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
started := time.Now()
|
||||||
|
|
||||||
|
next.ServeHTTP(w, r)
|
||||||
|
|
||||||
|
entry := &accessLogEntry{
|
||||||
|
Method: r.Method,
|
||||||
|
Host: r.Host,
|
||||||
|
Path: r.RequestURI,
|
||||||
|
IP: r.RemoteAddr,
|
||||||
|
ResponseTime: float64(time.Since(started)) / float64(time.Second),
|
||||||
|
}
|
||||||
|
|
||||||
|
// We will use the client's IP address as key. This will cause
|
||||||
|
// all the access log entries of the same IP address to end up
|
||||||
|
// on the same partition.
|
||||||
|
s.AccessLogProducer.Input() <- &sarama.ProducerMessage{
|
||||||
|
Topic: "access_log",
|
||||||
|
Key: sarama.StringEncoder(r.RemoteAddr),
|
||||||
|
Value: entry,
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func newDataCollector(brokerList []string) sarama.SyncProducer {
|
||||||
|
|
||||||
|
// For the data collector, we are looking for strong consistency semantics.
|
||||||
|
// Because we don't change the flush settings, sarama will try to produce messages
|
||||||
|
// as fast as possible to keep latency low.
|
||||||
|
config := sarama.NewConfig()
|
||||||
|
config.Producer.RequiredAcks = sarama.WaitForAll // Wait for all in-sync replicas to ack the message
|
||||||
|
config.Producer.Retry.Max = 10 // Retry up to 10 times to produce the message
|
||||||
|
config.Producer.Return.Successes = true
|
||||||
|
tlsConfig := createTlsConfiguration()
|
||||||
|
if tlsConfig != nil {
|
||||||
|
config.Net.TLS.Config = tlsConfig
|
||||||
|
config.Net.TLS.Enable = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// On the broker side, you may want to change the following settings to get
|
||||||
|
// stronger consistency guarantees:
|
||||||
|
// - For your broker, set `unclean.leader.election.enable` to false
|
||||||
|
// - For the topic, you could increase `min.insync.replicas`.
|
||||||
|
|
||||||
|
producer, err := sarama.NewSyncProducer(brokerList, config)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatalln("Failed to start Sarama producer:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return producer
|
||||||
|
}
|
||||||
|
|
||||||
|
func newAccessLogProducer(brokerList []string) sarama.AsyncProducer {
|
||||||
|
|
||||||
|
// For the access log, we are looking for AP semantics, with high throughput.
|
||||||
|
// By creating batches of compressed messages, we reduce network I/O at a cost of more latency.
|
||||||
|
config := sarama.NewConfig()
|
||||||
|
tlsConfig := createTlsConfiguration()
|
||||||
|
if tlsConfig != nil {
|
||||||
|
config.Net.TLS.Enable = true
|
||||||
|
config.Net.TLS.Config = tlsConfig
|
||||||
|
}
|
||||||
|
config.Producer.RequiredAcks = sarama.WaitForLocal // Only wait for the leader to ack
|
||||||
|
config.Producer.Compression = sarama.CompressionSnappy // Compress messages
|
||||||
|
config.Producer.Flush.Frequency = 500 * time.Millisecond // Flush batches every 500ms
|
||||||
|
|
||||||
|
producer, err := sarama.NewAsyncProducer(brokerList, config)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatalln("Failed to start Sarama producer:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// We will just log to STDOUT if we're not able to produce messages.
|
||||||
|
// Note: messages will only be returned here after all retry attempts are exhausted.
|
||||||
|
go func() {
|
||||||
|
for err := range producer.Errors() {
|
||||||
|
log.Println("Failed to write access log entry:", err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return producer
|
||||||
|
}
|
109
vendor/github.com/Shopify/sarama/examples/http_server/http_server_test.go
generated
vendored
Normal file
109
vendor/github.com/Shopify/sarama/examples/http_server/http_server_test.go
generated
vendored
Normal file
|
@ -0,0 +1,109 @@
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io"
|
||||||
|
"net/http"
|
||||||
|
"net/http/httptest"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
"github.com/Shopify/sarama/mocks"
|
||||||
|
)
|
||||||
|
|
||||||
|
// In normal operation, we expect one access log entry,
|
||||||
|
// and one data collector entry. Let's assume both will succeed.
|
||||||
|
// We should return a HTTP 200 status.
|
||||||
|
func TestCollectSuccessfully(t *testing.T) {
|
||||||
|
dataCollectorMock := mocks.NewSyncProducer(t, nil)
|
||||||
|
dataCollectorMock.ExpectSendMessageAndSucceed()
|
||||||
|
|
||||||
|
accessLogProducerMock := mocks.NewAsyncProducer(t, nil)
|
||||||
|
accessLogProducerMock.ExpectInputAndSucceed()
|
||||||
|
|
||||||
|
// Now, use dependency injection to use the mocks.
|
||||||
|
s := &Server{
|
||||||
|
DataCollector: dataCollectorMock,
|
||||||
|
AccessLogProducer: accessLogProducerMock,
|
||||||
|
}
|
||||||
|
|
||||||
|
// The Server's Close call is important; it will call Close on
|
||||||
|
// the two mock producers, which will then validate whether all
|
||||||
|
// expectations are resolved.
|
||||||
|
defer safeClose(t, s)
|
||||||
|
|
||||||
|
req, err := http.NewRequest("GET", "http://example.com/?data", nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
res := httptest.NewRecorder()
|
||||||
|
s.Handler().ServeHTTP(res, req)
|
||||||
|
|
||||||
|
if res.Code != 200 {
|
||||||
|
t.Errorf("Expected HTTP status 200, found %d", res.Code)
|
||||||
|
}
|
||||||
|
|
||||||
|
if string(res.Body.Bytes()) != "Your data is stored with unique identifier important/0/1" {
|
||||||
|
t.Error("Unexpected response body", res.Body)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now, let's see if we handle the case of not being able to produce
|
||||||
|
// to the data collector properly. In this case we should return a 500 status.
|
||||||
|
func TestCollectionFailure(t *testing.T) {
|
||||||
|
dataCollectorMock := mocks.NewSyncProducer(t, nil)
|
||||||
|
dataCollectorMock.ExpectSendMessageAndFail(sarama.ErrRequestTimedOut)
|
||||||
|
|
||||||
|
accessLogProducerMock := mocks.NewAsyncProducer(t, nil)
|
||||||
|
accessLogProducerMock.ExpectInputAndSucceed()
|
||||||
|
|
||||||
|
s := &Server{
|
||||||
|
DataCollector: dataCollectorMock,
|
||||||
|
AccessLogProducer: accessLogProducerMock,
|
||||||
|
}
|
||||||
|
defer safeClose(t, s)
|
||||||
|
|
||||||
|
req, err := http.NewRequest("GET", "http://example.com/?data", nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
res := httptest.NewRecorder()
|
||||||
|
s.Handler().ServeHTTP(res, req)
|
||||||
|
|
||||||
|
if res.Code != 500 {
|
||||||
|
t.Errorf("Expected HTTP status 500, found %d", res.Code)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// We don't expect any data collector calls because the path is wrong,
|
||||||
|
// so we are not setting any expectations on the dataCollectorMock. It
|
||||||
|
// will still generate an access log entry though.
|
||||||
|
func TestWrongPath(t *testing.T) {
|
||||||
|
dataCollectorMock := mocks.NewSyncProducer(t, nil)
|
||||||
|
|
||||||
|
accessLogProducerMock := mocks.NewAsyncProducer(t, nil)
|
||||||
|
accessLogProducerMock.ExpectInputAndSucceed()
|
||||||
|
|
||||||
|
s := &Server{
|
||||||
|
DataCollector: dataCollectorMock,
|
||||||
|
AccessLogProducer: accessLogProducerMock,
|
||||||
|
}
|
||||||
|
defer safeClose(t, s)
|
||||||
|
|
||||||
|
req, err := http.NewRequest("GET", "http://example.com/wrong?data", nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
res := httptest.NewRecorder()
|
||||||
|
|
||||||
|
s.Handler().ServeHTTP(res, req)
|
||||||
|
|
||||||
|
if res.Code != 404 {
|
||||||
|
t.Errorf("Expected HTTP status 404, found %d", res.Code)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func safeClose(t *testing.T, o io.Closer) {
|
||||||
|
if err := o.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
13
vendor/github.com/Shopify/sarama/mocks/README.md
generated
vendored
Normal file
13
vendor/github.com/Shopify/sarama/mocks/README.md
generated
vendored
Normal file
|
@ -0,0 +1,13 @@
|
||||||
|
# sarama/mocks
|
||||||
|
|
||||||
|
The `mocks` subpackage includes mock implementations that implement the interfaces of the major sarama types.
|
||||||
|
You can use them to test your sarama applications using dependency injection.
|
||||||
|
|
||||||
|
The following mock objects are available:
|
||||||
|
|
||||||
|
- [Consumer](https://godoc.org/github.com/Shopify/sarama/mocks#Consumer), which will create [PartitionConsumer](https://godoc.org/github.com/Shopify/sarama/mocks#PartitionConsumer) mocks.
|
||||||
|
- [AsyncProducer](https://godoc.org/github.com/Shopify/sarama/mocks#AsyncProducer)
|
||||||
|
- [SyncProducer](https://godoc.org/github.com/Shopify/sarama/mocks#SyncProducer)
|
||||||
|
|
||||||
|
The mocks allow you to set expectations on them. When you close the mocks, the expectations will be verified,
|
||||||
|
and the results will be reported to the `*testing.T` object you provided when creating the mock.
|
174
vendor/github.com/Shopify/sarama/mocks/async_producer.go
generated
vendored
Normal file
174
vendor/github.com/Shopify/sarama/mocks/async_producer.go
generated
vendored
Normal file
|
@ -0,0 +1,174 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
// AsyncProducer implements sarama's Producer interface for testing purposes.
|
||||||
|
// Before you can send messages to it's Input channel, you have to set expectations
|
||||||
|
// so it knows how to handle the input; it returns an error if the number of messages
|
||||||
|
// received is bigger then the number of expectations set. You can also set a
|
||||||
|
// function in each expectation so that the message value is checked by this function
|
||||||
|
// and an error is returned if the match fails.
|
||||||
|
type AsyncProducer struct {
|
||||||
|
l sync.Mutex
|
||||||
|
t ErrorReporter
|
||||||
|
expectations []*producerExpectation
|
||||||
|
closed chan struct{}
|
||||||
|
input chan *sarama.ProducerMessage
|
||||||
|
successes chan *sarama.ProducerMessage
|
||||||
|
errors chan *sarama.ProducerError
|
||||||
|
lastOffset int64
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewAsyncProducer instantiates a new Producer mock. The t argument should
|
||||||
|
// be the *testing.T instance of your test method. An error will be written to it if
|
||||||
|
// an expectation is violated. The config argument is used to determine whether it
|
||||||
|
// should ack successes on the Successes channel.
|
||||||
|
func NewAsyncProducer(t ErrorReporter, config *sarama.Config) *AsyncProducer {
|
||||||
|
if config == nil {
|
||||||
|
config = sarama.NewConfig()
|
||||||
|
}
|
||||||
|
mp := &AsyncProducer{
|
||||||
|
t: t,
|
||||||
|
closed: make(chan struct{}, 0),
|
||||||
|
expectations: make([]*producerExpectation, 0),
|
||||||
|
input: make(chan *sarama.ProducerMessage, config.ChannelBufferSize),
|
||||||
|
successes: make(chan *sarama.ProducerMessage, config.ChannelBufferSize),
|
||||||
|
errors: make(chan *sarama.ProducerError, config.ChannelBufferSize),
|
||||||
|
}
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
defer func() {
|
||||||
|
close(mp.successes)
|
||||||
|
close(mp.errors)
|
||||||
|
}()
|
||||||
|
|
||||||
|
for msg := range mp.input {
|
||||||
|
mp.l.Lock()
|
||||||
|
if mp.expectations == nil || len(mp.expectations) == 0 {
|
||||||
|
mp.expectations = nil
|
||||||
|
mp.t.Errorf("No more expectation set on this mock producer to handle the input message.")
|
||||||
|
} else {
|
||||||
|
expectation := mp.expectations[0]
|
||||||
|
mp.expectations = mp.expectations[1:]
|
||||||
|
if expectation.CheckFunction != nil {
|
||||||
|
if val, err := msg.Value.Encode(); err != nil {
|
||||||
|
mp.t.Errorf("Input message encoding failed: %s", err.Error())
|
||||||
|
mp.errors <- &sarama.ProducerError{Err: err, Msg: msg}
|
||||||
|
} else {
|
||||||
|
err = expectation.CheckFunction(val)
|
||||||
|
if err != nil {
|
||||||
|
mp.t.Errorf("Check function returned an error: %s", err.Error())
|
||||||
|
mp.errors <- &sarama.ProducerError{Err: err, Msg: msg}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if expectation.Result == errProduceSuccess {
|
||||||
|
mp.lastOffset++
|
||||||
|
if config.Producer.Return.Successes {
|
||||||
|
msg.Offset = mp.lastOffset
|
||||||
|
mp.successes <- msg
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if config.Producer.Return.Errors {
|
||||||
|
mp.errors <- &sarama.ProducerError{Err: expectation.Result, Msg: msg}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
mp.l.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
mp.l.Lock()
|
||||||
|
if len(mp.expectations) > 0 {
|
||||||
|
mp.t.Errorf("Expected to exhaust all expectations, but %d are left.", len(mp.expectations))
|
||||||
|
}
|
||||||
|
mp.l.Unlock()
|
||||||
|
|
||||||
|
close(mp.closed)
|
||||||
|
}()
|
||||||
|
|
||||||
|
return mp
|
||||||
|
}
|
||||||
|
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
// Implement Producer interface
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// AsyncClose corresponds with the AsyncClose method of sarama's Producer implementation.
|
||||||
|
// By closing a mock producer, you also tell it that no more input will be provided, so it will
|
||||||
|
// write an error to the test state if there's any remaining expectations.
|
||||||
|
func (mp *AsyncProducer) AsyncClose() {
|
||||||
|
close(mp.input)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close corresponds with the Close method of sarama's Producer implementation.
|
||||||
|
// By closing a mock producer, you also tell it that no more input will be provided, so it will
|
||||||
|
// write an error to the test state if there's any remaining expectations.
|
||||||
|
func (mp *AsyncProducer) Close() error {
|
||||||
|
mp.AsyncClose()
|
||||||
|
<-mp.closed
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Input corresponds with the Input method of sarama's Producer implementation.
|
||||||
|
// You have to set expectations on the mock producer before writing messages to the Input
|
||||||
|
// channel, so it knows how to handle them. If there is no more remaining expectations and
|
||||||
|
// a messages is written to the Input channel, the mock producer will write an error to the test
|
||||||
|
// state object.
|
||||||
|
func (mp *AsyncProducer) Input() chan<- *sarama.ProducerMessage {
|
||||||
|
return mp.input
|
||||||
|
}
|
||||||
|
|
||||||
|
// Successes corresponds with the Successes method of sarama's Producer implementation.
|
||||||
|
func (mp *AsyncProducer) Successes() <-chan *sarama.ProducerMessage {
|
||||||
|
return mp.successes
|
||||||
|
}
|
||||||
|
|
||||||
|
// Errors corresponds with the Errors method of sarama's Producer implementation.
|
||||||
|
func (mp *AsyncProducer) Errors() <-chan *sarama.ProducerError {
|
||||||
|
return mp.errors
|
||||||
|
}
|
||||||
|
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
// Setting expectations
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// ExpectInputWithCheckerFunctionAndSucceed sets an expectation on the mock producer that a message
|
||||||
|
// will be provided on the input channel. The mock producer will call the given function to check
|
||||||
|
// the message value. If an error is returned it will be made available on the Errors channel
|
||||||
|
// otherwise the mock will handle the message as if it produced successfully, i.e. it will make
|
||||||
|
// it available on the Successes channel if the Producer.Return.Successes setting is set to true.
|
||||||
|
func (mp *AsyncProducer) ExpectInputWithCheckerFunctionAndSucceed(cf ValueChecker) {
|
||||||
|
mp.l.Lock()
|
||||||
|
defer mp.l.Unlock()
|
||||||
|
mp.expectations = append(mp.expectations, &producerExpectation{Result: errProduceSuccess, CheckFunction: cf})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectInputWithCheckerFunctionAndFail sets an expectation on the mock producer that a message
|
||||||
|
// will be provided on the input channel. The mock producer will first call the given function to
|
||||||
|
// check the message value. If an error is returned it will be made available on the Errors channel
|
||||||
|
// otherwise the mock will handle the message as if it failed to produce successfully. This means
|
||||||
|
// it will make a ProducerError available on the Errors channel.
|
||||||
|
func (mp *AsyncProducer) ExpectInputWithCheckerFunctionAndFail(cf ValueChecker, err error) {
|
||||||
|
mp.l.Lock()
|
||||||
|
defer mp.l.Unlock()
|
||||||
|
mp.expectations = append(mp.expectations, &producerExpectation{Result: err, CheckFunction: cf})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectInputAndSucceed sets an expectation on the mock producer that a message will be provided
|
||||||
|
// on the input channel. The mock producer will handle the message as if it is produced successfully,
|
||||||
|
// i.e. it will make it available on the Successes channel if the Producer.Return.Successes setting
|
||||||
|
// is set to true.
|
||||||
|
func (mp *AsyncProducer) ExpectInputAndSucceed() {
|
||||||
|
mp.ExpectInputWithCheckerFunctionAndSucceed(nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectInputAndFail sets an expectation on the mock producer that a message will be provided
|
||||||
|
// on the input channel. The mock producer will handle the message as if it failed to produce
|
||||||
|
// successfully. This means it will make a ProducerError available on the Errors channel.
|
||||||
|
func (mp *AsyncProducer) ExpectInputAndFail(err error) {
|
||||||
|
mp.ExpectInputWithCheckerFunctionAndFail(nil, err)
|
||||||
|
}
|
132
vendor/github.com/Shopify/sarama/mocks/async_producer_test.go
generated
vendored
Normal file
132
vendor/github.com/Shopify/sarama/mocks/async_producer_test.go
generated
vendored
Normal file
|
@ -0,0 +1,132 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"regexp"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
func generateRegexpChecker(re string) func([]byte) error {
|
||||||
|
return func(val []byte) error {
|
||||||
|
matched, err := regexp.MatchString(re, string(val))
|
||||||
|
if err != nil {
|
||||||
|
return errors.New("Error while trying to match the input message with the expected pattern: " + err.Error())
|
||||||
|
}
|
||||||
|
if !matched {
|
||||||
|
return fmt.Errorf("No match between input value \"%s\" and expected pattern \"%s\"", val, re)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type testReporterMock struct {
|
||||||
|
errors []string
|
||||||
|
}
|
||||||
|
|
||||||
|
func newTestReporterMock() *testReporterMock {
|
||||||
|
return &testReporterMock{errors: make([]string, 0)}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (trm *testReporterMock) Errorf(format string, args ...interface{}) {
|
||||||
|
trm.errors = append(trm.errors, fmt.Sprintf(format, args...))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMockAsyncProducerImplementsAsyncProducerInterface(t *testing.T) {
|
||||||
|
var mp interface{} = &AsyncProducer{}
|
||||||
|
if _, ok := mp.(sarama.AsyncProducer); !ok {
|
||||||
|
t.Error("The mock producer should implement the sarama.Producer interface.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProducerReturnsExpectationsToChannels(t *testing.T) {
|
||||||
|
config := sarama.NewConfig()
|
||||||
|
config.Producer.Return.Successes = true
|
||||||
|
mp := NewAsyncProducer(t, config)
|
||||||
|
|
||||||
|
mp.ExpectInputAndSucceed()
|
||||||
|
mp.ExpectInputAndSucceed()
|
||||||
|
mp.ExpectInputAndFail(sarama.ErrOutOfBrokers)
|
||||||
|
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test 1"}
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test 2"}
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test 3"}
|
||||||
|
|
||||||
|
msg1 := <-mp.Successes()
|
||||||
|
msg2 := <-mp.Successes()
|
||||||
|
err1 := <-mp.Errors()
|
||||||
|
|
||||||
|
if msg1.Topic != "test 1" {
|
||||||
|
t.Error("Expected message 1 to be returned first")
|
||||||
|
}
|
||||||
|
|
||||||
|
if msg2.Topic != "test 2" {
|
||||||
|
t.Error("Expected message 2 to be returned second")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err1.Msg.Topic != "test 3" || err1.Err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Error("Expected message 3 to be returned as error")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := mp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProducerWithTooFewExpectations(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
mp := NewAsyncProducer(trm, nil)
|
||||||
|
mp.ExpectInputAndSucceed()
|
||||||
|
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test"}
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test"}
|
||||||
|
|
||||||
|
if err := mp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProducerWithTooManyExpectations(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
mp := NewAsyncProducer(trm, nil)
|
||||||
|
mp.ExpectInputAndSucceed()
|
||||||
|
mp.ExpectInputAndFail(sarama.ErrOutOfBrokers)
|
||||||
|
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test"}
|
||||||
|
if err := mp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProducerWithCheckerFunction(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
mp := NewAsyncProducer(trm, nil)
|
||||||
|
mp.ExpectInputWithCheckerFunctionAndSucceed(generateRegexpChecker("^tes"))
|
||||||
|
mp.ExpectInputWithCheckerFunctionAndSucceed(generateRegexpChecker("^tes$"))
|
||||||
|
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
mp.Input() <- &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
if err := mp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(mp.Errors()) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
|
||||||
|
err1 := <-mp.Errors()
|
||||||
|
if !strings.HasPrefix(err1.Err.Error(), "No match") {
|
||||||
|
t.Error("Expected to report a value check error, found: ", err1.Err)
|
||||||
|
}
|
||||||
|
}
|
315
vendor/github.com/Shopify/sarama/mocks/consumer.go
generated
vendored
Normal file
315
vendor/github.com/Shopify/sarama/mocks/consumer.go
generated
vendored
Normal file
|
@ -0,0 +1,315 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Consumer implements sarama's Consumer interface for testing purposes.
|
||||||
|
// Before you can start consuming from this consumer, you have to register
|
||||||
|
// topic/partitions using ExpectConsumePartition, and set expectations on them.
|
||||||
|
type Consumer struct {
|
||||||
|
l sync.Mutex
|
||||||
|
t ErrorReporter
|
||||||
|
config *sarama.Config
|
||||||
|
partitionConsumers map[string]map[int32]*PartitionConsumer
|
||||||
|
metadata map[string][]int32
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewConsumer returns a new mock Consumer instance. The t argument should
|
||||||
|
// be the *testing.T instance of your test method. An error will be written to it if
|
||||||
|
// an expectation is violated. The config argument can be set to nil.
|
||||||
|
func NewConsumer(t ErrorReporter, config *sarama.Config) *Consumer {
|
||||||
|
if config == nil {
|
||||||
|
config = sarama.NewConfig()
|
||||||
|
}
|
||||||
|
|
||||||
|
c := &Consumer{
|
||||||
|
t: t,
|
||||||
|
config: config,
|
||||||
|
partitionConsumers: make(map[string]map[int32]*PartitionConsumer),
|
||||||
|
}
|
||||||
|
return c
|
||||||
|
}
|
||||||
|
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
// Consumer interface implementation
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// ConsumePartition implements the ConsumePartition method from the sarama.Consumer interface.
|
||||||
|
// Before you can start consuming a partition, you have to set expectations on it using
|
||||||
|
// ExpectConsumePartition. You can only consume a partition once per consumer.
|
||||||
|
func (c *Consumer) ConsumePartition(topic string, partition int32, offset int64) (sarama.PartitionConsumer, error) {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
if c.partitionConsumers[topic] == nil || c.partitionConsumers[topic][partition] == nil {
|
||||||
|
c.t.Errorf("No expectations set for %s/%d", topic, partition)
|
||||||
|
return nil, errOutOfExpectations
|
||||||
|
}
|
||||||
|
|
||||||
|
pc := c.partitionConsumers[topic][partition]
|
||||||
|
if pc.consumed {
|
||||||
|
return nil, sarama.ConfigurationError("The topic/partition is already being consumed")
|
||||||
|
}
|
||||||
|
|
||||||
|
if pc.offset != AnyOffset && pc.offset != offset {
|
||||||
|
c.t.Errorf("Unexpected offset when calling ConsumePartition for %s/%d. Expected %d, got %d.", topic, partition, pc.offset, offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
pc.consumed = true
|
||||||
|
return pc, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Topics returns a list of topics, as registered with SetMetadata
|
||||||
|
func (c *Consumer) Topics() ([]string, error) {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
if c.metadata == nil {
|
||||||
|
c.t.Errorf("Unexpected call to Topics. Initialize the mock's topic metadata with SetMetadata.")
|
||||||
|
return nil, sarama.ErrOutOfBrokers
|
||||||
|
}
|
||||||
|
|
||||||
|
var result []string
|
||||||
|
for topic := range c.metadata {
|
||||||
|
result = append(result, topic)
|
||||||
|
}
|
||||||
|
return result, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Partitions returns the list of parititons for the given topic, as registered with SetMetadata
|
||||||
|
func (c *Consumer) Partitions(topic string) ([]int32, error) {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
if c.metadata == nil {
|
||||||
|
c.t.Errorf("Unexpected call to Partitions. Initialize the mock's topic metadata with SetMetadata.")
|
||||||
|
return nil, sarama.ErrOutOfBrokers
|
||||||
|
}
|
||||||
|
if c.metadata[topic] == nil {
|
||||||
|
return nil, sarama.ErrUnknownTopicOrPartition
|
||||||
|
}
|
||||||
|
|
||||||
|
return c.metadata[topic], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Consumer) HighWaterMarks() map[string]map[int32]int64 {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
hwms := make(map[string]map[int32]int64, len(c.partitionConsumers))
|
||||||
|
for topic, partitionConsumers := range c.partitionConsumers {
|
||||||
|
hwm := make(map[int32]int64, len(partitionConsumers))
|
||||||
|
for partition, pc := range partitionConsumers {
|
||||||
|
hwm[partition] = pc.HighWaterMarkOffset()
|
||||||
|
}
|
||||||
|
hwms[topic] = hwm
|
||||||
|
}
|
||||||
|
|
||||||
|
return hwms
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close implements the Close method from the sarama.Consumer interface. It will close
|
||||||
|
// all registered PartitionConsumer instances.
|
||||||
|
func (c *Consumer) Close() error {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
for _, partitions := range c.partitionConsumers {
|
||||||
|
for _, partitionConsumer := range partitions {
|
||||||
|
_ = partitionConsumer.Close()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
// Expectation API
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// SetTopicMetadata sets the clusters topic/partition metadata,
|
||||||
|
// which will be returned by Topics() and Partitions().
|
||||||
|
func (c *Consumer) SetTopicMetadata(metadata map[string][]int32) {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
c.metadata = metadata
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectConsumePartition will register a topic/partition, so you can set expectations on it.
|
||||||
|
// The registered PartitionConsumer will be returned, so you can set expectations
|
||||||
|
// on it using method chaining. Once a topic/partition is registered, you are
|
||||||
|
// expected to start consuming it using ConsumePartition. If that doesn't happen,
|
||||||
|
// an error will be written to the error reporter once the mock consumer is closed. It will
|
||||||
|
// also expect that the
|
||||||
|
func (c *Consumer) ExpectConsumePartition(topic string, partition int32, offset int64) *PartitionConsumer {
|
||||||
|
c.l.Lock()
|
||||||
|
defer c.l.Unlock()
|
||||||
|
|
||||||
|
if c.partitionConsumers[topic] == nil {
|
||||||
|
c.partitionConsumers[topic] = make(map[int32]*PartitionConsumer)
|
||||||
|
}
|
||||||
|
|
||||||
|
if c.partitionConsumers[topic][partition] == nil {
|
||||||
|
c.partitionConsumers[topic][partition] = &PartitionConsumer{
|
||||||
|
t: c.t,
|
||||||
|
topic: topic,
|
||||||
|
partition: partition,
|
||||||
|
offset: offset,
|
||||||
|
messages: make(chan *sarama.ConsumerMessage, c.config.ChannelBufferSize),
|
||||||
|
errors: make(chan *sarama.ConsumerError, c.config.ChannelBufferSize),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return c.partitionConsumers[topic][partition]
|
||||||
|
}
|
||||||
|
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
// PartitionConsumer mock type
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// PartitionConsumer implements sarama's PartitionConsumer interface for testing purposes.
|
||||||
|
// It is returned by the mock Consumers ConsumePartitionMethod, but only if it is
|
||||||
|
// registered first using the Consumer's ExpectConsumePartition method. Before consuming the
|
||||||
|
// Errors and Messages channel, you should specify what values will be provided on these
|
||||||
|
// channels using YieldMessage and YieldError.
|
||||||
|
type PartitionConsumer struct {
|
||||||
|
highWaterMarkOffset int64 // must be at the top of the struct because https://golang.org/pkg/sync/atomic/#pkg-note-BUG
|
||||||
|
l sync.Mutex
|
||||||
|
t ErrorReporter
|
||||||
|
topic string
|
||||||
|
partition int32
|
||||||
|
offset int64
|
||||||
|
messages chan *sarama.ConsumerMessage
|
||||||
|
errors chan *sarama.ConsumerError
|
||||||
|
singleClose sync.Once
|
||||||
|
consumed bool
|
||||||
|
errorsShouldBeDrained bool
|
||||||
|
messagesShouldBeDrained bool
|
||||||
|
}
|
||||||
|
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
// PartitionConsumer interface implementation
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// AsyncClose implements the AsyncClose method from the sarama.PartitionConsumer interface.
|
||||||
|
func (pc *PartitionConsumer) AsyncClose() {
|
||||||
|
pc.singleClose.Do(func() {
|
||||||
|
close(pc.messages)
|
||||||
|
close(pc.errors)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close implements the Close method from the sarama.PartitionConsumer interface. It will
|
||||||
|
// verify whether the partition consumer was actually started.
|
||||||
|
func (pc *PartitionConsumer) Close() error {
|
||||||
|
if !pc.consumed {
|
||||||
|
pc.t.Errorf("Expectations set on %s/%d, but no partition consumer was started.", pc.topic, pc.partition)
|
||||||
|
return errPartitionConsumerNotStarted
|
||||||
|
}
|
||||||
|
|
||||||
|
if pc.errorsShouldBeDrained && len(pc.errors) > 0 {
|
||||||
|
pc.t.Errorf("Expected the errors channel for %s/%d to be drained on close, but found %d errors.", pc.topic, pc.partition, len(pc.errors))
|
||||||
|
}
|
||||||
|
|
||||||
|
if pc.messagesShouldBeDrained && len(pc.messages) > 0 {
|
||||||
|
pc.t.Errorf("Expected the messages channel for %s/%d to be drained on close, but found %d messages.", pc.topic, pc.partition, len(pc.messages))
|
||||||
|
}
|
||||||
|
|
||||||
|
pc.AsyncClose()
|
||||||
|
|
||||||
|
var (
|
||||||
|
closeErr error
|
||||||
|
wg sync.WaitGroup
|
||||||
|
)
|
||||||
|
|
||||||
|
wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
defer wg.Done()
|
||||||
|
|
||||||
|
var errs = make(sarama.ConsumerErrors, 0)
|
||||||
|
for err := range pc.errors {
|
||||||
|
errs = append(errs, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(errs) > 0 {
|
||||||
|
closeErr = errs
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
defer wg.Done()
|
||||||
|
for range pc.messages {
|
||||||
|
// drain
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
wg.Wait()
|
||||||
|
return closeErr
|
||||||
|
}
|
||||||
|
|
||||||
|
// Errors implements the Errors method from the sarama.PartitionConsumer interface.
|
||||||
|
func (pc *PartitionConsumer) Errors() <-chan *sarama.ConsumerError {
|
||||||
|
return pc.errors
|
||||||
|
}
|
||||||
|
|
||||||
|
// Messages implements the Messages method from the sarama.PartitionConsumer interface.
|
||||||
|
func (pc *PartitionConsumer) Messages() <-chan *sarama.ConsumerMessage {
|
||||||
|
return pc.messages
|
||||||
|
}
|
||||||
|
|
||||||
|
func (pc *PartitionConsumer) HighWaterMarkOffset() int64 {
|
||||||
|
return atomic.LoadInt64(&pc.highWaterMarkOffset) + 1
|
||||||
|
}
|
||||||
|
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
// Expectation API
|
||||||
|
///////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// YieldMessage will yield a messages Messages channel of this partition consumer
|
||||||
|
// when it is consumed. By default, the mock consumer will not verify whether this
|
||||||
|
// message was consumed from the Messages channel, because there are legitimate
|
||||||
|
// reasons forthis not to happen. ou can call ExpectMessagesDrainedOnClose so it will
|
||||||
|
// verify that the channel is empty on close.
|
||||||
|
func (pc *PartitionConsumer) YieldMessage(msg *sarama.ConsumerMessage) {
|
||||||
|
pc.l.Lock()
|
||||||
|
defer pc.l.Unlock()
|
||||||
|
|
||||||
|
msg.Topic = pc.topic
|
||||||
|
msg.Partition = pc.partition
|
||||||
|
msg.Offset = atomic.AddInt64(&pc.highWaterMarkOffset, 1)
|
||||||
|
|
||||||
|
pc.messages <- msg
|
||||||
|
}
|
||||||
|
|
||||||
|
// YieldError will yield an error on the Errors channel of this partition consumer
|
||||||
|
// when it is consumed. By default, the mock consumer will not verify whether this error was
|
||||||
|
// consumed from the Errors channel, because there are legitimate reasons for this
|
||||||
|
// not to happen. You can call ExpectErrorsDrainedOnClose so it will verify that
|
||||||
|
// the channel is empty on close.
|
||||||
|
func (pc *PartitionConsumer) YieldError(err error) {
|
||||||
|
pc.errors <- &sarama.ConsumerError{
|
||||||
|
Topic: pc.topic,
|
||||||
|
Partition: pc.partition,
|
||||||
|
Err: err,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectMessagesDrainedOnClose sets an expectation on the partition consumer
|
||||||
|
// that the messages channel will be fully drained when Close is called. If this
|
||||||
|
// expectation is not met, an error is reported to the error reporter.
|
||||||
|
func (pc *PartitionConsumer) ExpectMessagesDrainedOnClose() {
|
||||||
|
pc.messagesShouldBeDrained = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectErrorsDrainedOnClose sets an expectation on the partition consumer
|
||||||
|
// that the errors channel will be fully drained when Close is called. If this
|
||||||
|
// expectation is not met, an error is reported to the error reporter.
|
||||||
|
func (pc *PartitionConsumer) ExpectErrorsDrainedOnClose() {
|
||||||
|
pc.errorsShouldBeDrained = true
|
||||||
|
}
|
249
vendor/github.com/Shopify/sarama/mocks/consumer_test.go
generated
vendored
Normal file
249
vendor/github.com/Shopify/sarama/mocks/consumer_test.go
generated
vendored
Normal file
|
@ -0,0 +1,249 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sort"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestMockConsumerImplementsConsumerInterface(t *testing.T) {
|
||||||
|
var c interface{} = &Consumer{}
|
||||||
|
if _, ok := c.(sarama.Consumer); !ok {
|
||||||
|
t.Error("The mock consumer should implement the sarama.Consumer interface.")
|
||||||
|
}
|
||||||
|
|
||||||
|
var pc interface{} = &PartitionConsumer{}
|
||||||
|
if _, ok := pc.(sarama.PartitionConsumer); !ok {
|
||||||
|
t.Error("The mock partitionconsumer should implement the sarama.PartitionConsumer interface.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerHandlesExpectations(t *testing.T) {
|
||||||
|
consumer := NewConsumer(t, nil)
|
||||||
|
defer func() {
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest).YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello world")})
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest).YieldError(sarama.ErrOutOfBrokers)
|
||||||
|
consumer.ExpectConsumePartition("test", 1, sarama.OffsetOldest).YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello world again")})
|
||||||
|
consumer.ExpectConsumePartition("other", 0, AnyOffset).YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello other")})
|
||||||
|
|
||||||
|
pc_test0, err := consumer.ConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
test0_msg := <-pc_test0.Messages()
|
||||||
|
if test0_msg.Topic != "test" || test0_msg.Partition != 0 || string(test0_msg.Value) != "hello world" {
|
||||||
|
t.Error("Message was not as expected:", test0_msg)
|
||||||
|
}
|
||||||
|
test0_err := <-pc_test0.Errors()
|
||||||
|
if test0_err.Err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Error("Expected sarama.ErrOutOfBrokers, found:", test0_err.Err)
|
||||||
|
}
|
||||||
|
|
||||||
|
pc_test1, err := consumer.ConsumePartition("test", 1, sarama.OffsetOldest)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
test1_msg := <-pc_test1.Messages()
|
||||||
|
if test1_msg.Topic != "test" || test1_msg.Partition != 1 || string(test1_msg.Value) != "hello world again" {
|
||||||
|
t.Error("Message was not as expected:", test1_msg)
|
||||||
|
}
|
||||||
|
|
||||||
|
pc_other0, err := consumer.ConsumePartition("other", 0, sarama.OffsetNewest)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
other0_msg := <-pc_other0.Messages()
|
||||||
|
if other0_msg.Topic != "other" || other0_msg.Partition != 0 || string(other0_msg.Value) != "hello other" {
|
||||||
|
t.Error("Message was not as expected:", other0_msg)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerReturnsNonconsumedErrorsOnClose(t *testing.T) {
|
||||||
|
consumer := NewConsumer(t, nil)
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest).YieldError(sarama.ErrOutOfBrokers)
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest).YieldError(sarama.ErrOutOfBrokers)
|
||||||
|
|
||||||
|
pc, err := consumer.ConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-pc.Messages():
|
||||||
|
t.Error("Did not epxect a message on the messages channel.")
|
||||||
|
case err := <-pc.Errors():
|
||||||
|
if err.Err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Error("Expected sarama.ErrOutOfBrokers, found", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
errs := pc.Close().(sarama.ConsumerErrors)
|
||||||
|
if len(errs) != 1 && errs[0].Err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Error("Expected Close to return the remaining sarama.ErrOutOfBrokers")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerWithoutExpectationsOnPartition(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
|
||||||
|
_, err := consumer.ConsumePartition("test", 1, sarama.OffsetOldest)
|
||||||
|
if err != errOutOfExpectations {
|
||||||
|
t.Error("Expected ConsumePartition to return errOutOfExpectations")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error("No error expected on close, but found:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Errorf("Expected an expectation failure to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerWithExpectationsOnUnconsumedPartition(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest).YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello world")})
|
||||||
|
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error("No error expected on close, but found:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Errorf("Expected an expectation failure to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerWithWrongOffsetExpectation(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
|
||||||
|
_, err := consumer.ConsumePartition("test", 0, sarama.OffsetNewest)
|
||||||
|
if err != nil {
|
||||||
|
t.Error("Did not expect error, found:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Errorf("Expected an expectation failure to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerViolatesMessagesDrainedExpectation(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
pcmock := consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
pcmock.YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello")})
|
||||||
|
pcmock.YieldMessage(&sarama.ConsumerMessage{Value: []byte("hello")})
|
||||||
|
pcmock.ExpectMessagesDrainedOnClose()
|
||||||
|
|
||||||
|
pc, err := consumer.ConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
if err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// consume first message, not second one
|
||||||
|
<-pc.Messages()
|
||||||
|
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Errorf("Expected an expectation failure to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerMeetsErrorsDrainedExpectation(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
|
||||||
|
pcmock := consumer.ExpectConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
pcmock.YieldError(sarama.ErrInvalidMessage)
|
||||||
|
pcmock.YieldError(sarama.ErrInvalidMessage)
|
||||||
|
pcmock.ExpectErrorsDrainedOnClose()
|
||||||
|
|
||||||
|
pc, err := consumer.ConsumePartition("test", 0, sarama.OffsetOldest)
|
||||||
|
if err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// consume first and second error,
|
||||||
|
<-pc.Errors()
|
||||||
|
<-pc.Errors()
|
||||||
|
|
||||||
|
if err := consumer.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 0 {
|
||||||
|
t.Errorf("Expected no expectation failures to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerTopicMetadata(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
|
||||||
|
consumer.SetTopicMetadata(map[string][]int32{
|
||||||
|
"test1": {0, 1, 2, 3},
|
||||||
|
"test2": {0, 1, 2, 3, 4, 5, 6, 7},
|
||||||
|
})
|
||||||
|
|
||||||
|
topics, err := consumer.Topics()
|
||||||
|
if err != nil {
|
||||||
|
t.Error(t)
|
||||||
|
}
|
||||||
|
|
||||||
|
sortedTopics := sort.StringSlice(topics)
|
||||||
|
sortedTopics.Sort()
|
||||||
|
if len(sortedTopics) != 2 || sortedTopics[0] != "test1" || sortedTopics[1] != "test2" {
|
||||||
|
t.Error("Unexpected topics returned:", sortedTopics)
|
||||||
|
}
|
||||||
|
|
||||||
|
partitions1, err := consumer.Partitions("test1")
|
||||||
|
if err != nil {
|
||||||
|
t.Error(t)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(partitions1) != 4 {
|
||||||
|
t.Error("Unexpected partitions returned:", len(partitions1))
|
||||||
|
}
|
||||||
|
|
||||||
|
partitions2, err := consumer.Partitions("test2")
|
||||||
|
if err != nil {
|
||||||
|
t.Error(t)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(partitions2) != 8 {
|
||||||
|
t.Error("Unexpected partitions returned:", len(partitions2))
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 0 {
|
||||||
|
t.Errorf("Expected no expectation failures to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConsumerUnexpectedTopicMetadata(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
consumer := NewConsumer(trm, nil)
|
||||||
|
|
||||||
|
if _, err := consumer.Topics(); err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Error("Expected sarama.ErrOutOfBrokers, found", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Errorf("Expected an expectation failure to be set on the error reporter.")
|
||||||
|
}
|
||||||
|
}
|
48
vendor/github.com/Shopify/sarama/mocks/mocks.go
generated
vendored
Normal file
48
vendor/github.com/Shopify/sarama/mocks/mocks.go
generated
vendored
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
/*
|
||||||
|
Package mocks provides mocks that can be used for testing applications
|
||||||
|
that use Sarama. The mock types provided by this package implement the
|
||||||
|
interfaces Sarama exports, so you can use them for dependency injection
|
||||||
|
in your tests.
|
||||||
|
|
||||||
|
All mock instances require you to set expectations on them before you
|
||||||
|
can use them. It will determine how the mock will behave. If an
|
||||||
|
expectation is not met, it will make your test fail.
|
||||||
|
|
||||||
|
NOTE: this package currently does not fall under the API stability
|
||||||
|
guarantee of Sarama as it is still considered experimental.
|
||||||
|
*/
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ErrorReporter is a simple interface that includes the testing.T methods we use to report
|
||||||
|
// expectation violations when using the mock objects.
|
||||||
|
type ErrorReporter interface {
|
||||||
|
Errorf(string, ...interface{})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ValueChecker is a function type to be set in each expectation of the producer mocks
|
||||||
|
// to check the value passed.
|
||||||
|
type ValueChecker func(val []byte) error
|
||||||
|
|
||||||
|
var (
|
||||||
|
errProduceSuccess error = nil
|
||||||
|
errOutOfExpectations = errors.New("No more expectations set on mock")
|
||||||
|
errPartitionConsumerNotStarted = errors.New("The partition consumer was never started")
|
||||||
|
)
|
||||||
|
|
||||||
|
const AnyOffset int64 = -1000
|
||||||
|
|
||||||
|
type producerExpectation struct {
|
||||||
|
Result error
|
||||||
|
CheckFunction ValueChecker
|
||||||
|
}
|
||||||
|
|
||||||
|
type consumerExpectation struct {
|
||||||
|
Err error
|
||||||
|
Msg *sarama.ConsumerMessage
|
||||||
|
}
|
146
vendor/github.com/Shopify/sarama/mocks/sync_producer.go
generated
vendored
Normal file
146
vendor/github.com/Shopify/sarama/mocks/sync_producer.go
generated
vendored
Normal file
|
@ -0,0 +1,146 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SyncProducer implements sarama's SyncProducer interface for testing purposes.
|
||||||
|
// Before you can use it, you have to set expectations on the mock SyncProducer
|
||||||
|
// to tell it how to handle calls to SendMessage, so you can easily test success
|
||||||
|
// and failure scenarios.
|
||||||
|
type SyncProducer struct {
|
||||||
|
l sync.Mutex
|
||||||
|
t ErrorReporter
|
||||||
|
expectations []*producerExpectation
|
||||||
|
lastOffset int64
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewSyncProducer instantiates a new SyncProducer mock. The t argument should
|
||||||
|
// be the *testing.T instance of your test method. An error will be written to it if
|
||||||
|
// an expectation is violated. The config argument is currently unused, but is
|
||||||
|
// maintained to be compatible with the async Producer.
|
||||||
|
func NewSyncProducer(t ErrorReporter, config *sarama.Config) *SyncProducer {
|
||||||
|
return &SyncProducer{
|
||||||
|
t: t,
|
||||||
|
expectations: make([]*producerExpectation, 0),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
// Implement SyncProducer interface
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// SendMessage corresponds with the SendMessage method of sarama's SyncProducer implementation.
|
||||||
|
// You have to set expectations on the mock producer before calling SendMessage, so it knows
|
||||||
|
// how to handle them. You can set a function in each expectation so that the message value
|
||||||
|
// checked by this function and an error is returned if the match fails.
|
||||||
|
// If there is no more remaining expectation when SendMessage is called,
|
||||||
|
// the mock producer will write an error to the test state object.
|
||||||
|
func (sp *SyncProducer) SendMessage(msg *sarama.ProducerMessage) (partition int32, offset int64, err error) {
|
||||||
|
sp.l.Lock()
|
||||||
|
defer sp.l.Unlock()
|
||||||
|
|
||||||
|
if len(sp.expectations) > 0 {
|
||||||
|
expectation := sp.expectations[0]
|
||||||
|
sp.expectations = sp.expectations[1:]
|
||||||
|
if expectation.CheckFunction != nil {
|
||||||
|
val, err := msg.Value.Encode()
|
||||||
|
if err != nil {
|
||||||
|
sp.t.Errorf("Input message encoding failed: %s", err.Error())
|
||||||
|
return -1, -1, err
|
||||||
|
}
|
||||||
|
|
||||||
|
errCheck := expectation.CheckFunction(val)
|
||||||
|
if errCheck != nil {
|
||||||
|
sp.t.Errorf("Check function returned an error: %s", errCheck.Error())
|
||||||
|
return -1, -1, errCheck
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if expectation.Result == errProduceSuccess {
|
||||||
|
sp.lastOffset++
|
||||||
|
msg.Offset = sp.lastOffset
|
||||||
|
return 0, msg.Offset, nil
|
||||||
|
}
|
||||||
|
return -1, -1, expectation.Result
|
||||||
|
}
|
||||||
|
sp.t.Errorf("No more expectation set on this mock producer to handle the input message.")
|
||||||
|
return -1, -1, errOutOfExpectations
|
||||||
|
}
|
||||||
|
|
||||||
|
// SendMessages corresponds with the SendMessages method of sarama's SyncProducer implementation.
|
||||||
|
// You have to set expectations on the mock producer before calling SendMessages, so it knows
|
||||||
|
// how to handle them. If there is no more remaining expectations when SendMessages is called,
|
||||||
|
// the mock producer will write an error to the test state object.
|
||||||
|
func (sp *SyncProducer) SendMessages(msgs []*sarama.ProducerMessage) error {
|
||||||
|
sp.l.Lock()
|
||||||
|
defer sp.l.Unlock()
|
||||||
|
|
||||||
|
if len(sp.expectations) >= len(msgs) {
|
||||||
|
expectations := sp.expectations[0 : len(msgs)-1]
|
||||||
|
sp.expectations = sp.expectations[len(msgs):]
|
||||||
|
|
||||||
|
for _, expectation := range expectations {
|
||||||
|
if expectation.Result != errProduceSuccess {
|
||||||
|
return expectation.Result
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
sp.t.Errorf("Insufficient expectations set on this mock producer to handle the input messages.")
|
||||||
|
return errOutOfExpectations
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close corresponds with the Close method of sarama's SyncProducer implementation.
|
||||||
|
// By closing a mock syncproducer, you also tell it that no more SendMessage calls will follow,
|
||||||
|
// so it will write an error to the test state if there's any remaining expectations.
|
||||||
|
func (sp *SyncProducer) Close() error {
|
||||||
|
sp.l.Lock()
|
||||||
|
defer sp.l.Unlock()
|
||||||
|
|
||||||
|
if len(sp.expectations) > 0 {
|
||||||
|
sp.t.Errorf("Expected to exhaust all expectations, but %d are left.", len(sp.expectations))
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
// Setting expectations
|
||||||
|
////////////////////////////////////////////////
|
||||||
|
|
||||||
|
// ExpectSendMessageWithCheckerFunctionAndSucceed sets an expectation on the mock producer that SendMessage
|
||||||
|
// will be called. The mock producer will first call the given function to check the message value.
|
||||||
|
// It will cascade the error of the function, if any, or handle the message as if it produced
|
||||||
|
// successfully, i.e. by returning a valid partition, and offset, and a nil error.
|
||||||
|
func (sp *SyncProducer) ExpectSendMessageWithCheckerFunctionAndSucceed(cf ValueChecker) {
|
||||||
|
sp.l.Lock()
|
||||||
|
defer sp.l.Unlock()
|
||||||
|
sp.expectations = append(sp.expectations, &producerExpectation{Result: errProduceSuccess, CheckFunction: cf})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectSendMessageWithCheckerFunctionAndFail sets an expectation on the mock producer that SendMessage will be
|
||||||
|
// called. The mock producer will first call the given function to check the message value.
|
||||||
|
// It will cascade the error of the function, if any, or handle the message as if it failed
|
||||||
|
// to produce successfully, i.e. by returning the provided error.
|
||||||
|
func (sp *SyncProducer) ExpectSendMessageWithCheckerFunctionAndFail(cf ValueChecker, err error) {
|
||||||
|
sp.l.Lock()
|
||||||
|
defer sp.l.Unlock()
|
||||||
|
sp.expectations = append(sp.expectations, &producerExpectation{Result: err, CheckFunction: cf})
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectSendMessageAndSucceed sets an expectation on the mock producer that SendMessage will be
|
||||||
|
// called. The mock producer will handle the message as if it produced successfully, i.e. by
|
||||||
|
// returning a valid partition, and offset, and a nil error.
|
||||||
|
func (sp *SyncProducer) ExpectSendMessageAndSucceed() {
|
||||||
|
sp.ExpectSendMessageWithCheckerFunctionAndSucceed(nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExpectSendMessageAndFail sets an expectation on the mock producer that SendMessage will be
|
||||||
|
// called. The mock producer will handle the message as if it failed to produce
|
||||||
|
// successfully, i.e. by returning the provided error.
|
||||||
|
func (sp *SyncProducer) ExpectSendMessageAndFail(err error) {
|
||||||
|
sp.ExpectSendMessageWithCheckerFunctionAndFail(nil, err)
|
||||||
|
}
|
124
vendor/github.com/Shopify/sarama/mocks/sync_producer_test.go
generated
vendored
Normal file
124
vendor/github.com/Shopify/sarama/mocks/sync_producer_test.go
generated
vendored
Normal file
|
@ -0,0 +1,124 @@
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import (
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestMockSyncProducerImplementsSyncProducerInterface(t *testing.T) {
|
||||||
|
var mp interface{} = &SyncProducer{}
|
||||||
|
if _, ok := mp.(sarama.SyncProducer); !ok {
|
||||||
|
t.Error("The mock async producer should implement the sarama.SyncProducer interface.")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestSyncProducerReturnsExpectationsToSendMessage(t *testing.T) {
|
||||||
|
sp := NewSyncProducer(t, nil)
|
||||||
|
defer func() {
|
||||||
|
if err := sp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
sp.ExpectSendMessageAndSucceed()
|
||||||
|
sp.ExpectSendMessageAndSucceed()
|
||||||
|
sp.ExpectSendMessageAndFail(sarama.ErrOutOfBrokers)
|
||||||
|
|
||||||
|
msg := &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
|
||||||
|
_, offset, err := sp.SendMessage(msg)
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("The first message should have been produced successfully, but got %s", err)
|
||||||
|
}
|
||||||
|
if offset != 1 || offset != msg.Offset {
|
||||||
|
t.Errorf("The first message should have been assigned offset 1, but got %d", msg.Offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, offset, err = sp.SendMessage(msg)
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("The second message should have been produced successfully, but got %s", err)
|
||||||
|
}
|
||||||
|
if offset != 2 || offset != msg.Offset {
|
||||||
|
t.Errorf("The second message should have been assigned offset 2, but got %d", offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, _, err = sp.SendMessage(msg)
|
||||||
|
if err != sarama.ErrOutOfBrokers {
|
||||||
|
t.Errorf("The third message should not have been produced successfully")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := sp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestSyncProducerWithTooManyExpectations(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
|
||||||
|
sp := NewSyncProducer(trm, nil)
|
||||||
|
sp.ExpectSendMessageAndSucceed()
|
||||||
|
sp.ExpectSendMessageAndFail(sarama.ErrOutOfBrokers)
|
||||||
|
|
||||||
|
msg := &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
if _, _, err := sp.SendMessage(msg); err != nil {
|
||||||
|
t.Error("No error expected on first SendMessage call", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := sp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestSyncProducerWithTooFewExpectations(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
|
||||||
|
sp := NewSyncProducer(trm, nil)
|
||||||
|
sp.ExpectSendMessageAndSucceed()
|
||||||
|
|
||||||
|
msg := &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
if _, _, err := sp.SendMessage(msg); err != nil {
|
||||||
|
t.Error("No error expected on first SendMessage call", err)
|
||||||
|
}
|
||||||
|
if _, _, err := sp.SendMessage(msg); err != errOutOfExpectations {
|
||||||
|
t.Error("errOutOfExpectations expected on second SendMessage call, found:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := sp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestSyncProducerWithCheckerFunction(t *testing.T) {
|
||||||
|
trm := newTestReporterMock()
|
||||||
|
|
||||||
|
sp := NewSyncProducer(trm, nil)
|
||||||
|
sp.ExpectSendMessageWithCheckerFunctionAndSucceed(generateRegexpChecker("^tes"))
|
||||||
|
sp.ExpectSendMessageWithCheckerFunctionAndSucceed(generateRegexpChecker("^tes$"))
|
||||||
|
|
||||||
|
msg := &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
if _, _, err := sp.SendMessage(msg); err != nil {
|
||||||
|
t.Error("No error expected on first SendMessage call, found: ", err)
|
||||||
|
}
|
||||||
|
msg = &sarama.ProducerMessage{Topic: "test", Value: sarama.StringEncoder("test")}
|
||||||
|
if _, _, err := sp.SendMessage(msg); err == nil || !strings.HasPrefix(err.Error(), "No match") {
|
||||||
|
t.Error("Error during value check expected on second SendMessage call, found:", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := sp.Close(); err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(trm.errors) != 1 {
|
||||||
|
t.Error("Expected to report an error")
|
||||||
|
}
|
||||||
|
}
|
10
vendor/github.com/Shopify/sarama/tools/README.md
generated
vendored
Normal file
10
vendor/github.com/Shopify/sarama/tools/README.md
generated
vendored
Normal file
|
@ -0,0 +1,10 @@
|
||||||
|
# Sarama tools
|
||||||
|
|
||||||
|
This folder contains applications that are useful for exploration of your Kafka cluster, or instrumentation.
|
||||||
|
Some of these tools mirror tools that ship with Kafka, but these tools won't require installing the JVM to function.
|
||||||
|
|
||||||
|
- [kafka-console-producer](./kafka-console-producer): a command line tool to produce a single message to your Kafka custer.
|
||||||
|
- [kafka-console-partitionconsumer](./kafka-console-partitionconsumer): (deprecated) a command line tool to consume a single partition of a topic on your Kafka cluster.
|
||||||
|
- [kafka-console-consumer](./kafka-console-consumer): a command line tool to consume arbitrary partitions of a topic on your Kafka cluster.
|
||||||
|
|
||||||
|
To install all tools, run `go get github.com/Shopify/sarama/tools/...`
|
2
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/.gitignore
generated
vendored
Normal file
2
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/.gitignore
generated
vendored
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
kafka-console-consumer
|
||||||
|
kafka-console-consumer.test
|
29
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/README.md
generated
vendored
Normal file
29
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/README.md
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
# kafka-console-consumer
|
||||||
|
|
||||||
|
A simple command line tool to consume partitions of a topic and print the
|
||||||
|
messages on the standard output.
|
||||||
|
|
||||||
|
### Installation
|
||||||
|
|
||||||
|
go get github.com/Shopify/sarama/tools/kafka-console-consumer
|
||||||
|
|
||||||
|
### Usage
|
||||||
|
|
||||||
|
# Minimum invocation
|
||||||
|
kafka-console-consumer -topic=test -brokers=kafka1:9092
|
||||||
|
|
||||||
|
# It will pick up a KAFKA_PEERS environment variable
|
||||||
|
export KAFKA_PEERS=kafka1:9092,kafka2:9092,kafka3:9092
|
||||||
|
kafka-console-consumer -topic=test
|
||||||
|
|
||||||
|
# You can specify the offset you want to start at. It can be either
|
||||||
|
# `oldest`, `newest`. The default is `newest`.
|
||||||
|
kafka-console-consumer -topic=test -offset=oldest
|
||||||
|
kafka-console-consumer -topic=test -offset=newest
|
||||||
|
|
||||||
|
# You can specify the partition(s) you want to consume as a comma-separated
|
||||||
|
# list. The default is `all`.
|
||||||
|
kafka-console-consumer -topic=test -partitions=1,2,3
|
||||||
|
|
||||||
|
# Display all command line options
|
||||||
|
kafka-console-consumer -help
|
145
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/kafka-console-consumer.go
generated
vendored
Normal file
145
vendor/github.com/Shopify/sarama/tools/kafka-console-consumer/kafka-console-consumer.go
generated
vendored
Normal file
|
@ -0,0 +1,145 @@
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"os/signal"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
brokerList = flag.String("brokers", os.Getenv("KAFKA_PEERS"), "The comma separated list of brokers in the Kafka cluster")
|
||||||
|
topic = flag.String("topic", "", "REQUIRED: the topic to consume")
|
||||||
|
partitions = flag.String("partitions", "all", "The partitions to consume, can be 'all' or comma-separated numbers")
|
||||||
|
offset = flag.String("offset", "newest", "The offset to start with. Can be `oldest`, `newest`")
|
||||||
|
verbose = flag.Bool("verbose", false, "Whether to turn on sarama logging")
|
||||||
|
bufferSize = flag.Int("buffer-size", 256, "The buffer size of the message channel.")
|
||||||
|
|
||||||
|
logger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
flag.Parse()
|
||||||
|
|
||||||
|
if *brokerList == "" {
|
||||||
|
printUsageErrorAndExit("You have to provide -brokers as a comma-separated list, or set the KAFKA_PEERS environment variable.")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *topic == "" {
|
||||||
|
printUsageErrorAndExit("-topic is required")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *verbose {
|
||||||
|
sarama.Logger = logger
|
||||||
|
}
|
||||||
|
|
||||||
|
var initialOffset int64
|
||||||
|
switch *offset {
|
||||||
|
case "oldest":
|
||||||
|
initialOffset = sarama.OffsetOldest
|
||||||
|
case "newest":
|
||||||
|
initialOffset = sarama.OffsetNewest
|
||||||
|
default:
|
||||||
|
printUsageErrorAndExit("-offset should be `oldest` or `newest`")
|
||||||
|
}
|
||||||
|
|
||||||
|
c, err := sarama.NewConsumer(strings.Split(*brokerList, ","), nil)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to start consumer: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
partitionList, err := getPartitions(c)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to get the list of partitions: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
messages = make(chan *sarama.ConsumerMessage, *bufferSize)
|
||||||
|
closing = make(chan struct{})
|
||||||
|
wg sync.WaitGroup
|
||||||
|
)
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
signals := make(chan os.Signal, 1)
|
||||||
|
signal.Notify(signals, os.Kill, os.Interrupt)
|
||||||
|
<-signals
|
||||||
|
logger.Println("Initiating shutdown of consumer...")
|
||||||
|
close(closing)
|
||||||
|
}()
|
||||||
|
|
||||||
|
for _, partition := range partitionList {
|
||||||
|
pc, err := c.ConsumePartition(*topic, partition, initialOffset)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to start consumer for partition %d: %s", partition, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
go func(pc sarama.PartitionConsumer) {
|
||||||
|
<-closing
|
||||||
|
pc.AsyncClose()
|
||||||
|
}(pc)
|
||||||
|
|
||||||
|
wg.Add(1)
|
||||||
|
go func(pc sarama.PartitionConsumer) {
|
||||||
|
defer wg.Done()
|
||||||
|
for message := range pc.Messages() {
|
||||||
|
messages <- message
|
||||||
|
}
|
||||||
|
}(pc)
|
||||||
|
}
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for msg := range messages {
|
||||||
|
fmt.Printf("Partition:\t%d\n", msg.Partition)
|
||||||
|
fmt.Printf("Offset:\t%d\n", msg.Offset)
|
||||||
|
fmt.Printf("Key:\t%s\n", string(msg.Key))
|
||||||
|
fmt.Printf("Value:\t%s\n", string(msg.Value))
|
||||||
|
fmt.Println()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
wg.Wait()
|
||||||
|
logger.Println("Done consuming topic", *topic)
|
||||||
|
close(messages)
|
||||||
|
|
||||||
|
if err := c.Close(); err != nil {
|
||||||
|
logger.Println("Failed to close consumer: ", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func getPartitions(c sarama.Consumer) ([]int32, error) {
|
||||||
|
if *partitions == "all" {
|
||||||
|
return c.Partitions(*topic)
|
||||||
|
}
|
||||||
|
|
||||||
|
tmp := strings.Split(*partitions, ",")
|
||||||
|
var pList []int32
|
||||||
|
for i := range tmp {
|
||||||
|
val, err := strconv.ParseInt(tmp[i], 10, 32)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
pList = append(pList, int32(val))
|
||||||
|
}
|
||||||
|
|
||||||
|
return pList, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func printErrorAndExit(code int, format string, values ...interface{}) {
|
||||||
|
fmt.Fprintf(os.Stderr, "ERROR: %s\n", fmt.Sprintf(format, values...))
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
os.Exit(code)
|
||||||
|
}
|
||||||
|
|
||||||
|
func printUsageErrorAndExit(format string, values ...interface{}) {
|
||||||
|
fmt.Fprintf(os.Stderr, "ERROR: %s\n", fmt.Sprintf(format, values...))
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
fmt.Fprintln(os.Stderr, "Available command line options:")
|
||||||
|
flag.PrintDefaults()
|
||||||
|
os.Exit(64)
|
||||||
|
}
|
2
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/.gitignore
generated
vendored
Normal file
2
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/.gitignore
generated
vendored
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
kafka-console-partitionconsumer
|
||||||
|
kafka-console-partitionconsumer.test
|
28
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/README.md
generated
vendored
Normal file
28
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/README.md
generated
vendored
Normal file
|
@ -0,0 +1,28 @@
|
||||||
|
# kafka-console-partitionconsumer
|
||||||
|
|
||||||
|
NOTE: this tool is deprecated in favour of the more general and more powerful
|
||||||
|
`kafka-console-consumer`.
|
||||||
|
|
||||||
|
A simple command line tool to consume a partition of a topic and print the messages
|
||||||
|
on the standard output.
|
||||||
|
|
||||||
|
### Installation
|
||||||
|
|
||||||
|
go get github.com/Shopify/sarama/tools/kafka-console-partitionconsumer
|
||||||
|
|
||||||
|
### Usage
|
||||||
|
|
||||||
|
# Minimum invocation
|
||||||
|
kafka-console-partitionconsumer -topic=test -partition=4 -brokers=kafka1:9092
|
||||||
|
|
||||||
|
# It will pick up a KAFKA_PEERS environment variable
|
||||||
|
export KAFKA_PEERS=kafka1:9092,kafka2:9092,kafka3:9092
|
||||||
|
kafka-console-partitionconsumer -topic=test -partition=4
|
||||||
|
|
||||||
|
# You can specify the offset you want to start at. It can be either
|
||||||
|
# `oldest`, `newest`, or a specific offset number
|
||||||
|
kafka-console-partitionconsumer -topic=test -partition=3 -offset=oldest
|
||||||
|
kafka-console-partitionconsumer -topic=test -partition=2 -offset=1337
|
||||||
|
|
||||||
|
# Display all command line options
|
||||||
|
kafka-console-partitionconsumer -help
|
102
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/kafka-console-partitionconsumer.go
generated
vendored
Normal file
102
vendor/github.com/Shopify/sarama/tools/kafka-console-partitionconsumer/kafka-console-partitionconsumer.go
generated
vendored
Normal file
|
@ -0,0 +1,102 @@
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"os/signal"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
brokerList = flag.String("brokers", os.Getenv("KAFKA_PEERS"), "The comma separated list of brokers in the Kafka cluster")
|
||||||
|
topic = flag.String("topic", "", "REQUIRED: the topic to consume")
|
||||||
|
partition = flag.Int("partition", -1, "REQUIRED: the partition to consume")
|
||||||
|
offset = flag.String("offset", "newest", "The offset to start with. Can be `oldest`, `newest`, or an actual offset")
|
||||||
|
verbose = flag.Bool("verbose", false, "Whether to turn on sarama logging")
|
||||||
|
|
||||||
|
logger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
flag.Parse()
|
||||||
|
|
||||||
|
if *brokerList == "" {
|
||||||
|
printUsageErrorAndExit("You have to provide -brokers as a comma-separated list, or set the KAFKA_PEERS environment variable.")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *topic == "" {
|
||||||
|
printUsageErrorAndExit("-topic is required")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *partition == -1 {
|
||||||
|
printUsageErrorAndExit("-partition is required")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *verbose {
|
||||||
|
sarama.Logger = logger
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
initialOffset int64
|
||||||
|
offsetError error
|
||||||
|
)
|
||||||
|
switch *offset {
|
||||||
|
case "oldest":
|
||||||
|
initialOffset = sarama.OffsetOldest
|
||||||
|
case "newest":
|
||||||
|
initialOffset = sarama.OffsetNewest
|
||||||
|
default:
|
||||||
|
initialOffset, offsetError = strconv.ParseInt(*offset, 10, 64)
|
||||||
|
}
|
||||||
|
|
||||||
|
if offsetError != nil {
|
||||||
|
printUsageErrorAndExit("Invalid initial offset: %s", *offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
c, err := sarama.NewConsumer(strings.Split(*brokerList, ","), nil)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to start consumer: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
pc, err := c.ConsumePartition(*topic, int32(*partition), initialOffset)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to start partition consumer: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
signals := make(chan os.Signal, 1)
|
||||||
|
signal.Notify(signals, os.Kill, os.Interrupt)
|
||||||
|
<-signals
|
||||||
|
pc.AsyncClose()
|
||||||
|
}()
|
||||||
|
|
||||||
|
for msg := range pc.Messages() {
|
||||||
|
fmt.Printf("Offset:\t%d\n", msg.Offset)
|
||||||
|
fmt.Printf("Key:\t%s\n", string(msg.Key))
|
||||||
|
fmt.Printf("Value:\t%s\n", string(msg.Value))
|
||||||
|
fmt.Println()
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := c.Close(); err != nil {
|
||||||
|
logger.Println("Failed to close consumer: ", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func printErrorAndExit(code int, format string, values ...interface{}) {
|
||||||
|
fmt.Fprintf(os.Stderr, "ERROR: %s\n", fmt.Sprintf(format, values...))
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
os.Exit(code)
|
||||||
|
}
|
||||||
|
|
||||||
|
func printUsageErrorAndExit(format string, values ...interface{}) {
|
||||||
|
fmt.Fprintf(os.Stderr, "ERROR: %s\n", fmt.Sprintf(format, values...))
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
fmt.Fprintln(os.Stderr, "Available command line options:")
|
||||||
|
flag.PrintDefaults()
|
||||||
|
os.Exit(64)
|
||||||
|
}
|
2
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/.gitignore
generated
vendored
Normal file
2
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/.gitignore
generated
vendored
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
kafka-console-producer
|
||||||
|
kafka-console-producer.test
|
34
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/README.md
generated
vendored
Normal file
34
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/README.md
generated
vendored
Normal file
|
@ -0,0 +1,34 @@
|
||||||
|
# kafka-console-producer
|
||||||
|
|
||||||
|
A simple command line tool to produce a single message to Kafka.
|
||||||
|
|
||||||
|
### Installation
|
||||||
|
|
||||||
|
go get github.com/Shopify/sarama/tools/kafka-console-producer
|
||||||
|
|
||||||
|
|
||||||
|
### Usage
|
||||||
|
|
||||||
|
# Minimum invocation
|
||||||
|
kafka-console-producer -topic=test -value=value -brokers=kafka1:9092
|
||||||
|
|
||||||
|
# It will pick up a KAFKA_PEERS environment variable
|
||||||
|
export KAFKA_PEERS=kafka1:9092,kafka2:9092,kafka3:9092
|
||||||
|
kafka-console-producer -topic=test -value=value
|
||||||
|
|
||||||
|
# It will read the value from stdin by using pipes
|
||||||
|
echo "hello world" | kafka-console-producer -topic=test
|
||||||
|
|
||||||
|
# Specify a key:
|
||||||
|
echo "hello world" | kafka-console-producer -topic=test -key=key
|
||||||
|
|
||||||
|
# Partitioning: by default, kafka-console-producer will partition as follows:
|
||||||
|
# - manual partitioning if a -partition is provided
|
||||||
|
# - hash partitioning by key if a -key is provided
|
||||||
|
# - random partioning otherwise.
|
||||||
|
#
|
||||||
|
# You can override this using the -partitioner argument:
|
||||||
|
echo "hello world" | kafka-console-producer -topic=test -key=key -partitioner=random
|
||||||
|
|
||||||
|
# Display all command line options
|
||||||
|
kafka-console-producer -help
|
124
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/kafka-console-producer.go
generated
vendored
Normal file
124
vendor/github.com/Shopify/sarama/tools/kafka-console-producer/kafka-console-producer.go
generated
vendored
Normal file
|
@ -0,0 +1,124 @@
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
"github.com/rcrowley/go-metrics"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
brokerList = flag.String("brokers", os.Getenv("KAFKA_PEERS"), "The comma separated list of brokers in the Kafka cluster. You can also set the KAFKA_PEERS environment variable")
|
||||||
|
topic = flag.String("topic", "", "REQUIRED: the topic to produce to")
|
||||||
|
key = flag.String("key", "", "The key of the message to produce. Can be empty.")
|
||||||
|
value = flag.String("value", "", "REQUIRED: the value of the message to produce. You can also provide the value on stdin.")
|
||||||
|
partitioner = flag.String("partitioner", "", "The partitioning scheme to use. Can be `hash`, `manual`, or `random`")
|
||||||
|
partition = flag.Int("partition", -1, "The partition to produce to.")
|
||||||
|
verbose = flag.Bool("verbose", false, "Turn on sarama logging to stderr")
|
||||||
|
showMetrics = flag.Bool("metrics", false, "Output metrics on successful publish to stderr")
|
||||||
|
silent = flag.Bool("silent", false, "Turn off printing the message's topic, partition, and offset to stdout")
|
||||||
|
|
||||||
|
logger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
flag.Parse()
|
||||||
|
|
||||||
|
if *brokerList == "" {
|
||||||
|
printUsageErrorAndExit("no -brokers specified. Alternatively, set the KAFKA_PEERS environment variable")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *topic == "" {
|
||||||
|
printUsageErrorAndExit("no -topic specified")
|
||||||
|
}
|
||||||
|
|
||||||
|
if *verbose {
|
||||||
|
sarama.Logger = logger
|
||||||
|
}
|
||||||
|
|
||||||
|
config := sarama.NewConfig()
|
||||||
|
config.Producer.RequiredAcks = sarama.WaitForAll
|
||||||
|
config.Producer.Return.Successes = true
|
||||||
|
|
||||||
|
switch *partitioner {
|
||||||
|
case "":
|
||||||
|
if *partition >= 0 {
|
||||||
|
config.Producer.Partitioner = sarama.NewManualPartitioner
|
||||||
|
} else {
|
||||||
|
config.Producer.Partitioner = sarama.NewHashPartitioner
|
||||||
|
}
|
||||||
|
case "hash":
|
||||||
|
config.Producer.Partitioner = sarama.NewHashPartitioner
|
||||||
|
case "random":
|
||||||
|
config.Producer.Partitioner = sarama.NewRandomPartitioner
|
||||||
|
case "manual":
|
||||||
|
config.Producer.Partitioner = sarama.NewManualPartitioner
|
||||||
|
if *partition == -1 {
|
||||||
|
printUsageErrorAndExit("-partition is required when partitioning manually")
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
printUsageErrorAndExit(fmt.Sprintf("Partitioner %s not supported.", *partitioner))
|
||||||
|
}
|
||||||
|
|
||||||
|
message := &sarama.ProducerMessage{Topic: *topic, Partition: int32(*partition)}
|
||||||
|
|
||||||
|
if *key != "" {
|
||||||
|
message.Key = sarama.StringEncoder(*key)
|
||||||
|
}
|
||||||
|
|
||||||
|
if *value != "" {
|
||||||
|
message.Value = sarama.StringEncoder(*value)
|
||||||
|
} else if stdinAvailable() {
|
||||||
|
bytes, err := ioutil.ReadAll(os.Stdin)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(66, "Failed to read data from the standard input: %s", err)
|
||||||
|
}
|
||||||
|
message.Value = sarama.ByteEncoder(bytes)
|
||||||
|
} else {
|
||||||
|
printUsageErrorAndExit("-value is required, or you have to provide the value on stdin")
|
||||||
|
}
|
||||||
|
|
||||||
|
producer, err := sarama.NewSyncProducer(strings.Split(*brokerList, ","), config)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to open Kafka producer: %s", err)
|
||||||
|
}
|
||||||
|
defer func() {
|
||||||
|
if err := producer.Close(); err != nil {
|
||||||
|
logger.Println("Failed to close Kafka producer cleanly:", err)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
partition, offset, err := producer.SendMessage(message)
|
||||||
|
if err != nil {
|
||||||
|
printErrorAndExit(69, "Failed to produce message: %s", err)
|
||||||
|
} else if !*silent {
|
||||||
|
fmt.Printf("topic=%s\tpartition=%d\toffset=%d\n", *topic, partition, offset)
|
||||||
|
}
|
||||||
|
if *showMetrics {
|
||||||
|
metrics.WriteOnce(config.MetricRegistry, os.Stderr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func printErrorAndExit(code int, format string, values ...interface{}) {
|
||||||
|
fmt.Fprintf(os.Stderr, "ERROR: %s\n", fmt.Sprintf(format, values...))
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
os.Exit(code)
|
||||||
|
}
|
||||||
|
|
||||||
|
func printUsageErrorAndExit(message string) {
|
||||||
|
fmt.Fprintln(os.Stderr, "ERROR:", message)
|
||||||
|
fmt.Fprintln(os.Stderr)
|
||||||
|
fmt.Fprintln(os.Stderr, "Available command line options:")
|
||||||
|
flag.PrintDefaults()
|
||||||
|
os.Exit(64)
|
||||||
|
}
|
||||||
|
|
||||||
|
func stdinAvailable() bool {
|
||||||
|
stat, _ := os.Stdin.Stat()
|
||||||
|
return (stat.Mode() & os.ModeCharDevice) == 0
|
||||||
|
}
|
22
vendor/github.com/Shopify/sarama/vagrant/boot_cluster.sh
generated
vendored
Normal file
22
vendor/github.com/Shopify/sarama/vagrant/boot_cluster.sh
generated
vendored
Normal file
|
@ -0,0 +1,22 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
# Launch and wait for toxiproxy
|
||||||
|
${REPOSITORY_ROOT}/vagrant/run_toxiproxy.sh &
|
||||||
|
while ! nc -q 1 localhost 2181 </dev/null; do echo "Waiting"; sleep 1; done
|
||||||
|
while ! nc -q 1 localhost 9092 </dev/null; do echo "Waiting"; sleep 1; done
|
||||||
|
|
||||||
|
# Launch and wait for Zookeeper
|
||||||
|
for i in 1 2 3 4 5; do
|
||||||
|
KAFKA_PORT=`expr $i + 9090`
|
||||||
|
cd ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT} && bin/zookeeper-server-start.sh -daemon config/zookeeper.properties
|
||||||
|
done
|
||||||
|
while ! nc -q 1 localhost 21805 </dev/null; do echo "Waiting"; sleep 1; done
|
||||||
|
|
||||||
|
# Launch and wait for Kafka
|
||||||
|
for i in 1 2 3 4 5; do
|
||||||
|
KAFKA_PORT=`expr $i + 9090`
|
||||||
|
cd ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT} && bin/kafka-server-start.sh -daemon config/server.properties
|
||||||
|
done
|
||||||
|
while ! nc -q 1 localhost 29095 </dev/null; do echo "Waiting"; sleep 1; done
|
8
vendor/github.com/Shopify/sarama/vagrant/create_topics.sh
generated
vendored
Normal file
8
vendor/github.com/Shopify/sarama/vagrant/create_topics.sh
generated
vendored
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
cd ${KAFKA_INSTALL_ROOT}/kafka-9092
|
||||||
|
bin/kafka-topics.sh --create --partitions 1 --replication-factor 3 --topic test.1 --zookeeper localhost:2181
|
||||||
|
bin/kafka-topics.sh --create --partitions 4 --replication-factor 3 --topic test.4 --zookeeper localhost:2181
|
||||||
|
bin/kafka-topics.sh --create --partitions 64 --replication-factor 3 --topic test.64 --zookeeper localhost:2181
|
49
vendor/github.com/Shopify/sarama/vagrant/install_cluster.sh
generated
vendored
Normal file
49
vendor/github.com/Shopify/sarama/vagrant/install_cluster.sh
generated
vendored
Normal file
|
@ -0,0 +1,49 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
TOXIPROXY_VERSION=2.0.0
|
||||||
|
|
||||||
|
mkdir -p ${KAFKA_INSTALL_ROOT}
|
||||||
|
if [ ! -f ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_VERSION}.tgz ]; then
|
||||||
|
wget --quiet http://apache.mirror.gtcomm.net/kafka/${KAFKA_VERSION}/kafka_2.11-${KAFKA_VERSION}.tgz -O ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_VERSION}.tgz
|
||||||
|
fi
|
||||||
|
if [ ! -f ${KAFKA_INSTALL_ROOT}/toxiproxy-${TOXIPROXY_VERSION} ]; then
|
||||||
|
wget --quiet https://github.com/Shopify/toxiproxy/releases/download/v${TOXIPROXY_VERSION}/toxiproxy-server-linux-amd64 -O ${KAFKA_INSTALL_ROOT}/toxiproxy-${TOXIPROXY_VERSION}
|
||||||
|
chmod +x ${KAFKA_INSTALL_ROOT}/toxiproxy-${TOXIPROXY_VERSION}
|
||||||
|
fi
|
||||||
|
rm -f ${KAFKA_INSTALL_ROOT}/toxiproxy
|
||||||
|
ln -s ${KAFKA_INSTALL_ROOT}/toxiproxy-${TOXIPROXY_VERSION} ${KAFKA_INSTALL_ROOT}/toxiproxy
|
||||||
|
|
||||||
|
for i in 1 2 3 4 5; do
|
||||||
|
ZK_PORT=`expr $i + 2180`
|
||||||
|
ZK_PORT_REAL=`expr $i + 21800`
|
||||||
|
KAFKA_PORT=`expr $i + 9090`
|
||||||
|
KAFKA_PORT_REAL=`expr $i + 29090`
|
||||||
|
|
||||||
|
# unpack kafka
|
||||||
|
mkdir -p ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}
|
||||||
|
tar xzf ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_VERSION}.tgz -C ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT} --strip-components 1
|
||||||
|
|
||||||
|
# broker configuration
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/server.properties ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/
|
||||||
|
sed -i s/KAFKAID/${KAFKA_PORT}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/server.properties
|
||||||
|
sed -i s/KAFKAPORT/${KAFKA_PORT_REAL}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/server.properties
|
||||||
|
sed -i s/KAFKA_HOSTNAME/${KAFKA_HOSTNAME}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/server.properties
|
||||||
|
sed -i s/ZK_PORT/${ZK_PORT}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/server.properties
|
||||||
|
|
||||||
|
KAFKA_DATADIR="${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/data"
|
||||||
|
mkdir -p ${KAFKA_DATADIR}
|
||||||
|
sed -i s#KAFKA_DATADIR#${KAFKA_DATADIR}#g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/server.properties
|
||||||
|
|
||||||
|
# zookeeper configuration
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/zookeeper.properties ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/
|
||||||
|
sed -i s/KAFKAID/${KAFKA_PORT}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/zookeeper.properties
|
||||||
|
sed -i s/ZK_PORT/${ZK_PORT_REAL}/g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/zookeeper.properties
|
||||||
|
|
||||||
|
ZK_DATADIR="${KAFKA_INSTALL_ROOT}/zookeeper-${ZK_PORT}"
|
||||||
|
mkdir -p ${ZK_DATADIR}
|
||||||
|
sed -i s#ZK_DATADIR#${ZK_DATADIR}#g ${KAFKA_INSTALL_ROOT}/kafka-${KAFKA_PORT}/config/zookeeper.properties
|
||||||
|
|
||||||
|
echo $i > ${KAFKA_INSTALL_ROOT}/zookeeper-${ZK_PORT}/myid
|
||||||
|
done
|
9
vendor/github.com/Shopify/sarama/vagrant/kafka.conf
generated
vendored
Normal file
9
vendor/github.com/Shopify/sarama/vagrant/kafka.conf
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
start on started zookeeper-ZK_PORT
|
||||||
|
stop on stopping zookeeper-ZK_PORT
|
||||||
|
|
||||||
|
# Use a script instead of exec (using env stanza leaks KAFKA_HEAP_OPTS from zookeeper)
|
||||||
|
script
|
||||||
|
sleep 2
|
||||||
|
export KAFKA_HEAP_OPTS="-Xmx320m"
|
||||||
|
exec /opt/kafka-KAFKAID/bin/kafka-server-start.sh /opt/kafka-KAFKAID/config/server.properties
|
||||||
|
end script
|
15
vendor/github.com/Shopify/sarama/vagrant/provision.sh
generated
vendored
Normal file
15
vendor/github.com/Shopify/sarama/vagrant/provision.sh
generated
vendored
Normal file
|
@ -0,0 +1,15 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
apt-get update
|
||||||
|
yes | apt-get install default-jre
|
||||||
|
|
||||||
|
export KAFKA_INSTALL_ROOT=/opt
|
||||||
|
export KAFKA_HOSTNAME=192.168.100.67
|
||||||
|
export KAFKA_VERSION=0.9.0.1
|
||||||
|
export REPOSITORY_ROOT=/vagrant
|
||||||
|
|
||||||
|
sh /vagrant/vagrant/install_cluster.sh
|
||||||
|
sh /vagrant/vagrant/setup_services.sh
|
||||||
|
sh /vagrant/vagrant/create_topics.sh
|
22
vendor/github.com/Shopify/sarama/vagrant/run_toxiproxy.sh
generated
vendored
Normal file
22
vendor/github.com/Shopify/sarama/vagrant/run_toxiproxy.sh
generated
vendored
Normal file
|
@ -0,0 +1,22 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
${KAFKA_INSTALL_ROOT}/toxiproxy -port 8474 -host 0.0.0.0 &
|
||||||
|
PID=$!
|
||||||
|
|
||||||
|
while ! nc -q 1 localhost 8474 </dev/null; do echo "Waiting"; sleep 1; done
|
||||||
|
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"zk1", "upstream":"localhost:21801", "listen":"0.0.0.0:2181"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"zk2", "upstream":"localhost:21802", "listen":"0.0.0.0:2182"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"zk3", "upstream":"localhost:21803", "listen":"0.0.0.0:2183"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"zk4", "upstream":"localhost:21804", "listen":"0.0.0.0:2184"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"zk5", "upstream":"localhost:21805", "listen":"0.0.0.0:2185"}' localhost:8474/proxies
|
||||||
|
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"kafka1", "upstream":"localhost:29091", "listen":"0.0.0.0:9091"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"kafka2", "upstream":"localhost:29092", "listen":"0.0.0.0:9092"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"kafka3", "upstream":"localhost:29093", "listen":"0.0.0.0:9093"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"kafka4", "upstream":"localhost:29094", "listen":"0.0.0.0:9094"}' localhost:8474/proxies
|
||||||
|
wget -O/dev/null -S --post-data='{"name":"kafka5", "upstream":"localhost:29095", "listen":"0.0.0.0:9095"}' localhost:8474/proxies
|
||||||
|
|
||||||
|
wait $PID
|
127
vendor/github.com/Shopify/sarama/vagrant/server.properties
generated
vendored
Normal file
127
vendor/github.com/Shopify/sarama/vagrant/server.properties
generated
vendored
Normal file
|
@ -0,0 +1,127 @@
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
# contributor license agreements. See the NOTICE file distributed with
|
||||||
|
# this work for additional information regarding copyright ownership.
|
||||||
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
# (the "License"); you may not use this file except in compliance with
|
||||||
|
# the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
# see kafka.server.KafkaConfig for additional details and defaults
|
||||||
|
|
||||||
|
############################# Server Basics #############################
|
||||||
|
|
||||||
|
# The id of the broker. This must be set to a unique integer for each broker.
|
||||||
|
broker.id=KAFKAID
|
||||||
|
reserved.broker.max.id=10000
|
||||||
|
|
||||||
|
############################# Socket Server Settings #############################
|
||||||
|
|
||||||
|
# The port the socket server listens on
|
||||||
|
port=KAFKAPORT
|
||||||
|
|
||||||
|
# Hostname the broker will bind to. If not set, the server will bind to all interfaces
|
||||||
|
host.name=localhost
|
||||||
|
|
||||||
|
# Hostname the broker will advertise to producers and consumers. If not set, it uses the
|
||||||
|
# value for "host.name" if configured. Otherwise, it will use the value returned from
|
||||||
|
# java.net.InetAddress.getCanonicalHostName().
|
||||||
|
advertised.host.name=KAFKA_HOSTNAME
|
||||||
|
advertised.port=KAFKAID
|
||||||
|
|
||||||
|
# The port to publish to ZooKeeper for clients to use. If this is not set,
|
||||||
|
# it will publish the same port that the broker binds to.
|
||||||
|
# advertised.port=<port accessible by clients>
|
||||||
|
|
||||||
|
# The number of threads handling network requests
|
||||||
|
num.network.threads=2
|
||||||
|
|
||||||
|
# The number of threads doing disk I/O
|
||||||
|
num.io.threads=8
|
||||||
|
|
||||||
|
# The send buffer (SO_SNDBUF) used by the socket server
|
||||||
|
socket.send.buffer.bytes=1048576
|
||||||
|
|
||||||
|
# The receive buffer (SO_RCVBUF) used by the socket server
|
||||||
|
socket.receive.buffer.bytes=1048576
|
||||||
|
|
||||||
|
# The maximum size of a request that the socket server will accept (protection against OOM)
|
||||||
|
socket.request.max.bytes=104857600
|
||||||
|
|
||||||
|
|
||||||
|
############################# Log Basics #############################
|
||||||
|
|
||||||
|
# A comma seperated list of directories under which to store log files
|
||||||
|
log.dirs=KAFKA_DATADIR
|
||||||
|
|
||||||
|
# The default number of log partitions per topic. More partitions allow greater
|
||||||
|
# parallelism for consumption, but this will also result in more files across
|
||||||
|
# the brokers.
|
||||||
|
num.partitions=2
|
||||||
|
|
||||||
|
# Create new topics with a replication factor of 2 so failover can be tested
|
||||||
|
# more easily.
|
||||||
|
default.replication.factor=2
|
||||||
|
|
||||||
|
auto.create.topics.enable=false
|
||||||
|
delete.topic.enable=true
|
||||||
|
|
||||||
|
############################# Log Flush Policy #############################
|
||||||
|
|
||||||
|
# Messages are immediately written to the filesystem but by default we only fsync() to sync
|
||||||
|
# the OS cache lazily. The following configurations control the flush of data to disk.
|
||||||
|
# There are a few important trade-offs here:
|
||||||
|
# 1. Durability: Unflushed data may be lost if you are not using replication.
|
||||||
|
# 2. Latency: Very large flush intervals may lead to latency spikes when the flush does occur as there will be a lot of data to flush.
|
||||||
|
# 3. Throughput: The flush is generally the most expensive operation, and a small flush interval may lead to exceessive seeks.
|
||||||
|
# The settings below allow one to configure the flush policy to flush data after a period of time or
|
||||||
|
# every N messages (or both). This can be done globally and overridden on a per-topic basis.
|
||||||
|
|
||||||
|
# The number of messages to accept before forcing a flush of data to disk
|
||||||
|
#log.flush.interval.messages=10000
|
||||||
|
|
||||||
|
# The maximum amount of time a message can sit in a log before we force a flush
|
||||||
|
#log.flush.interval.ms=1000
|
||||||
|
|
||||||
|
############################# Log Retention Policy #############################
|
||||||
|
|
||||||
|
# The following configurations control the disposal of log segments. The policy can
|
||||||
|
# be set to delete segments after a period of time, or after a given size has accumulated.
|
||||||
|
# A segment will be deleted whenever *either* of these criteria are met. Deletion always happens
|
||||||
|
# from the end of the log.
|
||||||
|
|
||||||
|
# The minimum age of a log file to be eligible for deletion
|
||||||
|
log.retention.hours=168
|
||||||
|
|
||||||
|
# A size-based retention policy for logs. Segments are pruned from the log as long as the remaining
|
||||||
|
# segments don't drop below log.retention.bytes.
|
||||||
|
log.retention.bytes=268435456
|
||||||
|
|
||||||
|
# The maximum size of a log segment file. When this size is reached a new log segment will be created.
|
||||||
|
log.segment.bytes=268435456
|
||||||
|
|
||||||
|
# The interval at which log segments are checked to see if they can be deleted according
|
||||||
|
# to the retention policies
|
||||||
|
log.retention.check.interval.ms=60000
|
||||||
|
|
||||||
|
# By default the log cleaner is disabled and the log retention policy will default to just delete segments after their retention expires.
|
||||||
|
# If log.cleaner.enable=true is set the cleaner will be enabled and individual logs can then be marked for log compaction.
|
||||||
|
log.cleaner.enable=false
|
||||||
|
|
||||||
|
############################# Zookeeper #############################
|
||||||
|
|
||||||
|
# Zookeeper connection string (see zookeeper docs for details).
|
||||||
|
# This is a comma separated host:port pairs, each corresponding to a zk
|
||||||
|
# server. e.g. "127.0.0.1:3000,127.0.0.1:3001,127.0.0.1:3002".
|
||||||
|
# You can also append an optional chroot string to the urls to specify the
|
||||||
|
# root directory for all kafka znodes.
|
||||||
|
zookeeper.connect=localhost:ZK_PORT
|
||||||
|
|
||||||
|
# Timeout in ms for connecting to zookeeper
|
||||||
|
zookeeper.session.timeout.ms=3000
|
||||||
|
zookeeper.connection.timeout.ms=3000
|
29
vendor/github.com/Shopify/sarama/vagrant/setup_services.sh
generated
vendored
Normal file
29
vendor/github.com/Shopify/sarama/vagrant/setup_services.sh
generated
vendored
Normal file
|
@ -0,0 +1,29 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -ex
|
||||||
|
|
||||||
|
stop toxiproxy || true
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/toxiproxy.conf /etc/init/toxiproxy.conf
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/run_toxiproxy.sh ${KAFKA_INSTALL_ROOT}/
|
||||||
|
start toxiproxy
|
||||||
|
|
||||||
|
for i in 1 2 3 4 5; do
|
||||||
|
ZK_PORT=`expr $i + 2180`
|
||||||
|
KAFKA_PORT=`expr $i + 9090`
|
||||||
|
|
||||||
|
stop zookeeper-${ZK_PORT} || true
|
||||||
|
|
||||||
|
# set up zk service
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/zookeeper.conf /etc/init/zookeeper-${ZK_PORT}.conf
|
||||||
|
sed -i s/KAFKAID/${KAFKA_PORT}/g /etc/init/zookeeper-${ZK_PORT}.conf
|
||||||
|
|
||||||
|
# set up kafka service
|
||||||
|
cp ${REPOSITORY_ROOT}/vagrant/kafka.conf /etc/init/kafka-${KAFKA_PORT}.conf
|
||||||
|
sed -i s/KAFKAID/${KAFKA_PORT}/g /etc/init/kafka-${KAFKA_PORT}.conf
|
||||||
|
sed -i s/ZK_PORT/${ZK_PORT}/g /etc/init/kafka-${KAFKA_PORT}.conf
|
||||||
|
|
||||||
|
start zookeeper-${ZK_PORT}
|
||||||
|
done
|
||||||
|
|
||||||
|
# Wait for the last kafka node to finish booting
|
||||||
|
while ! nc -q 1 localhost 29095 </dev/null; do echo "Waiting"; sleep 1; done
|
6
vendor/github.com/Shopify/sarama/vagrant/toxiproxy.conf
generated
vendored
Normal file
6
vendor/github.com/Shopify/sarama/vagrant/toxiproxy.conf
generated
vendored
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
start on started networking
|
||||||
|
stop on shutdown
|
||||||
|
|
||||||
|
env KAFKA_INSTALL_ROOT=/opt
|
||||||
|
|
||||||
|
exec /opt/run_toxiproxy.sh
|
7
vendor/github.com/Shopify/sarama/vagrant/zookeeper.conf
generated
vendored
Normal file
7
vendor/github.com/Shopify/sarama/vagrant/zookeeper.conf
generated
vendored
Normal file
|
@ -0,0 +1,7 @@
|
||||||
|
start on started toxiproxy
|
||||||
|
stop on stopping toxiproxy
|
||||||
|
|
||||||
|
script
|
||||||
|
export KAFKA_HEAP_OPTS="-Xmx192m"
|
||||||
|
exec /opt/kafka-KAFKAID/bin/zookeeper-server-start.sh /opt/kafka-KAFKAID/config/zookeeper.properties
|
||||||
|
end script
|
36
vendor/github.com/Shopify/sarama/vagrant/zookeeper.properties
generated
vendored
Normal file
36
vendor/github.com/Shopify/sarama/vagrant/zookeeper.properties
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
# contributor license agreements. See the NOTICE file distributed with
|
||||||
|
# this work for additional information regarding copyright ownership.
|
||||||
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
# (the "License"); you may not use this file except in compliance with
|
||||||
|
# the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
# the directory where the snapshot is stored.
|
||||||
|
dataDir=ZK_DATADIR
|
||||||
|
# the port at which the clients will connect
|
||||||
|
clientPort=ZK_PORT
|
||||||
|
# disable the per-ip limit on the number of connections since this is a non-production config
|
||||||
|
maxClientCnxns=0
|
||||||
|
|
||||||
|
# The number of milliseconds of each tick
|
||||||
|
tickTime=2000
|
||||||
|
|
||||||
|
# The number of ticks that the initial synchronization phase can take
|
||||||
|
initLimit=10
|
||||||
|
|
||||||
|
# The number of ticks that can pass between
|
||||||
|
# sending a request and getting an acknowledgement
|
||||||
|
syncLimit=5
|
||||||
|
|
||||||
|
server.1=localhost:2281:2381
|
||||||
|
server.2=localhost:2282:2382
|
||||||
|
server.3=localhost:2283:2383
|
||||||
|
server.4=localhost:2284:2384
|
||||||
|
server.5=localhost:2285:2385
|
54
vendor/github.com/apache/thrift/aclocal/ac_prog_bison.m4
generated
vendored
Normal file
54
vendor/github.com/apache/thrift/aclocal/ac_prog_bison.m4
generated
vendored
Normal file
|
@ -0,0 +1,54 @@
|
||||||
|
dnl
|
||||||
|
dnl Check Bison version
|
||||||
|
dnl AC_PROG_BISON([MIN_VERSION=2.4])
|
||||||
|
dnl
|
||||||
|
dnl Will define BISON_USE_PARSER_H_EXTENSION if Automake is < 1.11
|
||||||
|
dnl for use with .h includes.
|
||||||
|
dnl
|
||||||
|
|
||||||
|
AC_DEFUN([AC_PROG_BISON], [
|
||||||
|
if test "x$1" = "x" ; then
|
||||||
|
bison_required_version="2.4"
|
||||||
|
else
|
||||||
|
bison_required_version="$1"
|
||||||
|
fi
|
||||||
|
|
||||||
|
AC_CHECK_PROG(have_prog_bison, [bison], [yes],[no])
|
||||||
|
|
||||||
|
AC_DEFINE_UNQUOTED([BISON_VERSION], [0.0], [Bison version if bison is not available])
|
||||||
|
|
||||||
|
#Do not use *.h extension for parser header files, use newer *.hh
|
||||||
|
bison_use_parser_h_extension=false
|
||||||
|
|
||||||
|
if test "$have_prog_bison" = "yes" ; then
|
||||||
|
AC_MSG_CHECKING([for bison version >= $bison_required_version])
|
||||||
|
bison_version=`bison --version | head -n 1 | cut '-d ' -f 4`
|
||||||
|
AC_DEFINE_UNQUOTED([BISON_VERSION], [$bison_version], [Defines bison version])
|
||||||
|
if test "$bison_version" \< "$bison_required_version" ; then
|
||||||
|
BISON=:
|
||||||
|
AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([Bison version $bison_required_version or higher must be installed on the system!])
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
BISON=bison
|
||||||
|
AC_SUBST(BISON)
|
||||||
|
|
||||||
|
#Verify automake version 1.11 headers for yy files are .h, > 1.12 uses .hh
|
||||||
|
automake_version=`automake --version | head -n 1 | cut '-d ' -f 4`
|
||||||
|
AC_DEFINE_UNQUOTED([AUTOMAKE_VERSION], [$automake_version], [Defines automake version])
|
||||||
|
|
||||||
|
if test "$automake_version" \< "1.12" ; then
|
||||||
|
#Use *.h extension for parser header file
|
||||||
|
bison_use_parser_h_extension=true
|
||||||
|
echo "Automake version < 1.12"
|
||||||
|
AC_DEFINE([BISON_USE_PARSER_H_EXTENSION], [1], [Use *.h extension for parser header file])
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
BISON=:
|
||||||
|
AC_MSG_RESULT([NO])
|
||||||
|
fi
|
||||||
|
|
||||||
|
AM_CONDITIONAL([BISON_USE_PARSER_H_EXTENSION], [test x$bison_use_parser_h_extension = xtrue])
|
||||||
|
AC_SUBST(BISON)
|
||||||
|
])
|
301
vendor/github.com/apache/thrift/aclocal/ax_boost_base.m4
generated
vendored
Normal file
301
vendor/github.com/apache/thrift/aclocal/ax_boost_base.m4
generated
vendored
Normal file
|
@ -0,0 +1,301 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_boost_base.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_BOOST_BASE([MINIMUM-VERSION], [ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Test for the Boost C++ libraries of a particular version (or newer)
|
||||||
|
#
|
||||||
|
# If no path to the installed boost library is given the macro searchs
|
||||||
|
# under /usr, /usr/local, /opt and /opt/local and evaluates the
|
||||||
|
# $BOOST_ROOT environment variable. Further documentation is available at
|
||||||
|
# <http://randspringer.de/boost/index.html>.
|
||||||
|
#
|
||||||
|
# This macro calls:
|
||||||
|
#
|
||||||
|
# AC_SUBST(BOOST_CPPFLAGS) / AC_SUBST(BOOST_LDFLAGS)
|
||||||
|
#
|
||||||
|
# And sets:
|
||||||
|
#
|
||||||
|
# HAVE_BOOST
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2008 Thomas Porschberg <thomas@randspringer.de>
|
||||||
|
# Copyright (c) 2009 Peter Adolphs
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 42
|
||||||
|
|
||||||
|
# example boost program (need to pass version)
|
||||||
|
m4_define([_AX_BOOST_BASE_PROGRAM],
|
||||||
|
[AC_LANG_PROGRAM([[
|
||||||
|
#include <boost/version.hpp>
|
||||||
|
]],[[
|
||||||
|
(void) ((void)sizeof(char[1 - 2*!!((BOOST_VERSION) < ($1))]));
|
||||||
|
]])])
|
||||||
|
|
||||||
|
AC_DEFUN([AX_BOOST_BASE],
|
||||||
|
[
|
||||||
|
AC_ARG_WITH([boost],
|
||||||
|
[AS_HELP_STRING([--with-boost@<:@=ARG@:>@],
|
||||||
|
[use Boost library from a standard location (ARG=yes),
|
||||||
|
from the specified location (ARG=<path>),
|
||||||
|
or disable it (ARG=no)
|
||||||
|
@<:@ARG=yes@:>@ ])],
|
||||||
|
[
|
||||||
|
AS_CASE([$withval],
|
||||||
|
[no],[want_boost="no";_AX_BOOST_BASE_boost_path=""],
|
||||||
|
[yes],[want_boost="yes";_AX_BOOST_BASE_boost_path=""],
|
||||||
|
[want_boost="yes";_AX_BOOST_BASE_boost_path="$withval"])
|
||||||
|
],
|
||||||
|
[want_boost="yes"])
|
||||||
|
|
||||||
|
|
||||||
|
AC_ARG_WITH([boost-libdir],
|
||||||
|
[AS_HELP_STRING([--with-boost-libdir=LIB_DIR],
|
||||||
|
[Force given directory for boost libraries.
|
||||||
|
Note that this will override library path detection,
|
||||||
|
so use this parameter only if default library detection fails
|
||||||
|
and you know exactly where your boost libraries are located.])],
|
||||||
|
[
|
||||||
|
AS_IF([test -d "$withval"],
|
||||||
|
[_AX_BOOST_BASE_boost_lib_path="$withval"],
|
||||||
|
[AC_MSG_ERROR([--with-boost-libdir expected directory name])])
|
||||||
|
],
|
||||||
|
[_AX_BOOST_BASE_boost_lib_path=""])
|
||||||
|
|
||||||
|
BOOST_LDFLAGS=""
|
||||||
|
BOOST_CPPFLAGS=""
|
||||||
|
AS_IF([test "x$want_boost" = "xyes"],
|
||||||
|
[_AX_BOOST_BASE_RUNDETECT([$1],[$2],[$3])])
|
||||||
|
AC_SUBST(BOOST_CPPFLAGS)
|
||||||
|
AC_SUBST(BOOST_LDFLAGS)
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
# convert a version string in $2 to numeric and affect to polymorphic var $1
|
||||||
|
AC_DEFUN([_AX_BOOST_BASE_TONUMERICVERSION],[
|
||||||
|
AS_IF([test "x$2" = "x"],[_AX_BOOST_BASE_TONUMERICVERSION_req="1.20.0"],[_AX_BOOST_BASE_TONUMERICVERSION_req="$2"])
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION_req_shorten=`expr $_AX_BOOST_BASE_TONUMERICVERSION_req : '\([[0-9]]*\.[[0-9]]*\)'`
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION_req_major=`expr $_AX_BOOST_BASE_TONUMERICVERSION_req : '\([[0-9]]*\)'`
|
||||||
|
AS_IF([test "x$_AX_BOOST_BASE_TONUMERICVERSION_req_major" = "x"],
|
||||||
|
[AC_MSG_ERROR([You should at least specify libboost major version])])
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION_req_minor=`expr $_AX_BOOST_BASE_TONUMERICVERSION_req : '[[0-9]]*\.\([[0-9]]*\)'`
|
||||||
|
AS_IF([test "x$_AX_BOOST_BASE_TONUMERICVERSION_req_minor" = "x"],
|
||||||
|
[_AX_BOOST_BASE_TONUMERICVERSION_req_minor="0"])
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION_req_sub_minor=`expr $_AX_BOOST_BASE_TONUMERICVERSION_req : '[[0-9]]*\.[[0-9]]*\.\([[0-9]]*\)'`
|
||||||
|
AS_IF([test "X$_AX_BOOST_BASE_TONUMERICVERSION_req_sub_minor" = "X"],
|
||||||
|
[_AX_BOOST_BASE_TONUMERICVERSION_req_sub_minor="0"])
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION_RET=`expr $_AX_BOOST_BASE_TONUMERICVERSION_req_major \* 100000 \+ $_AX_BOOST_BASE_TONUMERICVERSION_req_minor \* 100 \+ $_AX_BOOST_BASE_TONUMERICVERSION_req_sub_minor`
|
||||||
|
AS_VAR_SET($1,$_AX_BOOST_BASE_TONUMERICVERSION_RET)
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Run the detection of boost should be run only if $want_boost
|
||||||
|
AC_DEFUN([_AX_BOOST_BASE_RUNDETECT],[
|
||||||
|
_AX_BOOST_BASE_TONUMERICVERSION(WANT_BOOST_VERSION,[$1])
|
||||||
|
succeeded=no
|
||||||
|
|
||||||
|
|
||||||
|
AC_REQUIRE([AC_CANONICAL_HOST])
|
||||||
|
dnl On 64-bit systems check for system libraries in both lib64 and lib.
|
||||||
|
dnl The former is specified by FHS, but e.g. Debian does not adhere to
|
||||||
|
dnl this (as it rises problems for generic multi-arch support).
|
||||||
|
dnl The last entry in the list is chosen by default when no libraries
|
||||||
|
dnl are found, e.g. when only header-only libraries are installed!
|
||||||
|
AS_CASE([${host_cpu}],
|
||||||
|
[x86_64],[libsubdirs="lib64 libx32 lib lib64"],
|
||||||
|
[ppc64|s390x|sparc64|aarch64|ppc64le],[libsubdirs="lib64 lib lib64"],
|
||||||
|
[libsubdirs="lib"],
|
||||||
|
)
|
||||||
|
|
||||||
|
dnl allow for real multi-arch paths e.g. /usr/lib/x86_64-linux-gnu. Give
|
||||||
|
dnl them priority over the other paths since, if libs are found there, they
|
||||||
|
dnl are almost assuredly the ones desired.
|
||||||
|
AS_CASE([${host_cpu}],
|
||||||
|
[i?86],[multiarch_libsubdir="lib/i386-${host_os}"],
|
||||||
|
[multiarch_libsubdir="lib/${host_cpu}-${host_os}"]
|
||||||
|
)
|
||||||
|
|
||||||
|
dnl first we check the system location for boost libraries
|
||||||
|
dnl this location ist chosen if boost libraries are installed with the --layout=system option
|
||||||
|
dnl or if you install boost with RPM
|
||||||
|
AS_IF([test "x$_AX_BOOST_BASE_boost_path" != "x"],[
|
||||||
|
AC_MSG_CHECKING([for boostlib >= $1 ($WANT_BOOST_VERSION) includes in "$_AX_BOOST_BASE_boost_path/include"])
|
||||||
|
AS_IF([test -d "$_AX_BOOST_BASE_boost_path/include" && test -r "$_AX_BOOST_BASE_boost_path/include"],[
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
BOOST_CPPFLAGS="-I$_AX_BOOST_BASE_boost_path/include"
|
||||||
|
for _AX_BOOST_BASE_boost_path_tmp in $multiarch_libsubdir $libsubdirs; do
|
||||||
|
AC_MSG_CHECKING([for boostlib >= $1 ($WANT_BOOST_VERSION) lib path in "$_AX_BOOST_BASE_boost_path/$_AX_BOOST_BASE_boost_path_tmp"])
|
||||||
|
AS_IF([test -d "$_AX_BOOST_BASE_boost_path/$_AX_BOOST_BASE_boost_path_tmp" && test -r "$_AX_BOOST_BASE_boost_path/$_AX_BOOST_BASE_boost_path_tmp" ],[
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
BOOST_LDFLAGS="-L$_AX_BOOST_BASE_boost_path/$_AX_BOOST_BASE_boost_path_tmp";
|
||||||
|
break;
|
||||||
|
],
|
||||||
|
[AC_MSG_RESULT([no])])
|
||||||
|
done],[
|
||||||
|
AC_MSG_RESULT([no])])
|
||||||
|
],[
|
||||||
|
if test X"$cross_compiling" = Xyes; then
|
||||||
|
search_libsubdirs=$multiarch_libsubdir
|
||||||
|
else
|
||||||
|
search_libsubdirs="$multiarch_libsubdir $libsubdirs"
|
||||||
|
fi
|
||||||
|
for _AX_BOOST_BASE_boost_path_tmp in /usr /usr/local /opt /opt/local ; do
|
||||||
|
if test -d "$_AX_BOOST_BASE_boost_path_tmp/include/boost" && test -r "$_AX_BOOST_BASE_boost_path_tmp/include/boost" ; then
|
||||||
|
for libsubdir in $search_libsubdirs ; do
|
||||||
|
if ls "$_AX_BOOST_BASE_boost_path_tmp/$libsubdir/libboost_"* >/dev/null 2>&1 ; then break; fi
|
||||||
|
done
|
||||||
|
BOOST_LDFLAGS="-L$_AX_BOOST_BASE_boost_path_tmp/$libsubdir"
|
||||||
|
BOOST_CPPFLAGS="-I$_AX_BOOST_BASE_boost_path_tmp/include"
|
||||||
|
break;
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl overwrite ld flags if we have required special directory with
|
||||||
|
dnl --with-boost-libdir parameter
|
||||||
|
AS_IF([test "x$_AX_BOOST_BASE_boost_lib_path" != "x"],
|
||||||
|
[BOOST_LDFLAGS="-L$_AX_BOOST_BASE_boost_lib_path"])
|
||||||
|
|
||||||
|
AC_MSG_CHECKING([for boostlib >= $1 ($WANT_BOOST_VERSION)])
|
||||||
|
CPPFLAGS_SAVED="$CPPFLAGS"
|
||||||
|
CPPFLAGS="$CPPFLAGS $BOOST_CPPFLAGS"
|
||||||
|
export CPPFLAGS
|
||||||
|
|
||||||
|
LDFLAGS_SAVED="$LDFLAGS"
|
||||||
|
LDFLAGS="$LDFLAGS $BOOST_LDFLAGS"
|
||||||
|
export LDFLAGS
|
||||||
|
|
||||||
|
AC_REQUIRE([AC_PROG_CXX])
|
||||||
|
AC_LANG_PUSH(C++)
|
||||||
|
AC_COMPILE_IFELSE([_AX_BOOST_BASE_PROGRAM($WANT_BOOST_VERSION)],[
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
succeeded=yes
|
||||||
|
found_system=yes
|
||||||
|
],[
|
||||||
|
])
|
||||||
|
AC_LANG_POP([C++])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
dnl if we found no boost with system layout we search for boost libraries
|
||||||
|
dnl built and installed without the --layout=system option or for a staged(not installed) version
|
||||||
|
if test "x$succeeded" != "xyes" ; then
|
||||||
|
CPPFLAGS="$CPPFLAGS_SAVED"
|
||||||
|
LDFLAGS="$LDFLAGS_SAVED"
|
||||||
|
BOOST_CPPFLAGS=
|
||||||
|
if test -z "$_AX_BOOST_BASE_boost_lib_path" ; then
|
||||||
|
BOOST_LDFLAGS=
|
||||||
|
fi
|
||||||
|
_version=0
|
||||||
|
if test -n "$_AX_BOOST_BASE_boost_path" ; then
|
||||||
|
if test -d "$_AX_BOOST_BASE_boost_path" && test -r "$_AX_BOOST_BASE_boost_path"; then
|
||||||
|
for i in `ls -d $_AX_BOOST_BASE_boost_path/include/boost-* 2>/dev/null`; do
|
||||||
|
_version_tmp=`echo $i | sed "s#$_AX_BOOST_BASE_boost_path##" | sed 's/\/include\/boost-//' | sed 's/_/./'`
|
||||||
|
V_CHECK=`expr $_version_tmp \> $_version`
|
||||||
|
if test "x$V_CHECK" = "x1" ; then
|
||||||
|
_version=$_version_tmp
|
||||||
|
fi
|
||||||
|
VERSION_UNDERSCORE=`echo $_version | sed 's/\./_/'`
|
||||||
|
BOOST_CPPFLAGS="-I$_AX_BOOST_BASE_boost_path/include/boost-$VERSION_UNDERSCORE"
|
||||||
|
done
|
||||||
|
dnl if nothing found search for layout used in Windows distributions
|
||||||
|
if test -z "$BOOST_CPPFLAGS"; then
|
||||||
|
if test -d "$_AX_BOOST_BASE_boost_path/boost" && test -r "$_AX_BOOST_BASE_boost_path/boost"; then
|
||||||
|
BOOST_CPPFLAGS="-I$_AX_BOOST_BASE_boost_path"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
dnl if we found something and BOOST_LDFLAGS was unset before
|
||||||
|
dnl (because "$_AX_BOOST_BASE_boost_lib_path" = ""), set it here.
|
||||||
|
if test -n "$BOOST_CPPFLAGS" && test -z "$BOOST_LDFLAGS"; then
|
||||||
|
for libsubdir in $libsubdirs ; do
|
||||||
|
if ls "$_AX_BOOST_BASE_boost_path/$libsubdir/libboost_"* >/dev/null 2>&1 ; then break; fi
|
||||||
|
done
|
||||||
|
BOOST_LDFLAGS="-L$_AX_BOOST_BASE_boost_path/$libsubdir"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
if test "x$cross_compiling" != "xyes" ; then
|
||||||
|
for _AX_BOOST_BASE_boost_path in /usr /usr/local /opt /opt/local ; do
|
||||||
|
if test -d "$_AX_BOOST_BASE_boost_path" && test -r "$_AX_BOOST_BASE_boost_path" ; then
|
||||||
|
for i in `ls -d $_AX_BOOST_BASE_boost_path/include/boost-* 2>/dev/null`; do
|
||||||
|
_version_tmp=`echo $i | sed "s#$_AX_BOOST_BASE_boost_path##" | sed 's/\/include\/boost-//' | sed 's/_/./'`
|
||||||
|
V_CHECK=`expr $_version_tmp \> $_version`
|
||||||
|
if test "x$V_CHECK" = "x1" ; then
|
||||||
|
_version=$_version_tmp
|
||||||
|
best_path=$_AX_BOOST_BASE_boost_path
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
VERSION_UNDERSCORE=`echo $_version | sed 's/\./_/'`
|
||||||
|
BOOST_CPPFLAGS="-I$best_path/include/boost-$VERSION_UNDERSCORE"
|
||||||
|
if test -z "$_AX_BOOST_BASE_boost_lib_path" ; then
|
||||||
|
for libsubdir in $libsubdirs ; do
|
||||||
|
if ls "$best_path/$libsubdir/libboost_"* >/dev/null 2>&1 ; then break; fi
|
||||||
|
done
|
||||||
|
BOOST_LDFLAGS="-L$best_path/$libsubdir"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
if test -n "$BOOST_ROOT" ; then
|
||||||
|
for libsubdir in $libsubdirs ; do
|
||||||
|
if ls "$BOOST_ROOT/stage/$libsubdir/libboost_"* >/dev/null 2>&1 ; then break; fi
|
||||||
|
done
|
||||||
|
if test -d "$BOOST_ROOT" && test -r "$BOOST_ROOT" && test -d "$BOOST_ROOT/stage/$libsubdir" && test -r "$BOOST_ROOT/stage/$libsubdir"; then
|
||||||
|
version_dir=`expr //$BOOST_ROOT : '.*/\(.*\)'`
|
||||||
|
stage_version=`echo $version_dir | sed 's/boost_//' | sed 's/_/./g'`
|
||||||
|
stage_version_shorten=`expr $stage_version : '\([[0-9]]*\.[[0-9]]*\)'`
|
||||||
|
V_CHECK=`expr $stage_version_shorten \>\= $_version`
|
||||||
|
if test "x$V_CHECK" = "x1" && test -z "$_AX_BOOST_BASE_boost_lib_path" ; then
|
||||||
|
AC_MSG_NOTICE(We will use a staged boost library from $BOOST_ROOT)
|
||||||
|
BOOST_CPPFLAGS="-I$BOOST_ROOT"
|
||||||
|
BOOST_LDFLAGS="-L$BOOST_ROOT/stage/$libsubdir"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
CPPFLAGS="$CPPFLAGS $BOOST_CPPFLAGS"
|
||||||
|
export CPPFLAGS
|
||||||
|
LDFLAGS="$LDFLAGS $BOOST_LDFLAGS"
|
||||||
|
export LDFLAGS
|
||||||
|
|
||||||
|
AC_LANG_PUSH(C++)
|
||||||
|
AC_COMPILE_IFELSE([_AX_BOOST_BASE_PROGRAM($WANT_BOOST_VERSION)],[
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
succeeded=yes
|
||||||
|
found_system=yes
|
||||||
|
],[
|
||||||
|
])
|
||||||
|
AC_LANG_POP([C++])
|
||||||
|
fi
|
||||||
|
|
||||||
|
if test "x$succeeded" != "xyes" ; then
|
||||||
|
if test "x$_version" = "x0" ; then
|
||||||
|
AC_MSG_NOTICE([[We could not detect the boost libraries (version $1 or higher). If you have a staged boost library (still not installed) please specify \$BOOST_ROOT in your environment and do not give a PATH to --with-boost option. If you are sure you have boost installed, then check your version number looking in <boost/version.hpp>. See http://randspringer.de/boost for more documentation.]])
|
||||||
|
else
|
||||||
|
AC_MSG_NOTICE([Your boost libraries seems to old (version $_version).])
|
||||||
|
fi
|
||||||
|
# execute ACTION-IF-NOT-FOUND (if present):
|
||||||
|
ifelse([$3], , :, [$3])
|
||||||
|
else
|
||||||
|
AC_DEFINE(HAVE_BOOST,,[define if the Boost library is available])
|
||||||
|
# execute ACTION-IF-FOUND (if present):
|
||||||
|
ifelse([$2], , :, [$2])
|
||||||
|
fi
|
||||||
|
|
||||||
|
CPPFLAGS="$CPPFLAGS_SAVED"
|
||||||
|
LDFLAGS="$LDFLAGS_SAVED"
|
||||||
|
|
||||||
|
])
|
124
vendor/github.com/apache/thrift/aclocal/ax_check_openssl.m4
generated
vendored
Normal file
124
vendor/github.com/apache/thrift/aclocal/ax_check_openssl.m4
generated
vendored
Normal file
|
@ -0,0 +1,124 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_check_openssl.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_CHECK_OPENSSL([action-if-found[, action-if-not-found]])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Look for OpenSSL in a number of default spots, or in a user-selected
|
||||||
|
# spot (via --with-openssl). Sets
|
||||||
|
#
|
||||||
|
# OPENSSL_INCLUDES to the include directives required
|
||||||
|
# OPENSSL_LIBS to the -l directives required
|
||||||
|
# OPENSSL_LDFLAGS to the -L or -R flags required
|
||||||
|
#
|
||||||
|
# and calls ACTION-IF-FOUND or ACTION-IF-NOT-FOUND appropriately
|
||||||
|
#
|
||||||
|
# This macro sets OPENSSL_INCLUDES such that source files should use the
|
||||||
|
# openssl/ directory in include directives:
|
||||||
|
#
|
||||||
|
# #include <openssl/hmac.h>
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2009,2010 Zmanda Inc. <http://www.zmanda.com/>
|
||||||
|
# Copyright (c) 2009,2010 Dustin J. Mitchell <dustin@zmanda.com>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 10
|
||||||
|
|
||||||
|
AU_ALIAS([CHECK_SSL], [AX_CHECK_OPENSSL])
|
||||||
|
AC_DEFUN([AX_CHECK_OPENSSL], [
|
||||||
|
found=false
|
||||||
|
AC_ARG_WITH([openssl],
|
||||||
|
[AS_HELP_STRING([--with-openssl=DIR],
|
||||||
|
[root of the OpenSSL directory])],
|
||||||
|
[
|
||||||
|
case "$withval" in
|
||||||
|
"" | y | ye | yes | n | no)
|
||||||
|
AC_MSG_ERROR([Invalid --with-openssl value])
|
||||||
|
;;
|
||||||
|
*) ssldirs="$withval"
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
], [
|
||||||
|
# if pkg-config is installed and openssl has installed a .pc file,
|
||||||
|
# then use that information and don't search ssldirs
|
||||||
|
AC_CHECK_TOOL([PKG_CONFIG], [pkg-config])
|
||||||
|
if test x"$PKG_CONFIG" != x""; then
|
||||||
|
OPENSSL_LDFLAGS=`$PKG_CONFIG openssl --libs-only-L 2>/dev/null`
|
||||||
|
if test $? = 0; then
|
||||||
|
OPENSSL_LIBS=`$PKG_CONFIG openssl --libs-only-l 2>/dev/null`
|
||||||
|
OPENSSL_INCLUDES=`$PKG_CONFIG openssl --cflags-only-I 2>/dev/null`
|
||||||
|
found=true
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
# no such luck; use some default ssldirs
|
||||||
|
if ! $found; then
|
||||||
|
ssldirs="/usr/local/ssl /usr/lib/ssl /usr/ssl /usr/pkg /usr/local /usr"
|
||||||
|
fi
|
||||||
|
]
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
# note that we #include <openssl/foo.h>, so the OpenSSL headers have to be in
|
||||||
|
# an 'openssl' subdirectory
|
||||||
|
|
||||||
|
if ! $found; then
|
||||||
|
OPENSSL_INCLUDES=
|
||||||
|
for ssldir in $ssldirs; do
|
||||||
|
AC_MSG_CHECKING([for openssl/ssl.h in $ssldir])
|
||||||
|
if test -f "$ssldir/include/openssl/ssl.h"; then
|
||||||
|
OPENSSL_INCLUDES="-I$ssldir/include"
|
||||||
|
OPENSSL_LDFLAGS="-L$ssldir/lib"
|
||||||
|
OPENSSL_LIBS="-lssl -lcrypto"
|
||||||
|
found=true
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
break
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT([no])
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
# if the file wasn't found, well, go ahead and try the link anyway -- maybe
|
||||||
|
# it will just work!
|
||||||
|
fi
|
||||||
|
|
||||||
|
# try the preprocessor and linker with our new flags,
|
||||||
|
# being careful not to pollute the global LIBS, LDFLAGS, and CPPFLAGS
|
||||||
|
|
||||||
|
AC_MSG_CHECKING([whether compiling and linking against OpenSSL works])
|
||||||
|
echo "Trying link with OPENSSL_LDFLAGS=$OPENSSL_LDFLAGS;" \
|
||||||
|
"OPENSSL_LIBS=$OPENSSL_LIBS; OPENSSL_INCLUDES=$OPENSSL_INCLUDES" >&AS_MESSAGE_LOG_FD
|
||||||
|
|
||||||
|
save_LIBS="$LIBS"
|
||||||
|
save_LDFLAGS="$LDFLAGS"
|
||||||
|
save_CPPFLAGS="$CPPFLAGS"
|
||||||
|
LDFLAGS="$LDFLAGS $OPENSSL_LDFLAGS"
|
||||||
|
LIBS="$OPENSSL_LIBS $LIBS"
|
||||||
|
CPPFLAGS="$OPENSSL_INCLUDES $CPPFLAGS"
|
||||||
|
AC_LINK_IFELSE(
|
||||||
|
[AC_LANG_PROGRAM([#include <openssl/ssl.h>], [SSL_new(NULL)])],
|
||||||
|
[
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
$1
|
||||||
|
], [
|
||||||
|
AC_MSG_RESULT([no])
|
||||||
|
$2
|
||||||
|
])
|
||||||
|
CPPFLAGS="$save_CPPFLAGS"
|
||||||
|
LDFLAGS="$save_LDFLAGS"
|
||||||
|
LIBS="$save_LIBS"
|
||||||
|
|
||||||
|
AC_SUBST([OPENSSL_INCLUDES])
|
||||||
|
AC_SUBST([OPENSSL_LIBS])
|
||||||
|
AC_SUBST([OPENSSL_LDFLAGS])
|
||||||
|
])
|
177
vendor/github.com/apache/thrift/aclocal/ax_compare_version.m4
generated
vendored
Normal file
177
vendor/github.com/apache/thrift/aclocal/ax_compare_version.m4
generated
vendored
Normal file
|
@ -0,0 +1,177 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_compare_version.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_COMPARE_VERSION(VERSION_A, OP, VERSION_B, [ACTION-IF-TRUE], [ACTION-IF-FALSE])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# This macro compares two version strings. Due to the various number of
|
||||||
|
# minor-version numbers that can exist, and the fact that string
|
||||||
|
# comparisons are not compatible with numeric comparisons, this is not
|
||||||
|
# necessarily trivial to do in a autoconf script. This macro makes doing
|
||||||
|
# these comparisons easy.
|
||||||
|
#
|
||||||
|
# The six basic comparisons are available, as well as checking equality
|
||||||
|
# limited to a certain number of minor-version levels.
|
||||||
|
#
|
||||||
|
# The operator OP determines what type of comparison to do, and can be one
|
||||||
|
# of:
|
||||||
|
#
|
||||||
|
# eq - equal (test A == B)
|
||||||
|
# ne - not equal (test A != B)
|
||||||
|
# le - less than or equal (test A <= B)
|
||||||
|
# ge - greater than or equal (test A >= B)
|
||||||
|
# lt - less than (test A < B)
|
||||||
|
# gt - greater than (test A > B)
|
||||||
|
#
|
||||||
|
# Additionally, the eq and ne operator can have a number after it to limit
|
||||||
|
# the test to that number of minor versions.
|
||||||
|
#
|
||||||
|
# eq0 - equal up to the length of the shorter version
|
||||||
|
# ne0 - not equal up to the length of the shorter version
|
||||||
|
# eqN - equal up to N sub-version levels
|
||||||
|
# neN - not equal up to N sub-version levels
|
||||||
|
#
|
||||||
|
# When the condition is true, shell commands ACTION-IF-TRUE are run,
|
||||||
|
# otherwise shell commands ACTION-IF-FALSE are run. The environment
|
||||||
|
# variable 'ax_compare_version' is always set to either 'true' or 'false'
|
||||||
|
# as well.
|
||||||
|
#
|
||||||
|
# Examples:
|
||||||
|
#
|
||||||
|
# AX_COMPARE_VERSION([3.15.7],[lt],[3.15.8])
|
||||||
|
# AX_COMPARE_VERSION([3.15],[lt],[3.15.8])
|
||||||
|
#
|
||||||
|
# would both be true.
|
||||||
|
#
|
||||||
|
# AX_COMPARE_VERSION([3.15.7],[eq],[3.15.8])
|
||||||
|
# AX_COMPARE_VERSION([3.15],[gt],[3.15.8])
|
||||||
|
#
|
||||||
|
# would both be false.
|
||||||
|
#
|
||||||
|
# AX_COMPARE_VERSION([3.15.7],[eq2],[3.15.8])
|
||||||
|
#
|
||||||
|
# would be true because it is only comparing two minor versions.
|
||||||
|
#
|
||||||
|
# AX_COMPARE_VERSION([3.15.7],[eq0],[3.15])
|
||||||
|
#
|
||||||
|
# would be true because it is only comparing the lesser number of minor
|
||||||
|
# versions of the two values.
|
||||||
|
#
|
||||||
|
# Note: The characters that separate the version numbers do not matter. An
|
||||||
|
# empty string is the same as version 0. OP is evaluated by autoconf, not
|
||||||
|
# configure, so must be a string, not a variable.
|
||||||
|
#
|
||||||
|
# The author would like to acknowledge Guido Draheim whose advice about
|
||||||
|
# the m4_case and m4_ifvaln functions make this macro only include the
|
||||||
|
# portions necessary to perform the specific comparison specified by the
|
||||||
|
# OP argument in the final configure script.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2008 Tim Toolan <toolan@ele.uri.edu>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 12
|
||||||
|
|
||||||
|
dnl #########################################################################
|
||||||
|
AC_DEFUN([AX_COMPARE_VERSION], [
|
||||||
|
AC_REQUIRE([AC_PROG_AWK])
|
||||||
|
|
||||||
|
# Used to indicate true or false condition
|
||||||
|
ax_compare_version=false
|
||||||
|
|
||||||
|
# Convert the two version strings to be compared into a format that
|
||||||
|
# allows a simple string comparison. The end result is that a version
|
||||||
|
# string of the form 1.12.5-r617 will be converted to the form
|
||||||
|
# 0001001200050617. In other words, each number is zero padded to four
|
||||||
|
# digits, and non digits are removed.
|
||||||
|
AS_VAR_PUSHDEF([A],[ax_compare_version_A])
|
||||||
|
A=`echo "$1" | sed -e 's/\([[0-9]]*\)/Z\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]][[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]][[0-9]][[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/[[^0-9]]//g'`
|
||||||
|
|
||||||
|
AS_VAR_PUSHDEF([B],[ax_compare_version_B])
|
||||||
|
B=`echo "$3" | sed -e 's/\([[0-9]]*\)/Z\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]][[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/Z\([[0-9]][[0-9]][[0-9]]\)Z/Z0\1Z/g' \
|
||||||
|
-e 's/[[^0-9]]//g'`
|
||||||
|
|
||||||
|
dnl # In the case of le, ge, lt, and gt, the strings are sorted as necessary
|
||||||
|
dnl # then the first line is used to determine if the condition is true.
|
||||||
|
dnl # The sed right after the echo is to remove any indented white space.
|
||||||
|
m4_case(m4_tolower($2),
|
||||||
|
[lt],[
|
||||||
|
ax_compare_version=`echo "x$A
|
||||||
|
x$B" | sed 's/^ *//' | sort -r | sed "s/x${A}/false/;s/x${B}/true/;1q"`
|
||||||
|
],
|
||||||
|
[gt],[
|
||||||
|
ax_compare_version=`echo "x$A
|
||||||
|
x$B" | sed 's/^ *//' | sort | sed "s/x${A}/false/;s/x${B}/true/;1q"`
|
||||||
|
],
|
||||||
|
[le],[
|
||||||
|
ax_compare_version=`echo "x$A
|
||||||
|
x$B" | sed 's/^ *//' | sort | sed "s/x${A}/true/;s/x${B}/false/;1q"`
|
||||||
|
],
|
||||||
|
[ge],[
|
||||||
|
ax_compare_version=`echo "x$A
|
||||||
|
x$B" | sed 's/^ *//' | sort -r | sed "s/x${A}/true/;s/x${B}/false/;1q"`
|
||||||
|
],[
|
||||||
|
dnl Split the operator from the subversion count if present.
|
||||||
|
m4_bmatch(m4_substr($2,2),
|
||||||
|
[0],[
|
||||||
|
# A count of zero means use the length of the shorter version.
|
||||||
|
# Determine the number of characters in A and B.
|
||||||
|
ax_compare_version_len_A=`echo "$A" | $AWK '{print(length)}'`
|
||||||
|
ax_compare_version_len_B=`echo "$B" | $AWK '{print(length)}'`
|
||||||
|
|
||||||
|
# Set A to no more than B's length and B to no more than A's length.
|
||||||
|
A=`echo "$A" | sed "s/\(.\{$ax_compare_version_len_B\}\).*/\1/"`
|
||||||
|
B=`echo "$B" | sed "s/\(.\{$ax_compare_version_len_A\}\).*/\1/"`
|
||||||
|
],
|
||||||
|
[[0-9]+],[
|
||||||
|
# A count greater than zero means use only that many subversions
|
||||||
|
A=`echo "$A" | sed "s/\(\([[0-9]]\{4\}\)\{m4_substr($2,2)\}\).*/\1/"`
|
||||||
|
B=`echo "$B" | sed "s/\(\([[0-9]]\{4\}\)\{m4_substr($2,2)\}\).*/\1/"`
|
||||||
|
],
|
||||||
|
[.+],[
|
||||||
|
AC_WARNING(
|
||||||
|
[illegal OP numeric parameter: $2])
|
||||||
|
],[])
|
||||||
|
|
||||||
|
# Pad zeros at end of numbers to make same length.
|
||||||
|
ax_compare_version_tmp_A="$A`echo $B | sed 's/./0/g'`"
|
||||||
|
B="$B`echo $A | sed 's/./0/g'`"
|
||||||
|
A="$ax_compare_version_tmp_A"
|
||||||
|
|
||||||
|
# Check for equality or inequality as necessary.
|
||||||
|
m4_case(m4_tolower(m4_substr($2,0,2)),
|
||||||
|
[eq],[
|
||||||
|
test "x$A" = "x$B" && ax_compare_version=true
|
||||||
|
],
|
||||||
|
[ne],[
|
||||||
|
test "x$A" != "x$B" && ax_compare_version=true
|
||||||
|
],[
|
||||||
|
AC_WARNING([illegal OP parameter: $2])
|
||||||
|
])
|
||||||
|
])
|
||||||
|
|
||||||
|
AS_VAR_POPDEF([A])dnl
|
||||||
|
AS_VAR_POPDEF([B])dnl
|
||||||
|
|
||||||
|
dnl # Execute ACTION-IF-TRUE / ACTION-IF-FALSE.
|
||||||
|
if test "$ax_compare_version" = "true" ; then
|
||||||
|
m4_ifvaln([$4],[$4],[:])dnl
|
||||||
|
m4_ifvaln([$5],[else $5])dnl
|
||||||
|
fi
|
||||||
|
]) dnl AX_COMPARE_VERSION
|
982
vendor/github.com/apache/thrift/aclocal/ax_cxx_compile_stdcxx.m4
generated
vendored
Normal file
982
vendor/github.com/apache/thrift/aclocal/ax_cxx_compile_stdcxx.m4
generated
vendored
Normal file
|
@ -0,0 +1,982 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_cxx_compile_stdcxx.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_CXX_COMPILE_STDCXX(VERSION, [ext|noext], [mandatory|optional])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Check for baseline language coverage in the compiler for the specified
|
||||||
|
# version of the C++ standard. If necessary, add switches to CXX and
|
||||||
|
# CXXCPP to enable support. VERSION may be '11' (for the C++11 standard)
|
||||||
|
# or '14' (for the C++14 standard).
|
||||||
|
#
|
||||||
|
# The second argument, if specified, indicates whether you insist on an
|
||||||
|
# extended mode (e.g. -std=gnu++11) or a strict conformance mode (e.g.
|
||||||
|
# -std=c++11). If neither is specified, you get whatever works, with
|
||||||
|
# preference for an extended mode.
|
||||||
|
#
|
||||||
|
# The third argument, if specified 'mandatory' or if left unspecified,
|
||||||
|
# indicates that baseline support for the specified C++ standard is
|
||||||
|
# required and that the macro should error out if no mode with that
|
||||||
|
# support is found. If specified 'optional', then configuration proceeds
|
||||||
|
# regardless, after defining HAVE_CXX${VERSION} if and only if a
|
||||||
|
# supporting mode is found.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2008 Benjamin Kosnik <bkoz@redhat.com>
|
||||||
|
# Copyright (c) 2012 Zack Weinberg <zackw@panix.com>
|
||||||
|
# Copyright (c) 2013 Roy Stogner <roystgnr@ices.utexas.edu>
|
||||||
|
# Copyright (c) 2014, 2015 Google Inc.; contributed by Alexey Sokolov <sokolov@google.com>
|
||||||
|
# Copyright (c) 2015 Paul Norman <penorman@mac.com>
|
||||||
|
# Copyright (c) 2015 Moritz Klammler <moritz@klammler.eu>
|
||||||
|
# Copyright (c) 2016 Krzesimir Nowak <qdlacz@gmail.com>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 7
|
||||||
|
|
||||||
|
dnl This macro is based on the code from the AX_CXX_COMPILE_STDCXX_11 macro
|
||||||
|
dnl (serial version number 13).
|
||||||
|
|
||||||
|
AX_REQUIRE_DEFINED([AC_MSG_WARN])
|
||||||
|
AC_DEFUN([AX_CXX_COMPILE_STDCXX], [dnl
|
||||||
|
m4_if([$1], [11], [ax_cxx_compile_alternatives="11 0x"],
|
||||||
|
[$1], [14], [ax_cxx_compile_alternatives="14 1y"],
|
||||||
|
[$1], [17], [ax_cxx_compile_alternatives="17 1z"],
|
||||||
|
[m4_fatal([invalid first argument `$1' to AX_CXX_COMPILE_STDCXX])])dnl
|
||||||
|
m4_if([$2], [], [],
|
||||||
|
[$2], [ext], [],
|
||||||
|
[$2], [noext], [],
|
||||||
|
[m4_fatal([invalid second argument `$2' to AX_CXX_COMPILE_STDCXX])])dnl
|
||||||
|
m4_if([$3], [], [ax_cxx_compile_cxx$1_required=true],
|
||||||
|
[$3], [mandatory], [ax_cxx_compile_cxx$1_required=true],
|
||||||
|
[$3], [optional], [ax_cxx_compile_cxx$1_required=false],
|
||||||
|
[m4_fatal([invalid third argument `$3' to AX_CXX_COMPILE_STDCXX])])
|
||||||
|
AC_LANG_PUSH([C++])dnl
|
||||||
|
ac_success=no
|
||||||
|
AC_CACHE_CHECK(whether $CXX supports C++$1 features by default,
|
||||||
|
ax_cv_cxx_compile_cxx$1,
|
||||||
|
[AC_COMPILE_IFELSE([AC_LANG_SOURCE([_AX_CXX_COMPILE_STDCXX_testbody_$1])],
|
||||||
|
[ax_cv_cxx_compile_cxx$1=yes],
|
||||||
|
[ax_cv_cxx_compile_cxx$1=no])])
|
||||||
|
if test x$ax_cv_cxx_compile_cxx$1 = xyes; then
|
||||||
|
ac_success=yes
|
||||||
|
fi
|
||||||
|
|
||||||
|
m4_if([$2], [noext], [], [dnl
|
||||||
|
if test x$ac_success = xno; then
|
||||||
|
for alternative in ${ax_cxx_compile_alternatives}; do
|
||||||
|
switch="-std=gnu++${alternative}"
|
||||||
|
cachevar=AS_TR_SH([ax_cv_cxx_compile_cxx$1_$switch])
|
||||||
|
AC_CACHE_CHECK(whether $CXX supports C++$1 features with $switch,
|
||||||
|
$cachevar,
|
||||||
|
[ac_save_CXX="$CXX"
|
||||||
|
CXX="$CXX $switch"
|
||||||
|
AC_COMPILE_IFELSE([AC_LANG_SOURCE([_AX_CXX_COMPILE_STDCXX_testbody_$1])],
|
||||||
|
[eval $cachevar=yes],
|
||||||
|
[eval $cachevar=no])
|
||||||
|
CXX="$ac_save_CXX"])
|
||||||
|
if eval test x\$$cachevar = xyes; then
|
||||||
|
CXX="$CXX $switch"
|
||||||
|
if test -n "$CXXCPP" ; then
|
||||||
|
CXXCPP="$CXXCPP $switch"
|
||||||
|
fi
|
||||||
|
ac_success=yes
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi])
|
||||||
|
|
||||||
|
m4_if([$2], [ext], [], [dnl
|
||||||
|
if test x$ac_success = xno; then
|
||||||
|
dnl HP's aCC needs +std=c++11 according to:
|
||||||
|
dnl http://h21007.www2.hp.com/portal/download/files/unprot/aCxx/PDF_Release_Notes/769149-001.pdf
|
||||||
|
dnl Cray's crayCC needs "-h std=c++11"
|
||||||
|
for alternative in ${ax_cxx_compile_alternatives}; do
|
||||||
|
for switch in -std=c++${alternative} +std=c++${alternative} "-h std=c++${alternative}"; do
|
||||||
|
cachevar=AS_TR_SH([ax_cv_cxx_compile_cxx$1_$switch])
|
||||||
|
AC_CACHE_CHECK(whether $CXX supports C++$1 features with $switch,
|
||||||
|
$cachevar,
|
||||||
|
[ac_save_CXX="$CXX"
|
||||||
|
CXX="$CXX $switch"
|
||||||
|
AC_COMPILE_IFELSE([AC_LANG_SOURCE([_AX_CXX_COMPILE_STDCXX_testbody_$1])],
|
||||||
|
[eval $cachevar=yes],
|
||||||
|
[eval $cachevar=no])
|
||||||
|
CXX="$ac_save_CXX"])
|
||||||
|
if eval test x\$$cachevar = xyes; then
|
||||||
|
CXX="$CXX $switch"
|
||||||
|
if test -n "$CXXCPP" ; then
|
||||||
|
CXXCPP="$CXXCPP $switch"
|
||||||
|
fi
|
||||||
|
ac_success=yes
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
if test x$ac_success = xyes; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi])
|
||||||
|
AC_LANG_POP([C++])
|
||||||
|
if test x$ax_cxx_compile_cxx$1_required = xtrue; then
|
||||||
|
if test x$ac_success = xno; then
|
||||||
|
AC_MSG_ERROR([*** A compiler with support for C++$1 language features is required.])
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
if test x$ac_success = xno; then
|
||||||
|
HAVE_CXX$1=0
|
||||||
|
AC_MSG_NOTICE([No compiler with C++$1 support was found])
|
||||||
|
else
|
||||||
|
HAVE_CXX$1=1
|
||||||
|
AC_DEFINE(HAVE_CXX$1,1,
|
||||||
|
[define if the compiler supports basic C++$1 syntax])
|
||||||
|
fi
|
||||||
|
AC_SUBST(HAVE_CXX$1)
|
||||||
|
m4_if([$1], [17], [AC_MSG_WARN([C++17 is not yet standardized, so the checks may change in incompatible ways anytime])])
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl Test body for checking C++11 support
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_11],
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_11
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
dnl Test body for checking C++14 support
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_14],
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_11
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_14
|
||||||
|
)
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_17],
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_11
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_14
|
||||||
|
_AX_CXX_COMPILE_STDCXX_testbody_new_in_17
|
||||||
|
)
|
||||||
|
|
||||||
|
dnl Tests for new features in C++11
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_new_in_11], [[
|
||||||
|
|
||||||
|
// If the compiler admits that it is not ready for C++11, why torture it?
|
||||||
|
// Hopefully, this will speed up the test.
|
||||||
|
|
||||||
|
#ifndef __cplusplus
|
||||||
|
|
||||||
|
#error "This is not a C++ compiler"
|
||||||
|
|
||||||
|
#elif __cplusplus < 201103L
|
||||||
|
|
||||||
|
#error "This is not a C++11 compiler"
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
namespace cxx11
|
||||||
|
{
|
||||||
|
|
||||||
|
namespace test_static_assert
|
||||||
|
{
|
||||||
|
|
||||||
|
template <typename T>
|
||||||
|
struct check
|
||||||
|
{
|
||||||
|
static_assert(sizeof(int) <= sizeof(T), "not big enough");
|
||||||
|
};
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_final_override
|
||||||
|
{
|
||||||
|
|
||||||
|
struct Base
|
||||||
|
{
|
||||||
|
virtual void f() {}
|
||||||
|
};
|
||||||
|
|
||||||
|
struct Derived : public Base
|
||||||
|
{
|
||||||
|
virtual void f() override {}
|
||||||
|
};
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_double_right_angle_brackets
|
||||||
|
{
|
||||||
|
|
||||||
|
template < typename T >
|
||||||
|
struct check {};
|
||||||
|
|
||||||
|
typedef check<void> single_type;
|
||||||
|
typedef check<check<void>> double_type;
|
||||||
|
typedef check<check<check<void>>> triple_type;
|
||||||
|
typedef check<check<check<check<void>>>> quadruple_type;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_decltype
|
||||||
|
{
|
||||||
|
|
||||||
|
int
|
||||||
|
f()
|
||||||
|
{
|
||||||
|
int a = 1;
|
||||||
|
decltype(a) b = 2;
|
||||||
|
return a + b;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_type_deduction
|
||||||
|
{
|
||||||
|
|
||||||
|
template < typename T1, typename T2 >
|
||||||
|
struct is_same
|
||||||
|
{
|
||||||
|
static const bool value = false;
|
||||||
|
};
|
||||||
|
|
||||||
|
template < typename T >
|
||||||
|
struct is_same<T, T>
|
||||||
|
{
|
||||||
|
static const bool value = true;
|
||||||
|
};
|
||||||
|
|
||||||
|
template < typename T1, typename T2 >
|
||||||
|
auto
|
||||||
|
add(T1 a1, T2 a2) -> decltype(a1 + a2)
|
||||||
|
{
|
||||||
|
return a1 + a2;
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
test(const int c, volatile int v)
|
||||||
|
{
|
||||||
|
static_assert(is_same<int, decltype(0)>::value == true, "");
|
||||||
|
static_assert(is_same<int, decltype(c)>::value == false, "");
|
||||||
|
static_assert(is_same<int, decltype(v)>::value == false, "");
|
||||||
|
auto ac = c;
|
||||||
|
auto av = v;
|
||||||
|
auto sumi = ac + av + 'x';
|
||||||
|
auto sumf = ac + av + 1.0;
|
||||||
|
static_assert(is_same<int, decltype(ac)>::value == true, "");
|
||||||
|
static_assert(is_same<int, decltype(av)>::value == true, "");
|
||||||
|
static_assert(is_same<int, decltype(sumi)>::value == true, "");
|
||||||
|
static_assert(is_same<int, decltype(sumf)>::value == false, "");
|
||||||
|
static_assert(is_same<int, decltype(add(c, v))>::value == true, "");
|
||||||
|
return (sumf > 0.0) ? sumi : add(c, v);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_noexcept
|
||||||
|
{
|
||||||
|
|
||||||
|
int f() { return 0; }
|
||||||
|
int g() noexcept { return 0; }
|
||||||
|
|
||||||
|
static_assert(noexcept(f()) == false, "");
|
||||||
|
static_assert(noexcept(g()) == true, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_constexpr
|
||||||
|
{
|
||||||
|
|
||||||
|
template < typename CharT >
|
||||||
|
unsigned long constexpr
|
||||||
|
strlen_c_r(const CharT *const s, const unsigned long acc) noexcept
|
||||||
|
{
|
||||||
|
return *s ? strlen_c_r(s + 1, acc + 1) : acc;
|
||||||
|
}
|
||||||
|
|
||||||
|
template < typename CharT >
|
||||||
|
unsigned long constexpr
|
||||||
|
strlen_c(const CharT *const s) noexcept
|
||||||
|
{
|
||||||
|
return strlen_c_r(s, 0UL);
|
||||||
|
}
|
||||||
|
|
||||||
|
static_assert(strlen_c("") == 0UL, "");
|
||||||
|
static_assert(strlen_c("1") == 1UL, "");
|
||||||
|
static_assert(strlen_c("example") == 7UL, "");
|
||||||
|
static_assert(strlen_c("another\0example") == 7UL, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_rvalue_references
|
||||||
|
{
|
||||||
|
|
||||||
|
template < int N >
|
||||||
|
struct answer
|
||||||
|
{
|
||||||
|
static constexpr int value = N;
|
||||||
|
};
|
||||||
|
|
||||||
|
answer<1> f(int&) { return answer<1>(); }
|
||||||
|
answer<2> f(const int&) { return answer<2>(); }
|
||||||
|
answer<3> f(int&&) { return answer<3>(); }
|
||||||
|
|
||||||
|
void
|
||||||
|
test()
|
||||||
|
{
|
||||||
|
int i = 0;
|
||||||
|
const int c = 0;
|
||||||
|
static_assert(decltype(f(i))::value == 1, "");
|
||||||
|
static_assert(decltype(f(c))::value == 2, "");
|
||||||
|
static_assert(decltype(f(0))::value == 3, "");
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_uniform_initialization
|
||||||
|
{
|
||||||
|
|
||||||
|
struct test
|
||||||
|
{
|
||||||
|
static const int zero {};
|
||||||
|
static const int one {1};
|
||||||
|
};
|
||||||
|
|
||||||
|
static_assert(test::zero == 0, "");
|
||||||
|
static_assert(test::one == 1, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_lambdas
|
||||||
|
{
|
||||||
|
|
||||||
|
void
|
||||||
|
test1()
|
||||||
|
{
|
||||||
|
auto lambda1 = [](){};
|
||||||
|
auto lambda2 = lambda1;
|
||||||
|
lambda1();
|
||||||
|
lambda2();
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
test2()
|
||||||
|
{
|
||||||
|
auto a = [](int i, int j){ return i + j; }(1, 2);
|
||||||
|
auto b = []() -> int { return '0'; }();
|
||||||
|
auto c = [=](){ return a + b; }();
|
||||||
|
auto d = [&](){ return c; }();
|
||||||
|
auto e = [a, &b](int x) mutable {
|
||||||
|
const auto identity = [](int y){ return y; };
|
||||||
|
for (auto i = 0; i < a; ++i)
|
||||||
|
a += b--;
|
||||||
|
return x + identity(a + b);
|
||||||
|
}(0);
|
||||||
|
return a + b + c + d + e;
|
||||||
|
}
|
||||||
|
|
||||||
|
int
|
||||||
|
test3()
|
||||||
|
{
|
||||||
|
const auto nullary = [](){ return 0; };
|
||||||
|
const auto unary = [](int x){ return x; };
|
||||||
|
using nullary_t = decltype(nullary);
|
||||||
|
using unary_t = decltype(unary);
|
||||||
|
const auto higher1st = [](nullary_t f){ return f(); };
|
||||||
|
const auto higher2nd = [unary](nullary_t f1){
|
||||||
|
return [unary, f1](unary_t f2){ return f2(unary(f1())); };
|
||||||
|
};
|
||||||
|
return higher1st(nullary) + higher2nd(nullary)(unary);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_variadic_templates
|
||||||
|
{
|
||||||
|
|
||||||
|
template <int...>
|
||||||
|
struct sum;
|
||||||
|
|
||||||
|
template <int N0, int... N1toN>
|
||||||
|
struct sum<N0, N1toN...>
|
||||||
|
{
|
||||||
|
static constexpr auto value = N0 + sum<N1toN...>::value;
|
||||||
|
};
|
||||||
|
|
||||||
|
template <>
|
||||||
|
struct sum<>
|
||||||
|
{
|
||||||
|
static constexpr auto value = 0;
|
||||||
|
};
|
||||||
|
|
||||||
|
static_assert(sum<>::value == 0, "");
|
||||||
|
static_assert(sum<1>::value == 1, "");
|
||||||
|
static_assert(sum<23>::value == 23, "");
|
||||||
|
static_assert(sum<1, 2>::value == 3, "");
|
||||||
|
static_assert(sum<5, 5, 11>::value == 21, "");
|
||||||
|
static_assert(sum<2, 3, 5, 7, 11, 13>::value == 41, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// http://stackoverflow.com/questions/13728184/template-aliases-and-sfinae
|
||||||
|
// Clang 3.1 fails with headers of libstd++ 4.8.3 when using std::function
|
||||||
|
// because of this.
|
||||||
|
namespace test_template_alias_sfinae
|
||||||
|
{
|
||||||
|
|
||||||
|
struct foo {};
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
using member = typename T::member_type;
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
void func(...) {}
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
void func(member<T>*) {}
|
||||||
|
|
||||||
|
void test();
|
||||||
|
|
||||||
|
void test() { func<foo>(0); }
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace cxx11
|
||||||
|
|
||||||
|
#endif // __cplusplus >= 201103L
|
||||||
|
|
||||||
|
]])
|
||||||
|
|
||||||
|
|
||||||
|
dnl Tests for new features in C++14
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_new_in_14], [[
|
||||||
|
|
||||||
|
// If the compiler admits that it is not ready for C++14, why torture it?
|
||||||
|
// Hopefully, this will speed up the test.
|
||||||
|
|
||||||
|
#ifndef __cplusplus
|
||||||
|
|
||||||
|
#error "This is not a C++ compiler"
|
||||||
|
|
||||||
|
#elif __cplusplus < 201402L
|
||||||
|
|
||||||
|
#error "This is not a C++14 compiler"
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
namespace cxx14
|
||||||
|
{
|
||||||
|
|
||||||
|
namespace test_polymorphic_lambdas
|
||||||
|
{
|
||||||
|
|
||||||
|
int
|
||||||
|
test()
|
||||||
|
{
|
||||||
|
const auto lambda = [](auto&&... args){
|
||||||
|
const auto istiny = [](auto x){
|
||||||
|
return (sizeof(x) == 1UL) ? 1 : 0;
|
||||||
|
};
|
||||||
|
const int aretiny[] = { istiny(args)... };
|
||||||
|
return aretiny[0];
|
||||||
|
};
|
||||||
|
return lambda(1, 1L, 1.0f, '1');
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_binary_literals
|
||||||
|
{
|
||||||
|
|
||||||
|
constexpr auto ivii = 0b0000000000101010;
|
||||||
|
static_assert(ivii == 42, "wrong value");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_generalized_constexpr
|
||||||
|
{
|
||||||
|
|
||||||
|
template < typename CharT >
|
||||||
|
constexpr unsigned long
|
||||||
|
strlen_c(const CharT *const s) noexcept
|
||||||
|
{
|
||||||
|
auto length = 0UL;
|
||||||
|
for (auto p = s; *p; ++p)
|
||||||
|
++length;
|
||||||
|
return length;
|
||||||
|
}
|
||||||
|
|
||||||
|
static_assert(strlen_c("") == 0UL, "");
|
||||||
|
static_assert(strlen_c("x") == 1UL, "");
|
||||||
|
static_assert(strlen_c("test") == 4UL, "");
|
||||||
|
static_assert(strlen_c("another\0test") == 7UL, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_lambda_init_capture
|
||||||
|
{
|
||||||
|
|
||||||
|
int
|
||||||
|
test()
|
||||||
|
{
|
||||||
|
auto x = 0;
|
||||||
|
const auto lambda1 = [a = x](int b){ return a + b; };
|
||||||
|
const auto lambda2 = [a = lambda1(x)](){ return a; };
|
||||||
|
return lambda2();
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_digit_separators
|
||||||
|
{
|
||||||
|
|
||||||
|
constexpr auto ten_million = 100'000'000;
|
||||||
|
static_assert(ten_million == 100000000, "");
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_return_type_deduction
|
||||||
|
{
|
||||||
|
|
||||||
|
auto f(int& x) { return x; }
|
||||||
|
decltype(auto) g(int& x) { return x; }
|
||||||
|
|
||||||
|
template < typename T1, typename T2 >
|
||||||
|
struct is_same
|
||||||
|
{
|
||||||
|
static constexpr auto value = false;
|
||||||
|
};
|
||||||
|
|
||||||
|
template < typename T >
|
||||||
|
struct is_same<T, T>
|
||||||
|
{
|
||||||
|
static constexpr auto value = true;
|
||||||
|
};
|
||||||
|
|
||||||
|
int
|
||||||
|
test()
|
||||||
|
{
|
||||||
|
auto x = 0;
|
||||||
|
static_assert(is_same<int, decltype(f(x))>::value, "");
|
||||||
|
static_assert(is_same<int&, decltype(g(x))>::value, "");
|
||||||
|
return x;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace cxx14
|
||||||
|
|
||||||
|
#endif // __cplusplus >= 201402L
|
||||||
|
|
||||||
|
]])
|
||||||
|
|
||||||
|
|
||||||
|
dnl Tests for new features in C++17
|
||||||
|
|
||||||
|
m4_define([_AX_CXX_COMPILE_STDCXX_testbody_new_in_17], [[
|
||||||
|
|
||||||
|
// If the compiler admits that it is not ready for C++17, why torture it?
|
||||||
|
// Hopefully, this will speed up the test.
|
||||||
|
|
||||||
|
#ifndef __cplusplus
|
||||||
|
|
||||||
|
#error "This is not a C++ compiler"
|
||||||
|
|
||||||
|
#elif __cplusplus <= 201402L
|
||||||
|
|
||||||
|
#error "This is not a C++17 compiler"
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
#if defined(__clang__)
|
||||||
|
#define REALLY_CLANG
|
||||||
|
#else
|
||||||
|
#if defined(__GNUC__)
|
||||||
|
#define REALLY_GCC
|
||||||
|
#endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include <initializer_list>
|
||||||
|
#include <utility>
|
||||||
|
#include <type_traits>
|
||||||
|
|
||||||
|
namespace cxx17
|
||||||
|
{
|
||||||
|
|
||||||
|
#if !defined(REALLY_CLANG)
|
||||||
|
namespace test_constexpr_lambdas
|
||||||
|
{
|
||||||
|
|
||||||
|
// TODO: test it with clang++ from git
|
||||||
|
|
||||||
|
constexpr int foo = [](){return 42;}();
|
||||||
|
|
||||||
|
}
|
||||||
|
#endif // !defined(REALLY_CLANG)
|
||||||
|
|
||||||
|
namespace test::nested_namespace::definitions
|
||||||
|
{
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_fold_expression
|
||||||
|
{
|
||||||
|
|
||||||
|
template<typename... Args>
|
||||||
|
int multiply(Args... args)
|
||||||
|
{
|
||||||
|
return (args * ... * 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
template<typename... Args>
|
||||||
|
bool all(Args... args)
|
||||||
|
{
|
||||||
|
return (args && ...);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_extended_static_assert
|
||||||
|
{
|
||||||
|
|
||||||
|
static_assert (true);
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_auto_brace_init_list
|
||||||
|
{
|
||||||
|
|
||||||
|
auto foo = {5};
|
||||||
|
auto bar {5};
|
||||||
|
|
||||||
|
static_assert(std::is_same<std::initializer_list<int>, decltype(foo)>::value);
|
||||||
|
static_assert(std::is_same<int, decltype(bar)>::value);
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_typename_in_template_template_parameter
|
||||||
|
{
|
||||||
|
|
||||||
|
template<template<typename> typename X> struct D;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_fallthrough_nodiscard_maybe_unused_attributes
|
||||||
|
{
|
||||||
|
|
||||||
|
int f1()
|
||||||
|
{
|
||||||
|
return 42;
|
||||||
|
}
|
||||||
|
|
||||||
|
[[nodiscard]] int f2()
|
||||||
|
{
|
||||||
|
[[maybe_unused]] auto unused = f1();
|
||||||
|
|
||||||
|
switch (f1())
|
||||||
|
{
|
||||||
|
case 17:
|
||||||
|
f1();
|
||||||
|
[[fallthrough]];
|
||||||
|
case 42:
|
||||||
|
f1();
|
||||||
|
}
|
||||||
|
return f1();
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_extended_aggregate_initialization
|
||||||
|
{
|
||||||
|
|
||||||
|
struct base1
|
||||||
|
{
|
||||||
|
int b1, b2 = 42;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct base2
|
||||||
|
{
|
||||||
|
base2() {
|
||||||
|
b3 = 42;
|
||||||
|
}
|
||||||
|
int b3;
|
||||||
|
};
|
||||||
|
|
||||||
|
struct derived : base1, base2
|
||||||
|
{
|
||||||
|
int d;
|
||||||
|
};
|
||||||
|
|
||||||
|
derived d1 {{1, 2}, {}, 4}; // full initialization
|
||||||
|
derived d2 {{}, {}, 4}; // value-initialized bases
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_general_range_based_for_loop
|
||||||
|
{
|
||||||
|
|
||||||
|
struct iter
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
|
int& operator* ()
|
||||||
|
{
|
||||||
|
return i;
|
||||||
|
}
|
||||||
|
|
||||||
|
const int& operator* () const
|
||||||
|
{
|
||||||
|
return i;
|
||||||
|
}
|
||||||
|
|
||||||
|
iter& operator++()
|
||||||
|
{
|
||||||
|
++i;
|
||||||
|
return *this;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
struct sentinel
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
};
|
||||||
|
|
||||||
|
bool operator== (const iter& i, const sentinel& s)
|
||||||
|
{
|
||||||
|
return i.i == s.i;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool operator!= (const iter& i, const sentinel& s)
|
||||||
|
{
|
||||||
|
return !(i == s);
|
||||||
|
}
|
||||||
|
|
||||||
|
struct range
|
||||||
|
{
|
||||||
|
iter begin() const
|
||||||
|
{
|
||||||
|
return {0};
|
||||||
|
}
|
||||||
|
|
||||||
|
sentinel end() const
|
||||||
|
{
|
||||||
|
return {5};
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
void f()
|
||||||
|
{
|
||||||
|
range r {};
|
||||||
|
|
||||||
|
for (auto i : r)
|
||||||
|
{
|
||||||
|
[[maybe_unused]] auto v = i;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_lambda_capture_asterisk_this_by_value
|
||||||
|
{
|
||||||
|
|
||||||
|
struct t
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
int foo()
|
||||||
|
{
|
||||||
|
return [*this]()
|
||||||
|
{
|
||||||
|
return i;
|
||||||
|
}();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_enum_class_construction
|
||||||
|
{
|
||||||
|
|
||||||
|
enum class byte : unsigned char
|
||||||
|
{};
|
||||||
|
|
||||||
|
byte foo {42};
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_constexpr_if
|
||||||
|
{
|
||||||
|
|
||||||
|
template <bool cond>
|
||||||
|
int f ()
|
||||||
|
{
|
||||||
|
if constexpr(cond)
|
||||||
|
{
|
||||||
|
return 13;
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
return 42;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
namespace test_selection_statement_with_initializer
|
||||||
|
{
|
||||||
|
|
||||||
|
int f()
|
||||||
|
{
|
||||||
|
return 13;
|
||||||
|
}
|
||||||
|
|
||||||
|
int f2()
|
||||||
|
{
|
||||||
|
if (auto i = f(); i > 0)
|
||||||
|
{
|
||||||
|
return 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
switch (auto i = f(); i + 4)
|
||||||
|
{
|
||||||
|
case 17:
|
||||||
|
return 2;
|
||||||
|
|
||||||
|
default:
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
#if !defined(REALLY_CLANG)
|
||||||
|
namespace test_template_argument_deduction_for_class_templates
|
||||||
|
{
|
||||||
|
|
||||||
|
// TODO: test it with clang++ from git
|
||||||
|
|
||||||
|
template <typename T1, typename T2>
|
||||||
|
struct pair
|
||||||
|
{
|
||||||
|
pair (T1 p1, T2 p2)
|
||||||
|
: m1 {p1},
|
||||||
|
m2 {p2}
|
||||||
|
{}
|
||||||
|
|
||||||
|
T1 m1;
|
||||||
|
T2 m2;
|
||||||
|
};
|
||||||
|
|
||||||
|
void f()
|
||||||
|
{
|
||||||
|
[[maybe_unused]] auto p = pair{13, 42u};
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
#endif // !defined(REALLY_CLANG)
|
||||||
|
|
||||||
|
namespace test_non_type_auto_template_parameters
|
||||||
|
{
|
||||||
|
|
||||||
|
template <auto n>
|
||||||
|
struct B
|
||||||
|
{};
|
||||||
|
|
||||||
|
B<5> b1;
|
||||||
|
B<'a'> b2;
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
#if !defined(REALLY_CLANG)
|
||||||
|
namespace test_structured_bindings
|
||||||
|
{
|
||||||
|
|
||||||
|
// TODO: test it with clang++ from git
|
||||||
|
|
||||||
|
int arr[2] = { 1, 2 };
|
||||||
|
std::pair<int, int> pr = { 1, 2 };
|
||||||
|
|
||||||
|
auto f1() -> int(&)[2]
|
||||||
|
{
|
||||||
|
return arr;
|
||||||
|
}
|
||||||
|
|
||||||
|
auto f2() -> std::pair<int, int>&
|
||||||
|
{
|
||||||
|
return pr;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct S
|
||||||
|
{
|
||||||
|
int x1 : 2;
|
||||||
|
volatile double y1;
|
||||||
|
};
|
||||||
|
|
||||||
|
S f3()
|
||||||
|
{
|
||||||
|
return {};
|
||||||
|
}
|
||||||
|
|
||||||
|
auto [ x1, y1 ] = f1();
|
||||||
|
auto& [ xr1, yr1 ] = f1();
|
||||||
|
auto [ x2, y2 ] = f2();
|
||||||
|
auto& [ xr2, yr2 ] = f2();
|
||||||
|
const auto [ x3, y3 ] = f3();
|
||||||
|
|
||||||
|
}
|
||||||
|
#endif // !defined(REALLY_CLANG)
|
||||||
|
|
||||||
|
#if !defined(REALLY_CLANG)
|
||||||
|
namespace test_exception_spec_type_system
|
||||||
|
{
|
||||||
|
|
||||||
|
// TODO: test it with clang++ from git
|
||||||
|
|
||||||
|
struct Good {};
|
||||||
|
struct Bad {};
|
||||||
|
|
||||||
|
void g1() noexcept;
|
||||||
|
void g2();
|
||||||
|
|
||||||
|
template<typename T>
|
||||||
|
Bad
|
||||||
|
f(T*, T*);
|
||||||
|
|
||||||
|
template<typename T1, typename T2>
|
||||||
|
Good
|
||||||
|
f(T1*, T2*);
|
||||||
|
|
||||||
|
static_assert (std::is_same_v<Good, decltype(f(g1, g2))>);
|
||||||
|
|
||||||
|
}
|
||||||
|
#endif // !defined(REALLY_CLANG)
|
||||||
|
|
||||||
|
namespace test_inline_variables
|
||||||
|
{
|
||||||
|
|
||||||
|
template<class T> void f(T)
|
||||||
|
{}
|
||||||
|
|
||||||
|
template<class T> inline T g(T)
|
||||||
|
{
|
||||||
|
return T{};
|
||||||
|
}
|
||||||
|
|
||||||
|
template<> inline void f<>(int)
|
||||||
|
{}
|
||||||
|
|
||||||
|
template<> int g<>(int)
|
||||||
|
{
|
||||||
|
return 5;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace cxx17
|
||||||
|
|
||||||
|
#endif // __cplusplus <= 201402L
|
||||||
|
|
||||||
|
]])
|
39
vendor/github.com/apache/thrift/aclocal/ax_cxx_compile_stdcxx_11.m4
generated
vendored
Normal file
39
vendor/github.com/apache/thrift/aclocal/ax_cxx_compile_stdcxx_11.m4
generated
vendored
Normal file
|
@ -0,0 +1,39 @@
|
||||||
|
# =============================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_cxx_compile_stdcxx_11.html
|
||||||
|
# =============================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_CXX_COMPILE_STDCXX_11([ext|noext], [mandatory|optional])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Check for baseline language coverage in the compiler for the C++11
|
||||||
|
# standard; if necessary, add switches to CXX and CXXCPP to enable
|
||||||
|
# support.
|
||||||
|
#
|
||||||
|
# This macro is a convenience alias for calling the AX_CXX_COMPILE_STDCXX
|
||||||
|
# macro with the version set to C++11. The two optional arguments are
|
||||||
|
# forwarded literally as the second and third argument respectively.
|
||||||
|
# Please see the documentation for the AX_CXX_COMPILE_STDCXX macro for
|
||||||
|
# more information. If you want to use this macro, you also need to
|
||||||
|
# download the ax_cxx_compile_stdcxx.m4 file.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2008 Benjamin Kosnik <bkoz@redhat.com>
|
||||||
|
# Copyright (c) 2012 Zack Weinberg <zackw@panix.com>
|
||||||
|
# Copyright (c) 2013 Roy Stogner <roystgnr@ices.utexas.edu>
|
||||||
|
# Copyright (c) 2014, 2015 Google Inc.; contributed by Alexey Sokolov <sokolov@google.com>
|
||||||
|
# Copyright (c) 2015 Paul Norman <penorman@mac.com>
|
||||||
|
# Copyright (c) 2015 Moritz Klammler <moritz@klammler.eu>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 18
|
||||||
|
|
||||||
|
AX_REQUIRE_DEFINED([AX_CXX_COMPILE_STDCXX])
|
||||||
|
AC_DEFUN([AX_CXX_COMPILE_STDCXX_11], [AX_CXX_COMPILE_STDCXX([11], [$1], [$2])])
|
107
vendor/github.com/apache/thrift/aclocal/ax_dmd.m4
generated
vendored
Normal file
107
vendor/github.com/apache/thrift/aclocal/ax_dmd.m4
generated
vendored
Normal file
|
@ -0,0 +1,107 @@
|
||||||
|
dnl @synopsis AX_DMD
|
||||||
|
dnl
|
||||||
|
dnl Test for the presence of a DMD-compatible D2 compiler, and (optionally)
|
||||||
|
dnl specified modules on the import path.
|
||||||
|
dnl
|
||||||
|
dnl If "DMD" is defined in the environment, that will be the only
|
||||||
|
dnl dmd command tested. Otherwise, a hard-coded list will be used.
|
||||||
|
dnl
|
||||||
|
dnl After AX_DMD runs, the shell variables "success" and "ax_dmd" are set to
|
||||||
|
dnl "yes" or "no", and "DMD" is set to the appropriate command. Furthermore,
|
||||||
|
dnl "dmd_optlink" will be set to "yes" or "no" depending on whether OPTLINK is
|
||||||
|
dnl used as the linker (DMD/Windows), and "dmd_of_dirsep" will be set to the
|
||||||
|
dnl directory separator to use when passing -of to DMD (OPTLINK requires a
|
||||||
|
dnl backslash).
|
||||||
|
dnl
|
||||||
|
dnl AX_CHECK_D_MODULE must be run after AX_DMD. It tests for the presence of a
|
||||||
|
dnl module in the import path of the chosen compiler, and sets the shell
|
||||||
|
dnl variable "success" to "yes" or "no".
|
||||||
|
dnl
|
||||||
|
dnl @category D
|
||||||
|
dnl @version 2011-05-31
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copyright (C) 2011 David Nadlinger
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_DMD],
|
||||||
|
[
|
||||||
|
dnl Hard-coded default commands to test.
|
||||||
|
DMD_PROGS="dmd,gdmd,ldmd"
|
||||||
|
|
||||||
|
dnl Allow the user to specify an alternative.
|
||||||
|
if test -n "$DMD" ; then
|
||||||
|
DMD_PROGS="$DMD"
|
||||||
|
fi
|
||||||
|
|
||||||
|
AC_MSG_CHECKING(for DMD)
|
||||||
|
|
||||||
|
# std.algorithm as a quick way to check for D2/Phobos.
|
||||||
|
echo "import std.algorithm; void main() {}" > configtest_ax_dmd.d
|
||||||
|
success=no
|
||||||
|
oIFS="$IFS"
|
||||||
|
|
||||||
|
IFS=","
|
||||||
|
for DMD in $DMD_PROGS ; do
|
||||||
|
IFS="$oIFS"
|
||||||
|
|
||||||
|
echo "Running \"$DMD configtest_ax_dmd.d\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $DMD configtest_ax_dmd.d >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
success=yes
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
if test "$success" != "yes" ; then
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
DMD=""
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
fi
|
||||||
|
|
||||||
|
ax_dmd="$success"
|
||||||
|
|
||||||
|
# Test whether OPTLINK is used by trying if DMD accepts -L/? without
|
||||||
|
# erroring out.
|
||||||
|
if test "$success" == "yes" ; then
|
||||||
|
AC_MSG_CHECKING(whether DMD uses OPTLINK)
|
||||||
|
echo "Running \”$DMD -L/? configtest_ax_dmd.d\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $DMD -L/? configtest_ax_dmd.d >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
dmd_optlink="yes"
|
||||||
|
|
||||||
|
# This actually produces double slashes in the final configure
|
||||||
|
# output, but at least it works.
|
||||||
|
dmd_of_dirsep="\\\\"
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
dmd_optlink="no"
|
||||||
|
dmd_of_dirsep="/"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f configtest_ax_dmd*
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_CHECK_D_MODULE],
|
||||||
|
[
|
||||||
|
AC_MSG_CHECKING(for D module [$1])
|
||||||
|
|
||||||
|
echo "import $1; void main() {}" > configtest_ax_dmd.d
|
||||||
|
|
||||||
|
echo "Running \"$DMD configtest_ax_dmd.d\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $DMD -c configtest_ax_dmd.d >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
success=yes
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
success=no
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f configtest_ax_dmd*
|
||||||
|
])
|
129
vendor/github.com/apache/thrift/aclocal/ax_javac_and_java.m4
generated
vendored
Normal file
129
vendor/github.com/apache/thrift/aclocal/ax_javac_and_java.m4
generated
vendored
Normal file
|
@ -0,0 +1,129 @@
|
||||||
|
dnl @synopsis AX_JAVAC_AND_JAVA
|
||||||
|
dnl @synopsis AX_CHECK_JAVA_CLASS(CLASSNAME)
|
||||||
|
dnl
|
||||||
|
dnl Test for the presence of a JDK, and (optionally) specific classes.
|
||||||
|
dnl
|
||||||
|
dnl If "JAVA" is defined in the environment, that will be the only
|
||||||
|
dnl java command tested. Otherwise, a hard-coded list will be used.
|
||||||
|
dnl Similarly for "JAVAC".
|
||||||
|
dnl
|
||||||
|
dnl AX_JAVAC_AND_JAVA does not currently support testing for a particular
|
||||||
|
dnl Java version, testing for only one of "java" and "javac", or
|
||||||
|
dnl compiling or running user-provided Java code.
|
||||||
|
dnl
|
||||||
|
dnl After AX_JAVAC_AND_JAVA runs, the shell variables "success" and
|
||||||
|
dnl "ax_javac_and_java" are set to "yes" or "no", and "JAVAC" and
|
||||||
|
dnl "JAVA" are set to the appropriate commands.
|
||||||
|
dnl
|
||||||
|
dnl AX_CHECK_JAVA_CLASS must be run after AX_JAVAC_AND_JAVA.
|
||||||
|
dnl It tests for the presence of a class based on a fully-qualified name.
|
||||||
|
dnl It sets the shell variable "success" to "yes" or "no".
|
||||||
|
dnl
|
||||||
|
dnl @category Java
|
||||||
|
dnl @version 2009-02-09
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_JAVAC_AND_JAVA],
|
||||||
|
[
|
||||||
|
|
||||||
|
dnl Hard-coded default commands to test.
|
||||||
|
JAVAC_PROGS="javac,jikes,gcj -C"
|
||||||
|
JAVA_PROGS="java,kaffe"
|
||||||
|
|
||||||
|
dnl Allow the user to specify an alternative.
|
||||||
|
if test -n "$JAVAC" ; then
|
||||||
|
JAVAC_PROGS="$JAVAC"
|
||||||
|
fi
|
||||||
|
if test -n "$JAVA" ; then
|
||||||
|
JAVA_PROGS="$JAVA"
|
||||||
|
fi
|
||||||
|
|
||||||
|
AC_MSG_CHECKING(for javac and java)
|
||||||
|
|
||||||
|
echo "public class configtest_ax_javac_and_java { public static void main(String args@<:@@:>@) { } }" > configtest_ax_javac_and_java.java
|
||||||
|
success=no
|
||||||
|
oIFS="$IFS"
|
||||||
|
|
||||||
|
IFS=","
|
||||||
|
for JAVAC in $JAVAC_PROGS ; do
|
||||||
|
IFS="$oIFS"
|
||||||
|
|
||||||
|
echo "Running \"$JAVAC configtest_ax_javac_and_java.java\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $JAVAC configtest_ax_javac_and_java.java >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
|
||||||
|
# prevent $JAVA VM issues with UTF-8 path names (THRIFT-3271)
|
||||||
|
oLC_ALL="$LC_ALL"
|
||||||
|
LC_ALL=""
|
||||||
|
|
||||||
|
IFS=","
|
||||||
|
for JAVA in $JAVA_PROGS ; do
|
||||||
|
IFS="$oIFS"
|
||||||
|
|
||||||
|
echo "Running \"$JAVA configtest_ax_javac_and_java\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $JAVA configtest_ax_javac_and_java >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
success=yes
|
||||||
|
break 2
|
||||||
|
fi
|
||||||
|
|
||||||
|
done
|
||||||
|
|
||||||
|
# restore LC_ALL
|
||||||
|
LC_ALL="$oLC_ALL"
|
||||||
|
oLC_ALL=""
|
||||||
|
|
||||||
|
fi
|
||||||
|
|
||||||
|
done
|
||||||
|
|
||||||
|
rm -f configtest_ax_javac_and_java.java configtest_ax_javac_and_java.class
|
||||||
|
|
||||||
|
if test "$success" != "yes" ; then
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
JAVAC=""
|
||||||
|
JAVA=""
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
fi
|
||||||
|
|
||||||
|
ax_javac_and_java="$success"
|
||||||
|
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_CHECK_JAVA_CLASS],
|
||||||
|
[
|
||||||
|
AC_MSG_CHECKING(for Java class [$1])
|
||||||
|
|
||||||
|
echo "import $1; public class configtest_ax_javac_and_java { public static void main(String args@<:@@:>@) { } }" > configtest_ax_javac_and_java.java
|
||||||
|
|
||||||
|
echo "Running \"$JAVAC configtest_ax_javac_and_java.java\"" >&AS_MESSAGE_LOG_FD
|
||||||
|
if $JAVAC configtest_ax_javac_and_java.java >&AS_MESSAGE_LOG_FD 2>&1 ; then
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
success=yes
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
success=no
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f configtest_ax_javac_and_java.java configtest_ax_javac_and_java.class
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_CHECK_ANT_VERSION],
|
||||||
|
[
|
||||||
|
AC_MSG_CHECKING(for ant version > $2)
|
||||||
|
ANT_VALID=`expr "x$(printf "$2\n$($1 -version 2>/dev/null | sed -n 's/.*version \(@<:@0-9\.@:>@*\).*/\1/p')" | sort -t '.' -k 1,1 -k 2,2 -k 3,3 -g | sed -n 1p)" = "x$2"`
|
||||||
|
if test "x$ANT_VALID" = "x1" ; then
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
ANT=""
|
||||||
|
fi
|
||||||
|
])
|
||||||
|
|
194
vendor/github.com/apache/thrift/aclocal/ax_lib_event.m4
generated
vendored
Normal file
194
vendor/github.com/apache/thrift/aclocal/ax_lib_event.m4
generated
vendored
Normal file
|
@ -0,0 +1,194 @@
|
||||||
|
dnl @synopsis AX_LIB_EVENT([MINIMUM-VERSION])
|
||||||
|
dnl
|
||||||
|
dnl Test for the libevent library of a particular version (or newer).
|
||||||
|
dnl
|
||||||
|
dnl If no path to the installed libevent is given, the macro will first try
|
||||||
|
dnl using no -I or -L flags, then searches under /usr, /usr/local, /opt,
|
||||||
|
dnl and /opt/libevent.
|
||||||
|
dnl If these all fail, it will try the $LIBEVENT_ROOT environment variable.
|
||||||
|
dnl
|
||||||
|
dnl This macro requires that #include <sys/types.h> works and defines u_char.
|
||||||
|
dnl
|
||||||
|
dnl This macro calls:
|
||||||
|
dnl AC_SUBST(LIBEVENT_CPPFLAGS)
|
||||||
|
dnl AC_SUBST(LIBEVENT_LDFLAGS)
|
||||||
|
dnl AC_SUBST(LIBEVENT_LIBS)
|
||||||
|
dnl
|
||||||
|
dnl And (if libevent is found):
|
||||||
|
dnl AC_DEFINE(HAVE_LIBEVENT)
|
||||||
|
dnl
|
||||||
|
dnl It also leaves the shell variables "success" and "ax_have_libevent"
|
||||||
|
dnl set to "yes" or "no".
|
||||||
|
dnl
|
||||||
|
dnl NOTE: This macro does not currently work for cross-compiling,
|
||||||
|
dnl but it can be easily modified to allow it. (grep "cross").
|
||||||
|
dnl
|
||||||
|
dnl @category InstalledPackages
|
||||||
|
dnl @category C
|
||||||
|
dnl @version 2007-09-12
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
dnl Input: ax_libevent_path, WANT_LIBEVENT_VERSION
|
||||||
|
dnl Output: success=yes/no
|
||||||
|
AC_DEFUN([AX_LIB_EVENT_DO_CHECK],
|
||||||
|
[
|
||||||
|
# Save our flags.
|
||||||
|
CPPFLAGS_SAVED="$CPPFLAGS"
|
||||||
|
LDFLAGS_SAVED="$LDFLAGS"
|
||||||
|
LIBS_SAVED="$LIBS"
|
||||||
|
LD_LIBRARY_PATH_SAVED="$LD_LIBRARY_PATH"
|
||||||
|
|
||||||
|
# Set our flags if we are checking a specific directory.
|
||||||
|
if test -n "$ax_libevent_path" ; then
|
||||||
|
LIBEVENT_CPPFLAGS="-I$ax_libevent_path/include"
|
||||||
|
LIBEVENT_LDFLAGS="-L$ax_libevent_path/lib"
|
||||||
|
LD_LIBRARY_PATH="$ax_libevent_path/lib:$LD_LIBRARY_PATH"
|
||||||
|
else
|
||||||
|
LIBEVENT_CPPFLAGS=""
|
||||||
|
LIBEVENT_LDFLAGS=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Required flag for libevent.
|
||||||
|
LIBEVENT_LIBS="-levent"
|
||||||
|
|
||||||
|
# Prepare the environment for compilation.
|
||||||
|
CPPFLAGS="$CPPFLAGS $LIBEVENT_CPPFLAGS"
|
||||||
|
LDFLAGS="$LDFLAGS $LIBEVENT_LDFLAGS"
|
||||||
|
LIBS="$LIBS $LIBEVENT_LIBS"
|
||||||
|
export CPPFLAGS
|
||||||
|
export LDFLAGS
|
||||||
|
export LIBS
|
||||||
|
export LD_LIBRARY_PATH
|
||||||
|
|
||||||
|
success=no
|
||||||
|
|
||||||
|
# Compile, link, and run the program. This checks:
|
||||||
|
# - event.h is available for including.
|
||||||
|
# - event_get_version() is available for linking.
|
||||||
|
# - The event version string is lexicographically greater
|
||||||
|
# than the required version.
|
||||||
|
AC_LANG_PUSH([C])
|
||||||
|
dnl This can be changed to AC_LINK_IFELSE if you are cross-compiling,
|
||||||
|
dnl but then the version cannot be checked.
|
||||||
|
AC_LINK_IFELSE([AC_LANG_PROGRAM([[
|
||||||
|
#include <sys/types.h>
|
||||||
|
#include <event.h>
|
||||||
|
]], [[
|
||||||
|
const char* lib_version = event_get_version();
|
||||||
|
const char* wnt_version = "$WANT_LIBEVENT_VERSION";
|
||||||
|
int lib_digits;
|
||||||
|
int wnt_digits;
|
||||||
|
for (;;) {
|
||||||
|
/* If we reached the end of the want version. We have it. */
|
||||||
|
if (*wnt_version == '\0' || *wnt_version == '-') {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* If the want version continues but the lib version does not, */
|
||||||
|
/* we are missing a letter. We don't have it. */
|
||||||
|
if (*lib_version == '\0' || *lib_version == '-') {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
/* In the 1.4 version numbering style, if there are more digits */
|
||||||
|
/* in one version than the other, that one is higher. */
|
||||||
|
for (lib_digits = 0;
|
||||||
|
lib_version[lib_digits] >= '0' &&
|
||||||
|
lib_version[lib_digits] <= '9';
|
||||||
|
lib_digits++)
|
||||||
|
;
|
||||||
|
for (wnt_digits = 0;
|
||||||
|
wnt_version[wnt_digits] >= '0' &&
|
||||||
|
wnt_version[wnt_digits] <= '9';
|
||||||
|
wnt_digits++)
|
||||||
|
;
|
||||||
|
if (lib_digits > wnt_digits) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
if (lib_digits < wnt_digits) {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
/* If we have greater than what we want. We have it. */
|
||||||
|
if (*lib_version > *wnt_version) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* If we have less, we don't. */
|
||||||
|
if (*lib_version < *wnt_version) {
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
lib_version++;
|
||||||
|
wnt_version++;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
]])], [
|
||||||
|
success=yes
|
||||||
|
])
|
||||||
|
AC_LANG_POP([C])
|
||||||
|
|
||||||
|
# Restore flags.
|
||||||
|
CPPFLAGS="$CPPFLAGS_SAVED"
|
||||||
|
LDFLAGS="$LDFLAGS_SAVED"
|
||||||
|
LIBS="$LIBS_SAVED"
|
||||||
|
LD_LIBRARY_PATH="$LD_LIBRARY_PATH_SAVED"
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_LIB_EVENT],
|
||||||
|
[
|
||||||
|
|
||||||
|
dnl Allow search path to be overridden on the command line.
|
||||||
|
AC_ARG_WITH([libevent],
|
||||||
|
AS_HELP_STRING([--with-libevent@<:@=DIR@:>@], [use libevent [default=yes]. Optionally specify the root prefix dir where libevent is installed]),
|
||||||
|
[
|
||||||
|
if test "x$withval" = "xno"; then
|
||||||
|
want_libevent="no"
|
||||||
|
elif test "x$withval" = "xyes"; then
|
||||||
|
want_libevent="yes"
|
||||||
|
ax_libevent_path=""
|
||||||
|
else
|
||||||
|
want_libevent="yes"
|
||||||
|
ax_libevent_path="$withval"
|
||||||
|
fi
|
||||||
|
],
|
||||||
|
[ want_libevent="yes" ; ax_libevent_path="" ])
|
||||||
|
|
||||||
|
|
||||||
|
if test "$want_libevent" = "yes"; then
|
||||||
|
WANT_LIBEVENT_VERSION=ifelse([$1], ,1.2,$1)
|
||||||
|
|
||||||
|
AC_MSG_CHECKING(for libevent >= $WANT_LIBEVENT_VERSION)
|
||||||
|
|
||||||
|
# Run tests.
|
||||||
|
if test -n "$ax_libevent_path"; then
|
||||||
|
AX_LIB_EVENT_DO_CHECK
|
||||||
|
else
|
||||||
|
for ax_libevent_path in "" $lt_sysroot/usr $lt_sysroot/usr/local $lt_sysroot/opt $lt_sysroot/opt/local $lt_sysroot/opt/libevent "$LIBEVENT_ROOT" ; do
|
||||||
|
AX_LIB_EVENT_DO_CHECK
|
||||||
|
if test "$success" = "yes"; then
|
||||||
|
break;
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
if test "$success" != "yes" ; then
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
LIBEVENT_CPPFLAGS=""
|
||||||
|
LIBEVENT_LDFLAGS=""
|
||||||
|
LIBEVENT_LIBS=""
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
AC_DEFINE(HAVE_LIBEVENT,,[define if libevent is available])
|
||||||
|
ax_have_libevent_[]m4_translit([$1], [.], [_])="yes"
|
||||||
|
fi
|
||||||
|
|
||||||
|
ax_have_libevent="$success"
|
||||||
|
|
||||||
|
AC_SUBST(LIBEVENT_CPPFLAGS)
|
||||||
|
AC_SUBST(LIBEVENT_LDFLAGS)
|
||||||
|
AC_SUBST(LIBEVENT_LIBS)
|
||||||
|
fi
|
||||||
|
|
||||||
|
])
|
173
vendor/github.com/apache/thrift/aclocal/ax_lib_zlib.m4
generated
vendored
Normal file
173
vendor/github.com/apache/thrift/aclocal/ax_lib_zlib.m4
generated
vendored
Normal file
|
@ -0,0 +1,173 @@
|
||||||
|
dnl @synopsis AX_LIB_ZLIB([MINIMUM-VERSION])
|
||||||
|
dnl
|
||||||
|
dnl Test for the libz library of a particular version (or newer).
|
||||||
|
dnl
|
||||||
|
dnl If no path to the installed zlib is given, the macro will first try
|
||||||
|
dnl using no -I or -L flags, then searches under /usr, /usr/local, /opt,
|
||||||
|
dnl and /opt/zlib.
|
||||||
|
dnl If these all fail, it will try the $ZLIB_ROOT environment variable.
|
||||||
|
dnl
|
||||||
|
dnl This macro calls:
|
||||||
|
dnl AC_SUBST(ZLIB_CPPFLAGS)
|
||||||
|
dnl AC_SUBST(ZLIB_LDFLAGS)
|
||||||
|
dnl AC_SUBST(ZLIB_LIBS)
|
||||||
|
dnl
|
||||||
|
dnl And (if zlib is found):
|
||||||
|
dnl AC_DEFINE(HAVE_ZLIB)
|
||||||
|
dnl
|
||||||
|
dnl It also leaves the shell variables "success" and "ax_have_zlib"
|
||||||
|
dnl set to "yes" or "no".
|
||||||
|
dnl
|
||||||
|
dnl NOTE: This macro does not currently work for cross-compiling,
|
||||||
|
dnl but it can be easily modified to allow it. (grep "cross").
|
||||||
|
dnl
|
||||||
|
dnl @category InstalledPackages
|
||||||
|
dnl @category C
|
||||||
|
dnl @version 2007-09-12
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
dnl Input: ax_zlib_path, WANT_ZLIB_VERSION
|
||||||
|
dnl Output: success=yes/no
|
||||||
|
AC_DEFUN([AX_LIB_ZLIB_DO_CHECK],
|
||||||
|
[
|
||||||
|
# Save our flags.
|
||||||
|
CPPFLAGS_SAVED="$CPPFLAGS"
|
||||||
|
LDFLAGS_SAVED="$LDFLAGS"
|
||||||
|
LIBS_SAVED="$LIBS"
|
||||||
|
LD_LIBRARY_PATH_SAVED="$LD_LIBRARY_PATH"
|
||||||
|
|
||||||
|
# Set our flags if we are checking a specific directory.
|
||||||
|
if test -n "$ax_zlib_path" ; then
|
||||||
|
ZLIB_CPPFLAGS="-I$ax_zlib_path/include"
|
||||||
|
ZLIB_LDFLAGS="-L$ax_zlib_path/lib"
|
||||||
|
LD_LIBRARY_PATH="$ax_zlib_path/lib:$LD_LIBRARY_PATH"
|
||||||
|
else
|
||||||
|
ZLIB_CPPFLAGS=""
|
||||||
|
ZLIB_LDFLAGS=""
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Required flag for zlib.
|
||||||
|
ZLIB_LIBS="-lz"
|
||||||
|
|
||||||
|
# Prepare the environment for compilation.
|
||||||
|
CPPFLAGS="$CPPFLAGS $ZLIB_CPPFLAGS"
|
||||||
|
LDFLAGS="$LDFLAGS $ZLIB_LDFLAGS"
|
||||||
|
LIBS="$LIBS $ZLIB_LIBS"
|
||||||
|
export CPPFLAGS
|
||||||
|
export LDFLAGS
|
||||||
|
export LIBS
|
||||||
|
export LD_LIBRARY_PATH
|
||||||
|
|
||||||
|
success=no
|
||||||
|
|
||||||
|
# Compile, link, and run the program. This checks:
|
||||||
|
# - zlib.h is available for including.
|
||||||
|
# - zlibVersion() is available for linking.
|
||||||
|
# - ZLIB_VERNUM is greater than or equal to the desired version.
|
||||||
|
# - ZLIB_VERSION (defined in zlib.h) matches zlibVersion()
|
||||||
|
# (defined in the library).
|
||||||
|
AC_LANG_PUSH([C])
|
||||||
|
dnl This can be changed to AC_LINK_IFELSE if you are cross-compiling.
|
||||||
|
AC_LINK_IFELSE([AC_LANG_PROGRAM([[
|
||||||
|
#include <zlib.h>
|
||||||
|
#if ZLIB_VERNUM >= 0x$WANT_ZLIB_VERSION
|
||||||
|
#else
|
||||||
|
# error zlib is too old
|
||||||
|
#endif
|
||||||
|
]], [[
|
||||||
|
const char* lib_version = zlibVersion();
|
||||||
|
const char* hdr_version = ZLIB_VERSION;
|
||||||
|
for (;;) {
|
||||||
|
if (*lib_version != *hdr_version) {
|
||||||
|
/* If this happens, your zlib header doesn't match your zlib */
|
||||||
|
/* library. That is really bad. */
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
if (*lib_version == '\0') {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
lib_version++;
|
||||||
|
hdr_version++;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
]])], [
|
||||||
|
success=yes
|
||||||
|
])
|
||||||
|
AC_LANG_POP([C])
|
||||||
|
|
||||||
|
# Restore flags.
|
||||||
|
CPPFLAGS="$CPPFLAGS_SAVED"
|
||||||
|
LDFLAGS="$LDFLAGS_SAVED"
|
||||||
|
LIBS="$LIBS_SAVED"
|
||||||
|
LD_LIBRARY_PATH="$LD_LIBRARY_PATH_SAVED"
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFUN([AX_LIB_ZLIB],
|
||||||
|
[
|
||||||
|
|
||||||
|
dnl Allow search path to be overridden on the command line.
|
||||||
|
AC_ARG_WITH([zlib],
|
||||||
|
AS_HELP_STRING([--with-zlib@<:@=DIR@:>@], [use zlib (default is yes) - it is possible to specify an alternate root directory for zlib]),
|
||||||
|
[
|
||||||
|
if test "x$withval" = "xno"; then
|
||||||
|
want_zlib="no"
|
||||||
|
elif test "x$withval" = "xyes"; then
|
||||||
|
want_zlib="yes"
|
||||||
|
ax_zlib_path=""
|
||||||
|
else
|
||||||
|
want_zlib="yes"
|
||||||
|
ax_zlib_path="$withval"
|
||||||
|
fi
|
||||||
|
],
|
||||||
|
[want_zlib="yes" ; ax_zlib_path="" ])
|
||||||
|
|
||||||
|
|
||||||
|
if test "$want_zlib" = "yes"; then
|
||||||
|
# Parse out the version.
|
||||||
|
zlib_version_req=ifelse([$1], ,1.2.3,$1)
|
||||||
|
zlib_version_req_major=`expr $zlib_version_req : '\([[0-9]]*\)'`
|
||||||
|
zlib_version_req_minor=`expr $zlib_version_req : '[[0-9]]*\.\([[0-9]]*\)'`
|
||||||
|
zlib_version_req_patch=`expr $zlib_version_req : '[[0-9]]*\.[[0-9]]*\.\([[0-9]]*\)'`
|
||||||
|
if test -z "$zlib_version_req_patch" ; then
|
||||||
|
zlib_version_req_patch="0"
|
||||||
|
fi
|
||||||
|
WANT_ZLIB_VERSION=`expr $zlib_version_req_major \* 1000 \+ $zlib_version_req_minor \* 100 \+ $zlib_version_req_patch \* 10`
|
||||||
|
|
||||||
|
AC_MSG_CHECKING(for zlib >= $zlib_version_req)
|
||||||
|
|
||||||
|
# Run tests.
|
||||||
|
if test -n "$ax_zlib_path"; then
|
||||||
|
AX_LIB_ZLIB_DO_CHECK
|
||||||
|
else
|
||||||
|
for ax_zlib_path in "" /usr /usr/local /opt /opt/zlib "$ZLIB_ROOT" ; do
|
||||||
|
AX_LIB_ZLIB_DO_CHECK
|
||||||
|
if test "$success" = "yes"; then
|
||||||
|
break;
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
if test "$success" != "yes" ; then
|
||||||
|
AC_MSG_RESULT(no)
|
||||||
|
ZLIB_CPPFLAGS=""
|
||||||
|
ZLIB_LDFLAGS=""
|
||||||
|
ZLIB_LIBS=""
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(yes)
|
||||||
|
AC_DEFINE(HAVE_ZLIB,,[define if zlib is available])
|
||||||
|
fi
|
||||||
|
|
||||||
|
ax_have_zlib="$success"
|
||||||
|
|
||||||
|
AC_SUBST(ZLIB_CPPFLAGS)
|
||||||
|
AC_SUBST(ZLIB_LDFLAGS)
|
||||||
|
AC_SUBST(ZLIB_LIBS)
|
||||||
|
fi
|
||||||
|
|
||||||
|
])
|
664
vendor/github.com/apache/thrift/aclocal/ax_lua.m4
generated
vendored
Normal file
664
vendor/github.com/apache/thrift/aclocal/ax_lua.m4
generated
vendored
Normal file
|
@ -0,0 +1,664 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_lua.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_PROG_LUA[([MINIMUM-VERSION], [TOO-BIG-VERSION], [ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])]
|
||||||
|
# AX_LUA_HEADERS[([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])]
|
||||||
|
# AX_LUA_LIBS[([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])]
|
||||||
|
# AX_LUA_READLINE[([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])]
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Detect a Lua interpreter, optionally specifying a minimum and maximum
|
||||||
|
# version number. Set up important Lua paths, such as the directories in
|
||||||
|
# which to install scripts and modules (shared libraries).
|
||||||
|
#
|
||||||
|
# Also detect Lua headers and libraries. The Lua version contained in the
|
||||||
|
# header is checked to match the Lua interpreter version exactly. When
|
||||||
|
# searching for Lua libraries, the version number is used as a suffix.
|
||||||
|
# This is done with the goal of supporting multiple Lua installs (5.1,
|
||||||
|
# 5.2, and 5.3 side-by-side).
|
||||||
|
#
|
||||||
|
# A note on compatibility with previous versions: This file has been
|
||||||
|
# mostly rewritten for serial 18. Most developers should be able to use
|
||||||
|
# these macros without needing to modify configure.ac. Care has been taken
|
||||||
|
# to preserve each macro's behavior, but there are some differences:
|
||||||
|
#
|
||||||
|
# 1) AX_WITH_LUA is deprecated; it now expands to the exact same thing as
|
||||||
|
# AX_PROG_LUA with no arguments.
|
||||||
|
#
|
||||||
|
# 2) AX_LUA_HEADERS now checks that the version number defined in lua.h
|
||||||
|
# matches the interpreter version. AX_LUA_HEADERS_VERSION is therefore
|
||||||
|
# unnecessary, so it is deprecated and does not expand to anything.
|
||||||
|
#
|
||||||
|
# 3) The configure flag --with-lua-suffix no longer exists; the user
|
||||||
|
# should instead specify the LUA precious variable on the command line.
|
||||||
|
# See the AX_PROG_LUA description for details.
|
||||||
|
#
|
||||||
|
# Please read the macro descriptions below for more information.
|
||||||
|
#
|
||||||
|
# This file was inspired by Andrew Dalke's and James Henstridge's
|
||||||
|
# python.m4 and Tom Payne's, Matthieu Moy's, and Reuben Thomas's ax_lua.m4
|
||||||
|
# (serial 17). Basically, this file is a mash-up of those two files. I
|
||||||
|
# like to think it combines the best of the two!
|
||||||
|
#
|
||||||
|
# AX_PROG_LUA: Search for the Lua interpreter, and set up important Lua
|
||||||
|
# paths. Adds precious variable LUA, which may contain the path of the Lua
|
||||||
|
# interpreter. If LUA is blank, the user's path is searched for an
|
||||||
|
# suitable interpreter.
|
||||||
|
#
|
||||||
|
# If MINIMUM-VERSION is supplied, then only Lua interpreters with a
|
||||||
|
# version number greater or equal to MINIMUM-VERSION will be accepted. If
|
||||||
|
# TOO-BIG-VERSION is also supplied, then only Lua interpreters with a
|
||||||
|
# version number greater or equal to MINIMUM-VERSION and less than
|
||||||
|
# TOO-BIG-VERSION will be accepted.
|
||||||
|
#
|
||||||
|
# The Lua version number, LUA_VERSION, is found from the interpreter, and
|
||||||
|
# substituted. LUA_PLATFORM is also found, but not currently supported (no
|
||||||
|
# standard representation).
|
||||||
|
#
|
||||||
|
# Finally, the macro finds four paths:
|
||||||
|
#
|
||||||
|
# luadir Directory to install Lua scripts.
|
||||||
|
# pkgluadir $luadir/$PACKAGE
|
||||||
|
# luaexecdir Directory to install Lua modules.
|
||||||
|
# pkgluaexecdir $luaexecdir/$PACKAGE
|
||||||
|
#
|
||||||
|
# These paths are found based on $prefix, $exec_prefix, Lua's
|
||||||
|
# package.path, and package.cpath. The first path of package.path
|
||||||
|
# beginning with $prefix is selected as luadir. The first path of
|
||||||
|
# package.cpath beginning with $exec_prefix is used as luaexecdir. This
|
||||||
|
# should work on all reasonable Lua installations. If a path cannot be
|
||||||
|
# determined, a default path is used. Of course, the user can override
|
||||||
|
# these later when invoking make.
|
||||||
|
#
|
||||||
|
# luadir Default: $prefix/share/lua/$LUA_VERSION
|
||||||
|
# luaexecdir Default: $exec_prefix/lib/lua/$LUA_VERSION
|
||||||
|
#
|
||||||
|
# These directories can be used by Automake as install destinations. The
|
||||||
|
# variable name minus 'dir' needs to be used as a prefix to the
|
||||||
|
# appropriate Automake primary, e.g. lua_SCRIPS or luaexec_LIBRARIES.
|
||||||
|
#
|
||||||
|
# If an acceptable Lua interpreter is found, then ACTION-IF-FOUND is
|
||||||
|
# performed, otherwise ACTION-IF-NOT-FOUND is preformed. If ACTION-IF-NOT-
|
||||||
|
# FOUND is blank, then it will default to printing an error. To prevent
|
||||||
|
# the default behavior, give ':' as an action.
|
||||||
|
#
|
||||||
|
# AX_LUA_HEADERS: Search for Lua headers. Requires that AX_PROG_LUA be
|
||||||
|
# expanded before this macro. Adds precious variable LUA_INCLUDE, which
|
||||||
|
# may contain Lua specific include flags, e.g. -I/usr/include/lua5.1. If
|
||||||
|
# LUA_INCLUDE is blank, then this macro will attempt to find suitable
|
||||||
|
# flags.
|
||||||
|
#
|
||||||
|
# LUA_INCLUDE can be used by Automake to compile Lua modules or
|
||||||
|
# executables with embedded interpreters. The *_CPPFLAGS variables should
|
||||||
|
# be used for this purpose, e.g. myprog_CPPFLAGS = $(LUA_INCLUDE).
|
||||||
|
#
|
||||||
|
# This macro searches for the header lua.h (and others). The search is
|
||||||
|
# performed with a combination of CPPFLAGS, CPATH, etc, and LUA_INCLUDE.
|
||||||
|
# If the search is unsuccessful, then some common directories are tried.
|
||||||
|
# If the headers are then found, then LUA_INCLUDE is set accordingly.
|
||||||
|
#
|
||||||
|
# The paths automatically searched are:
|
||||||
|
#
|
||||||
|
# * /usr/include/luaX.Y
|
||||||
|
# * /usr/include/lua/X.Y
|
||||||
|
# * /usr/include/luaXY
|
||||||
|
# * /usr/local/include/luaX.Y
|
||||||
|
# * /usr/local/include/lua-X.Y
|
||||||
|
# * /usr/local/include/lua/X.Y
|
||||||
|
# * /usr/local/include/luaXY
|
||||||
|
#
|
||||||
|
# (Where X.Y is the Lua version number, e.g. 5.1.)
|
||||||
|
#
|
||||||
|
# The Lua version number found in the headers is always checked to match
|
||||||
|
# the Lua interpreter's version number. Lua headers with mismatched
|
||||||
|
# version numbers are not accepted.
|
||||||
|
#
|
||||||
|
# If headers are found, then ACTION-IF-FOUND is performed, otherwise
|
||||||
|
# ACTION-IF-NOT-FOUND is performed. If ACTION-IF-NOT-FOUND is blank, then
|
||||||
|
# it will default to printing an error. To prevent the default behavior,
|
||||||
|
# set the action to ':'.
|
||||||
|
#
|
||||||
|
# AX_LUA_LIBS: Search for Lua libraries. Requires that AX_PROG_LUA be
|
||||||
|
# expanded before this macro. Adds precious variable LUA_LIB, which may
|
||||||
|
# contain Lua specific linker flags, e.g. -llua5.1. If LUA_LIB is blank,
|
||||||
|
# then this macro will attempt to find suitable flags.
|
||||||
|
#
|
||||||
|
# LUA_LIB can be used by Automake to link Lua modules or executables with
|
||||||
|
# embedded interpreters. The *_LIBADD and *_LDADD variables should be used
|
||||||
|
# for this purpose, e.g. mymod_LIBADD = $(LUA_LIB).
|
||||||
|
#
|
||||||
|
# This macro searches for the Lua library. More technically, it searches
|
||||||
|
# for a library containing the function lua_load. The search is performed
|
||||||
|
# with a combination of LIBS, LIBRARY_PATH, and LUA_LIB.
|
||||||
|
#
|
||||||
|
# If the search determines that some linker flags are missing, then those
|
||||||
|
# flags will be added to LUA_LIB.
|
||||||
|
#
|
||||||
|
# If libraries are found, then ACTION-IF-FOUND is performed, otherwise
|
||||||
|
# ACTION-IF-NOT-FOUND is performed. If ACTION-IF-NOT-FOUND is blank, then
|
||||||
|
# it will default to printing an error. To prevent the default behavior,
|
||||||
|
# set the action to ':'.
|
||||||
|
#
|
||||||
|
# AX_LUA_READLINE: Search for readline headers and libraries. Requires the
|
||||||
|
# AX_LIB_READLINE macro, which is provided by ax_lib_readline.m4 from the
|
||||||
|
# Autoconf Archive.
|
||||||
|
#
|
||||||
|
# If a readline compatible library is found, then ACTION-IF-FOUND is
|
||||||
|
# performed, otherwise ACTION-IF-NOT-FOUND is performed.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2015 Reuben Thomas <rrt@sc3d.org>
|
||||||
|
# Copyright (c) 2014 Tim Perkins <tprk77@gmail.com>
|
||||||
|
#
|
||||||
|
# This program is free software: you can redistribute it and/or modify it
|
||||||
|
# under the terms of the GNU General Public License as published by the
|
||||||
|
# Free Software Foundation, either version 3 of the License, or (at your
|
||||||
|
# option) any later version.
|
||||||
|
#
|
||||||
|
# This program is distributed in the hope that it will be useful, but
|
||||||
|
# WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General
|
||||||
|
# Public License for more details.
|
||||||
|
#
|
||||||
|
# You should have received a copy of the GNU General Public License along
|
||||||
|
# with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
#
|
||||||
|
# As a special exception, the respective Autoconf Macro's copyright owner
|
||||||
|
# gives unlimited permission to copy, distribute and modify the configure
|
||||||
|
# scripts that are the output of Autoconf when processing the Macro. You
|
||||||
|
# need not follow the terms of the GNU General Public License when using
|
||||||
|
# or distributing such scripts, even though portions of the text of the
|
||||||
|
# Macro appear in them. The GNU General Public License (GPL) does govern
|
||||||
|
# all other use of the material that constitutes the Autoconf Macro.
|
||||||
|
#
|
||||||
|
# This special exception to the GPL applies to versions of the Autoconf
|
||||||
|
# Macro released by the Autoconf Archive. When you make and distribute a
|
||||||
|
# modified version of the Autoconf Macro, you may extend this special
|
||||||
|
# exception to the GPL to apply to your modified version as well.
|
||||||
|
|
||||||
|
#serial 40
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl AX_PROG_LUA([MINIMUM-VERSION], [TOO-BIG-VERSION],
|
||||||
|
dnl [ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([AX_PROG_LUA],
|
||||||
|
[
|
||||||
|
dnl Check for required tools.
|
||||||
|
AC_REQUIRE([AC_PROG_GREP])
|
||||||
|
AC_REQUIRE([AC_PROG_SED])
|
||||||
|
|
||||||
|
dnl Make LUA a precious variable.
|
||||||
|
AC_ARG_VAR([LUA], [The Lua interpreter, e.g. /usr/bin/lua5.1])
|
||||||
|
|
||||||
|
dnl Find a Lua interpreter.
|
||||||
|
m4_define_default([_AX_LUA_INTERPRETER_LIST],
|
||||||
|
[lua lua5.3 lua53 lua5.2 lua52 lua5.1 lua51 lua50])
|
||||||
|
|
||||||
|
m4_if([$1], [],
|
||||||
|
[ dnl No version check is needed. Find any Lua interpreter.
|
||||||
|
AS_IF([test "x$LUA" = 'x'],
|
||||||
|
[AC_PATH_PROGS([LUA], [_AX_LUA_INTERPRETER_LIST], [:])])
|
||||||
|
ax_display_LUA='lua'
|
||||||
|
|
||||||
|
AS_IF([test "x$LUA" != 'x:'],
|
||||||
|
[ dnl At least check if this is a Lua interpreter.
|
||||||
|
AC_MSG_CHECKING([if $LUA is a Lua interpreter])
|
||||||
|
_AX_LUA_CHK_IS_INTRP([$LUA],
|
||||||
|
[AC_MSG_RESULT([yes])],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([not a Lua interpreter])
|
||||||
|
])
|
||||||
|
])
|
||||||
|
],
|
||||||
|
[ dnl A version check is needed.
|
||||||
|
AS_IF([test "x$LUA" != 'x'],
|
||||||
|
[ dnl Check if this is a Lua interpreter.
|
||||||
|
AC_MSG_CHECKING([if $LUA is a Lua interpreter])
|
||||||
|
_AX_LUA_CHK_IS_INTRP([$LUA],
|
||||||
|
[AC_MSG_RESULT([yes])],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([not a Lua interpreter])
|
||||||
|
])
|
||||||
|
dnl Check the version.
|
||||||
|
m4_if([$2], [],
|
||||||
|
[_ax_check_text="whether $LUA version >= $1"],
|
||||||
|
[_ax_check_text="whether $LUA version >= $1, < $2"])
|
||||||
|
AC_MSG_CHECKING([$_ax_check_text])
|
||||||
|
_AX_LUA_CHK_VER([$LUA], [$1], [$2],
|
||||||
|
[AC_MSG_RESULT([yes])],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([version is out of range for specified LUA])])
|
||||||
|
ax_display_LUA=$LUA
|
||||||
|
],
|
||||||
|
[ dnl Try each interpreter until we find one that satisfies VERSION.
|
||||||
|
m4_if([$2], [],
|
||||||
|
[_ax_check_text="for a Lua interpreter with version >= $1"],
|
||||||
|
[_ax_check_text="for a Lua interpreter with version >= $1, < $2"])
|
||||||
|
AC_CACHE_CHECK([$_ax_check_text],
|
||||||
|
[ax_cv_pathless_LUA],
|
||||||
|
[ for ax_cv_pathless_LUA in _AX_LUA_INTERPRETER_LIST none; do
|
||||||
|
test "x$ax_cv_pathless_LUA" = 'xnone' && break
|
||||||
|
_AX_LUA_CHK_IS_INTRP([$ax_cv_pathless_LUA], [], [continue])
|
||||||
|
_AX_LUA_CHK_VER([$ax_cv_pathless_LUA], [$1], [$2], [break])
|
||||||
|
done
|
||||||
|
])
|
||||||
|
dnl Set $LUA to the absolute path of $ax_cv_pathless_LUA.
|
||||||
|
AS_IF([test "x$ax_cv_pathless_LUA" = 'xnone'],
|
||||||
|
[LUA=':'],
|
||||||
|
[AC_PATH_PROG([LUA], [$ax_cv_pathless_LUA])])
|
||||||
|
ax_display_LUA=$ax_cv_pathless_LUA
|
||||||
|
])
|
||||||
|
])
|
||||||
|
|
||||||
|
AS_IF([test "x$LUA" = 'x:'],
|
||||||
|
[ dnl Run any user-specified action, or abort.
|
||||||
|
m4_default([$4], [AC_MSG_ERROR([cannot find suitable Lua interpreter])])
|
||||||
|
],
|
||||||
|
[ dnl Query Lua for its version number.
|
||||||
|
AC_CACHE_CHECK([for $ax_display_LUA version],
|
||||||
|
[ax_cv_lua_version],
|
||||||
|
[ dnl Get the interpreter version in X.Y format. This should work for
|
||||||
|
dnl interpreters version 5.0 and beyond.
|
||||||
|
ax_cv_lua_version=[`$LUA -e '
|
||||||
|
-- return a version number in X.Y format
|
||||||
|
local _, _, ver = string.find(_VERSION, "^Lua (%d+%.%d+)")
|
||||||
|
print(ver)'`]
|
||||||
|
])
|
||||||
|
AS_IF([test "x$ax_cv_lua_version" = 'x'],
|
||||||
|
[AC_MSG_ERROR([invalid Lua version number])])
|
||||||
|
AC_SUBST([LUA_VERSION], [$ax_cv_lua_version])
|
||||||
|
AC_SUBST([LUA_SHORT_VERSION], [`echo "$LUA_VERSION" | $SED 's|\.||'`])
|
||||||
|
|
||||||
|
dnl The following check is not supported:
|
||||||
|
dnl At times (like when building shared libraries) you may want to know
|
||||||
|
dnl which OS platform Lua thinks this is.
|
||||||
|
AC_CACHE_CHECK([for $ax_display_LUA platform],
|
||||||
|
[ax_cv_lua_platform],
|
||||||
|
[ax_cv_lua_platform=[`$LUA -e 'print("unknown")'`]])
|
||||||
|
AC_SUBST([LUA_PLATFORM], [$ax_cv_lua_platform])
|
||||||
|
|
||||||
|
dnl Use the values of $prefix and $exec_prefix for the corresponding
|
||||||
|
dnl values of LUA_PREFIX and LUA_EXEC_PREFIX. These are made distinct
|
||||||
|
dnl variables so they can be overridden if need be. However, the general
|
||||||
|
dnl consensus is that you shouldn't need this ability.
|
||||||
|
AC_SUBST([LUA_PREFIX], ['${prefix}'])
|
||||||
|
AC_SUBST([LUA_EXEC_PREFIX], ['${exec_prefix}'])
|
||||||
|
|
||||||
|
dnl Lua provides no way to query the script directory, and instead
|
||||||
|
dnl provides LUA_PATH. However, we should be able to make a safe educated
|
||||||
|
dnl guess. If the built-in search path contains a directory which is
|
||||||
|
dnl prefixed by $prefix, then we can store scripts there. The first
|
||||||
|
dnl matching path will be used.
|
||||||
|
AC_CACHE_CHECK([for $ax_display_LUA script directory],
|
||||||
|
[ax_cv_lua_luadir],
|
||||||
|
[ AS_IF([test "x$prefix" = 'xNONE'],
|
||||||
|
[ax_lua_prefix=$ac_default_prefix],
|
||||||
|
[ax_lua_prefix=$prefix])
|
||||||
|
|
||||||
|
dnl Initialize to the default path.
|
||||||
|
ax_cv_lua_luadir="$LUA_PREFIX/share/lua/$LUA_VERSION"
|
||||||
|
|
||||||
|
dnl Try to find a path with the prefix.
|
||||||
|
_AX_LUA_FND_PRFX_PTH([$LUA], [$ax_lua_prefix], [script])
|
||||||
|
AS_IF([test "x$ax_lua_prefixed_path" != 'x'],
|
||||||
|
[ dnl Fix the prefix.
|
||||||
|
_ax_strip_prefix=`echo "$ax_lua_prefix" | $SED 's|.|.|g'`
|
||||||
|
ax_cv_lua_luadir=`echo "$ax_lua_prefixed_path" | \
|
||||||
|
$SED "s|^$_ax_strip_prefix|$LUA_PREFIX|"`
|
||||||
|
])
|
||||||
|
])
|
||||||
|
AC_SUBST([luadir], [$ax_cv_lua_luadir])
|
||||||
|
AC_SUBST([pkgluadir], [\${luadir}/$PACKAGE])
|
||||||
|
|
||||||
|
dnl Lua provides no way to query the module directory, and instead
|
||||||
|
dnl provides LUA_PATH. However, we should be able to make a safe educated
|
||||||
|
dnl guess. If the built-in search path contains a directory which is
|
||||||
|
dnl prefixed by $exec_prefix, then we can store modules there. The first
|
||||||
|
dnl matching path will be used.
|
||||||
|
AC_CACHE_CHECK([for $ax_display_LUA module directory],
|
||||||
|
[ax_cv_lua_luaexecdir],
|
||||||
|
[ AS_IF([test "x$exec_prefix" = 'xNONE'],
|
||||||
|
[ax_lua_exec_prefix=$ax_lua_prefix],
|
||||||
|
[ax_lua_exec_prefix=$exec_prefix])
|
||||||
|
|
||||||
|
dnl Initialize to the default path.
|
||||||
|
ax_cv_lua_luaexecdir="$LUA_EXEC_PREFIX/lib/lua/$LUA_VERSION"
|
||||||
|
|
||||||
|
dnl Try to find a path with the prefix.
|
||||||
|
_AX_LUA_FND_PRFX_PTH([$LUA],
|
||||||
|
[$ax_lua_exec_prefix], [module])
|
||||||
|
AS_IF([test "x$ax_lua_prefixed_path" != 'x'],
|
||||||
|
[ dnl Fix the prefix.
|
||||||
|
_ax_strip_prefix=`echo "$ax_lua_exec_prefix" | $SED 's|.|.|g'`
|
||||||
|
ax_cv_lua_luaexecdir=`echo "$ax_lua_prefixed_path" | \
|
||||||
|
$SED "s|^$_ax_strip_prefix|$LUA_EXEC_PREFIX|"`
|
||||||
|
])
|
||||||
|
])
|
||||||
|
AC_SUBST([luaexecdir], [$ax_cv_lua_luaexecdir])
|
||||||
|
AC_SUBST([pkgluaexecdir], [\${luaexecdir}/$PACKAGE])
|
||||||
|
|
||||||
|
dnl Run any user specified action.
|
||||||
|
$3
|
||||||
|
])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl AX_WITH_LUA is now the same thing as AX_PROG_LUA.
|
||||||
|
AC_DEFUN([AX_WITH_LUA],
|
||||||
|
[
|
||||||
|
AC_MSG_WARN([[$0 is deprecated, please use AX_PROG_LUA instead]])
|
||||||
|
AX_PROG_LUA
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl _AX_LUA_CHK_IS_INTRP(PROG, [ACTION-IF-TRUE], [ACTION-IF-FALSE])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([_AX_LUA_CHK_IS_INTRP],
|
||||||
|
[
|
||||||
|
dnl A minimal Lua factorial to prove this is an interpreter. This should work
|
||||||
|
dnl for Lua interpreters version 5.0 and beyond.
|
||||||
|
_ax_lua_factorial=[`$1 2>/dev/null -e '
|
||||||
|
-- a simple factorial
|
||||||
|
function fact (n)
|
||||||
|
if n == 0 then
|
||||||
|
return 1
|
||||||
|
else
|
||||||
|
return n * fact(n-1)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
print("fact(5) is " .. fact(5))'`]
|
||||||
|
AS_IF([test "$_ax_lua_factorial" = 'fact(5) is 120'],
|
||||||
|
[$2], [$3])
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl _AX_LUA_CHK_VER(PROG, MINIMUM-VERSION, [TOO-BIG-VERSION],
|
||||||
|
dnl [ACTION-IF-TRUE], [ACTION-IF-FALSE])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([_AX_LUA_CHK_VER],
|
||||||
|
[
|
||||||
|
dnl Check that the Lua version is within the bounds. Only the major and minor
|
||||||
|
dnl version numbers are considered. This should work for Lua interpreters
|
||||||
|
dnl version 5.0 and beyond.
|
||||||
|
_ax_lua_good_version=[`$1 -e '
|
||||||
|
-- a script to compare versions
|
||||||
|
function verstr2num(verstr)
|
||||||
|
local _, _, majorver, minorver = string.find(verstr, "^(%d+)%.(%d+)")
|
||||||
|
if majorver and minorver then
|
||||||
|
return tonumber(majorver) * 100 + tonumber(minorver)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
local minver = verstr2num("$2")
|
||||||
|
local _, _, trimver = string.find(_VERSION, "^Lua (.*)")
|
||||||
|
local ver = verstr2num(trimver)
|
||||||
|
local maxver = verstr2num("$3") or 1e9
|
||||||
|
if minver <= ver and ver < maxver then
|
||||||
|
print("yes")
|
||||||
|
else
|
||||||
|
print("no")
|
||||||
|
end'`]
|
||||||
|
AS_IF([test "x$_ax_lua_good_version" = "xyes"],
|
||||||
|
[$4], [$5])
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl _AX_LUA_FND_PRFX_PTH(PROG, PREFIX, SCRIPT-OR-MODULE-DIR)
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([_AX_LUA_FND_PRFX_PTH],
|
||||||
|
[
|
||||||
|
dnl Get the script or module directory by querying the Lua interpreter,
|
||||||
|
dnl filtering on the given prefix, and selecting the shallowest path. If no
|
||||||
|
dnl path is found matching the prefix, the result will be an empty string.
|
||||||
|
dnl The third argument determines the type of search, it can be 'script' or
|
||||||
|
dnl 'module'. Supplying 'script' will perform the search with package.path
|
||||||
|
dnl and LUA_PATH, and supplying 'module' will search with package.cpath and
|
||||||
|
dnl LUA_CPATH. This is done for compatibility with Lua 5.0.
|
||||||
|
|
||||||
|
ax_lua_prefixed_path=[`$1 -e '
|
||||||
|
-- get the path based on search type
|
||||||
|
local searchtype = "$3"
|
||||||
|
local paths = ""
|
||||||
|
if searchtype == "script" then
|
||||||
|
paths = (package and package.path) or LUA_PATH
|
||||||
|
elseif searchtype == "module" then
|
||||||
|
paths = (package and package.cpath) or LUA_CPATH
|
||||||
|
end
|
||||||
|
-- search for the prefix
|
||||||
|
local prefix = "'$2'"
|
||||||
|
local minpath = ""
|
||||||
|
local mindepth = 1e9
|
||||||
|
string.gsub(paths, "(@<:@^;@:>@+)",
|
||||||
|
function (path)
|
||||||
|
path = string.gsub(path, "%?.*$", "")
|
||||||
|
path = string.gsub(path, "/@<:@^/@:>@*$", "")
|
||||||
|
if string.find(path, prefix) then
|
||||||
|
local depth = string.len(string.gsub(path, "@<:@^/@:>@", ""))
|
||||||
|
if depth < mindepth then
|
||||||
|
minpath = path
|
||||||
|
mindepth = depth
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end)
|
||||||
|
print(minpath)'`]
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl AX_LUA_HEADERS([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([AX_LUA_HEADERS],
|
||||||
|
[
|
||||||
|
dnl Check for LUA_VERSION.
|
||||||
|
AC_MSG_CHECKING([if LUA_VERSION is defined])
|
||||||
|
AS_IF([test "x$LUA_VERSION" != 'x'],
|
||||||
|
[AC_MSG_RESULT([yes])],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([cannot check Lua headers without knowing LUA_VERSION])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Make LUA_INCLUDE a precious variable.
|
||||||
|
AC_ARG_VAR([LUA_INCLUDE], [The Lua includes, e.g. -I/usr/include/lua5.1])
|
||||||
|
|
||||||
|
dnl Some default directories to search.
|
||||||
|
LUA_SHORT_VERSION=`echo "$LUA_VERSION" | $SED 's|\.||'`
|
||||||
|
m4_define_default([_AX_LUA_INCLUDE_LIST],
|
||||||
|
[ /usr/include/lua$LUA_VERSION \
|
||||||
|
/usr/include/lua-$LUA_VERSION \
|
||||||
|
/usr/include/lua/$LUA_VERSION \
|
||||||
|
/usr/include/lua$LUA_SHORT_VERSION \
|
||||||
|
/usr/local/include/lua$LUA_VERSION \
|
||||||
|
/usr/local/include/lua-$LUA_VERSION \
|
||||||
|
/usr/local/include/lua/$LUA_VERSION \
|
||||||
|
/usr/local/include/lua$LUA_SHORT_VERSION \
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Try to find the headers.
|
||||||
|
_ax_lua_saved_cppflags=$CPPFLAGS
|
||||||
|
CPPFLAGS="$CPPFLAGS $LUA_INCLUDE"
|
||||||
|
AC_CHECK_HEADERS([lua.h lualib.h lauxlib.h luaconf.h])
|
||||||
|
CPPFLAGS=$_ax_lua_saved_cppflags
|
||||||
|
|
||||||
|
dnl Try some other directories if LUA_INCLUDE was not set.
|
||||||
|
AS_IF([test "x$LUA_INCLUDE" = 'x' &&
|
||||||
|
test "x$ac_cv_header_lua_h" != 'xyes'],
|
||||||
|
[ dnl Try some common include paths.
|
||||||
|
for _ax_include_path in _AX_LUA_INCLUDE_LIST; do
|
||||||
|
test ! -d "$_ax_include_path" && continue
|
||||||
|
|
||||||
|
AC_MSG_CHECKING([for Lua headers in])
|
||||||
|
AC_MSG_RESULT([$_ax_include_path])
|
||||||
|
|
||||||
|
AS_UNSET([ac_cv_header_lua_h])
|
||||||
|
AS_UNSET([ac_cv_header_lualib_h])
|
||||||
|
AS_UNSET([ac_cv_header_lauxlib_h])
|
||||||
|
AS_UNSET([ac_cv_header_luaconf_h])
|
||||||
|
|
||||||
|
_ax_lua_saved_cppflags=$CPPFLAGS
|
||||||
|
CPPFLAGS="$CPPFLAGS -I$_ax_include_path"
|
||||||
|
AC_CHECK_HEADERS([lua.h lualib.h lauxlib.h luaconf.h])
|
||||||
|
CPPFLAGS=$_ax_lua_saved_cppflags
|
||||||
|
|
||||||
|
AS_IF([test "x$ac_cv_header_lua_h" = 'xyes'],
|
||||||
|
[ LUA_INCLUDE="-I$_ax_include_path"
|
||||||
|
break
|
||||||
|
])
|
||||||
|
done
|
||||||
|
])
|
||||||
|
|
||||||
|
AS_IF([test "x$ac_cv_header_lua_h" = 'xyes'],
|
||||||
|
[ dnl Make a program to print LUA_VERSION defined in the header.
|
||||||
|
dnl TODO It would be really nice if we could do this without compiling a
|
||||||
|
dnl program, then it would work when cross compiling. But I'm not sure how
|
||||||
|
dnl to do this reliably. For now, assume versions match when cross compiling.
|
||||||
|
|
||||||
|
AS_IF([test "x$cross_compiling" != 'xyes'],
|
||||||
|
[ AC_CACHE_CHECK([for Lua header version],
|
||||||
|
[ax_cv_lua_header_version],
|
||||||
|
[ _ax_lua_saved_cppflags=$CPPFLAGS
|
||||||
|
CPPFLAGS="$CPPFLAGS $LUA_INCLUDE"
|
||||||
|
AC_RUN_IFELSE(
|
||||||
|
[ AC_LANG_SOURCE([[
|
||||||
|
#include <lua.h>
|
||||||
|
#include <stdlib.h>
|
||||||
|
#include <stdio.h>
|
||||||
|
int main(int argc, char ** argv)
|
||||||
|
{
|
||||||
|
if(argc > 1) printf("%s", LUA_VERSION);
|
||||||
|
exit(EXIT_SUCCESS);
|
||||||
|
}
|
||||||
|
]])
|
||||||
|
],
|
||||||
|
[ ax_cv_lua_header_version=`./conftest$EXEEXT p | \
|
||||||
|
$SED -n "s|^Lua \(@<:@0-9@:>@\{1,\}\.@<:@0-9@:>@\{1,\}\).\{0,\}|\1|p"`
|
||||||
|
],
|
||||||
|
[ax_cv_lua_header_version='unknown'])
|
||||||
|
CPPFLAGS=$_ax_lua_saved_cppflags
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Compare this to the previously found LUA_VERSION.
|
||||||
|
AC_MSG_CHECKING([if Lua header version matches $LUA_VERSION])
|
||||||
|
AS_IF([test "x$ax_cv_lua_header_version" = "x$LUA_VERSION"],
|
||||||
|
[ AC_MSG_RESULT([yes])
|
||||||
|
ax_header_version_match='yes'
|
||||||
|
],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
ax_header_version_match='no'
|
||||||
|
])
|
||||||
|
],
|
||||||
|
[ AC_MSG_WARN([cross compiling so assuming header version number matches])
|
||||||
|
ax_header_version_match='yes'
|
||||||
|
])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Was LUA_INCLUDE specified?
|
||||||
|
AS_IF([test "x$ax_header_version_match" != 'xyes' &&
|
||||||
|
test "x$LUA_INCLUDE" != 'x'],
|
||||||
|
[AC_MSG_ERROR([cannot find headers for specified LUA_INCLUDE])])
|
||||||
|
|
||||||
|
dnl Test the final result and run user code.
|
||||||
|
AS_IF([test "x$ax_header_version_match" = 'xyes'], [$1],
|
||||||
|
[m4_default([$2], [AC_MSG_ERROR([cannot find Lua includes])])])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl AX_LUA_HEADERS_VERSION no longer exists, use AX_LUA_HEADERS.
|
||||||
|
AC_DEFUN([AX_LUA_HEADERS_VERSION],
|
||||||
|
[
|
||||||
|
AC_MSG_WARN([[$0 is deprecated, please use AX_LUA_HEADERS instead]])
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl AX_LUA_LIBS([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([AX_LUA_LIBS],
|
||||||
|
[
|
||||||
|
dnl TODO Should this macro also check various -L flags?
|
||||||
|
|
||||||
|
dnl Check for LUA_VERSION.
|
||||||
|
AC_MSG_CHECKING([if LUA_VERSION is defined])
|
||||||
|
AS_IF([test "x$LUA_VERSION" != 'x'],
|
||||||
|
[AC_MSG_RESULT([yes])],
|
||||||
|
[ AC_MSG_RESULT([no])
|
||||||
|
AC_MSG_ERROR([cannot check Lua libs without knowing LUA_VERSION])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Make LUA_LIB a precious variable.
|
||||||
|
AC_ARG_VAR([LUA_LIB], [The Lua library, e.g. -llua5.1])
|
||||||
|
|
||||||
|
AS_IF([test "x$LUA_LIB" != 'x'],
|
||||||
|
[ dnl Check that LUA_LIBS works.
|
||||||
|
_ax_lua_saved_libs=$LIBS
|
||||||
|
LIBS="$LIBS $LUA_LIB"
|
||||||
|
AC_SEARCH_LIBS([lua_load], [],
|
||||||
|
[_ax_found_lua_libs='yes'],
|
||||||
|
[_ax_found_lua_libs='no'])
|
||||||
|
LIBS=$_ax_lua_saved_libs
|
||||||
|
|
||||||
|
dnl Check the result.
|
||||||
|
AS_IF([test "x$_ax_found_lua_libs" != 'xyes'],
|
||||||
|
[AC_MSG_ERROR([cannot find libs for specified LUA_LIB])])
|
||||||
|
],
|
||||||
|
[ dnl First search for extra libs.
|
||||||
|
_ax_lua_extra_libs=''
|
||||||
|
|
||||||
|
_ax_lua_saved_libs=$LIBS
|
||||||
|
LIBS="$LIBS $LUA_LIB"
|
||||||
|
AC_SEARCH_LIBS([exp], [m])
|
||||||
|
AC_SEARCH_LIBS([dlopen], [dl])
|
||||||
|
LIBS=$_ax_lua_saved_libs
|
||||||
|
|
||||||
|
AS_IF([test "x$ac_cv_search_exp" != 'xno' &&
|
||||||
|
test "x$ac_cv_search_exp" != 'xnone required'],
|
||||||
|
[_ax_lua_extra_libs="$_ax_lua_extra_libs $ac_cv_search_exp"])
|
||||||
|
|
||||||
|
AS_IF([test "x$ac_cv_search_dlopen" != 'xno' &&
|
||||||
|
test "x$ac_cv_search_dlopen" != 'xnone required'],
|
||||||
|
[_ax_lua_extra_libs="$_ax_lua_extra_libs $ac_cv_search_dlopen"])
|
||||||
|
|
||||||
|
dnl Try to find the Lua libs.
|
||||||
|
_ax_lua_saved_libs=$LIBS
|
||||||
|
LIBS="$LIBS $LUA_LIB"
|
||||||
|
AC_SEARCH_LIBS([lua_load],
|
||||||
|
[ lua$LUA_VERSION \
|
||||||
|
lua$LUA_SHORT_VERSION \
|
||||||
|
lua-$LUA_VERSION \
|
||||||
|
lua-$LUA_SHORT_VERSION \
|
||||||
|
lua \
|
||||||
|
],
|
||||||
|
[_ax_found_lua_libs='yes'],
|
||||||
|
[_ax_found_lua_libs='no'],
|
||||||
|
[$_ax_lua_extra_libs])
|
||||||
|
LIBS=$_ax_lua_saved_libs
|
||||||
|
|
||||||
|
AS_IF([test "x$ac_cv_search_lua_load" != 'xno' &&
|
||||||
|
test "x$ac_cv_search_lua_load" != 'xnone required'],
|
||||||
|
[LUA_LIB="$ac_cv_search_lua_load $_ax_lua_extra_libs"])
|
||||||
|
])
|
||||||
|
|
||||||
|
dnl Test the result and run user code.
|
||||||
|
AS_IF([test "x$_ax_found_lua_libs" = 'xyes'], [$1],
|
||||||
|
[m4_default([$2], [AC_MSG_ERROR([cannot find Lua libs])])])
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
dnl =========================================================================
|
||||||
|
dnl AX_LUA_READLINE([ACTION-IF-FOUND], [ACTION-IF-NOT-FOUND])
|
||||||
|
dnl =========================================================================
|
||||||
|
AC_DEFUN([AX_LUA_READLINE],
|
||||||
|
[
|
||||||
|
AX_LIB_READLINE
|
||||||
|
AS_IF([test "x$ac_cv_header_readline_readline_h" != 'x' &&
|
||||||
|
test "x$ac_cv_header_readline_history_h" != 'x'],
|
||||||
|
[ LUA_LIBS_CFLAGS="-DLUA_USE_READLINE $LUA_LIBS_CFLAGS"
|
||||||
|
$1
|
||||||
|
],
|
||||||
|
[$2])
|
||||||
|
])
|
61
vendor/github.com/apache/thrift/aclocal/ax_prog_dotnetcore_version.m4
generated
vendored
Normal file
61
vendor/github.com/apache/thrift/aclocal/ax_prog_dotnetcore_version.m4
generated
vendored
Normal file
|
@ -0,0 +1,61 @@
|
||||||
|
# ===============================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_prog_dotnetcore_version.html
|
||||||
|
# ===============================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_PROG_DOTNETCORE_VERSION([VERSION],[ACTION-IF-TRUE],[ACTION-IF-FALSE])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Makes sure that .NET Core supports the version indicated. If true the
|
||||||
|
# shell commands in ACTION-IF-TRUE are executed. If not the shell commands
|
||||||
|
# in ACTION-IF-FALSE are run. The $dotnetcore_version variable will be
|
||||||
|
# filled with the detected version.
|
||||||
|
#
|
||||||
|
# This macro uses the $DOTNETCORE variable to perform the check. If
|
||||||
|
# $DOTNETCORE is not set prior to calling this macro, the macro will fail.
|
||||||
|
#
|
||||||
|
# Example:
|
||||||
|
#
|
||||||
|
# AC_PATH_PROG([DOTNETCORE],[dotnet])
|
||||||
|
# AC_PROG_DOTNETCORE_VERSION([1.0.2],[ ... ],[ ... ])
|
||||||
|
#
|
||||||
|
# Searches for .NET Core, then checks if at least version 1.0.2 is
|
||||||
|
# present.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2016 Jens Geyer <jensg@apache.org>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 2
|
||||||
|
|
||||||
|
AC_DEFUN([AX_PROG_DOTNETCORE_VERSION],[
|
||||||
|
AC_REQUIRE([AC_PROG_SED])
|
||||||
|
|
||||||
|
AS_IF([test -n "$DOTNETCORE"],[
|
||||||
|
ax_dotnetcore_version="$1"
|
||||||
|
|
||||||
|
AC_MSG_CHECKING([for .NET Core version])
|
||||||
|
dotnetcore_version=`$DOTNETCORE --version 2>&1 | $SED -e 's/\(@<:@0-9@:>@*\.@<:@0-9@:>@*\.@<:@0-9@:>@*\)\(.*\)/\1/'`
|
||||||
|
AC_MSG_RESULT($dotnetcore_version)
|
||||||
|
|
||||||
|
AC_SUBST([DOTNETCORE_VERSION],[$dotnetcore_version])
|
||||||
|
|
||||||
|
AX_COMPARE_VERSION([$ax_dotnetcore_version],[le],[$dotnetcore_version],[
|
||||||
|
:
|
||||||
|
$2
|
||||||
|
],[
|
||||||
|
:
|
||||||
|
$3
|
||||||
|
])
|
||||||
|
],[
|
||||||
|
AC_MSG_WARN([could not find .NET Core])
|
||||||
|
$3
|
||||||
|
])
|
||||||
|
])
|
60
vendor/github.com/apache/thrift/aclocal/ax_prog_haxe_version.m4
generated
vendored
Normal file
60
vendor/github.com/apache/thrift/aclocal/ax_prog_haxe_version.m4
generated
vendored
Normal file
|
@ -0,0 +1,60 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_prog_haxe_version.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_PROG_HAXE_VERSION([VERSION],[ACTION-IF-TRUE],[ACTION-IF-FALSE])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Makes sure that haxe supports the version indicated. If true the shell
|
||||||
|
# commands in ACTION-IF-TRUE are executed. If not the shell commands in
|
||||||
|
# ACTION-IF-FALSE are run. The $HAXE_VERSION variable will be filled with
|
||||||
|
# the detected version.
|
||||||
|
#
|
||||||
|
# This macro uses the $HAXE variable to perform the check. If $HAXE is not
|
||||||
|
# set prior to calling this macro, the macro will fail.
|
||||||
|
#
|
||||||
|
# Example:
|
||||||
|
#
|
||||||
|
# AC_PATH_PROG([HAXE],[haxe])
|
||||||
|
# AC_PROG_HAXE_VERSION([3.1.3],[ ... ],[ ... ])
|
||||||
|
#
|
||||||
|
# Searches for Haxe, then checks if at least version 3.1.3 is present.
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2015 Jens Geyer <jensg@apache.org>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 2
|
||||||
|
|
||||||
|
AC_DEFUN([AX_PROG_HAXE_VERSION],[
|
||||||
|
AC_REQUIRE([AC_PROG_SED])
|
||||||
|
|
||||||
|
AS_IF([test -n "$HAXE"],[
|
||||||
|
ax_haxe_version="$1"
|
||||||
|
|
||||||
|
AC_MSG_CHECKING([for haxe version])
|
||||||
|
haxe_version=`$HAXE -version 2>&1 | $SED -e 's/^.* \( @<:@0-9@:>@*\.@<:@0-9@:>@*\.@<:@0-9@:>@*\) .*/\1/'`
|
||||||
|
AC_MSG_RESULT($haxe_version)
|
||||||
|
|
||||||
|
AC_SUBST([HAXE_VERSION],[$haxe_version])
|
||||||
|
|
||||||
|
AX_COMPARE_VERSION([$ax_haxe_version],[le],[$haxe_version],[
|
||||||
|
:
|
||||||
|
$2
|
||||||
|
],[
|
||||||
|
:
|
||||||
|
$3
|
||||||
|
])
|
||||||
|
],[
|
||||||
|
AC_MSG_WARN([could not find Haxe])
|
||||||
|
$3
|
||||||
|
])
|
||||||
|
])
|
77
vendor/github.com/apache/thrift/aclocal/ax_prog_perl_modules.m4
generated
vendored
Normal file
77
vendor/github.com/apache/thrift/aclocal/ax_prog_perl_modules.m4
generated
vendored
Normal file
|
@ -0,0 +1,77 @@
|
||||||
|
# ===========================================================================
|
||||||
|
# https://www.gnu.org/software/autoconf-archive/ax_prog_perl_modules.html
|
||||||
|
# ===========================================================================
|
||||||
|
#
|
||||||
|
# SYNOPSIS
|
||||||
|
#
|
||||||
|
# AX_PROG_PERL_MODULES([MODULES], [ACTION-IF-TRUE], [ACTION-IF-FALSE])
|
||||||
|
#
|
||||||
|
# DESCRIPTION
|
||||||
|
#
|
||||||
|
# Checks to see if the given perl modules are available. If true the shell
|
||||||
|
# commands in ACTION-IF-TRUE are executed. If not the shell commands in
|
||||||
|
# ACTION-IF-FALSE are run. Note if $PERL is not set (for example by
|
||||||
|
# calling AC_CHECK_PROG, or AC_PATH_PROG), AC_CHECK_PROG(PERL, perl, perl)
|
||||||
|
# will be run.
|
||||||
|
#
|
||||||
|
# MODULES is a space separated list of module names. To check for a
|
||||||
|
# minimum version of a module, append the version number to the module
|
||||||
|
# name, separated by an equals sign.
|
||||||
|
#
|
||||||
|
# Example:
|
||||||
|
#
|
||||||
|
# AX_PROG_PERL_MODULES( Text::Wrap Net::LDAP=1.0.3, ,
|
||||||
|
# AC_MSG_WARN(Need some Perl modules)
|
||||||
|
#
|
||||||
|
# LICENSE
|
||||||
|
#
|
||||||
|
# Copyright (c) 2009 Dean Povey <povey@wedgetail.com>
|
||||||
|
#
|
||||||
|
# Copying and distribution of this file, with or without modification, are
|
||||||
|
# permitted in any medium without royalty provided the copyright notice
|
||||||
|
# and this notice are preserved. This file is offered as-is, without any
|
||||||
|
# warranty.
|
||||||
|
|
||||||
|
#serial 8
|
||||||
|
|
||||||
|
AU_ALIAS([AC_PROG_PERL_MODULES], [AX_PROG_PERL_MODULES])
|
||||||
|
AC_DEFUN([AX_PROG_PERL_MODULES],[dnl
|
||||||
|
|
||||||
|
m4_define([ax_perl_modules])
|
||||||
|
m4_foreach([ax_perl_module], m4_split(m4_normalize([$1])),
|
||||||
|
[
|
||||||
|
m4_append([ax_perl_modules],
|
||||||
|
[']m4_bpatsubst(ax_perl_module,=,[ ])[' ])
|
||||||
|
])
|
||||||
|
|
||||||
|
# Make sure we have perl
|
||||||
|
if test -z "$PERL"; then
|
||||||
|
AC_CHECK_PROG(PERL,perl,perl)
|
||||||
|
fi
|
||||||
|
|
||||||
|
if test "x$PERL" != x; then
|
||||||
|
ax_perl_modules_failed=0
|
||||||
|
for ax_perl_module in ax_perl_modules; do
|
||||||
|
AC_MSG_CHECKING(for perl module $ax_perl_module)
|
||||||
|
|
||||||
|
# Would be nice to log result here, but can't rely on autoconf internals
|
||||||
|
$PERL -e "use $ax_perl_module; exit" > /dev/null 2>&1
|
||||||
|
if test $? -ne 0; then
|
||||||
|
AC_MSG_RESULT(no);
|
||||||
|
ax_perl_modules_failed=1
|
||||||
|
else
|
||||||
|
AC_MSG_RESULT(ok);
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
# Run optional shell commands
|
||||||
|
if test "$ax_perl_modules_failed" = 0; then
|
||||||
|
:
|
||||||
|
$2
|
||||||
|
else
|
||||||
|
:
|
||||||
|
$3
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
AC_MSG_WARN(could not find perl)
|
||||||
|
fi])dnl
|
127
vendor/github.com/apache/thrift/aclocal/ax_signed_right_shift.m4
generated
vendored
Normal file
127
vendor/github.com/apache/thrift/aclocal/ax_signed_right_shift.m4
generated
vendored
Normal file
|
@ -0,0 +1,127 @@
|
||||||
|
dnl @synopsis AX_SIGNED_RIGHT_SHIFT
|
||||||
|
dnl
|
||||||
|
dnl Tests the behavior of a right shift on a negative signed int.
|
||||||
|
dnl
|
||||||
|
dnl This macro calls:
|
||||||
|
dnl AC_DEFINE(SIGNED_RIGHT_SHIFT_IS)
|
||||||
|
dnl AC_DEFINE(ARITHMETIC_RIGHT_SHIFT)
|
||||||
|
dnl AC_DEFINE(LOGICAL_RIGHT_SHIFT)
|
||||||
|
dnl AC_DEFINE(UNKNOWN_RIGHT_SHIFT)
|
||||||
|
dnl
|
||||||
|
dnl SIGNED_RIGHT_SHIFT_IS will be equal to one of the other macros.
|
||||||
|
dnl It also leaves the shell variables "ax_signed_right_shift"
|
||||||
|
dnl set to "arithmetic", "logical", or "unknown".
|
||||||
|
dnl
|
||||||
|
dnl NOTE: This macro does not work for cross-compiling.
|
||||||
|
dnl
|
||||||
|
dnl @category C
|
||||||
|
dnl @version 2009-03-25
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
AC_DEFUN([AX_SIGNED_RIGHT_SHIFT],
|
||||||
|
[
|
||||||
|
|
||||||
|
AC_MSG_CHECKING(the behavior of a signed right shift)
|
||||||
|
|
||||||
|
success_arithmetic=no
|
||||||
|
AC_RUN_IFELSE([AC_LANG_PROGRAM([[]], [[
|
||||||
|
return
|
||||||
|
/* 0xffffffff */
|
||||||
|
-1 >> 1 != -1 ||
|
||||||
|
-1 >> 2 != -1 ||
|
||||||
|
-1 >> 3 != -1 ||
|
||||||
|
-1 >> 4 != -1 ||
|
||||||
|
-1 >> 8 != -1 ||
|
||||||
|
-1 >> 16 != -1 ||
|
||||||
|
-1 >> 24 != -1 ||
|
||||||
|
-1 >> 31 != -1 ||
|
||||||
|
/* 0x80000000 */
|
||||||
|
(-2147483647 - 1) >> 1 != -1073741824 ||
|
||||||
|
(-2147483647 - 1) >> 2 != -536870912 ||
|
||||||
|
(-2147483647 - 1) >> 3 != -268435456 ||
|
||||||
|
(-2147483647 - 1) >> 4 != -134217728 ||
|
||||||
|
(-2147483647 - 1) >> 8 != -8388608 ||
|
||||||
|
(-2147483647 - 1) >> 16 != -32768 ||
|
||||||
|
(-2147483647 - 1) >> 24 != -128 ||
|
||||||
|
(-2147483647 - 1) >> 31 != -1 ||
|
||||||
|
/* 0x90800000 */
|
||||||
|
-1870659584 >> 1 != -935329792 ||
|
||||||
|
-1870659584 >> 2 != -467664896 ||
|
||||||
|
-1870659584 >> 3 != -233832448 ||
|
||||||
|
-1870659584 >> 4 != -116916224 ||
|
||||||
|
-1870659584 >> 8 != -7307264 ||
|
||||||
|
-1870659584 >> 16 != -28544 ||
|
||||||
|
-1870659584 >> 24 != -112 ||
|
||||||
|
-1870659584 >> 31 != -1 ||
|
||||||
|
0;
|
||||||
|
]])], [
|
||||||
|
success_arithmetic=yes
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
success_logical=no
|
||||||
|
AC_RUN_IFELSE([AC_LANG_PROGRAM([[]], [[
|
||||||
|
return
|
||||||
|
/* 0xffffffff */
|
||||||
|
-1 >> 1 != (signed)((unsigned)-1 >> 1) ||
|
||||||
|
-1 >> 2 != (signed)((unsigned)-1 >> 2) ||
|
||||||
|
-1 >> 3 != (signed)((unsigned)-1 >> 3) ||
|
||||||
|
-1 >> 4 != (signed)((unsigned)-1 >> 4) ||
|
||||||
|
-1 >> 8 != (signed)((unsigned)-1 >> 8) ||
|
||||||
|
-1 >> 16 != (signed)((unsigned)-1 >> 16) ||
|
||||||
|
-1 >> 24 != (signed)((unsigned)-1 >> 24) ||
|
||||||
|
-1 >> 31 != (signed)((unsigned)-1 >> 31) ||
|
||||||
|
/* 0x80000000 */
|
||||||
|
(-2147483647 - 1) >> 1 != (signed)((unsigned)(-2147483647 - 1) >> 1) ||
|
||||||
|
(-2147483647 - 1) >> 2 != (signed)((unsigned)(-2147483647 - 1) >> 2) ||
|
||||||
|
(-2147483647 - 1) >> 3 != (signed)((unsigned)(-2147483647 - 1) >> 3) ||
|
||||||
|
(-2147483647 - 1) >> 4 != (signed)((unsigned)(-2147483647 - 1) >> 4) ||
|
||||||
|
(-2147483647 - 1) >> 8 != (signed)((unsigned)(-2147483647 - 1) >> 8) ||
|
||||||
|
(-2147483647 - 1) >> 16 != (signed)((unsigned)(-2147483647 - 1) >> 16) ||
|
||||||
|
(-2147483647 - 1) >> 24 != (signed)((unsigned)(-2147483647 - 1) >> 24) ||
|
||||||
|
(-2147483647 - 1) >> 31 != (signed)((unsigned)(-2147483647 - 1) >> 31) ||
|
||||||
|
/* 0x90800000 */
|
||||||
|
-1870659584 >> 1 != (signed)((unsigned)-1870659584 >> 1) ||
|
||||||
|
-1870659584 >> 2 != (signed)((unsigned)-1870659584 >> 2) ||
|
||||||
|
-1870659584 >> 3 != (signed)((unsigned)-1870659584 >> 3) ||
|
||||||
|
-1870659584 >> 4 != (signed)((unsigned)-1870659584 >> 4) ||
|
||||||
|
-1870659584 >> 8 != (signed)((unsigned)-1870659584 >> 8) ||
|
||||||
|
-1870659584 >> 16 != (signed)((unsigned)-1870659584 >> 16) ||
|
||||||
|
-1870659584 >> 24 != (signed)((unsigned)-1870659584 >> 24) ||
|
||||||
|
-1870659584 >> 31 != (signed)((unsigned)-1870659584 >> 31) ||
|
||||||
|
0;
|
||||||
|
]])], [
|
||||||
|
success_logical=yes
|
||||||
|
])
|
||||||
|
|
||||||
|
|
||||||
|
AC_DEFINE([ARITHMETIC_RIGHT_SHIFT], 1, [Possible value for SIGNED_RIGHT_SHIFT_IS])
|
||||||
|
AC_DEFINE([LOGICAL_RIGHT_SHIFT], 2, [Possible value for SIGNED_RIGHT_SHIFT_IS])
|
||||||
|
AC_DEFINE([UNKNOWN_RIGHT_SHIFT], 3, [Possible value for SIGNED_RIGHT_SHIFT_IS])
|
||||||
|
|
||||||
|
if test "$success_arithmetic" = "yes" && test "$success_logical" = "yes" ; then
|
||||||
|
AC_MSG_ERROR("Right shift appears to be both arithmetic and logical!")
|
||||||
|
elif test "$success_arithmetic" = "yes" ; then
|
||||||
|
ax_signed_right_shift=arithmetic
|
||||||
|
AC_DEFINE([SIGNED_RIGHT_SHIFT_IS], 1,
|
||||||
|
[Indicates the effect of the right shift operator
|
||||||
|
on negative signed integers])
|
||||||
|
elif test "$success_logical" = "yes" ; then
|
||||||
|
ax_signed_right_shift=logical
|
||||||
|
AC_DEFINE([SIGNED_RIGHT_SHIFT_IS], 2,
|
||||||
|
[Indicates the effect of the right shift operator
|
||||||
|
on negative signed integers])
|
||||||
|
else
|
||||||
|
ax_signed_right_shift=unknown
|
||||||
|
AC_DEFINE([SIGNED_RIGHT_SHIFT_IS], 3,
|
||||||
|
[Indicates the effect of the right shift operator
|
||||||
|
on negative signed integers])
|
||||||
|
fi
|
||||||
|
|
||||||
|
AC_MSG_RESULT($ax_signed_right_shift)
|
||||||
|
])
|
28
vendor/github.com/apache/thrift/aclocal/ax_thrift_internal.m4
generated
vendored
Normal file
28
vendor/github.com/apache/thrift/aclocal/ax_thrift_internal.m4
generated
vendored
Normal file
|
@ -0,0 +1,28 @@
|
||||||
|
dnl @synopsis AX_THRIFT_GEN(SHORT_LANGUAGE, LONG_LANGUAGE, DEFAULT)
|
||||||
|
dnl @synopsis AX_THRIFT_LIB(SHORT_LANGUAGE, LONG_LANGUAGE, DEFAULT)
|
||||||
|
dnl
|
||||||
|
dnl Allow a particular language generator to be disabled.
|
||||||
|
dnl Allow a particular language library to be disabled.
|
||||||
|
dnl
|
||||||
|
dnl These macros have poor error handling and are poorly documented.
|
||||||
|
dnl They are intended only for internal use by the Thrift compiler.
|
||||||
|
dnl
|
||||||
|
dnl @version 2008-02-20
|
||||||
|
dnl @license AllPermissive
|
||||||
|
dnl
|
||||||
|
dnl Copyright (C) 2009 David Reiss
|
||||||
|
dnl Copying and distribution of this file, with or without modification,
|
||||||
|
dnl are permitted in any medium without royalty provided the copyright
|
||||||
|
dnl notice and this notice are preserved.
|
||||||
|
|
||||||
|
AC_DEFUN([AX_THRIFT_LIB],
|
||||||
|
[
|
||||||
|
AC_ARG_WITH($1,
|
||||||
|
AC_HELP_STRING([--with-$1], [build the $2 library @<:@default=$3@:>@]),
|
||||||
|
[with_$1="$withval"],
|
||||||
|
[with_$1=$3]
|
||||||
|
)
|
||||||
|
have_$1=no
|
||||||
|
dnl What we do here is going to vary from library to library,
|
||||||
|
dnl so we can't really generalize (yet!).
|
||||||
|
])
|
36
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-build.bat
generated
vendored
Normal file
36
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-build.bat
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_build.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
|
||||||
|
SET CMAKEARGS=^
|
||||||
|
-G'%GENERATOR%' ^
|
||||||
|
-DCMAKE_BUILD_TYPE=%CONFIGURATION% ^
|
||||||
|
-DCMAKE_INSTALL_PREFIX=%INSTDIR_MSYS% ^
|
||||||
|
-DCMAKE_MAKE_PROGRAM=/mingw64/bin/mingw32-make ^
|
||||||
|
-DCMAKE_C_COMPILER=x86_64-w64-mingw32-gcc.exe ^
|
||||||
|
-DCMAKE_CXX_COMPILER=x86_64-w64-mingw32-g++.exe ^
|
||||||
|
-DWITH_LIBEVENT=OFF ^
|
||||||
|
-DWITH_PYTHON=OFF ^
|
||||||
|
-DWITH_SHARED_LIB=OFF ^
|
||||||
|
-DWITH_STATIC_LIB=ON
|
||||||
|
|
||||||
|
@ECHO ON
|
||||||
|
%BASH% -lc "mkdir -p %BUILDDIR_MSYS% && cd %BUILDDIR_MSYS% && cmake.exe %SRCDIR_MSYS% %CMAKEARGS% && cmake --build . --config %CONFIGURATION% --target install" || EXIT /B
|
||||||
|
@ECHO OFF
|
21
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-install.bat
generated
vendored
Normal file
21
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-install.bat
generated
vendored
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Appveyor install script for MinGW
|
||||||
|
:: Installs (or builds) third party packages we need
|
||||||
|
::
|
||||||
|
|
||||||
|
:: Same as the MSYS installation requirements
|
||||||
|
CALL build\appveyor\MSYS-appveyor-install.bat
|
16
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-test.bat
generated
vendored
Normal file
16
vendor/github.com/apache/thrift/build/appveyor/MING-appveyor-test.bat
generated
vendored
Normal file
|
@ -0,0 +1,16 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
:: Same as MSYS2
|
||||||
|
CALL build\appveyor\MSYS-appveyor-test.bat
|
47
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-build.bat
generated
vendored
Normal file
47
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-build.bat
generated
vendored
Normal file
|
@ -0,0 +1,47 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_build.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
MKDIR "%BUILDDIR%" || EXIT /B
|
||||||
|
CD "%BUILDDIR%" || EXIT /B
|
||||||
|
|
||||||
|
@ECHO ON
|
||||||
|
cmake "%SRCDIR%" ^
|
||||||
|
-G"%GENERATOR%" ^
|
||||||
|
-DBISON_EXECUTABLE=C:\ProgramData\chocolatey\lib\winflexbison3\tools\win_bison.exe ^
|
||||||
|
-DBOOST_ROOT="%BOOST_ROOT%" ^
|
||||||
|
-DBOOST_LIBRARYDIR="%BOOST_LIBRARYDIR%" ^
|
||||||
|
-DCMAKE_BUILD_TYPE="%CONFIGURATION%" ^
|
||||||
|
-DCMAKE_INSTALL_PREFIX="%INSTDIR%" ^
|
||||||
|
-DFLEX_EXECUTABLE=C:\ProgramData\chocolatey\lib\winflexbison3\tools\win_flex.exe ^
|
||||||
|
-DINTTYPES_ROOT="%WIN3P%\msinttypes" ^
|
||||||
|
-DLIBEVENT_ROOT="%WIN3P%\libevent-%LIBEVENT_VERSION%-stable" ^
|
||||||
|
-DOPENSSL_ROOT_DIR="%OPENSSL_ROOT%" ^
|
||||||
|
-DOPENSSL_USE_STATIC_LIBS=OFF ^
|
||||||
|
-DZLIB_LIBRARY="%WIN3P%\zlib-inst\lib\zlib%ZLIB_LIB_SUFFIX%.lib" ^
|
||||||
|
-DZLIB_ROOT="%WIN3P%\zlib-inst" ^
|
||||||
|
-DWITH_PYTHON=%WITH_PYTHON% ^
|
||||||
|
-DWITH_%THREADMODEL%THREADS=ON ^
|
||||||
|
-DWITH_SHARED_LIB=OFF ^
|
||||||
|
-DWITH_STATIC_LIB=ON || EXIT /B
|
||||||
|
@ECHO OFF
|
||||||
|
|
||||||
|
cmake --build . ^
|
||||||
|
--config "%CONFIGURATION%" ^
|
||||||
|
--target INSTALL || EXIT /B
|
72
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-install.bat
generated
vendored
Normal file
72
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-install.bat
generated
vendored
Normal file
|
@ -0,0 +1,72 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Appveyor install script for MSVC
|
||||||
|
:: Installs (or builds) third party packages we need
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_install.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
CALL cl_showenv.bat || EXIT /B
|
||||||
|
MKDIR "%WIN3P%" || EXIT /B
|
||||||
|
|
||||||
|
:: Install ant - this also installs the latest JDK as a dependency
|
||||||
|
:: The installation of JDK requires us to pick up PATH and JAVE_HOME from the registry
|
||||||
|
cinst -c "%BUILDCACHE%" -y ant || EXIT /B
|
||||||
|
|
||||||
|
:: Install bison and flex
|
||||||
|
cinst -c "%BUILDCACHE%" -y winflexbison3 || EXIT /B
|
||||||
|
|
||||||
|
:: zlib
|
||||||
|
CD "%APPVEYOR_SCRIPTS%" || EXIT /B
|
||||||
|
call build-zlib.bat || EXIT /B
|
||||||
|
|
||||||
|
:: libevent
|
||||||
|
CD "%APPVEYOR_SCRIPTS%" || EXIT /B
|
||||||
|
call build-libevent.bat || EXIT /B
|
||||||
|
|
||||||
|
:: python packages
|
||||||
|
pip install backports.ssl_match_hostname ^
|
||||||
|
ipaddress ^
|
||||||
|
tornado ^
|
||||||
|
twisted || EXIT /B
|
||||||
|
|
||||||
|
:: msinttypes - for MSVC2010 only
|
||||||
|
SET MSINTTYPESURL=https://storage.googleapis.com/google-code-archive-downloads/v2/code.google.com/msinttypes/msinttypes-r26.zip
|
||||||
|
IF "%COMPILER%" == "vc100" (
|
||||||
|
MKDIR "%WIN3P%\msinttypes" || EXIT /B
|
||||||
|
CD "%WIN3P%\msinttypes" || EXIT /B
|
||||||
|
appveyor DownloadFile "%MSINTTYPESURL%" || EXIT /B
|
||||||
|
7z x "msinttypes-r26.zip" || EXIT /B
|
||||||
|
)
|
||||||
|
|
||||||
|
:: appveyor build slaves do not have MSVC2010 Boost installed
|
||||||
|
IF "%COMPILER%" == "vc100" (
|
||||||
|
SET BITS=64
|
||||||
|
IF "%PLATFORM%" == "x86" (
|
||||||
|
SET BITS=32
|
||||||
|
)
|
||||||
|
SET BOOSTEXEURL=https://downloads.sourceforge.net/project/boost/boost-binaries/%BOOST_VERSION%/boost_%BOOST_VERSION:.=_%-msvc-10.0-!BITS!.exe
|
||||||
|
SET BOOSTEXE=C:\projects\thrift\buildcache\boost_%BOOST_VERSION:.=_%-msvc-10.0-!BITS!.exe
|
||||||
|
appveyor DownloadFile "!BOOSTEXEURL!" -FileName "!BOOSTEXE!" || EXIT /B
|
||||||
|
"!BOOSTEXE!" /dir=C:\Libraries\boost_%BOOST_VERSION:.=_% /silent || EXIT /B
|
||||||
|
)
|
||||||
|
|
||||||
|
:: Haskell (GHC) and cabal
|
||||||
|
cinst -c "%BUILDCACHE%" -y ghc || EXIT /B
|
25
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-test.bat
generated
vendored
Normal file
25
vendor/github.com/apache/thrift/build/appveyor/MSVC-appveyor-test.bat
generated
vendored
Normal file
|
@ -0,0 +1,25 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_test.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
CD "%BUILDDIR%" || EXIT /B
|
||||||
|
|
||||||
|
:: Add directories to the path to find DLLs of third party libraries so tests run
|
||||||
|
SET PATH=%BOOST_LIBRARYDIR%;%OPENSSL_ROOT%\bin;%WIN3P%\zlib-inst\bin;%PATH%
|
||||||
|
|
||||||
|
ctest -C %CONFIGURATION% --timeout 300 -VV -E "(%DISABLED_TESTS%)" || EXIT /B
|
47
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-build.bat
generated
vendored
Normal file
47
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-build.bat
generated
vendored
Normal file
|
@ -0,0 +1,47 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_build.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
|
||||||
|
SET BASH=C:\msys64\usr\bin\bash
|
||||||
|
SET CMAKE=/c/msys64/mingw64/bin/cmake.exe
|
||||||
|
|
||||||
|
@ECHO ON
|
||||||
|
SET CMAKEARGS=-G\"%GENERATOR%\" ^
|
||||||
|
-DBoost_DEBUG=ON ^
|
||||||
|
-DBoost_NAMESPACE=libboost ^
|
||||||
|
-DBOOST_INCLUDEDIR=%BOOST_INCLUDEDIR% ^
|
||||||
|
-DBOOST_LIBRARYDIR=%BOOST_LIBRARYDIR% ^
|
||||||
|
-DCMAKE_BUILD_TYPE=%CONFIGURATION% ^
|
||||||
|
-DCMAKE_C_COMPILER=gcc.exe ^
|
||||||
|
-DCMAKE_CXX_COMPILER=g++.exe ^
|
||||||
|
-DCMAKE_MAKE_PROGRAM=make.exe ^
|
||||||
|
-DCMAKE_INSTALL_PREFIX=%INSTDIR_MSYS% ^
|
||||||
|
-DOPENSSL_LIBRARIES=%OPENSSL_LIBRARIES% ^
|
||||||
|
-DOPENSSL_ROOT_DIR=%OPENSSL_ROOT% ^
|
||||||
|
-DOPENSSL_USE_STATIC_LIBS=ON ^
|
||||||
|
-DWITH_BOOST_STATIC=ON ^
|
||||||
|
-DWITH_JAVA=OFF ^
|
||||||
|
-DWITH_LIBEVENT=OFF ^
|
||||||
|
-DWITH_PYTHON=%WITH_PYTHON% ^
|
||||||
|
-DWITH_SHARED_LIB=OFF ^
|
||||||
|
-DWITH_STATIC_LIB=ON
|
||||||
|
|
||||||
|
%BASH% -lc "mkdir %BUILDDIR_MSYS% && cd %BUILDDIR_MSYS% && %CMAKE% %SRCDIR_MSYS% %CMAKEARGS% && %CMAKE% --build . --config %CONFIGURATION% --target install" || EXIT /B
|
||||||
|
@ECHO OFF
|
41
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-install.bat
generated
vendored
Normal file
41
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-install.bat
generated
vendored
Normal file
|
@ -0,0 +1,41 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Appveyor install script for MSYS
|
||||||
|
:: Installs (or builds) third party packages we need
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_install.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
CALL cl_showenv.bat || EXIT /B
|
||||||
|
|
||||||
|
SET PACKAGES=^
|
||||||
|
--needed -S bison flex ^
|
||||||
|
make ^
|
||||||
|
mingw-w64-x86_64-boost ^
|
||||||
|
mingw-w64-x86_64-cmake ^
|
||||||
|
mingw-w64-x86_64-openssl ^
|
||||||
|
mingw-w64-x86_64-toolchain ^
|
||||||
|
mingw-w64-x86_64-zlib
|
||||||
|
|
||||||
|
:: omitting libevent-devel for now it is version 2.1.4 and doesn't play nice with MinGW
|
||||||
|
|
||||||
|
%BASH% -lc "pacman --noconfirm -Syu" || EXIT /B
|
||||||
|
%BASH% -lc "pacman --noconfirm -Su" || EXIT /B
|
||||||
|
%BASH% -lc "pacman --noconfirm %PACKAGES%" || EXIT /B
|
26
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-test.bat
generated
vendored
Normal file
26
vendor/github.com/apache/thrift/build/appveyor/MSYS-appveyor-test.bat
generated
vendored
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
CD build\appveyor || EXIT /B
|
||||||
|
CALL cl_banner_test.bat || EXIT /B
|
||||||
|
CALL cl_setenv.bat || EXIT /B
|
||||||
|
CD "%BUILDDIR%" || EXIT /B
|
||||||
|
|
||||||
|
:: randomly fails on mingw; see Jira THRIFT-4106
|
||||||
|
SET DISABLED_TESTS=concurrency_test
|
||||||
|
|
||||||
|
%BASH% -lc "cd %BUILDDIR_MSYS% && ctest.exe -C %CONFIGURATION% --timeout 300 -VV -E '(%DISABLED_TESTS%)'" || EXIT /B
|
34
vendor/github.com/apache/thrift/build/appveyor/README.md
generated
vendored
Normal file
34
vendor/github.com/apache/thrift/build/appveyor/README.md
generated
vendored
Normal file
|
@ -0,0 +1,34 @@
|
||||||
|
<!---
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
-->
|
||||||
|
|
||||||
|
# Appveyor Build
|
||||||
|
|
||||||
|
Appveyor is capable of building MSVC 2010 through 2015 as well as
|
||||||
|
having the latest MSYS2/MinGW 64-bit environment. It has many versions
|
||||||
|
of boost and python installed as well. See what appveyor has
|
||||||
|
[installed on build workers](https://www.appveyor.com/docs/installed-software/).
|
||||||
|
|
||||||
|
We run a matrix build on Appveyor and build the following combinations:
|
||||||
|
|
||||||
|
* MinGW x64 (gcc 6.3.0)
|
||||||
|
* MSVC 2010 x86, an older boost, an older python
|
||||||
|
* MSVC 2015 x86/x64, the latest boost, the latest python
|
||||||
|
* MSYS2 x64 (gcc 6.3.0) - this is a work in progress
|
||||||
|
|
||||||
|
The Appveyor script takes the first four letters from the PROFILE specified in
|
||||||
|
the environment stanza and runs these scripts in order:
|
||||||
|
|
||||||
|
????-appveyor-install.bat will install third party libraries and set up the environment
|
||||||
|
????-appveyor-build.bat will build with cmake
|
||||||
|
????-appveyor-test.bat will run ctest
|
30
vendor/github.com/apache/thrift/build/appveyor/build-libevent.bat
generated
vendored
Normal file
30
vendor/github.com/apache/thrift/build/appveyor/build-libevent.bat
generated
vendored
Normal file
|
@ -0,0 +1,30 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
SET URLFILE=libevent-%LIBEVENT_VERSION%-stable.tar.gz
|
||||||
|
SET URL=https://github.com/libevent/libevent/releases/download/release-%LIBEVENT_VERSION%-stable/%URLFILE%
|
||||||
|
|
||||||
|
CD %WIN3P% || EXIT /B
|
||||||
|
appveyor DownloadFile %URL% || EXIT /B
|
||||||
|
7z x %URLFILE% -so | 7z x -si -ttar > nul || EXIT /B
|
||||||
|
CD "libevent-%LIBEVENT_VERSION%-stable" || EXIT /B
|
||||||
|
nmake -f Makefile.nmake || EXIT /B
|
||||||
|
mkdir lib || EXIT /B
|
||||||
|
move *.lib lib\ || EXIT /B
|
||||||
|
move WIN32-Code\event2\* include\event2\ || EXIT /B
|
||||||
|
move *.h include\ || EXIT /B
|
||||||
|
|
||||||
|
ENDLOCAL
|
49
vendor/github.com/apache/thrift/build/appveyor/build-zlib.bat
generated
vendored
Normal file
49
vendor/github.com/apache/thrift/build/appveyor/build-zlib.bat
generated
vendored
Normal file
|
@ -0,0 +1,49 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
SET PACKAGE=zlib-%ZLIB_VERSION%
|
||||||
|
SET BUILDDIR=%WIN3P%\zlib-build
|
||||||
|
SET INSTDIR=%WIN3P%\zlib-inst
|
||||||
|
SET SRCDIR=%WIN3P%\%PACKAGE%
|
||||||
|
SET URLFILE=%PACKAGE%.tar.gz
|
||||||
|
|
||||||
|
:: This allows us to tolerate when the current version is archived
|
||||||
|
SET URL=http://zlib.net/%URLFILE%
|
||||||
|
SET FURL=http://zlib.net/fossils/%URLFILE%
|
||||||
|
|
||||||
|
:: Download
|
||||||
|
CD "%WIN3P%" || EXIT /B
|
||||||
|
appveyor DownloadFile "%URL%"
|
||||||
|
IF ERRORLEVEL 1 (
|
||||||
|
appveyor DownloadFile "%FURL%" || EXIT /B
|
||||||
|
)
|
||||||
|
7z x "%URLFILE%" -so | 7z x -si -ttar > nul || EXIT /B
|
||||||
|
|
||||||
|
:: Generate
|
||||||
|
MKDIR "%BUILDDIR%" || EXIT /B
|
||||||
|
CD "%BUILDDIR%" || EXIT /B
|
||||||
|
cmake "%SRCDIR%" ^
|
||||||
|
-G"NMake Makefiles" ^
|
||||||
|
-DCMAKE_INSTALL_PREFIX="%INSTDIR%" ^
|
||||||
|
-DCMAKE_BUILD_TYPE="%CONFIGURATION%" || EXIT /B
|
||||||
|
|
||||||
|
:: Build
|
||||||
|
nmake /fMakefile install || EXIT /B
|
||||||
|
IF "%CONFIGURATION%" == "Debug" (
|
||||||
|
COPY "%BUILDDIR%\zlibd.pdb" "%INSTDIR%\bin\" || EXIT /B
|
||||||
|
)
|
||||||
|
|
||||||
|
ENDLOCAL
|
24
vendor/github.com/apache/thrift/build/appveyor/cl_banner_apache_thrift.bat
generated
vendored
Normal file
24
vendor/github.com/apache/thrift/build/appveyor/cl_banner_apache_thrift.bat
generated
vendored
Normal file
|
@ -0,0 +1,24 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
:: A visual indicator in a large log helps you locate things when scanning
|
||||||
|
:: http://www.patorjk.com/software/taag/#p=display&f=Small%20Slant&t=Apache%20Thrift
|
||||||
|
|
||||||
|
ECHO/
|
||||||
|
ECHO ___ __ ________ _ _____
|
||||||
|
ECHO / _ | ___ ___ _____/ / ___ /_ __/ / ____(_) _/ /_
|
||||||
|
ECHO / __ |/ _ \/ _ `/ __/ _ \/ -_) / / / _ \/ __/ / _/ __/
|
||||||
|
ECHO /_/ |_/ .__/\_,_/\__/_//_/\__/ /_/ /_//_/_/ /_/_/ \__/
|
||||||
|
ECHO /_/
|
||||||
|
ECHO/
|
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_build.bat
generated
vendored
Normal file
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_build.bat
generated
vendored
Normal file
|
@ -0,0 +1,23 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
:: A visual indicator in a large log helps you locate things when scanning
|
||||||
|
:: http://www.patorjk.com/software/taag/#p=display&f=Small%20Slant&t=Build
|
||||||
|
|
||||||
|
ECHO/
|
||||||
|
ECHO ___ _ __ __
|
||||||
|
ECHO / _ )__ __(_) /__/ /
|
||||||
|
ECHO / _ / // / / / _ / @@@ BUILD
|
||||||
|
ECHO /____/\_,_/_/_/\_,_/
|
||||||
|
ECHO/
|
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_install.bat
generated
vendored
Normal file
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_install.bat
generated
vendored
Normal file
|
@ -0,0 +1,23 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
:: A visual indicator in a large log helps you locate things when scanning
|
||||||
|
:: http://www.patorjk.com/software/taag/#p=display&f=Small%20Slant&t=Install
|
||||||
|
|
||||||
|
ECHO/
|
||||||
|
ECHO ____ __ ____
|
||||||
|
ECHO / _/__ ___ / /____ _/ / /
|
||||||
|
ECHO _/ // _ \(_-^</ __/ _ \/ / / @@@ INSTALL
|
||||||
|
ECHO /___/_//_/___/\__/\_,_/_/_/
|
||||||
|
ECHO/
|
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_test.bat
generated
vendored
Normal file
23
vendor/github.com/apache/thrift/build/appveyor/cl_banner_test.bat
generated
vendored
Normal file
|
@ -0,0 +1,23 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
:: A visual indicator in a large log helps you locate things when scanning
|
||||||
|
:: http://www.patorjk.com/software/taag/#p=display&f=Small%20Slant&t=Test
|
||||||
|
|
||||||
|
ECHO/
|
||||||
|
ECHO ______ __
|
||||||
|
ECHO /_ __/__ ___ / /_
|
||||||
|
ECHO / / / -_^|_-^</ __/ @@@ TEST
|
||||||
|
ECHO /_/ \__/___/\__/
|
||||||
|
ECHO/
|
63
vendor/github.com/apache/thrift/build/appveyor/cl_setcompiler.bat
generated
vendored
Normal file
63
vendor/github.com/apache/thrift/build/appveyor/cl_setcompiler.bat
generated
vendored
Normal file
|
@ -0,0 +1,63 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Detect the compiler edition we're building in.
|
||||||
|
:: Set the COMPILER environment variable to one of:
|
||||||
|
:: gcc = MinGW / MSYS2 and gcc toolchain
|
||||||
|
:: vc100 = Visual Studio 2010
|
||||||
|
:: vc110 = Visual Studio 2012
|
||||||
|
:: vc120 = Visual Studio 2013
|
||||||
|
:: vc140 = Visual Studio 2015
|
||||||
|
:: vc150 = Visual Studio 2017
|
||||||
|
::
|
||||||
|
:: Honors any existing COMPILER environment variable
|
||||||
|
:: setting instead of overwriting it, to allow it
|
||||||
|
:: to be forced if needed.
|
||||||
|
::
|
||||||
|
:: Sets ERRORLEVEL to 0 if COMPILER can be determined,
|
||||||
|
:: to 1 if it cannot.
|
||||||
|
::
|
||||||
|
|
||||||
|
IF DEFINED COMPILER (
|
||||||
|
ECHO [warn ] using existing environment variable COMPILER
|
||||||
|
EXIT /B 0
|
||||||
|
)
|
||||||
|
|
||||||
|
IF NOT "%PROFILE:~0,4%" == "MSVC" (
|
||||||
|
SET COMPILER=gcc
|
||||||
|
) ELSE (
|
||||||
|
CALL :CHECK 16
|
||||||
|
IF !ERRORLEVEL! == 0 (SET COMPILER=vc100)
|
||||||
|
CALL :CHECK 17
|
||||||
|
IF !ERRORLEVEL! == 0 (SET COMPILER=vc110)
|
||||||
|
CALL :CHECK 18
|
||||||
|
IF !ERRORLEVEL! == 0 (SET COMPILER=vc120)
|
||||||
|
CALL :CHECK 19.00
|
||||||
|
IF !ERRORLEVEL! == 0 (SET COMPILER=vc140)
|
||||||
|
CALL :CHECK 19.10
|
||||||
|
IF !ERRORLEVEL! == 0 (SET COMPILER=vc150)
|
||||||
|
)
|
||||||
|
|
||||||
|
IF NOT DEFINED COMPILER (
|
||||||
|
ECHO [error] unable to determine the compiler edition
|
||||||
|
EXIT /B 1
|
||||||
|
)
|
||||||
|
|
||||||
|
ECHO [info ] detected compiler edition %COMPILER%
|
||||||
|
EXIT /B 0
|
||||||
|
|
||||||
|
:CHECK
|
||||||
|
cl /? 2>&1 | findstr /C:"Version %1%." > nul
|
||||||
|
EXIT /B
|
92
vendor/github.com/apache/thrift/build/appveyor/cl_setenv.bat
generated
vendored
Normal file
92
vendor/github.com/apache/thrift/build/appveyor/cl_setenv.bat
generated
vendored
Normal file
|
@ -0,0 +1,92 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
IF "%PROFILE%" == "MSVC2010" (
|
||||||
|
CALL "C:\Program Files (x86)\Microsoft Visual Studio 10.0\VC\vcvarsall.bat" %PLATFORM%
|
||||||
|
) ELSE IF "%PROFILE%" == "MSVC2012" (
|
||||||
|
CALL "C:\Program Files (x86)\Microsoft Visual Studio 11.0\VC\vcvarsall.bat" %PLATFORM%
|
||||||
|
) ELSE IF "%PROFILE%" == "MSVC2013" (
|
||||||
|
CALL "C:\Program Files (x86)\Microsoft Visual Studio 12.0\VC\vcvarsall.bat" %PLATFORM%
|
||||||
|
) ELSE IF "%PROFILE%" == "MSVC2015" (
|
||||||
|
CALL "C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat" %PLATFORM%
|
||||||
|
) ELSE IF "%PROFILE%" == "MSVC2017" (
|
||||||
|
CALL "C:\Program Files (x86)\Microsoft Visual Studio\2017\Enterprise\Common7\Tools\VsDevCmd.bat" %PLATFORM%
|
||||||
|
) ELSE IF "%PROFILE%" == "MINGW" (
|
||||||
|
SET MSYS2_PATH_TYPE=stock
|
||||||
|
) ELSE IF "%PROFILE%" == "MSYS" (
|
||||||
|
SET MSYS2_PATH_TYPE=stock
|
||||||
|
) ELSE (
|
||||||
|
ECHO Unsupported PROFILE=%PROFILE% or PLATFORM=%PLATFORM%
|
||||||
|
EXIT /B 1
|
||||||
|
)
|
||||||
|
|
||||||
|
CALL cl_setcompiler.bat || EXIT /B
|
||||||
|
CALL cl_setgenerator.bat || EXIT /B
|
||||||
|
|
||||||
|
SET APPVEYOR_SCRIPTS=%APPVEYOR_BUILD_FOLDER%\build\appveyor
|
||||||
|
SET BUILDCACHE=%APPVEYOR_BUILD_FOLDER%\buildcache
|
||||||
|
SET BUILDDIR=%APPVEYOR_BUILD_FOLDER%\local-thrift-build
|
||||||
|
SET INSTDIR=%APPVEYOR_BUILD_FOLDER%\local-thrift-inst
|
||||||
|
SET SRCDIR=%APPVEYOR_BUILD_FOLDER%
|
||||||
|
|
||||||
|
: PLATFORM is x64 or x86, but we want x86 to become "32" when we strip it down for paths:
|
||||||
|
SET NORM_PLATFORM=%PLATFORM:~-2,2%
|
||||||
|
IF "%NORM_PLATFORM%" == "86" (SET NORM_PLATFORM=32)
|
||||||
|
|
||||||
|
:: FindBoost needs forward slashes so cmake doesn't see something as an escaped character
|
||||||
|
SET BOOST_ROOT=C:/Libraries/boost_%BOOST_VERSION:.=_%
|
||||||
|
SET BOOST_LIBRARYDIR=%BOOST_ROOT%/lib%NORM_PLATFORM%-msvc-%COMPILER:~-3,2%.0
|
||||||
|
SET OPENSSL_ROOT=C:\OpenSSL-Win%NORM_PLATFORM%
|
||||||
|
SET WIN3P=%APPVEYOR_BUILD_FOLDER%\thirdparty
|
||||||
|
|
||||||
|
:: MSVC2010 doesn't "do" std::thread
|
||||||
|
IF "%COMPILER%" == "vc100" (
|
||||||
|
SET THREADMODEL=BOOST
|
||||||
|
) ELSE (
|
||||||
|
SET THREADMODEL=STD
|
||||||
|
)
|
||||||
|
|
||||||
|
IF "%PYTHON_VERSION%" == "" (
|
||||||
|
SET WITH_PYTHON=OFF
|
||||||
|
) ELSE (
|
||||||
|
SET WITH_PYTHON=ON
|
||||||
|
SET PATH=C:\Python%PYTHON_VERSION:.=%\scripts;C:\Python%PYTHON_VERSION:.=%;!PATH!
|
||||||
|
)
|
||||||
|
IF "%CONFIGURATION%" == "Debug" (SET ZLIB_LIB_SUFFIX=d)
|
||||||
|
|
||||||
|
IF NOT "%QT_VERSION%" == "" (
|
||||||
|
IF /i "%PLATFORM%" == "x64" SET QTEXT=_64
|
||||||
|
SET PATH=C:\Qt\%QT_VERSION%\%PROFILE%!QTEXT!\bin;!PATH!
|
||||||
|
)
|
||||||
|
|
||||||
|
IF NOT "%PROFILE:~0,4%" == "MSVC" (
|
||||||
|
|
||||||
|
SET BASH=C:\msys64\usr\bin\bash.exe
|
||||||
|
SET BOOST_ROOT=
|
||||||
|
SET BOOST_INCLUDEDIR=/mingw64/include
|
||||||
|
SET BOOST_LIBRARYDIR=/mingw64/lib
|
||||||
|
SET OPENSSL_LIBRARIES=/mingw64/lib
|
||||||
|
SET OPENSSL_ROOT=/mingw64
|
||||||
|
SET WIN3P=
|
||||||
|
|
||||||
|
!BASH! -lc "sed -i '/export PATH=\/mingw64\/bin/d' ~/.bash_profile && echo 'export PATH=/mingw64/bin:$PATH' >> ~/.bash_profile" || EXIT /B
|
||||||
|
|
||||||
|
)
|
||||||
|
|
||||||
|
SET BUILDDIR_MSYS=%BUILDDIR:\=/%
|
||||||
|
SET BUILDDIR_MSYS=/c%BUILDDIR_MSYS:~2%
|
||||||
|
SET INSTDIR_MSYS=%INSTDIR:\=/%
|
||||||
|
SET INSTDIR_MSYS=/c%INSTDIR_MSYS:~2%
|
||||||
|
SET SRCDIR_MSYS=%SRCDIR:\=/%
|
||||||
|
SET SRCDIR_MSYS=/c%SRCDIR_MSYS:~2%
|
74
vendor/github.com/apache/thrift/build/appveyor/cl_setgenerator.bat
generated
vendored
Normal file
74
vendor/github.com/apache/thrift/build/appveyor/cl_setgenerator.bat
generated
vendored
Normal file
|
@ -0,0 +1,74 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Detect the compiler edition we're building in and then
|
||||||
|
:: set the GENERATOR environment variable to one of:
|
||||||
|
::
|
||||||
|
:: Visual Studio 15 2017 [arch] = Generates Visual Studio 2017 project files.
|
||||||
|
:: Optional [arch] can be "Win64" or "ARM".
|
||||||
|
:: Visual Studio 14 2015 [arch] = Generates Visual Studio 2015 project files.
|
||||||
|
:: Optional [arch] can be "Win64" or "ARM".
|
||||||
|
:: Visual Studio 12 2013 [arch] = Generates Visual Studio 2013 project files.
|
||||||
|
:: Optional [arch] can be "Win64" or "ARM".
|
||||||
|
:: Visual Studio 11 2012 [arch] = Generates Visual Studio 2012 project files.
|
||||||
|
:: Optional [arch] can be "Win64" or "ARM".
|
||||||
|
:: Visual Studio 10 2010 [arch] = Generates Visual Studio 2010 project files.
|
||||||
|
:: Optional [arch] can be "Win64" or "IA64".
|
||||||
|
:: MinGW Makefiles = Generates makefiles for MinGW
|
||||||
|
:: MSYS Makefiles = Generates makefiles for MSYS
|
||||||
|
::
|
||||||
|
:: Honors any existing GENERATOR environment variable
|
||||||
|
:: setting instead of overwriting it, to allow it
|
||||||
|
:: to be forced if needed.
|
||||||
|
::
|
||||||
|
:: Sets ERRORLEVEL to 0 if GENERATOR can be determined,
|
||||||
|
:: to 1 if it cannot.
|
||||||
|
::
|
||||||
|
|
||||||
|
IF DEFINED GENERATOR (
|
||||||
|
ECHO [warn ] using existing environment variable GENERATOR
|
||||||
|
EXIT /B 0
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
IF "%PROFILE:~0,4%" == "MING" (
|
||||||
|
SET GENERATOR=MinGW Makefiles
|
||||||
|
) ELSE IF "%PROFILE:~0,4%" == "MSYS" (
|
||||||
|
SET GENERATOR=MSYS Makefiles
|
||||||
|
) ELSE (
|
||||||
|
IF /i "%PLATFORM%" == "x64" SET GENARCH= Win64
|
||||||
|
CALL :CHECK 16
|
||||||
|
IF !ERRORLEVEL! == 0 SET GENERATOR=Visual Studio 10 2010!GENARCH!
|
||||||
|
CALL :CHECK 17
|
||||||
|
IF !ERRORLEVEL! == 0 SET GENERATOR=Visual Studio 11 2012!GENARCH!
|
||||||
|
CALL :CHECK 18
|
||||||
|
IF !ERRORLEVEL! == 0 SET GENERATOR=Visual Studio 12 2013!GENARCH!
|
||||||
|
CALL :CHECK 19.00
|
||||||
|
IF !ERRORLEVEL! == 0 SET GENERATOR=Visual Studio 14 2015!GENARCH!
|
||||||
|
CALL :CHECK 19.10
|
||||||
|
IF !ERRORLEVEL! == 0 SET GENERATOR=Visual Studio 15 2017!GENARCH!
|
||||||
|
)
|
||||||
|
|
||||||
|
IF NOT DEFINED GENERATOR (
|
||||||
|
ECHO [error] unable to determine the CMake generator to use
|
||||||
|
EXIT /B 1
|
||||||
|
)
|
||||||
|
|
||||||
|
ECHO [info ] using CMake generator %GENERATOR%
|
||||||
|
EXIT /B 0
|
||||||
|
|
||||||
|
:CHECK
|
||||||
|
cl /? 2>&1 | findstr /C:"Version %1%." > nul
|
||||||
|
EXIT /B
|
67
vendor/github.com/apache/thrift/build/appveyor/cl_showenv.bat
generated
vendored
Normal file
67
vendor/github.com/apache/thrift/build/appveyor/cl_showenv.bat
generated
vendored
Normal file
|
@ -0,0 +1,67 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
ECHO/
|
||||||
|
ECHO ===============================================================================
|
||||||
|
IF "%PROFILE:~0,4%" == "MSVC" (
|
||||||
|
ECHO Versions
|
||||||
|
ECHO -------------------------------------------------------------------------------
|
||||||
|
ECHO boost = %BOOST_VERSION%
|
||||||
|
ECHO libevent = %LIBEVENT_VERSION%
|
||||||
|
ECHO python = %PYTHON_VERSION%
|
||||||
|
ECHO qt = %QT_VERSION%
|
||||||
|
ECHO zlib = %ZLIB_VERSION%
|
||||||
|
ECHO/
|
||||||
|
)
|
||||||
|
ECHO Appveyor Variables
|
||||||
|
ECHO -------------------------------------------------------------------------------
|
||||||
|
ECHO APPVEYOR_BUILD_FOLDER = %APPVEYOR_BUILD_FOLDER%
|
||||||
|
ECHO CONFIGURATION = %CONFIGURATION%
|
||||||
|
ECHO PLATFORM = %PLATFORM%
|
||||||
|
ECHO PROFILE = %PROFILE%
|
||||||
|
ECHO/
|
||||||
|
ECHO Our Variables
|
||||||
|
ECHO -------------------------------------------------------------------------------
|
||||||
|
ECHO APPVEYOR_SCRIPTS = %APPVEYOR_SCRIPTS%
|
||||||
|
ECHO BOOST_ROOT = %BOOST_ROOT%
|
||||||
|
ECHO BOOST_INCLUDEDIR = %BOOST_INCLUDEDIR%
|
||||||
|
ECHO BOOST_LIBRARYDIR = %BOOST_LIBRARYDIR%
|
||||||
|
ECHO BUILDCACHE = %BUILDCACHE%
|
||||||
|
ECHO BUILDDIR = %BUILDDIR%
|
||||||
|
ECHO COMPILER = %COMPILER%
|
||||||
|
ECHO GENERATOR = %GENERATOR%
|
||||||
|
ECHO INSTDIR = %INSTDIR%
|
||||||
|
ECHO JAVA_HOME = %JAVA_HOME%
|
||||||
|
ECHO OPENSSL_ROOT = %OPENSSL_ROOT%
|
||||||
|
ECHO SRCDIR = %SRCDIR%
|
||||||
|
ECHO WIN3P = %WIN3P%
|
||||||
|
ECHO WITH_PYTHON = %WITH_PYTHON%
|
||||||
|
ECHO ZLIB_STATIC_SUFFIX = %ZLIB_STATIC_SUFFIX%
|
||||||
|
IF NOT "%PROFILE:~0,4%" == "MSVC" (
|
||||||
|
ECHO/
|
||||||
|
ECHO MSYS2/MinGW
|
||||||
|
ECHO -------------------------------------------------------------------------------
|
||||||
|
ECHO BUILDDIR_MSYS = %BUILDDIR_MSYS%
|
||||||
|
ECHO INSTDIR_MSYS = %INSTDIR_MSYS%
|
||||||
|
ECHO MSYS2_PATH_TYPE = %MSYS2_PATH_TYPE%
|
||||||
|
ECHO SRCDIR_MSYS = %SRCDIR_MSYS%
|
||||||
|
ECHO PATH =
|
||||||
|
C:\msys64\usr\bin\bash -lc "echo $PATH"
|
||||||
|
)
|
||||||
|
ECHO/
|
||||||
|
ECHO Windows PATH
|
||||||
|
ECHO -------------------------------------------------------------------------------
|
||||||
|
ECHO %PATH%
|
||||||
|
ECHO ===============================================================================
|
||||||
|
ECHO/
|
35
vendor/github.com/apache/thrift/build/appveyor/simulate-appveyor.bat
generated
vendored
Normal file
35
vendor/github.com/apache/thrift/build/appveyor/simulate-appveyor.bat
generated
vendored
Normal file
|
@ -0,0 +1,35 @@
|
||||||
|
::
|
||||||
|
:: Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
:: you may not use this file except in compliance with the License.
|
||||||
|
:: You may obtain a copy of the License at
|
||||||
|
::
|
||||||
|
:: http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
::
|
||||||
|
:: Unless required by applicable law or agreed to in writing, software
|
||||||
|
:: distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
:: WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
:: See the License for the specific language governing permissions and
|
||||||
|
:: limitations under the License.
|
||||||
|
::
|
||||||
|
|
||||||
|
::
|
||||||
|
:: Helps build thrift by pretending to be appveyor
|
||||||
|
:: Usage:
|
||||||
|
:: cd build\appveyor
|
||||||
|
:: simulate-appveyor.bat [Debug|Release] [x86|x64] [MINGW|MSVC2015]
|
||||||
|
::
|
||||||
|
|
||||||
|
@ECHO OFF
|
||||||
|
SETLOCAL EnableDelayedExpansion
|
||||||
|
|
||||||
|
SET APPVEYOR_BUILD_FOLDER=%~dp0..\..
|
||||||
|
SET CONFIGURATION=%1
|
||||||
|
SET PLATFORM=%2
|
||||||
|
SET PROFILE=%3
|
||||||
|
|
||||||
|
CD %APPVEYOR_BUILD_FOLDER%
|
||||||
|
CALL build\appveyor\%PROFILE:~0,4%-appveyor-install.bat || EXIT /B
|
||||||
|
CD %APPVEYOR_BUILD_FOLDER%
|
||||||
|
CALL build\appveyor\%PROFILE:~0,4%-appveyor-build.bat || EXIT /B
|
||||||
|
CD %APPVEYOR_BUILD_FOLDER%
|
||||||
|
CALL build\appveyor\%PROFILE:~0,4%-appveyor-test.bat
|
68
vendor/github.com/apache/thrift/build/cmake/CPackConfig.cmake
generated
vendored
Normal file
68
vendor/github.com/apache/thrift/build/cmake/CPackConfig.cmake
generated
vendored
Normal file
|
@ -0,0 +1,68 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
#TODO: Should we bundle system libraries for DLLs?
|
||||||
|
#include(InstallRequiredSystemLibraries)
|
||||||
|
|
||||||
|
# For help take a look at:
|
||||||
|
# http://www.cmake.org/Wiki/CMake:CPackConfiguration
|
||||||
|
|
||||||
|
### general settings
|
||||||
|
set(CPACK_PACKAGE_NAME "thrift")
|
||||||
|
set(CPACK_PACKAGE_VERSION "${PACKAGE_VERSION}")
|
||||||
|
set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "Apache Thrift")
|
||||||
|
set(CPACK_PACKAGE_DESCRIPTION_FILE "${CMAKE_CURRENT_SOURCE_DIR}/README.md")
|
||||||
|
set(CPACK_RESOURCE_FILE_LICENSE "${CMAKE_CURRENT_SOURCE_DIR}/LICENSE")
|
||||||
|
set(CPACK_PACKAGE_VENDOR "Apache Software Foundation")
|
||||||
|
set(CPACK_PACKAGE_CONTACT "dev@thrift.apache.org")
|
||||||
|
set(CPACK_PACKAGE_INSTALL_DIRECTORY "${CPACK_PACKAGE_NAME}")
|
||||||
|
set(CPACK_SYSTEM_NAME "${CMAKE_SYSTEM_NAME}")
|
||||||
|
|
||||||
|
### versions
|
||||||
|
set(CPACK_PACKAGE_VERSION_MAJOR ${thrift_VERSION_MAJOR})
|
||||||
|
set(CPACK_PACKAGE_VERSION_MINOR ${thrift_VERSION_MINOR})
|
||||||
|
set(CPACK_PACKAGE_VERSION_PATCH ${thrift_VERSION_PATCH})
|
||||||
|
|
||||||
|
### source generator
|
||||||
|
set(CPACK_SOURCE_GENERATOR "TGZ")
|
||||||
|
set(CPACK_SOURCE_IGNORE_FILES "~$;[.]swp$;/[.]svn/;/[.]git/;.gitignore;/build/;tags;cscope.*")
|
||||||
|
set(CPACK_SOURCE_PACKAGE_FILE_NAME "${CPACK_PACKAGE_NAME}-${CPACK_PACKAGE_VERSION}")
|
||||||
|
|
||||||
|
### zip generator
|
||||||
|
set(CPACK_GENERATOR "ZIP")
|
||||||
|
set(CPACK_PACKAGE_INSTALL_DIRECTORY "thrift")
|
||||||
|
|
||||||
|
|
||||||
|
if(CMAKE_SYSTEM_NAME STREQUAL "Windows")
|
||||||
|
set(CPACK_GENERATOR "NSIS")
|
||||||
|
set(CPACK_NSIS_HELP_LINK "http://thrift.apache.org")
|
||||||
|
set(CPACK_NSIS_MENU_LINKS
|
||||||
|
"http://thrift.apache.org" "Apache Thrift - Web Site"
|
||||||
|
"https://issues.apache.org/jira/browse/THRIFT" "Apache Thrift - Issues")
|
||||||
|
set(CPACK_NSIS_CONTACT ${CPACK_PACKAGE_CONTACT})
|
||||||
|
set(CPACK_NSIS_MODIFY_PATH "ON")
|
||||||
|
set(CPACK_PACKAGE_INSTALL_DIRECTORY "${CPACK_PACKAGE_NAME}")
|
||||||
|
else()
|
||||||
|
set(CPACK_GENERATOR "DEB" )
|
||||||
|
set(CPACK_DEBIAN_PACKAGE_MAINTAINER ${CPACK_PACKAGE_CONTACT})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
|
||||||
|
include(CPack)
|
76
vendor/github.com/apache/thrift/build/cmake/ConfigureChecks.cmake
generated
vendored
Normal file
76
vendor/github.com/apache/thrift/build/cmake/ConfigureChecks.cmake
generated
vendored
Normal file
|
@ -0,0 +1,76 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
include(CheckFunctionExists)
|
||||||
|
include(CheckIncludeFile)
|
||||||
|
include(CheckIncludeFiles)
|
||||||
|
include(CheckSymbolExists)
|
||||||
|
|
||||||
|
if (Inttypes_FOUND)
|
||||||
|
# This allows the inttypes.h and stdint.h checks to succeed on platforms that
|
||||||
|
# do not natively provide there.
|
||||||
|
set (CMAKE_REQUIRED_INCLUDES ${INTTYPES_INCLUDE_DIRS})
|
||||||
|
endif ()
|
||||||
|
|
||||||
|
check_include_file(arpa/inet.h HAVE_ARPA_INET_H)
|
||||||
|
check_include_file(fcntl.h HAVE_FCNTL_H)
|
||||||
|
check_include_file(getopt.h HAVE_GETOPT_H)
|
||||||
|
check_include_file(inttypes.h HAVE_INTTYPES_H)
|
||||||
|
check_include_file(netdb.h HAVE_NETDB_H)
|
||||||
|
check_include_file(netinet/in.h HAVE_NETINET_IN_H)
|
||||||
|
check_include_file(stdint.h HAVE_STDINT_H)
|
||||||
|
check_include_file(unistd.h HAVE_UNISTD_H)
|
||||||
|
check_include_file(pthread.h HAVE_PTHREAD_H)
|
||||||
|
check_include_file(sys/time.h HAVE_SYS_TIME_H)
|
||||||
|
check_include_file(sys/param.h HAVE_SYS_PARAM_H)
|
||||||
|
check_include_file(sys/resource.h HAVE_SYS_RESOURCE_H)
|
||||||
|
check_include_file(sys/socket.h HAVE_SYS_SOCKET_H)
|
||||||
|
check_include_file(sys/stat.h HAVE_SYS_STAT_H)
|
||||||
|
check_include_file(sys/un.h HAVE_SYS_UN_H)
|
||||||
|
check_include_file(sys/poll.h HAVE_SYS_POLL_H)
|
||||||
|
check_include_file(sys/select.h HAVE_SYS_SELECT_H)
|
||||||
|
check_include_file(sched.h HAVE_SCHED_H)
|
||||||
|
check_include_file(string.h HAVE_STRING_H)
|
||||||
|
check_include_file(strings.h HAVE_STRINGS_H)
|
||||||
|
|
||||||
|
check_function_exists(gethostbyname HAVE_GETHOSTBYNAME)
|
||||||
|
check_function_exists(gethostbyname_r HAVE_GETHOSTBYNAME_R)
|
||||||
|
check_function_exists(strerror_r HAVE_STRERROR_R)
|
||||||
|
check_function_exists(sched_get_priority_max HAVE_SCHED_GET_PRIORITY_MAX)
|
||||||
|
check_function_exists(sched_get_priority_min HAVE_SCHED_GET_PRIORITY_MIN)
|
||||||
|
|
||||||
|
include(CheckCSourceCompiles)
|
||||||
|
include(CheckCXXSourceCompiles)
|
||||||
|
|
||||||
|
check_cxx_source_compiles(
|
||||||
|
"
|
||||||
|
#include <string.h>
|
||||||
|
int main(){char b;char *a = strerror_r(0, &b, 0); return(0);}
|
||||||
|
"
|
||||||
|
STRERROR_R_CHAR_P)
|
||||||
|
|
||||||
|
|
||||||
|
set(PACKAGE ${PACKAGE_NAME})
|
||||||
|
set(PACKAGE_STRING "${PACKAGE_NAME} ${PACKAGE_VERSION}")
|
||||||
|
set(VERSION ${thrift_VERSION})
|
||||||
|
|
||||||
|
# generate a config.h file
|
||||||
|
configure_file("${CMAKE_CURRENT_SOURCE_DIR}/build/cmake/config.h.in" "${CMAKE_CURRENT_BINARY_DIR}/thrift/config.h")
|
||||||
|
|
||||||
|
include_directories("${CMAKE_CURRENT_BINARY_DIR}")
|
93
vendor/github.com/apache/thrift/build/cmake/DefineCMakeDefaults.cmake
generated
vendored
Normal file
93
vendor/github.com/apache/thrift/build/cmake/DefineCMakeDefaults.cmake
generated
vendored
Normal file
|
@ -0,0 +1,93 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# Always include srcdir and builddir in include path
|
||||||
|
# This saves typing ${CMAKE_CURRENT_SOURCE_DIR} ${CMAKE_CURRENT_BINARY} in
|
||||||
|
# about every subdir
|
||||||
|
# since cmake 2.4.0
|
||||||
|
set(CMAKE_INCLUDE_CURRENT_DIR ON)
|
||||||
|
|
||||||
|
# Put the include dirs which are in the source or build tree
|
||||||
|
# before all other include dirs, so the headers in the sources
|
||||||
|
# are preferred over the already installed ones
|
||||||
|
# since cmake 2.4.1
|
||||||
|
set(CMAKE_INCLUDE_DIRECTORIES_PROJECT_BEFORE ON)
|
||||||
|
|
||||||
|
# Use colored output
|
||||||
|
# since cmake 2.4.0
|
||||||
|
set(CMAKE_COLOR_MAKEFILE ON)
|
||||||
|
|
||||||
|
# Define the generic version of the libraries here
|
||||||
|
set(GENERIC_LIB_VERSION "0.1.0")
|
||||||
|
set(GENERIC_LIB_SOVERSION "0")
|
||||||
|
|
||||||
|
# Set the default build type to release with debug info
|
||||||
|
if (NOT CMAKE_BUILD_TYPE)
|
||||||
|
set(CMAKE_BUILD_TYPE RelWithDebInfo
|
||||||
|
CACHE STRING
|
||||||
|
"Choose the type of build, options are: None Debug Release RelWithDebInfo MinSizeRel."
|
||||||
|
)
|
||||||
|
endif (NOT CMAKE_BUILD_TYPE)
|
||||||
|
|
||||||
|
# Create the compile command database for clang by default
|
||||||
|
set(CMAKE_EXPORT_COMPILE_COMMANDS ON)
|
||||||
|
|
||||||
|
# Put the libraries and binaries that get built into directories at the
|
||||||
|
# top of the build tree rather than in hard-to-find leaf
|
||||||
|
# directories. This simplifies manual testing and the use of the build
|
||||||
|
# tree rather than installed thrift libraries.
|
||||||
|
set(CMAKE_LIBRARY_OUTPUT_DIRECTORY ${CMAKE_BINARY_DIR}/lib)
|
||||||
|
set(CMAKE_ARCHIVE_OUTPUT_DIRECTORY ${CMAKE_BINARY_DIR}/lib)
|
||||||
|
set(CMAKE_RUNTIME_OUTPUT_DIRECTORY ${CMAKE_BINARY_DIR}/bin)
|
||||||
|
|
||||||
|
#
|
||||||
|
# "rpath" support.
|
||||||
|
# See http://www.itk.org/Wiki/index.php?title=CMake_RPATH_handling
|
||||||
|
#
|
||||||
|
# On MacOSX, for shared libraries, enable rpath support.
|
||||||
|
set(CMAKE_MACOSX_RPATH TRUE)
|
||||||
|
#
|
||||||
|
# On any OS, for executables, allow linking with shared libraries in non-system
|
||||||
|
# locations and running the executables without LD_PRELOAD or similar.
|
||||||
|
# This requires the library to be built with rpath support.
|
||||||
|
set(CMAKE_INSTALL_RPATH_USE_LINK_PATH TRUE)
|
||||||
|
|
||||||
|
#
|
||||||
|
# C++ Language Level Defaults - this depends on the compiler capabilities
|
||||||
|
#
|
||||||
|
if (NOT DEFINED CMAKE_CXX_STANDARD)
|
||||||
|
if (MSVC AND MSVC_VERSION LESS 1800)
|
||||||
|
# MSVC 2012 and earlier don't support template aliases so you have to use C++98
|
||||||
|
set(CMAKE_CXX_STANDARD 98)
|
||||||
|
message(STATUS "Setting C++98 as the default language level (for an older MSVC compiler).")
|
||||||
|
else()
|
||||||
|
set(CMAKE_CXX_STANDARD 11) # C++11
|
||||||
|
message(STATUS "Setting C++11 as the default language level.")
|
||||||
|
endif()
|
||||||
|
message(STATUS "To specify a different C++ language level, set CMAKE_CXX_STANDARD")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if (NOT DEFINED CMAKE_CXX_STANDARD_REQUIRED)
|
||||||
|
set(CMAKE_CXX_STANDARD_REQUIRED OFF) # can degrade to C++98 if compiler does not support C++11
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if (NOT DEFINED CMAKE_CXX_EXTENSIONS)
|
||||||
|
set(CMAKE_CXX_EXTENSIONS OFF) # use standards compliant language level for portability
|
||||||
|
endif()
|
26
vendor/github.com/apache/thrift/build/cmake/DefineInstallationPaths.cmake
generated
vendored
Normal file
26
vendor/github.com/apache/thrift/build/cmake/DefineInstallationPaths.cmake
generated
vendored
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# Define the default install paths
|
||||||
|
set(BIN_INSTALL_DIR "bin" CACHE PATH "The binary install dir (default: bin)")
|
||||||
|
set(LIB_INSTALL_DIR "lib${LIB_SUFFIX}" CACHE PATH "The library install dir (default: lib${LIB_SUFFIX})")
|
||||||
|
set(INCLUDE_INSTALL_DIR "include" CACHE PATH "The library install dir (default: include)")
|
||||||
|
set(CMAKE_INSTALL_DIR "cmake" CACHE PATH "The subdirectory to install cmake config files (default: cmake)")
|
||||||
|
set(DOC_INSTALL_DIR "share/doc" CACHE PATH "The subdirectory to install documentation files (default: share/doc)")
|
219
vendor/github.com/apache/thrift/build/cmake/DefineOptions.cmake
generated
vendored
Normal file
219
vendor/github.com/apache/thrift/build/cmake/DefineOptions.cmake
generated
vendored
Normal file
|
@ -0,0 +1,219 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
include(CMakeDependentOption)
|
||||||
|
|
||||||
|
set(THRIFT_COMPILER "" CACHE FILEPATH "External Thrift compiler to use during build")
|
||||||
|
|
||||||
|
# Additional components
|
||||||
|
option(BUILD_COMPILER "Build Thrift compiler" ON)
|
||||||
|
|
||||||
|
if(BUILD_COMPILER OR EXISTS ${THRIFT_COMPILER})
|
||||||
|
set(HAVE_COMPILER ON)
|
||||||
|
endif()
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_TESTING "Build with unit tests" ON "HAVE_COMPILER" OFF)
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_EXAMPLES "Build examples" ON "HAVE_COMPILER" OFF)
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_TUTORIALS "Build Thrift tutorials" ON "HAVE_COMPILER" OFF)
|
||||||
|
option(BUILD_LIBRARIES "Build Thrift libraries" ON)
|
||||||
|
|
||||||
|
# Libraries to build
|
||||||
|
|
||||||
|
# Each language library can be enabled or disabled using the WITH_<LANG> flag.
|
||||||
|
# By default CMake checks if the required dependencies for a language are present
|
||||||
|
# and enables the library if all are found. This means the default is to build as
|
||||||
|
# much as possible but leaving out libraries if their dependencies are not met.
|
||||||
|
|
||||||
|
option(WITH_BOOST_FUNCTIONAL "Use boost/tr1/functional.hpp even under C++11 or later" OFF)
|
||||||
|
if (WITH_BOOST_FUNCTIONAL)
|
||||||
|
add_definitions(-DFORCE_BOOST_FUNCTIONAL)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
option(WITH_BOOST_SMART_PTR "Use boost/smart_ptr.hpp even under C++11 or later" OFF)
|
||||||
|
if (WITH_BOOST_SMART_PTR)
|
||||||
|
add_definitions(-DFORCE_BOOST_SMART_PTR)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
option(WITH_BOOST_STATIC "Build with Boost static link library" OFF)
|
||||||
|
set(Boost_USE_STATIC_LIBS ${WITH_BOOST_STATIC})
|
||||||
|
if (NOT WITH_BOOST_STATIC)
|
||||||
|
add_definitions(-DBOOST_ALL_DYN_LINK)
|
||||||
|
add_definitions(-DBOOST_TEST_DYN_LINK)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# C++
|
||||||
|
option(WITH_CPP "Build C++ Thrift library" ON)
|
||||||
|
if(WITH_CPP)
|
||||||
|
find_package(Boost 1.53 QUIET)
|
||||||
|
# NOTE: Currently the following options are C++ specific,
|
||||||
|
# but in future other libraries might reuse them.
|
||||||
|
# So they are not dependent on WITH_CPP but setting them without WITH_CPP currently
|
||||||
|
# has no effect.
|
||||||
|
if(ZLIB_LIBRARY)
|
||||||
|
# FindZLIB.cmake does not normalize path so we need to do it ourselves.
|
||||||
|
file(TO_CMAKE_PATH ${ZLIB_LIBRARY} ZLIB_LIBRARY)
|
||||||
|
endif()
|
||||||
|
find_package(ZLIB QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_ZLIB "Build with ZLIB support" ON
|
||||||
|
"ZLIB_FOUND" OFF)
|
||||||
|
find_package(Libevent QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_LIBEVENT "Build with libevent support" ON
|
||||||
|
"Libevent_FOUND" OFF)
|
||||||
|
find_package(Qt4 QUIET COMPONENTS QtCore QtNetwork)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_QT4 "Build with Qt4 support" ON
|
||||||
|
"QT4_FOUND" OFF)
|
||||||
|
find_package(Qt5 QUIET COMPONENTS Core Network)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_QT5 "Build with Qt5 support" ON
|
||||||
|
"Qt5_FOUND" OFF)
|
||||||
|
if(${WITH_QT4} AND ${WITH_QT5} AND ${CMAKE_MAJOR_VERSION} LESS 3)
|
||||||
|
# cmake < 3.0.0 causes conflict when building both Qt4 and Qt5
|
||||||
|
set(WITH_QT4 OFF)
|
||||||
|
endif()
|
||||||
|
find_package(OpenSSL QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_OPENSSL "Build with OpenSSL support" ON
|
||||||
|
"OPENSSL_FOUND" OFF)
|
||||||
|
option(WITH_STDTHREADS "Build with C++ std::thread support" OFF)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_BOOSTTHREADS "Build with Boost threads support" OFF
|
||||||
|
"NOT WITH_STDTHREADS;Boost_FOUND" OFF)
|
||||||
|
endif()
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_CPP "Build C++ library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_CPP;Boost_FOUND" OFF)
|
||||||
|
CMAKE_DEPENDENT_OPTION(WITH_PLUGIN "Build compiler plugin support" OFF
|
||||||
|
"BUILD_COMPILER;BUILD_CPP" OFF)
|
||||||
|
|
||||||
|
# C GLib
|
||||||
|
option(WITH_C_GLIB "Build C (GLib) Thrift library" ON)
|
||||||
|
if(WITH_C_GLIB)
|
||||||
|
find_package(GLIB QUIET COMPONENTS gobject)
|
||||||
|
endif()
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_C_GLIB "Build C (GLib) library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_C_GLIB;GLIB_FOUND" OFF)
|
||||||
|
|
||||||
|
if(BUILD_CPP)
|
||||||
|
set(boost_components)
|
||||||
|
if(WITH_BOOSTTHREADS OR BUILD_TESTING)
|
||||||
|
list(APPEND boost_components system thread)
|
||||||
|
endif()
|
||||||
|
if(BUILD_TESTING)
|
||||||
|
list(APPEND boost_components unit_test_framework filesystem chrono program_options)
|
||||||
|
endif()
|
||||||
|
if(boost_components)
|
||||||
|
find_package(Boost 1.53 REQUIRED COMPONENTS ${boost_components})
|
||||||
|
endif()
|
||||||
|
elseif(BUILD_C_GLIB AND BUILD_TESTING)
|
||||||
|
find_package(Boost 1.53 REQUIRED)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# Java
|
||||||
|
option(WITH_JAVA "Build Java Thrift library" ON)
|
||||||
|
if(ANDROID)
|
||||||
|
find_package(Gradle QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_JAVA "Build Java library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_JAVA;GRADLE_FOUND" OFF)
|
||||||
|
else()
|
||||||
|
find_package(Java QUIET)
|
||||||
|
find_package(Ant QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_JAVA "Build Java library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_JAVA;JAVA_FOUND;ANT_FOUND" OFF)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# Python
|
||||||
|
option(WITH_PYTHON "Build Python Thrift library" ON)
|
||||||
|
find_package(PythonInterp QUIET) # for Python executable
|
||||||
|
find_package(PythonLibs QUIET) # for Python.h
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_PYTHON "Build Python library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_PYTHON;PYTHONLIBS_FOUND" OFF)
|
||||||
|
|
||||||
|
# Haskell
|
||||||
|
option(WITH_HASKELL "Build Haskell Thrift library" ON)
|
||||||
|
find_package(GHC QUIET)
|
||||||
|
find_package(Cabal QUIET)
|
||||||
|
CMAKE_DEPENDENT_OPTION(BUILD_HASKELL "Build GHC library" ON
|
||||||
|
"BUILD_LIBRARIES;WITH_HASKELL;GHC_FOUND;CABAL_FOUND" OFF)
|
||||||
|
|
||||||
|
# Common library options
|
||||||
|
option(WITH_SHARED_LIB "Build shared libraries" ON)
|
||||||
|
option(WITH_STATIC_LIB "Build static libraries" ON)
|
||||||
|
if (NOT WITH_SHARED_LIB AND NOT WITH_STATIC_LIB)
|
||||||
|
message(FATAL_ERROR "Cannot build with both shared and static outputs disabled!")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
#NOTE: C++ compiler options are defined in the lib/cpp/CMakeLists.txt
|
||||||
|
|
||||||
|
# Visual Studio only options
|
||||||
|
if(MSVC)
|
||||||
|
option(WITH_MT "Build using MT instead of MD (MSVC only)" OFF)
|
||||||
|
endif(MSVC)
|
||||||
|
|
||||||
|
macro(MESSAGE_DEP flag summary)
|
||||||
|
if(NOT ${flag})
|
||||||
|
message(STATUS " - ${summary}")
|
||||||
|
endif()
|
||||||
|
endmacro(MESSAGE_DEP flag summary)
|
||||||
|
|
||||||
|
macro(PRINT_CONFIG_SUMMARY)
|
||||||
|
message(STATUS "----------------------------------------------------------")
|
||||||
|
message(STATUS "Thrift version: ${thrift_VERSION} (${thrift_VERSION_MAJOR}.${thrift_VERSION_MINOR}.${thrift_VERSION_PATCH})")
|
||||||
|
message(STATUS "Thrift package version: ${PACKAGE_VERSION}")
|
||||||
|
message(STATUS "Build configuration Summary")
|
||||||
|
message(STATUS " Build Thrift compiler: ${BUILD_COMPILER}")
|
||||||
|
message(STATUS " Build compiler plugin support: ${WITH_PLUGIN}")
|
||||||
|
message(STATUS " Build with unit tests: ${BUILD_TESTING}")
|
||||||
|
MESSAGE_DEP(HAVE_COMPILER "Disabled because BUILD_THRIFT=OFF and no valid THRIFT_COMPILER is given")
|
||||||
|
message(STATUS " Build examples: ${BUILD_EXAMPLES}")
|
||||||
|
MESSAGE_DEP(HAVE_COMPILER "Disabled because BUILD_THRIFT=OFF and no valid THRIFT_COMPILER is given")
|
||||||
|
message(STATUS " Build Thrift libraries: ${BUILD_LIBRARIES}")
|
||||||
|
message(STATUS " Language libraries:")
|
||||||
|
message(STATUS " Build C++ library: ${BUILD_CPP}")
|
||||||
|
MESSAGE_DEP(WITH_CPP "Disabled by WITH_CPP=OFF")
|
||||||
|
MESSAGE_DEP(Boost_FOUND "Boost headers missing")
|
||||||
|
message(STATUS " C++ Language Level: ${CXX_LANGUAGE_LEVEL}")
|
||||||
|
message(STATUS " Build C (GLib) library: ${BUILD_C_GLIB}")
|
||||||
|
MESSAGE_DEP(WITH_C_GLIB "Disabled by WITH_C_GLIB=OFF")
|
||||||
|
MESSAGE_DEP(GLIB_FOUND "GLib missing")
|
||||||
|
message(STATUS " Build Java library: ${BUILD_JAVA}")
|
||||||
|
MESSAGE_DEP(WITH_JAVA "Disabled by WITH_JAVA=OFF")
|
||||||
|
if(ANDROID)
|
||||||
|
MESSAGE_DEP(GRADLE_FOUND "Gradle missing")
|
||||||
|
else()
|
||||||
|
MESSAGE_DEP(JAVA_FOUND "Java Runtime missing")
|
||||||
|
MESSAGE_DEP(ANT_FOUND "Ant missing")
|
||||||
|
endif()
|
||||||
|
message(STATUS " Build Python library: ${BUILD_PYTHON}")
|
||||||
|
MESSAGE_DEP(WITH_PYTHON "Disabled by WITH_PYTHON=OFF")
|
||||||
|
MESSAGE_DEP(PYTHONLIBS_FOUND "Python libraries missing")
|
||||||
|
message(STATUS " Build Haskell library: ${BUILD_HASKELL}")
|
||||||
|
MESSAGE_DEP(WITH_HASKELL "Disabled by WITH_HASKELL=OFF")
|
||||||
|
MESSAGE_DEP(GHC_FOUND "GHC missing")
|
||||||
|
MESSAGE_DEP(CABAL_FOUND "Cabal missing")
|
||||||
|
message(STATUS " Library features:")
|
||||||
|
message(STATUS " Build shared libraries: ${WITH_SHARED_LIB}")
|
||||||
|
message(STATUS " Build static libraries: ${WITH_STATIC_LIB}")
|
||||||
|
message(STATUS " Build with Boost static link library: ${WITH_BOOST_STATIC}")
|
||||||
|
message(STATUS " Build with Boost thread support: ${WITH_BOOSTTHREADS}")
|
||||||
|
message(STATUS " Build with boost/tr1/functional (forced) ${WITH_BOOST_FUNCTIONAL}")
|
||||||
|
message(STATUS " Build with boost/smart_ptr (forced) ${WITH_BOOST_SMART_PTR}")
|
||||||
|
message(STATUS " Build with C++ std::thread support: ${WITH_STDTHREADS}")
|
||||||
|
message(STATUS " Build with libevent support: ${WITH_LIBEVENT}")
|
||||||
|
message(STATUS " Build with OpenSSL support: ${WITH_OPENSSL}")
|
||||||
|
message(STATUS " Build with Qt4 support: ${WITH_QT4}")
|
||||||
|
message(STATUS " Build with Qt5 support: ${WITH_QT5}")
|
||||||
|
message(STATUS " Build with ZLIB support: ${WITH_ZLIB}")
|
||||||
|
message(STATUS "----------------------------------------------------------")
|
||||||
|
endmacro(PRINT_CONFIG_SUMMARY)
|
131
vendor/github.com/apache/thrift/build/cmake/DefinePlatformSpecifc.cmake
generated
vendored
Normal file
131
vendor/github.com/apache/thrift/build/cmake/DefinePlatformSpecifc.cmake
generated
vendored
Normal file
|
@ -0,0 +1,131 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
# Uncomment this to show some basic cmake variables about platforms
|
||||||
|
# include (NewPlatformDebug)
|
||||||
|
|
||||||
|
# Visual Studio specific options
|
||||||
|
if(MSVC)
|
||||||
|
#For visual studio the library naming is as following:
|
||||||
|
# Dynamic libraries:
|
||||||
|
# - thrift.dll for release library
|
||||||
|
# - thriftd.dll for debug library
|
||||||
|
#
|
||||||
|
# Static libraries:
|
||||||
|
# - thriftmd.lib for /MD release build
|
||||||
|
# - thriftmt.lib for /MT release build
|
||||||
|
#
|
||||||
|
# - thriftmdd.lib for /MD debug build
|
||||||
|
# - thriftmtd.lib for /MT debug build
|
||||||
|
#
|
||||||
|
# the same holds for other libraries like libthriftz etc.
|
||||||
|
|
||||||
|
# For Debug build types, append a "d" to the library names.
|
||||||
|
set(CMAKE_DEBUG_POSTFIX "d" CACHE STRING "Set debug library postfix" FORCE)
|
||||||
|
set(CMAKE_RELEASE_POSTFIX "" CACHE STRING "Set release library postfix" FORCE)
|
||||||
|
set(CMAKE_RELWITHDEBINFO_POSTFIX "" CACHE STRING "Set release library postfix" FORCE)
|
||||||
|
|
||||||
|
# Build using /MT option instead of /MD if the WITH_MT options is set
|
||||||
|
if(WITH_MT)
|
||||||
|
set(CompilerFlags
|
||||||
|
CMAKE_CXX_FLAGS
|
||||||
|
CMAKE_CXX_FLAGS_DEBUG
|
||||||
|
CMAKE_CXX_FLAGS_RELEASE
|
||||||
|
CMAKE_CXX_FLAGS_RELWITHDEBINFO
|
||||||
|
CMAKE_C_FLAGS
|
||||||
|
CMAKE_C_FLAGS_DEBUG
|
||||||
|
CMAKE_C_FLAGS_RELEASE
|
||||||
|
CMAKE_C_FLAGS_RELWITHDEBINFO
|
||||||
|
)
|
||||||
|
foreach(CompilerFlag ${CompilerFlags})
|
||||||
|
string(REPLACE "/MD" "/MT" ${CompilerFlag} "${${CompilerFlag}}")
|
||||||
|
endforeach()
|
||||||
|
set(STATIC_POSTFIX "mt" CACHE STRING "Set static library postfix" FORCE)
|
||||||
|
else(WITH_MT)
|
||||||
|
set(STATIC_POSTFIX "md" CACHE STRING "Set static library postfix" FORCE)
|
||||||
|
endif(WITH_MT)
|
||||||
|
|
||||||
|
# Disable Windows.h definition of macros for min and max
|
||||||
|
add_definitions("-DNOMINMAX")
|
||||||
|
|
||||||
|
# Disable boost auto linking pragmas - cmake includes the right files
|
||||||
|
add_definitions("-DBOOST_ALL_NO_LIB")
|
||||||
|
|
||||||
|
# Windows build does not know how to make a shared library yet
|
||||||
|
# as there are no __declspec(dllexport) or exports files in the project.
|
||||||
|
if (WITH_SHARED_LIB)
|
||||||
|
message (FATAL_ERROR "Windows build does not support shared library output yet, please set -DWITH_SHARED_LIB=off")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
add_definitions("/MP") # parallel build
|
||||||
|
add_definitions("/W3") # warning level 3
|
||||||
|
|
||||||
|
# VS2010 does not provide inttypes which we need for "PRId64" used in many places
|
||||||
|
find_package(Inttypes)
|
||||||
|
if (Inttypes_FOUND)
|
||||||
|
include_directories(${INTTYPES_INCLUDE_DIRS})
|
||||||
|
# OpenSSL conflicts with the definition of PRId64 unless it is defined first
|
||||||
|
add_definitions("/FIinttypes.h")
|
||||||
|
endif ()
|
||||||
|
elseif(UNIX)
|
||||||
|
find_program( MEMORYCHECK_COMMAND valgrind )
|
||||||
|
set( MEMORYCHECK_COMMAND_OPTIONS "--gen-suppressions=all --leak-check=full" )
|
||||||
|
set( MEMORYCHECK_SUPPRESSIONS_FILE "${PROJECT_SOURCE_DIR}/test/valgrind.suppress" )
|
||||||
|
endif()
|
||||||
|
|
||||||
|
add_definitions("-D__STDC_FORMAT_MACROS")
|
||||||
|
|
||||||
|
# WITH_*THREADS selects which threading library to use
|
||||||
|
if(WITH_BOOSTTHREADS)
|
||||||
|
add_definitions("-DUSE_BOOST_THREAD=1")
|
||||||
|
elseif(WITH_STDTHREADS)
|
||||||
|
add_definitions("-DUSE_STD_THREAD=1")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# C++ Language Level
|
||||||
|
set(CXX_LANGUAGE_LEVEL "C++${CMAKE_CXX_STANDARD}")
|
||||||
|
if (CMAKE_CXX_STANDARD_REQUIRED)
|
||||||
|
string(CONCAT CXX_LANGUAGE_LEVEL "${CXX_LANGUAGE_LEVEL} [compiler must support it]")
|
||||||
|
else()
|
||||||
|
string(CONCAT CXX_LANGUAGE_LEVEL "${CXX_LANGUAGE_LEVEL} [fallback to earlier if compiler does not support it]")
|
||||||
|
endif()
|
||||||
|
if (CMAKE_CXX_EXTENSIONS)
|
||||||
|
string(CONCAT CXX_LANGUAGE_LEVEL "${CXX_LANGUAGE_LEVEL} [with compiler-specific extensions]")
|
||||||
|
else()
|
||||||
|
if ((CMAKE_CXX_COMPILER_ID MATCHES "GNU" OR CMAKE_CXX_COMPILER_ID MATCHES "Clang") AND NOT MINGW)
|
||||||
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -Wno-variadic-macros -Wno-long-long")
|
||||||
|
endif()
|
||||||
|
if ((CMAKE_CXX_COMPILER_ID MATCHES "Clang") AND NOT MINGW)
|
||||||
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -Wno-c++11-long-long")
|
||||||
|
endif()
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if (CMAKE_CXX_COMPILER_ID MATCHES "Clang")
|
||||||
|
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -Wno-deprecated-register")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# Building WITH_PLUGIN requires boost memory operations, for now, and gcc >= 4.8
|
||||||
|
if (WITH_PLUGIN)
|
||||||
|
if (CMAKE_CXX_COMPILER_ID MATCHES "GNU" AND CMAKE_CXX_COMPILER_VERSION VERSION_LESS "4.8")
|
||||||
|
message(SEND_ERROR "Thrift compiler plug-in support is not possible with older gcc ( < 4.8 ) compiler")
|
||||||
|
endif()
|
||||||
|
message(STATUS "Forcing use of boost::smart_ptr to build WITH_PLUGIN")
|
||||||
|
add_definitions("-DFORCE_BOOST_SMART_PTR=1")
|
||||||
|
endif()
|
||||||
|
|
30
vendor/github.com/apache/thrift/build/cmake/FindAnt.cmake
generated
vendored
Normal file
30
vendor/github.com/apache/thrift/build/cmake/FindAnt.cmake
generated
vendored
Normal file
|
@ -0,0 +1,30 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# ANT_FOUND - system has Ant
|
||||||
|
# Ant_EXECUTABLE - the Ant executable
|
||||||
|
#
|
||||||
|
# It will search the environment variable ANT_HOME if it is set
|
||||||
|
|
||||||
|
include(FindPackageHandleStandardArgs)
|
||||||
|
|
||||||
|
find_program(Ant_EXECUTABLE NAMES ant PATHS $ENV{ANT_HOME}/bin)
|
||||||
|
find_package_handle_standard_args(Ant DEFAULT_MSG Ant_EXECUTABLE)
|
||||||
|
mark_as_advanced(Ant_EXECUTABLE)
|
30
vendor/github.com/apache/thrift/build/cmake/FindCabal.cmake
generated
vendored
Normal file
30
vendor/github.com/apache/thrift/build/cmake/FindCabal.cmake
generated
vendored
Normal file
|
@ -0,0 +1,30 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# Cabal_FOUND - system has Cabal
|
||||||
|
# Cabal - the Cabal executable
|
||||||
|
#
|
||||||
|
# It will search the environment variable CABAL_HOME if it is set
|
||||||
|
|
||||||
|
include(FindPackageHandleStandardArgs)
|
||||||
|
|
||||||
|
find_program(CABAL NAMES cabal PATHS $ENV{HOME}/.cabal/bin $ENV{CABAL_HOME}/bin)
|
||||||
|
find_package_handle_standard_args(CABAL DEFAULT_MSG CABAL)
|
||||||
|
mark_as_advanced(CABAL)
|
36
vendor/github.com/apache/thrift/build/cmake/FindGHC.cmake
generated
vendored
Normal file
36
vendor/github.com/apache/thrift/build/cmake/FindGHC.cmake
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# GHC_FOUND - system has GHC
|
||||||
|
# GHC - the GHC executable
|
||||||
|
# RUN_HASKELL_FOUND - system has runhaskell
|
||||||
|
# RUN_HASKELL - the runhaskell executable
|
||||||
|
#
|
||||||
|
# It will search the environment variable GHC_HOME if it is set
|
||||||
|
|
||||||
|
include(FindPackageHandleStandardArgs)
|
||||||
|
|
||||||
|
find_program(GHC NAMES ghc PATHS $ENV{GHC_HOME}/bin)
|
||||||
|
find_package_handle_standard_args(GHC DEFAULT_MSG GHC)
|
||||||
|
mark_as_advanced(GHC)
|
||||||
|
|
||||||
|
find_program(RUN_HASKELL NAMES runhaskell PATHS $ENV{GHC_HOME}/bin)
|
||||||
|
find_package_handle_standard_args(RUN_HASKELL DEFAULT_MSG RUN_HASKELL)
|
||||||
|
mark_as_advanced(RUN_HASKELL)
|
122
vendor/github.com/apache/thrift/build/cmake/FindGLIB.cmake
generated
vendored
Normal file
122
vendor/github.com/apache/thrift/build/cmake/FindGLIB.cmake
generated
vendored
Normal file
|
@ -0,0 +1,122 @@
|
||||||
|
# - Try to find Glib and its components (gio, gobject etc)
|
||||||
|
# Once done, this will define
|
||||||
|
#
|
||||||
|
# GLIB_FOUND - system has Glib
|
||||||
|
# GLIB_INCLUDE_DIRS - the Glib include directories
|
||||||
|
# GLIB_LIBRARIES - link these to use Glib
|
||||||
|
#
|
||||||
|
# Optionally, the COMPONENTS keyword can be passed to find_package()
|
||||||
|
# and Glib components can be looked for. Currently, the following
|
||||||
|
# components can be used, and they define the following variables if
|
||||||
|
# found:
|
||||||
|
#
|
||||||
|
# gio: GLIB_GIO_LIBRARIES
|
||||||
|
# gobject: GLIB_GOBJECT_LIBRARIES
|
||||||
|
# gmodule: GLIB_GMODULE_LIBRARIES
|
||||||
|
# gthread: GLIB_GTHREAD_LIBRARIES
|
||||||
|
#
|
||||||
|
# Note that the respective _INCLUDE_DIR variables are not set, since
|
||||||
|
# all headers are in the same directory as GLIB_INCLUDE_DIRS.
|
||||||
|
#
|
||||||
|
# Copyright (C) 2012 Raphael Kubo da Costa <rakuco@webkit.org>
|
||||||
|
#
|
||||||
|
# Redistribution and use in source and binary forms, with or without
|
||||||
|
# modification, are permitted provided that the following conditions
|
||||||
|
# are met:
|
||||||
|
# 1. Redistributions of source code must retain the above copyright
|
||||||
|
# notice, this list of conditions and the following disclaimer.
|
||||||
|
# 2. Redistributions in binary form must reproduce the above copyright
|
||||||
|
# notice, this list of conditions and the following disclaimer in the
|
||||||
|
# documentation and/or other materials provided with the distribution.
|
||||||
|
#
|
||||||
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER AND ITS CONTRIBUTORS ``AS
|
||||||
|
# IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
||||||
|
# THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
||||||
|
# PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR ITS
|
||||||
|
# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
||||||
|
# EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
||||||
|
# PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
|
||||||
|
# OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
||||||
|
# WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
|
||||||
|
# OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
|
||||||
|
# ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
find_package(PkgConfig)
|
||||||
|
pkg_check_modules(PC_GLIB QUIET glib-2.0)
|
||||||
|
|
||||||
|
find_library(GLIB_LIBRARIES
|
||||||
|
NAMES glib-2.0
|
||||||
|
HINTS ${PC_GLIB_LIBDIR}
|
||||||
|
${PC_GLIB_LIBRARY_DIRS}
|
||||||
|
)
|
||||||
|
|
||||||
|
# Files in glib's main include path may include glibconfig.h, which,
|
||||||
|
# for some odd reason, is normally in $LIBDIR/glib-2.0/include.
|
||||||
|
get_filename_component(_GLIB_LIBRARY_DIR ${GLIB_LIBRARIES} PATH)
|
||||||
|
find_path(GLIBCONFIG_INCLUDE_DIR
|
||||||
|
NAMES glibconfig.h
|
||||||
|
HINTS ${PC_LIBDIR} ${PC_LIBRARY_DIRS} ${_GLIB_LIBRARY_DIR}
|
||||||
|
${PC_GLIB_INCLUDEDIR} ${PC_GLIB_INCLUDE_DIRS}
|
||||||
|
PATH_SUFFIXES glib-2.0/include
|
||||||
|
)
|
||||||
|
|
||||||
|
find_path(GLIB_INCLUDE_DIR
|
||||||
|
NAMES glib.h
|
||||||
|
HINTS ${PC_GLIB_INCLUDEDIR}
|
||||||
|
${PC_GLIB_INCLUDE_DIRS}
|
||||||
|
PATH_SUFFIXES glib-2.0
|
||||||
|
)
|
||||||
|
|
||||||
|
set(GLIB_INCLUDE_DIRS ${GLIB_INCLUDE_DIR} ${GLIBCONFIG_INCLUDE_DIR})
|
||||||
|
|
||||||
|
if(GLIBCONFIG_INCLUDE_DIR)
|
||||||
|
# Version detection
|
||||||
|
file(READ "${GLIBCONFIG_INCLUDE_DIR}/glibconfig.h" GLIBCONFIG_H_CONTENTS)
|
||||||
|
string(REGEX MATCH "#define GLIB_MAJOR_VERSION ([0-9]+)" _dummy "${GLIBCONFIG_H_CONTENTS}")
|
||||||
|
set(GLIB_VERSION_MAJOR "${CMAKE_MATCH_1}")
|
||||||
|
string(REGEX MATCH "#define GLIB_MINOR_VERSION ([0-9]+)" _dummy "${GLIBCONFIG_H_CONTENTS}")
|
||||||
|
set(GLIB_VERSION_MINOR "${CMAKE_MATCH_1}")
|
||||||
|
string(REGEX MATCH "#define GLIB_MICRO_VERSION ([0-9]+)" _dummy "${GLIBCONFIG_H_CONTENTS}")
|
||||||
|
set(GLIB_VERSION_MICRO "${CMAKE_MATCH_1}")
|
||||||
|
set(GLIB_VERSION "${GLIB_VERSION_MAJOR}.${GLIB_VERSION_MINOR}.${GLIB_VERSION_MICRO}")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# Additional Glib components. We only look for libraries, as not all of them
|
||||||
|
# have corresponding headers and all headers are installed alongside the main
|
||||||
|
# glib ones.
|
||||||
|
foreach (_component ${GLIB_FIND_COMPONENTS})
|
||||||
|
if (${_component} STREQUAL "gio")
|
||||||
|
find_library(GLIB_GIO_LIBRARIES NAMES gio-2.0 HINTS ${_GLIB_LIBRARY_DIR})
|
||||||
|
set(ADDITIONAL_REQUIRED_VARS ${ADDITIONAL_REQUIRED_VARS} GLIB_GIO_LIBRARIES)
|
||||||
|
elseif (${_component} STREQUAL "gobject")
|
||||||
|
find_library(GLIB_GOBJECT_LIBRARIES NAMES gobject-2.0 HINTS ${_GLIB_LIBRARY_DIR})
|
||||||
|
set(ADDITIONAL_REQUIRED_VARS ${ADDITIONAL_REQUIRED_VARS} GLIB_GOBJECT_LIBRARIES)
|
||||||
|
elseif (${_component} STREQUAL "gmodule")
|
||||||
|
find_library(GLIB_GMODULE_LIBRARIES NAMES gmodule-2.0 HINTS ${_GLIB_LIBRARY_DIR})
|
||||||
|
set(ADDITIONAL_REQUIRED_VARS ${ADDITIONAL_REQUIRED_VARS} GLIB_GMODULE_LIBRARIES)
|
||||||
|
elseif (${_component} STREQUAL "gthread")
|
||||||
|
find_library(GLIB_GTHREAD_LIBRARIES NAMES gthread-2.0 HINTS ${_GLIB_LIBRARY_DIR})
|
||||||
|
set(ADDITIONAL_REQUIRED_VARS ${ADDITIONAL_REQUIRED_VARS} GLIB_GTHREAD_LIBRARIES)
|
||||||
|
elseif (${_component} STREQUAL "gio-unix")
|
||||||
|
# gio-unix is compiled as part of the gio library, but the include paths
|
||||||
|
# are separate from the shared glib ones. Since this is currently only used
|
||||||
|
# by WebKitGTK+ we don't go to extraordinary measures beyond pkg-config.
|
||||||
|
pkg_check_modules(GIO_UNIX QUIET gio-unix-2.0)
|
||||||
|
endif ()
|
||||||
|
endforeach ()
|
||||||
|
|
||||||
|
include(FindPackageHandleStandardArgs)
|
||||||
|
FIND_PACKAGE_HANDLE_STANDARD_ARGS(GLIB REQUIRED_VARS GLIB_INCLUDE_DIRS GLIB_LIBRARIES ${ADDITIONAL_REQUIRED_VARS}
|
||||||
|
VERSION_VAR GLIB_VERSION)
|
||||||
|
|
||||||
|
mark_as_advanced(
|
||||||
|
GLIBCONFIG_INCLUDE_DIR
|
||||||
|
GLIB_GIO_LIBRARIES
|
||||||
|
GLIB_GIO_UNIX_LIBRARIES
|
||||||
|
GLIB_GMODULE_LIBRARIES
|
||||||
|
GLIB_GOBJECT_LIBRARIES
|
||||||
|
GLIB_GTHREAD_LIBRARIES
|
||||||
|
GLIB_INCLUDE_DIR
|
||||||
|
GLIB_INCLUDE_DIRS
|
||||||
|
GLIB_LIBRARIES
|
||||||
|
)
|
30
vendor/github.com/apache/thrift/build/cmake/FindGradle.cmake
generated
vendored
Normal file
30
vendor/github.com/apache/thrift/build/cmake/FindGradle.cmake
generated
vendored
Normal file
|
@ -0,0 +1,30 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
# GRADLE_FOUND - system has Gradle
|
||||||
|
# GRADLE_EXECUTABLE - the Gradle executable
|
||||||
|
#
|
||||||
|
# It will search the environment variable ANT_HOME if it is set
|
||||||
|
|
||||||
|
include(FindPackageHandleStandardArgs)
|
||||||
|
|
||||||
|
find_program(GRADLE_EXECUTABLE NAMES gradle PATHS $ENV{GRADLE_HOME}/bin NO_CMAKE_FIND_ROOT_PATH)
|
||||||
|
find_package_handle_standard_args(Gradle DEFAULT_MSG GRADLE_EXECUTABLE)
|
||||||
|
mark_as_advanced(GRADLE_EXECUTABLE)
|
41
vendor/github.com/apache/thrift/build/cmake/FindInttypes.cmake
generated
vendored
Normal file
41
vendor/github.com/apache/thrift/build/cmake/FindInttypes.cmake
generated
vendored
Normal file
|
@ -0,0 +1,41 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
# find msinttypes on compilers that don't provide it, for example
|
||||||
|
# VS2010
|
||||||
|
|
||||||
|
# Usage:
|
||||||
|
# Provide INTTYPES_ROOT if you need it
|
||||||
|
# Result: INTTYPES_INCLUDE_DIRS, where to find inttypes.h
|
||||||
|
# Result: Inttypes_FOUND, If false, inttypes.h was not found
|
||||||
|
|
||||||
|
find_path(INTTYPES_INCLUDE_DIRS inttypes.h HINTS ${INTTYPES_ROOT})
|
||||||
|
if (INTTYPES_INCLUDE_DIRS)
|
||||||
|
set(Inttypes_FOUND TRUE)
|
||||||
|
else ()
|
||||||
|
set(Inttypes_FOUND FALSE)
|
||||||
|
if (Inttypes_FIND_REQUIRED)
|
||||||
|
message(FATAL_ERROR "Could NOT find inttypes.h")
|
||||||
|
endif ()
|
||||||
|
message(STATUS "inttypes.h NOT found")
|
||||||
|
endif ()
|
||||||
|
|
||||||
|
mark_as_advanced(
|
||||||
|
INTTYPES_INCLUDE_DIRS
|
||||||
|
)
|
45
vendor/github.com/apache/thrift/build/cmake/FindLibevent.cmake
generated
vendored
Normal file
45
vendor/github.com/apache/thrift/build/cmake/FindLibevent.cmake
generated
vendored
Normal file
|
@ -0,0 +1,45 @@
|
||||||
|
# find LibEvent
|
||||||
|
# an event notification library (http://libevent.org/)
|
||||||
|
#
|
||||||
|
# Usage:
|
||||||
|
# LIBEVENT_INCLUDE_DIRS, where to find LibEvent headers
|
||||||
|
# LIBEVENT_LIBRARIES, LibEvent libraries
|
||||||
|
# Libevent_FOUND, If false, do not try to use libevent
|
||||||
|
|
||||||
|
set(LIBEVENT_ROOT CACHE PATH "Root directory of libevent installation")
|
||||||
|
set(LibEvent_EXTRA_PREFIXES /usr/local /opt/local "$ENV{HOME}" ${LIBEVENT_ROOT})
|
||||||
|
foreach(prefix ${LibEvent_EXTRA_PREFIXES})
|
||||||
|
list(APPEND LibEvent_INCLUDE_PATHS "${prefix}/include")
|
||||||
|
list(APPEND LibEvent_LIBRARIES_PATHS "${prefix}/lib")
|
||||||
|
endforeach()
|
||||||
|
|
||||||
|
# Looking for "event.h" will find the Platform SDK include dir on windows
|
||||||
|
# so we also look for a peer header like evhttp.h to get the right path
|
||||||
|
find_path(LIBEVENT_INCLUDE_DIRS evhttp.h event.h PATHS ${LibEvent_INCLUDE_PATHS})
|
||||||
|
|
||||||
|
# "lib" prefix is needed on Windows in some cases
|
||||||
|
# newer versions of libevent use three libraries
|
||||||
|
find_library(LIBEVENT_LIBRARIES NAMES event event_core event_extra libevent PATHS ${LibEvent_LIBRARIES_PATHS})
|
||||||
|
|
||||||
|
if (LIBEVENT_LIBRARIES AND LIBEVENT_INCLUDE_DIRS)
|
||||||
|
set(Libevent_FOUND TRUE)
|
||||||
|
set(LIBEVENT_LIBRARIES ${LIBEVENT_LIBRARIES})
|
||||||
|
else ()
|
||||||
|
set(Libevent_FOUND FALSE)
|
||||||
|
endif ()
|
||||||
|
|
||||||
|
if (Libevent_FOUND)
|
||||||
|
if (NOT Libevent_FIND_QUIETLY)
|
||||||
|
message(STATUS "Found libevent: ${LIBEVENT_LIBRARIES}")
|
||||||
|
endif ()
|
||||||
|
else ()
|
||||||
|
if (LibEvent_FIND_REQUIRED)
|
||||||
|
message(FATAL_ERROR "Could NOT find libevent.")
|
||||||
|
endif ()
|
||||||
|
message(STATUS "libevent NOT found.")
|
||||||
|
endif ()
|
||||||
|
|
||||||
|
mark_as_advanced(
|
||||||
|
LIBEVENT_LIBRARIES
|
||||||
|
LIBEVENT_INCLUDE_DIRS
|
||||||
|
)
|
44
vendor/github.com/apache/thrift/build/cmake/NewPlatformDebug.cmake
generated
vendored
Normal file
44
vendor/github.com/apache/thrift/build/cmake/NewPlatformDebug.cmake
generated
vendored
Normal file
|
@ -0,0 +1,44 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
#
|
||||||
|
# For debugging new platforms, just to see what some environment flags are...
|
||||||
|
#
|
||||||
|
macro(SHOWFLAG flag)
|
||||||
|
message(STATUS "${flag} = ${${flag}}")
|
||||||
|
endmacro(SHOWFLAG)
|
||||||
|
|
||||||
|
set(NEWPLATFORMDEBUG ON)
|
||||||
|
|
||||||
|
if(NEWPLATFORMDEBUG)
|
||||||
|
SHOWFLAG("APPLE")
|
||||||
|
SHOWFLAG("BORLAND")
|
||||||
|
SHOWFLAG("CMAKE_C_COMPILER_ID")
|
||||||
|
SHOWFLAG("CMAKE_CXX_COMPILER_ID")
|
||||||
|
SHOWFLAG("CMAKE_COMPILER_IS_GNUCC")
|
||||||
|
SHOWFLAG("CMAKE_COMPILER_IS_GNUCXX")
|
||||||
|
SHOWFLAG("CYGWIN")
|
||||||
|
SHOWFLAG("MINGW")
|
||||||
|
SHOWFLAG("MSVC")
|
||||||
|
SHOWFLAG("MSVC_VERSION")
|
||||||
|
SHOWFLAG("MSYS")
|
||||||
|
SHOWFLAG("UNIX")
|
||||||
|
SHOWFLAG("WATCOM")
|
||||||
|
SHOWFLAG("WIN32")
|
||||||
|
endif(NEWPLATFORMDEBUG)
|
63
vendor/github.com/apache/thrift/build/cmake/README-MSYS2.md
generated
vendored
Normal file
63
vendor/github.com/apache/thrift/build/cmake/README-MSYS2.md
generated
vendored
Normal file
|
@ -0,0 +1,63 @@
|
||||||
|
<!---
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
-->
|
||||||
|
|
||||||
|
# Building thrift on Windows (MinGW64/MSYS2)
|
||||||
|
|
||||||
|
Thrift uses cmake to make it easier to build the project on multiple platforms, however to build a fully functional and production ready thrift on Windows requires a number of third party libraries to be obtained. Once third party libraries are ready, the right combination of options must be passed to cmake in order to generate the correct environment.
|
||||||
|
|
||||||
|
> Note: libevent and libevent-devel do not work with this toolchain as they do not properly detect mingw64 and expect some headers to exist that do not, so the non-blocking server is not currently built into this solution.
|
||||||
|
|
||||||
|
## MSYS2
|
||||||
|
|
||||||
|
Download and fully upgrade msys2 following the instructions at:
|
||||||
|
|
||||||
|
https://msys2.github.io/
|
||||||
|
|
||||||
|
Install the necessary toolchain items for C++:
|
||||||
|
|
||||||
|
$ pacman --needed -S bison flex make mingw-w64-x86_64-openssl \
|
||||||
|
mingw-w64-x86_64-boost mingw-w64-x86_64-cmake \
|
||||||
|
mingw-w64-x86_64-toolchain mingw-w64-x86_64-zlib
|
||||||
|
|
||||||
|
Update your msys2 bash path to include /mingw64/bin by adding a line to your ~/.bash_profiles using this command:
|
||||||
|
|
||||||
|
echo "export PATH=/mingw64/bin:\$PATH" >> ~/.bash_profile
|
||||||
|
|
||||||
|
After that, close your shell and open a new one.
|
||||||
|
|
||||||
|
Use cmake to create a MinGW makefile, out of tree (assumes you are in the top level of the thrift source tree):
|
||||||
|
|
||||||
|
mkdir ../thrift-build
|
||||||
|
cd ../thrift-build
|
||||||
|
cmake -G"MinGW Makefiles" -DCMAKE_MAKE_PROGRAM=/mingw64/bin/mingw32-make \
|
||||||
|
-DCMAKE_C_COMPILER=x86_64-w64-mingw32-gcc.exe \
|
||||||
|
-DCMAKE_CXX_COMPILER=x86_64-w64-mingw32-g++.exe \
|
||||||
|
-DWITH_BOOSTTHREADS=ON -DWITH_LIBEVENT=OFF \
|
||||||
|
-DWITH_SHARED_LIB=OFF -DWITH_STATIC_LIB=ON \
|
||||||
|
-DWITH_JAVA=OFF -DWITH_PYTHON=OFF -DWITH_PERL=OFF \
|
||||||
|
../thrift
|
||||||
|
|
||||||
|
Build thrift (inside thrift-build):
|
||||||
|
|
||||||
|
cmake --build .
|
||||||
|
|
||||||
|
Run the tests (inside thrift-build):
|
||||||
|
|
||||||
|
ctest
|
||||||
|
|
||||||
|
> If you run into issues, check Apache Jira THRIFT-4046 for patches relating to MinGW64/MSYS2 builds.
|
||||||
|
|
||||||
|
## Tested With
|
||||||
|
|
||||||
|
msys2 64-bit 2016-10-26 distribution
|
60
vendor/github.com/apache/thrift/build/cmake/README.md
generated
vendored
Normal file
60
vendor/github.com/apache/thrift/build/cmake/README.md
generated
vendored
Normal file
|
@ -0,0 +1,60 @@
|
||||||
|
# Apache Thrift - CMake build
|
||||||
|
|
||||||
|
## Goal
|
||||||
|
Extend Apache Thrift's *make cross* approach to the build system.
|
||||||
|
|
||||||
|
Due to growing the field of operating system support, a proper executable
|
||||||
|
and library detection mechanism running on as much platforms as possible
|
||||||
|
becomes required. The other aspect to simplify the release process and
|
||||||
|
package generation process.
|
||||||
|
|
||||||
|
As nice side benefit of CMake is the generation of development environment
|
||||||
|
specific soultion files. => No solution files within source tree.
|
||||||
|
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
just do this:
|
||||||
|
|
||||||
|
mkdir cmake-build && cd cmake-build
|
||||||
|
cmake ..
|
||||||
|
|
||||||
|
if you use a specific toolchain pass it to cmake, the same for options:
|
||||||
|
|
||||||
|
cmake -DCMAKE_TOOLCHAIN_FILE=../build/cmake/mingw32-toolchain.cmake ..
|
||||||
|
cmake -DCMAKE_C_COMPILER=clang-3.5 -DCMAKE_CXX_COMPILER=clang++-3.5 ..
|
||||||
|
cmake -DTHRIFT_COMPILER_HS=OFF ..
|
||||||
|
cmake -DWITH_ZLIB=ON ..
|
||||||
|
|
||||||
|
or on Windows
|
||||||
|
|
||||||
|
cmake -G "Visual Studio 12 2013 Win64" \
|
||||||
|
-DBOOST_ROOT=C:/3rdparty/boost_1_58_0 \
|
||||||
|
-DZLIB_ROOT=C:/3rdparty/zlib128-dll \
|
||||||
|
-DWITH_SHARED_LIB=off -DWITH_BOOSTTHREADS=ON ..
|
||||||
|
|
||||||
|
and open the development environment you like with the solution or do this:
|
||||||
|
|
||||||
|
make
|
||||||
|
make check
|
||||||
|
make cross
|
||||||
|
make dist
|
||||||
|
|
||||||
|
to generate an installer and distribution package do this:
|
||||||
|
|
||||||
|
cpack
|
||||||
|
|
||||||
|
## TODO
|
||||||
|
* git hash or tag based versioning depending on source state
|
||||||
|
* build tutorial
|
||||||
|
* build test
|
||||||
|
* with/without language lib/<lang>/
|
||||||
|
* enable/disable
|
||||||
|
* make cross
|
||||||
|
* make dist (create an alias to make package_source)
|
||||||
|
* make doc
|
||||||
|
* cpack (C++ and make dist only ?)
|
||||||
|
* thrift-compiler
|
||||||
|
* libthrift
|
||||||
|
* tutorial
|
||||||
|
* test
|
||||||
|
* merge into /README.md
|
105
vendor/github.com/apache/thrift/build/cmake/ThriftMacros.cmake
generated
vendored
Normal file
105
vendor/github.com/apache/thrift/build/cmake/ThriftMacros.cmake
generated
vendored
Normal file
|
@ -0,0 +1,105 @@
|
||||||
|
#
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
#
|
||||||
|
|
||||||
|
|
||||||
|
set(CMAKE_DEBUG_POSTFIX "d" CACHE STRING "Set debug library postfix" FORCE)
|
||||||
|
|
||||||
|
|
||||||
|
macro(ADD_LIBRARY_THRIFT name)
|
||||||
|
|
||||||
|
if(WITH_SHARED_LIB)
|
||||||
|
add_library(${name} SHARED ${ARGN})
|
||||||
|
set_target_properties(${name} PROPERTIES
|
||||||
|
OUTPUT_NAME ${name}
|
||||||
|
VERSION ${thrift_VERSION}
|
||||||
|
SOVERSION ${thrift_VERSION} )
|
||||||
|
#set_target_properties(${name} PROPERTIES PUBLIC_HEADER "${thriftcpp_HEADERS}")
|
||||||
|
install(TARGETS ${name}
|
||||||
|
RUNTIME DESTINATION "${BIN_INSTALL_DIR}"
|
||||||
|
LIBRARY DESTINATION "${LIB_INSTALL_DIR}"
|
||||||
|
ARCHIVE DESTINATION "${LIB_INSTALL_DIR}"
|
||||||
|
PUBLIC_HEADER DESTINATION "${INCLUDE_INSTALL_DIR}")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if(WITH_STATIC_LIB)
|
||||||
|
add_library(${name}_static STATIC ${ARGN})
|
||||||
|
set_target_properties(${name}_static PROPERTIES
|
||||||
|
OUTPUT_NAME ${name}${STATIC_POSTFIX}
|
||||||
|
VERSION ${thrift_VERSION}
|
||||||
|
SOVERSION ${thrift_VERSION} )
|
||||||
|
install(TARGETS ${name}_static
|
||||||
|
RUNTIME DESTINATION "${BIN_INSTALL_DIR}"
|
||||||
|
LIBRARY DESTINATION "${LIB_INSTALL_DIR}"
|
||||||
|
ARCHIVE DESTINATION "${LIB_INSTALL_DIR}"
|
||||||
|
PUBLIC_HEADER DESTINATION "${INCLUDE_INSTALL_DIR}")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
endmacro(ADD_LIBRARY_THRIFT)
|
||||||
|
|
||||||
|
|
||||||
|
macro(TARGET_INCLUDE_DIRECTORIES_THRIFT name)
|
||||||
|
|
||||||
|
if(WITH_SHARED_LIB)
|
||||||
|
target_include_directories(${name} ${ARGN})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if(WITH_STATIC_LIB)
|
||||||
|
target_include_directories(${name}_static ${ARGN})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
endmacro(TARGET_INCLUDE_DIRECTORIES_THRIFT)
|
||||||
|
|
||||||
|
|
||||||
|
macro(TARGET_LINK_LIBRARIES_THRIFT name)
|
||||||
|
|
||||||
|
if(WITH_SHARED_LIB)
|
||||||
|
target_link_libraries(${name} ${ARGN})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if(WITH_STATIC_LIB)
|
||||||
|
target_link_libraries(${name}_static ${ARGN})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
endmacro(TARGET_LINK_LIBRARIES_THRIFT)
|
||||||
|
|
||||||
|
|
||||||
|
macro(LINK_AGAINST_THRIFT_LIBRARY target libname)
|
||||||
|
|
||||||
|
if (WITH_SHARED_LIB)
|
||||||
|
target_link_libraries(${target} ${libname})
|
||||||
|
elseif (WITH_STATIC_LIB)
|
||||||
|
target_link_libraries(${target} ${libname}_static)
|
||||||
|
else()
|
||||||
|
message(FATAL "Not linking with shared or static libraries?")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
endmacro(LINK_AGAINST_THRIFT_LIBRARY)
|
||||||
|
|
||||||
|
|
||||||
|
macro(TARGET_LINK_LIBRARIES_THRIFT_AGAINST_THRIFT_LIBRARY target libname)
|
||||||
|
|
||||||
|
if(WITH_SHARED_LIB)
|
||||||
|
target_link_libraries(${target} ${ARGN} ${libname})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
if(WITH_STATIC_LIB)
|
||||||
|
target_link_libraries(${target}_static ${ARGN} ${libname}_static)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
endmacro(TARGET_LINK_LIBRARIES_THRIFT_AGAINST_THRIFT_LIBRARY)
|
26
vendor/github.com/apache/thrift/build/cmake/android-toolchain.cmake
generated
vendored
Normal file
26
vendor/github.com/apache/thrift/build/cmake/android-toolchain.cmake
generated
vendored
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
set(ANDROID_NDK "/opt/android-ndk" CACHE)
|
||||||
|
set(ANDROID_PLATFORM "android-15" CACHE)
|
||||||
|
set(ANDROID_ARCH "arch-arm" CACHE)
|
||||||
|
set(ANDROID_TOOL_ARCH "android-arm" CACHE)
|
||||||
|
set(ANDROID_CPU "armeabi-v7a" CACHE)
|
||||||
|
set(ANDROID_GCC_VERSION 4.9 CACHE)
|
||||||
|
set(HOST_ARCH linux-x86_64 CACHE)
|
||||||
|
|
||||||
|
set(CMAKE_SYSTEM_NAME Android)
|
||||||
|
set(ANDROID_SYSROOT "${ANDROID_NDK}/platforms/${ANDROID_PLATFORM}/${ANDROID_ARCH}")
|
||||||
|
set(ANDROID_TRIPLET arm-linux-androideabi)
|
||||||
|
set(ANDROID_STL "${ANDROID_NDK}/sources/cxx-stl/gnu-libstd++/${ANDROID_GCC_VERSION}")
|
||||||
|
|
||||||
|
set(_COMPILER_ROOT ${ANDROID_NDK}/prebuilt/${ANDROID_TRIPLET}-${ANDROID_GCC_VERSION}/prebuilt/${HOST_ARCH})
|
||||||
|
set(CMAKE_C_COMPILER ${_COMPILER_ROOT}/bin/${ANDROID_TRIPLET}-gcc)
|
||||||
|
set(CMAKE_CXCX_COMPILER ${_COMPILER_ROOT}/bin/${ANDROID_TRIPLET}-g++)
|
||||||
|
|
||||||
|
include_directories(
|
||||||
|
${ANDROID_STL}/include
|
||||||
|
${ANDROID_STL}/libs/${ANDROID_CPU}/include)
|
||||||
|
|
||||||
|
set(CMAKE_FIND_ROOT_PATH ${ANDROID_SYSROOT})
|
||||||
|
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER)
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)
|
154
vendor/github.com/apache/thrift/build/cmake/config.h.in
generated
vendored
Normal file
154
vendor/github.com/apache/thrift/build/cmake/config.h.in
generated
vendored
Normal file
|
@ -0,0 +1,154 @@
|
||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing,
|
||||||
|
* software distributed under the License is distributed on an
|
||||||
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
* KIND, either express or implied. See the License for the
|
||||||
|
* specific language governing permissions and limitations
|
||||||
|
* under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* config.h generated by CMake from config.h.in */
|
||||||
|
|
||||||
|
#ifndef CONFIG_H
|
||||||
|
#define CONFIG_H
|
||||||
|
|
||||||
|
|
||||||
|
/* Name of package */
|
||||||
|
#cmakedefine PACKAGE "${PACKAGE}"
|
||||||
|
|
||||||
|
/* Define to the address where bug reports for this package should be sent. */
|
||||||
|
#cmakedefine PACKAGE_BUGREPORT "${PACKAGE_BUGREPORT}"
|
||||||
|
|
||||||
|
/* Define to the full name of this package. */
|
||||||
|
#cmakedefine PACKAGE_NAME "${PACKAGE_NAME}"
|
||||||
|
|
||||||
|
/* Define to the one symbol short name of this package. */
|
||||||
|
#cmakedefine PACKAGE_TARNAME "${PACKAGE_TARNAME}"
|
||||||
|
|
||||||
|
/* Define to the home page for this package. */
|
||||||
|
#cmakedefine PACKAGE_URL "${PACKAGE_URL}"
|
||||||
|
|
||||||
|
/* Define to the version of this package. */
|
||||||
|
#define PACKAGE_VERSION "${PACKAGE_VERSION}"
|
||||||
|
|
||||||
|
/* Define to the full name and version of this package. */
|
||||||
|
#define PACKAGE_STRING "${PACKAGE_STRING}"
|
||||||
|
|
||||||
|
/************************** DEFINES *************************/
|
||||||
|
|
||||||
|
/* Define if the AI_ADDRCONFIG symbol is unavailable */
|
||||||
|
#cmakedefine AI_ADDRCONFIG 0
|
||||||
|
|
||||||
|
/* Possible value for SIGNED_RIGHT_SHIFT_IS */
|
||||||
|
/* TODO: This is just set to 1 for the moment
|
||||||
|
port the macro aclocal/ax_signed_right_shift.m4 to CMake to make this work */
|
||||||
|
#define ARITHMETIC_RIGHT_SHIFT 1
|
||||||
|
|
||||||
|
/* Indicates the effect of the right shift operator on negative signed
|
||||||
|
integers */
|
||||||
|
/* TODO: This is just set to 1 for the moment */
|
||||||
|
#define SIGNED_RIGHT_SHIFT_IS 1
|
||||||
|
|
||||||
|
/* Use *.h extension for parser header file */
|
||||||
|
/* TODO: This might now be necessary anymore as it is set only for automake < 1.11
|
||||||
|
see: aclocal/ac_prog_bison.m4 */
|
||||||
|
#cmakedefine BISON_USE_PARSER_H_EXTENSION 1
|
||||||
|
|
||||||
|
/* replaces POSIX pthread by boost::thread */
|
||||||
|
#cmakedefine USE_BOOST_THREAD 1
|
||||||
|
|
||||||
|
/* replaces POSIX pthread by std::thread */
|
||||||
|
#cmakedefine USE_STD_THREAD 1
|
||||||
|
|
||||||
|
/* Define to 1 if strerror_r returns char *. */
|
||||||
|
#cmakedefine STRERROR_R_CHAR_P 1
|
||||||
|
|
||||||
|
|
||||||
|
/************************** HEADER FILES *************************/
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <arpa/inet.h> header file. */
|
||||||
|
#cmakedefine HAVE_ARPA_INET_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <fcntl.h> header file. */
|
||||||
|
#cmakedefine HAVE_FCNTL_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <inttypes.h> header file. */
|
||||||
|
#cmakedefine HAVE_INTTYPES_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <netdb.h> header file. */
|
||||||
|
#cmakedefine HAVE_NETDB_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <netinet/in.h> header file. */
|
||||||
|
#cmakedefine HAVE_NETINET_IN_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <stdint.h> header file. */
|
||||||
|
#cmakedefine HAVE_STDINT_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <unistd.h> header file. */
|
||||||
|
#cmakedefine HAVE_UNISTD_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <pthread.h> header file. */
|
||||||
|
#cmakedefine HAVE_PTHREAD_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/time.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_TIME_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/param.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_PARAM_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/resource.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_RESOURCE_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/socket.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_SOCKET_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/stat.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_STAT_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/un.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_UN_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/poll.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_POLL_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sys/select.h> header file. */
|
||||||
|
#cmakedefine HAVE_SYS_SELECT_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <sched.h> header file. */
|
||||||
|
#cmakedefine HAVE_SCHED_H 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the <strings.h> header file. */
|
||||||
|
#define HAVE_STRINGS_H 1
|
||||||
|
|
||||||
|
/*************************** FUNCTIONS ***************************/
|
||||||
|
|
||||||
|
/* Define to 1 if you have the `gethostbyname' function. */
|
||||||
|
#cmakedefine HAVE_GETHOSTBYNAME 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the `gethostbyname_r' function. */
|
||||||
|
#cmakedefine HAVE_GETHOSTBYNAME_R 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the `strerror_r' function. */
|
||||||
|
#cmakedefine HAVE_STRERROR_R 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the `sched_get_priority_max' function. */
|
||||||
|
#cmakedefine HAVE_SCHED_GET_PRIORITY_MAX 1
|
||||||
|
|
||||||
|
/* Define to 1 if you have the `sched_get_priority_min' function. */
|
||||||
|
#cmakedefine HAVE_SCHED_GET_PRIORITY_MIN 1
|
||||||
|
|
||||||
|
|
||||||
|
/* Define to 1 if strerror_r returns char *. */
|
||||||
|
#cmakedefine STRERROR_R_CHAR_P 1
|
||||||
|
|
||||||
|
#endif
|
24
vendor/github.com/apache/thrift/build/cmake/mingw32-toolchain.cmake
generated
vendored
Normal file
24
vendor/github.com/apache/thrift/build/cmake/mingw32-toolchain.cmake
generated
vendored
Normal file
|
@ -0,0 +1,24 @@
|
||||||
|
# CMake mingw32 cross compile toolchain file
|
||||||
|
|
||||||
|
# the name of the target operating system
|
||||||
|
SET(CMAKE_SYSTEM_NAME Windows)
|
||||||
|
|
||||||
|
# which compilers to use for C and C++
|
||||||
|
SET(CMAKE_C_COMPILER i586-mingw32msvc-gcc)
|
||||||
|
SET(CMAKE_CXX_COMPILER i586-mingw32msvc-g++)
|
||||||
|
SET(CMAKE_RC_COMPILER i586-mingw32msvc-windres)
|
||||||
|
|
||||||
|
# here is the target environment located
|
||||||
|
SET(CMAKE_FIND_ROOT_PATH /usr/i586-mingw32msvc)
|
||||||
|
|
||||||
|
# adjust the default behaviour of the FIND_XXX() commands:
|
||||||
|
# search headers and libraries in the target environment, search
|
||||||
|
# programs in the host environment
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER)
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)
|
||||||
|
set(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)
|
||||||
|
|
||||||
|
set(BUILD_SHARED_LIBS OFF)
|
||||||
|
SET(CMAKE_EXE_LINKER_FLAGS "-static")
|
||||||
|
set(CMAKE_C_FLAGS ${CMAKE_C_FLAGS} "-static-libgcc")
|
||||||
|
set(CMAKE_CXX_FLAGS ${CMAKE_CXX_FLAGS} "-static-libstdc++")
|
168
vendor/github.com/apache/thrift/build/docker/README.md
generated
vendored
Normal file
168
vendor/github.com/apache/thrift/build/docker/README.md
generated
vendored
Normal file
|
@ -0,0 +1,168 @@
|
||||||
|
# Docker Integration #
|
||||||
|
|
||||||
|
Due to the large number of language requirements to build Apache Thrift, docker containers are used to build and test the project on a variety of platforms to provide maximum test coverage.
|
||||||
|
|
||||||
|
## Travis CI Integration ##
|
||||||
|
|
||||||
|
The Travis CI scripts use the following environment variables and logic to determine their behavior.
|
||||||
|
|
||||||
|
### Environment Variables ###
|
||||||
|
|
||||||
|
| Variable | Default | Usage |
|
||||||
|
| -------- | ----- | ------- |
|
||||||
|
| `DISTRO` | `ubuntu-xenial` | Set by various build jobs in `.travis.yml` to run builds in different containers. Not intended to be set externally.|
|
||||||
|
| `DOCKER_REPO` | `thrift/thrift-build` | The name of the Docker Hub repository to obtain and store docker images. |
|
||||||
|
| `DOCKER_USER` | `<none>` | The Docker Hub account name containing the repository. |
|
||||||
|
| `DOCKER_PASS` | `<none>` | The Docker Hub account password to use when pushing new tags. |
|
||||||
|
|
||||||
|
For example, the default docker image that is used in builds if no overrides are specified would be: `thrift/thrift-build:ubuntu-xenial`
|
||||||
|
|
||||||
|
### Forks ###
|
||||||
|
|
||||||
|
If you have forked the Apache Thrift repository and you would like to use your own Docker Hub account to store thrift build images, you can use the Travis CI web interface to set the `DOCKER_USER`, `DOCKER_PASS`, and `DOCKER_REPO` variables in a secure manner. Your fork builds will then pull, push, and tag the docker images in your account.
|
||||||
|
|
||||||
|
### Logic ###
|
||||||
|
|
||||||
|
The Travis CI build runs in two phases - first the docker images are rebuilt for each of the three supported containers if they do not match the Dockerfile that was used to build the most recent tag. If a `DOCKER_PASS` environment variable is specified, the docker stage builds will attempt to log into Docker Hub and push the resulting tags.
|
||||||
|
|
||||||
|
## Supported Containers ##
|
||||||
|
|
||||||
|
The Travis CI (continuous integration) builds use the Ubuntu Trusty, Xenial, and Artful images to maximize language level coverage.
|
||||||
|
|
||||||
|
### Ubuntu ###
|
||||||
|
|
||||||
|
* trusty (legacy)
|
||||||
|
* xenial (stable)
|
||||||
|
* artful (latest)
|
||||||
|
|
||||||
|
## Unsupported Containers ##
|
||||||
|
|
||||||
|
These containers may be in various states, and may not build everything.
|
||||||
|
|
||||||
|
### CentOS ###
|
||||||
|
* 7.3
|
||||||
|
* make check in lib/py may hang in test_sslsocket - root cause unknown
|
||||||
|
|
||||||
|
### Debian ###
|
||||||
|
|
||||||
|
* jessie
|
||||||
|
* stretch
|
||||||
|
* make check in lib/cpp fails due to https://svn.boost.org/trac10/ticket/12507
|
||||||
|
|
||||||
|
## Building like Travis CI does, locally ##
|
||||||
|
|
||||||
|
We recommend you build locally the same way Travis CI does, so that when you submit your pull request you will run into fewer surprises. To make it a little easier, put the following into your `~/.bash_aliases` file:
|
||||||
|
|
||||||
|
# Kill all running containers.
|
||||||
|
alias dockerkillall='docker kill $(docker ps -q)'
|
||||||
|
|
||||||
|
# Delete all stopped containers.
|
||||||
|
alias dockercleanc='printf "\n>>> Deleting stopped containers\n\n" && docker rm $(docker ps -a -q)'
|
||||||
|
|
||||||
|
# Delete all untagged images.
|
||||||
|
alias dockercleani='printf "\n>>> Deleting untagged images\n\n" && docker rmi $(docker images -q -f dangling=true)'
|
||||||
|
|
||||||
|
# Delete all stopped containers and untagged images.
|
||||||
|
alias dockerclean='dockercleanc || true && dockercleani'
|
||||||
|
|
||||||
|
# Build a thrift docker image (run from top level of git repo): argument #1 is image type (ubuntu, centos, etc).
|
||||||
|
function dockerbuild
|
||||||
|
{
|
||||||
|
docker build -t $1 build/docker/$1
|
||||||
|
}
|
||||||
|
|
||||||
|
# Run a thrift docker image: argument #1 is image type (ubuntu, centos, etc).
|
||||||
|
function dockerrun
|
||||||
|
{
|
||||||
|
docker run -v $(pwd):/thrift/src -it $1 /bin/bash
|
||||||
|
}
|
||||||
|
|
||||||
|
To pull down the current image being used to build (the same way Travis CI does it) - if it is out of date in any way it will build a new one for you:
|
||||||
|
|
||||||
|
thrift$ DOCKER_REPO=thrift/thrift-build DISTRO=ubuntu-xenial build/docker/refresh.sh
|
||||||
|
|
||||||
|
To run all unit tests (just like Travis CI):
|
||||||
|
|
||||||
|
thrift$ dockerrun ubuntu-xenial
|
||||||
|
root@8caf56b0ce7b:/thrift/src# build/docker/scripts/autotools.sh
|
||||||
|
|
||||||
|
To run the cross tests (just like Travis CI):
|
||||||
|
|
||||||
|
thrift$ dockerrun ubuntu-xenial
|
||||||
|
root@8caf56b0ce7b:/thrift/src# build/docker/scripts/cross-test.sh
|
||||||
|
|
||||||
|
When you are done, you want to clean up occasionally so that docker isn't using lots of extra disk space:
|
||||||
|
|
||||||
|
thrift$ dockerclean
|
||||||
|
|
||||||
|
You need to run the docker commands from the root of the git repository for them to work.
|
||||||
|
|
||||||
|
When you are done in the root docker shell you can `exit` to go back to your user host shell. Once the unit tests and cross test passes locally, then submit he changes, and squash the pull request to one commit to make it easier to merge. Thanks. I am going to update the docker README.md with this information so others can leverage it too. Now you are building like Travis CI does!
|
||||||
|
|
||||||
|
## Raw Commands for Building with Docker ##
|
||||||
|
|
||||||
|
If you do not want to use the same scripts Travis CI does, you can do it manually:
|
||||||
|
|
||||||
|
Build the image:
|
||||||
|
|
||||||
|
thrift$ docker build -t thrift build/docker/ubuntu-xenial
|
||||||
|
|
||||||
|
Open a command prompt in the image:
|
||||||
|
|
||||||
|
thrift$ docker run -v $(pwd):/thrift/src -it thrift /bin/bash
|
||||||
|
|
||||||
|
## Core Tool Versions per Dockerfile ##
|
||||||
|
|
||||||
|
Last updated: October 1, 2017
|
||||||
|
|
||||||
|
| Tool | ubuntu-trusty | ubuntu-xenial | ubuntu-artful | Notes |
|
||||||
|
| :-------- | :------------ | :------------ | :------------ | :---- |
|
||||||
|
| ant | 1.9.3 | 1.9.6 | 1.9.9 | |
|
||||||
|
| autoconf | 2.69 | 2.69 | 2.69 | |
|
||||||
|
| automake | 1.14.1 | 1.15 | 1.15 | |
|
||||||
|
| bison | 3.0.2 | 3.0.4 | 3.0.4 | |
|
||||||
|
| boost | 1.54.0 | 1.58.0 | 1.63.0 | artful: stock boost 1.62.0 has problems running unit tests |
|
||||||
|
| cmake | 3.2.2 | 3.5.1 | 3.9.1 | |
|
||||||
|
| cppcheck | 1.61 | 1.72 | 1.80 | |
|
||||||
|
| flex | 2.5.35 | 2.6.0 | 2.6.1 | |
|
||||||
|
| glibc | 2.19 | 2.23 | 2.26 | |
|
||||||
|
| libevent | 2.0.21 | 2.0.21 | 2.1 | |
|
||||||
|
| libstdc++ | 4.8.4 | 5.4.0 | 7.2.0 | |
|
||||||
|
| make | 3.81 | 4.1 | 4.1 | |
|
||||||
|
| openssl | 1.0.1f | 1.0.2g | 1.0.2g | |
|
||||||
|
| qt5 | 5.2.1 | 5.5.1 | 5.9.1 | |
|
||||||
|
|
||||||
|
## Compiler/Language Versions per Dockerfile ##
|
||||||
|
|
||||||
|
Last updated: October 1, 2017
|
||||||
|
|
||||||
|
| Language | ubuntu-trusty | ubuntu-xenial | ubuntu-artful | Notes |
|
||||||
|
| :-------- | :------------ | :------------ | :------------ | :---- |
|
||||||
|
| as3 | | | | Not in CI |
|
||||||
|
| C++ gcc | 4.8.4 | 5.4.0 | 7.2.0 | |
|
||||||
|
| C++ clang | 3.4 | 3.8 | 4.0 | |
|
||||||
|
| C# (mono) | 3.2.8.0 | 4.2.1 | 4.6.2.7 | |
|
||||||
|
| c_glib | 2.40.2 | 2.48.2 | 2.54.0 | |
|
||||||
|
| cocoa | | | | Not in CI |
|
||||||
|
| d | 2.070.2 | 2.073.2 | 2.076.0 | |
|
||||||
|
| dart | 1.20.1 | 1.24.2 | | artful: apt repo not compatible with apt 1.4? |
|
||||||
|
| delphi | | | | Not in CI |
|
||||||
|
| dotnet | | 2.0.0 | | Needs to be added to artful |
|
||||||
|
| erlang | R16B03 | 18.3 | 20.0.4 | |
|
||||||
|
| go | 1.2.1 | 1.6.2 | 1.8.3 | |
|
||||||
|
| haskell | 7.6.3 | 7.10.3 | 8.0.2 | |
|
||||||
|
| haxe | | 3.2.1 | 3.4.2 | disabled in trusty builds - cores on install v3.0.0, disabled in artful builds - see THRIFT-4352 |
|
||||||
|
| java | 1.7.0_151 | 1.8.0_131 | 1.8.0_144 | |
|
||||||
|
| js | | | | Unsure how to look for version info? |
|
||||||
|
| lua | 5.1.5 | 5.2.4 | 5.3.3 | |
|
||||||
|
| nodejs | | 4.2.6 | 8.8.1 | trusty has node.js 0.10.0 which is too old |
|
||||||
|
| ocaml | | 4.02.3 | 4.04.0 | |
|
||||||
|
| perl | 5.18.2 | 5.22.1 | 5.26.0 | |
|
||||||
|
| php | 5.5.9 | 7.0.22 | 7.1.8 | |
|
||||||
|
| python | 2.7.6 | 2.7.12 | 2.7.14 | |
|
||||||
|
| python3 | 3.4.3 | 3.5.2 | 3.6.3 | |
|
||||||
|
| ruby | 1.9.3p484 | 2.3.1p112 | 2.3.3p222 | |
|
||||||
|
| rust | 1.15.1 | 1.15.1 | 1.18.0 | |
|
||||||
|
| smalltalk | | | | Not in CI |
|
||||||
|
| swift | | | | Not in CI |
|
||||||
|
|
59
vendor/github.com/apache/thrift/build/docker/Vagrantfile
generated
vendored
Normal file
59
vendor/github.com/apache/thrift/build/docker/Vagrantfile
generated
vendored
Normal file
|
@ -0,0 +1,59 @@
|
||||||
|
# -*- mode: ruby -*-
|
||||||
|
# vi: set ft=ruby :
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
# Base system bootstrap script
|
||||||
|
$bootstrap_script = <<__BOOTSTRAP__
|
||||||
|
echo "Provisioning defaults"
|
||||||
|
|
||||||
|
sudo apt-get update -y
|
||||||
|
sudo apt-get upgrade -y
|
||||||
|
|
||||||
|
# Install default packages
|
||||||
|
sudo apt-get install -y build-essential curl git
|
||||||
|
|
||||||
|
# Install latest Docker version
|
||||||
|
sudo curl -sSL https://get.docker.io/gpg | sudo apt-key add -
|
||||||
|
sudo echo "deb http://get.docker.io/ubuntu docker main" > /etc/apt/sources.list.d/docker.list
|
||||||
|
sudo apt-get update -y
|
||||||
|
sudo apt-get install -y linux-image-extra-`uname -r` aufs-tools
|
||||||
|
sudo apt-get install -y lxc-docker
|
||||||
|
|
||||||
|
echo "Finished provisioning defaults"
|
||||||
|
__BOOTSTRAP__
|
||||||
|
|
||||||
|
Vagrant.configure("2") do |config|
|
||||||
|
config.vm.box = "trusty64"
|
||||||
|
config.vm.box_url = "https://cloud-images.ubuntu.com/vagrant/trusty/current/trusty-server-cloudimg-amd64-vagrant-disk1.box"
|
||||||
|
config.ssh.forward_agent = true
|
||||||
|
|
||||||
|
config.vm.provider :virtualbox do |vbox|
|
||||||
|
vbox.customize ["modifyvm", :id, "--memory", "1024"]
|
||||||
|
vbox.customize ["modifyvm", :id, "--cpus", "2"]
|
||||||
|
end
|
||||||
|
|
||||||
|
# Setup the default bootstrap script for our ubuntu base box image
|
||||||
|
config.vm.provision "shell", inline: $bootstrap_script
|
||||||
|
|
||||||
|
# Setup the custom docker image from our Ubuntu Dockerfile
|
||||||
|
config.vm.provision "docker" do |d|
|
||||||
|
d.build_image "/vagrant/ubuntu", args: "-t thrift"
|
||||||
|
end
|
||||||
|
|
||||||
|
# Setup the custom docker image from our Centos Dockerfile
|
||||||
|
#config.vm.provision "docker" do |d|
|
||||||
|
# d.build_image "/vagrant/centos", args: "-t thrift-centos"
|
||||||
|
#end
|
||||||
|
|
||||||
|
end
|
199
vendor/github.com/apache/thrift/build/docker/centos-7.3/Dockerfile
generated
vendored
Normal file
199
vendor/github.com/apache/thrift/build/docker/centos-7.3/Dockerfile
generated
vendored
Normal file
|
@ -0,0 +1,199 @@
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
# Apache Thrift Docker build environment for CentOS
|
||||||
|
#
|
||||||
|
# Known missing client libraries:
|
||||||
|
# - dotnet (will update to 2.0.0 separately)
|
||||||
|
# - haxe (not in centos)
|
||||||
|
|
||||||
|
FROM centos:7.3.1611
|
||||||
|
MAINTAINER Apache Thrift <dev@thrift.apache.org>
|
||||||
|
|
||||||
|
RUN yum install -y epel-release
|
||||||
|
|
||||||
|
# General dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
autoconf \
|
||||||
|
bison \
|
||||||
|
bison-devel \
|
||||||
|
clang \
|
||||||
|
clang-analyzer \
|
||||||
|
cmake3 \
|
||||||
|
curl \
|
||||||
|
flex \
|
||||||
|
gcc \
|
||||||
|
gcc-c++ \
|
||||||
|
gdb \
|
||||||
|
git \
|
||||||
|
libtool \
|
||||||
|
m4 \
|
||||||
|
make \
|
||||||
|
tar \
|
||||||
|
unzip \
|
||||||
|
valgrind \
|
||||||
|
wget && \
|
||||||
|
ln -s /usr/bin/cmake3 /usr/bin/cmake && \
|
||||||
|
ln -s /usr/bin/cpack3 /usr/bin/cpack && \
|
||||||
|
ln -s /usr/bin/ctest3 /usr/bin/ctest
|
||||||
|
|
||||||
|
# C++ dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
boost-devel-static \
|
||||||
|
zlib-devel \
|
||||||
|
openssl-devel \
|
||||||
|
libevent-devel && \
|
||||||
|
cd /usr/lib64 && \
|
||||||
|
ln -s libboost_thread-mt.a libboost_thread.a
|
||||||
|
|
||||||
|
# C# Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
mono-core \
|
||||||
|
mono-devel \
|
||||||
|
mono-web-devel \
|
||||||
|
mono-extras
|
||||||
|
|
||||||
|
# D Dependencies
|
||||||
|
RUN yum install -y http://downloads.dlang.org/releases/2.x/2.076.0/dmd-2.076.0-0.fedora.x86_64.rpm xdg-utils
|
||||||
|
RUN curl -sSL https://github.com/D-Programming-Deimos/openssl/archive/master.tar.gz| tar xz && \
|
||||||
|
curl -sSL https://github.com/D-Programming-Deimos/libevent/archive/master.tar.gz| tar xz && \
|
||||||
|
mkdir -p /usr/include/dmd/druntime/import/deimos /usr/include/dmd/druntime/import/C && \
|
||||||
|
mv libevent-master/deimos/* openssl-master/deimos/* /usr/include/dmd/druntime/import/deimos/ && \
|
||||||
|
mv libevent-master/C/* openssl-master/C/* /usr/include/dmd/druntime/import/C/ && \
|
||||||
|
rm -rf libevent-master openssl-master
|
||||||
|
|
||||||
|
# Dart
|
||||||
|
RUN cd /usr/local && \
|
||||||
|
wget -q https://storage.googleapis.com/dart-archive/channels/stable/release/1.24.2/sdk/dartsdk-linux-x64-release.zip && \
|
||||||
|
unzip -q dartsdk-linux-x64-release.zip && \
|
||||||
|
rm dartsdk-linux-x64-release.zip
|
||||||
|
ENV PATH /usr/local/dart-sdk/bin:$PATH
|
||||||
|
|
||||||
|
# Erlang Dependencies
|
||||||
|
RUN curl -sSL http://packages.erlang-solutions.com/rpm/centos/erlang_solutions.repo -o /etc/yum.repos.d/erlang_solutions.repo && \
|
||||||
|
yum install -y \
|
||||||
|
erlang-kernel \
|
||||||
|
erlang-erts \
|
||||||
|
erlang-stdlib \
|
||||||
|
erlang-eunit \
|
||||||
|
erlang-rebar \
|
||||||
|
erlang-tools
|
||||||
|
|
||||||
|
# GLibC Dependencies
|
||||||
|
RUN yum install -y glib2-devel
|
||||||
|
|
||||||
|
# Go Dependencies
|
||||||
|
RUN curl -sSL https://storage.googleapis.com/golang/go1.9.linux-amd64.tar.gz | tar -C /usr/local/ -xz
|
||||||
|
ENV PATH /usr/local/go/bin:$PATH
|
||||||
|
|
||||||
|
# Haskell Dependencies
|
||||||
|
RUN yum -y install haskell-platform
|
||||||
|
|
||||||
|
# Haxe Dependencies
|
||||||
|
# Not in debian/stretch
|
||||||
|
|
||||||
|
# Java Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
ant \
|
||||||
|
junit \
|
||||||
|
ant-junit \
|
||||||
|
java-1.8.0-openjdk-devel
|
||||||
|
|
||||||
|
# Lua Dependencies
|
||||||
|
# Lua in epel is too old (5.1.4, need 5.2) so we get the latest
|
||||||
|
RUN yum install -y readline-devel && \
|
||||||
|
wget -q http://www.lua.org/ftp/lua-5.3.4.tar.gz && \
|
||||||
|
tar xzf lua-5.3.4.tar.gz && \
|
||||||
|
cd lua-5.3.4 && \
|
||||||
|
sed -i 's/CFLAGS= /CFLAGS= -fPIC /g' src/Makefile && \
|
||||||
|
make linux && \
|
||||||
|
make install && \
|
||||||
|
cd .. && \
|
||||||
|
rm -rf lua-5*
|
||||||
|
|
||||||
|
# MinGW Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
mingw32-binutils \
|
||||||
|
mingw32-crt \
|
||||||
|
mingw32-nsis
|
||||||
|
|
||||||
|
# Node.js Dependencies
|
||||||
|
# Work around epel issue where they removed http-parser that nodejs depends on!
|
||||||
|
RUN yum -y install https://opensource.enda.eu/packages/http-parser-2.7.1-3.el7.x86_64.rpm
|
||||||
|
RUN yum install -y \
|
||||||
|
nodejs \
|
||||||
|
npm
|
||||||
|
|
||||||
|
# Ocaml Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
ocaml \
|
||||||
|
ocaml-ocamldoc && \
|
||||||
|
wget -q https://raw.github.com/ocaml/opam/master/shell/opam_installer.sh -O - | sh -s /usr/local/bin && \
|
||||||
|
opam init --yes && \
|
||||||
|
opam install --yes oasis && \
|
||||||
|
echo '. /root/.opam/opam-init/init.sh > /dev/null 2> /dev/null || true' >> ~/.bashrc
|
||||||
|
|
||||||
|
# Perl Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
perl \
|
||||||
|
perl-version \
|
||||||
|
perl-Bit-Vector \
|
||||||
|
perl-Class-Accessor \
|
||||||
|
perl-ExtUtils-MakeMaker \
|
||||||
|
perl-Test-Simple \
|
||||||
|
perl-IO-Socket-SSL \
|
||||||
|
perl-Net-SSLeay \
|
||||||
|
perl-Crypt-SSLeay
|
||||||
|
|
||||||
|
# PHP Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
php \
|
||||||
|
php-devel \
|
||||||
|
php-pear \
|
||||||
|
re2c \
|
||||||
|
php-phpunit-PHPUnit \
|
||||||
|
bzip2
|
||||||
|
|
||||||
|
# Python Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
python \
|
||||||
|
python-devel \
|
||||||
|
python-pip \
|
||||||
|
python-setuptools \
|
||||||
|
python34 \
|
||||||
|
python34-devel \
|
||||||
|
python34-pip \
|
||||||
|
python34-setuptools
|
||||||
|
RUN pip2 install --upgrade pip
|
||||||
|
RUN pip2 install --upgrade backports.ssl_match_hostname ipaddress setuptools six tornado tornado-testing twisted virtualenv zope-interface
|
||||||
|
RUN pip3 install --upgrade pip
|
||||||
|
RUN pip3 install --upgrade backports.ssl_match_hostname ipaddress setuptools six tornado tornado-testing twisted virtualenv zope-interface
|
||||||
|
|
||||||
|
# Ruby Dependencies
|
||||||
|
RUN yum install -y \
|
||||||
|
ruby \
|
||||||
|
ruby-devel \
|
||||||
|
rubygems && \
|
||||||
|
gem install bundler rake
|
||||||
|
|
||||||
|
# Rust
|
||||||
|
RUN curl https://sh.rustup.rs -sSf | sh -s -- -y --default-toolchain 1.17.0
|
||||||
|
ENV PATH /root/.cargo/bin:$PATH
|
||||||
|
|
||||||
|
# Clean up
|
||||||
|
RUN rm -rf /tmp/* && \
|
||||||
|
yum clean all
|
||||||
|
|
||||||
|
ENV THRIFT_ROOT /thrift
|
||||||
|
RUN mkdir -p $THRIFT_ROOT/src
|
||||||
|
COPY Dockerfile $THRIFT_ROOT/
|
||||||
|
WORKDIR $THRIFT_ROOT/src
|
204
vendor/github.com/apache/thrift/build/docker/debian-jessie/Dockerfile
generated
vendored
Normal file
204
vendor/github.com/apache/thrift/build/docker/debian-jessie/Dockerfile
generated
vendored
Normal file
|
@ -0,0 +1,204 @@
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
# Apache Thrift Docker build environment for Debian
|
||||||
|
#
|
||||||
|
# Known missing client libraries:
|
||||||
|
# - dotnetcore
|
||||||
|
# - rust
|
||||||
|
|
||||||
|
FROM buildpack-deps:jessie-scm
|
||||||
|
MAINTAINER Apache Thrift <dev@thrift.apache.org>
|
||||||
|
|
||||||
|
ENV DEBIAN_FRONTEND noninteractive
|
||||||
|
|
||||||
|
# Add apt sources
|
||||||
|
# jessie-backports for cmake and some ruby bits
|
||||||
|
RUN echo "deb http://ftp.debian.org/debian jessie-backports main" > /etc/apt/sources.list.d/jessie-backports.list
|
||||||
|
|
||||||
|
# Dart
|
||||||
|
RUN curl https://dl-ssl.google.com/linux/linux_signing_key.pub | apt-key add - && \
|
||||||
|
curl https://storage.googleapis.com/download.dartlang.org/linux/debian/dart_stable.list > /etc/apt/sources.list.d/dart_stable.list && \
|
||||||
|
sed -i /etc/apt/sources.list.d/dart_stable.list -e 's/https:/http:/g'
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# General dependencies` \
|
||||||
|
bison \
|
||||||
|
build-essential \
|
||||||
|
clang \
|
||||||
|
debhelper \
|
||||||
|
flex \
|
||||||
|
pkg-config && \
|
||||||
|
apt-get -t jessie-backports install -y --no-install-recommends cmake
|
||||||
|
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
`# C++ dependencies` \
|
||||||
|
libboost-dev \
|
||||||
|
libboost-filesystem-dev \
|
||||||
|
libboost-program-options-dev \
|
||||||
|
libboost-system-dev \
|
||||||
|
libboost-test-dev \
|
||||||
|
libboost-thread-dev \
|
||||||
|
libevent-dev \
|
||||||
|
libssl-dev \
|
||||||
|
qt5-default \
|
||||||
|
qtbase5-dev \
|
||||||
|
qtbase5-dev-tools
|
||||||
|
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
`# Java dependencies` \
|
||||||
|
ant \
|
||||||
|
ant-optional \
|
||||||
|
openjdk-7-jdk \
|
||||||
|
maven
|
||||||
|
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
`# Python dependencies` \
|
||||||
|
python-all \
|
||||||
|
python-all-dbg \
|
||||||
|
python-all-dev \
|
||||||
|
python-pip \
|
||||||
|
python-setuptools \
|
||||||
|
python-twisted \
|
||||||
|
python-zope.interface \
|
||||||
|
python3-all \
|
||||||
|
python3-all-dbg \
|
||||||
|
python3-all-dev \
|
||||||
|
python3-setuptools \
|
||||||
|
python3-pip
|
||||||
|
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
`# Ruby dependencies` \
|
||||||
|
ruby \
|
||||||
|
ruby-dev \
|
||||||
|
`# Perl dependencies` \
|
||||||
|
libbit-vector-perl \
|
||||||
|
libclass-accessor-class-perl \
|
||||||
|
libcrypt-ssleay-perl \
|
||||||
|
libio-socket-ssl-perl \
|
||||||
|
libnet-ssleay-perl
|
||||||
|
|
||||||
|
RUN apt-get -t jessie-backports install -y ruby-bundler
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
`# Php dependencies` \
|
||||||
|
php5 \
|
||||||
|
php5-dev \
|
||||||
|
php5-cli \
|
||||||
|
php-pear \
|
||||||
|
re2c \
|
||||||
|
phpunit \
|
||||||
|
`# GlibC dependencies` \
|
||||||
|
libglib2.0-dev
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# Erlang dependencies` \
|
||||||
|
erlang-base \
|
||||||
|
erlang-eunit \
|
||||||
|
erlang-dev \
|
||||||
|
erlang-tools \
|
||||||
|
rebar
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# Haskell dependencies` \
|
||||||
|
ghc \
|
||||||
|
cabal-install \
|
||||||
|
`# Haxe dependencies` \
|
||||||
|
neko \
|
||||||
|
neko-dev \
|
||||||
|
libneko0
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# Node.js dependencies` \
|
||||||
|
nodejs \
|
||||||
|
nodejs-dev \
|
||||||
|
nodejs-legacy \
|
||||||
|
npm
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# CSharp dependencies` \
|
||||||
|
libmono-system-web2.0-cil \
|
||||||
|
mono-devel
|
||||||
|
|
||||||
|
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||||
|
`# D dependencies` \
|
||||||
|
xdg-utils \
|
||||||
|
`# Dart dependencies` \
|
||||||
|
dart \
|
||||||
|
`# Lua dependencies` \
|
||||||
|
lua5.2 \
|
||||||
|
lua5.2-dev \
|
||||||
|
`# MinGW dependencies` \
|
||||||
|
mingw32 \
|
||||||
|
mingw32-binutils \
|
||||||
|
`# mingw32-runtime` \
|
||||||
|
nsis \
|
||||||
|
`# Clean up` \
|
||||||
|
&& rm -rf /var/cache/apt/* && \
|
||||||
|
rm -rf /var/lib/apt/lists/* && \
|
||||||
|
rm -rf /tmp/* && \
|
||||||
|
rm -rf /var/tmp/*
|
||||||
|
|
||||||
|
# Ruby
|
||||||
|
RUN gem install bundler --no-ri --no-rdoc
|
||||||
|
|
||||||
|
# Python optional dependencies
|
||||||
|
RUN pip2 install -U ipaddress backports.ssl_match_hostname tornado
|
||||||
|
RUN pip3 install -U backports.ssl_match_hostname tornado
|
||||||
|
|
||||||
|
# Go
|
||||||
|
RUN curl -sSL https://storage.googleapis.com/golang/go1.4.3.linux-amd64.tar.gz | tar -C /usr/local/ -xz
|
||||||
|
ENV PATH /usr/local/go/bin:$PATH
|
||||||
|
|
||||||
|
# Haxe
|
||||||
|
RUN mkdir -p /usr/lib/haxe && \
|
||||||
|
wget -O - https://github.com/HaxeFoundation/haxe/releases/download/3.2.1/haxe-3.2.1-linux64.tar.gz | \
|
||||||
|
tar -C /usr/lib/haxe --strip-components=1 -xz && \
|
||||||
|
ln -s /usr/lib/haxe/haxe /usr/bin/haxe && \
|
||||||
|
ln -s /usr/lib/haxe/haxelib /usr/bin/haxelib && \
|
||||||
|
mkdir -p /usr/lib/haxe/lib && \
|
||||||
|
chmod -R 777 /usr/lib/haxe/lib && \
|
||||||
|
haxelib setup /usr/lib/haxe/lib && \
|
||||||
|
haxelib install hxcpp
|
||||||
|
|
||||||
|
# D
|
||||||
|
RUN curl -sSL http://downloads.dlang.org/releases/2.x/2.070.0/dmd_2.070.0-0_amd64.deb -o /tmp/dmd_2.070.0-0_amd64.deb && \
|
||||||
|
dpkg -i /tmp/dmd_2.070.0-0_amd64.deb && \
|
||||||
|
rm /tmp/dmd_2.070.0-0_amd64.deb && \
|
||||||
|
curl -sSL https://github.com/D-Programming-Deimos/openssl/archive/master.tar.gz| tar xz && \
|
||||||
|
curl -sSL https://github.com/D-Programming-Deimos/libevent/archive/master.tar.gz| tar xz && \
|
||||||
|
mkdir -p /usr/include/dmd/druntime/import/deimos /usr/include/dmd/druntime/import/C && \
|
||||||
|
mv libevent-master/deimos/* openssl-master/deimos/* /usr/include/dmd/druntime/import/deimos/ && \
|
||||||
|
mv libevent-master/C/* openssl-master/C/* /usr/include/dmd/druntime/import/C/ && \
|
||||||
|
rm -rf libevent-master openssl-master && \
|
||||||
|
echo 'gcc -Wl,--no-as-needed $*' > /usr/local/bin/gcc-dmd && \
|
||||||
|
chmod 755 /usr/local/bin/gcc-dmd && \
|
||||||
|
echo 'CC=/usr/local/bin/gcc-dmd' >> /etc/dmd.conf
|
||||||
|
|
||||||
|
# Dart
|
||||||
|
ENV PATH /usr/lib/dart/bin:$PATH
|
||||||
|
|
||||||
|
# OCaml
|
||||||
|
RUN echo 'deb http://ppa.launchpad.net/avsm/ppa/ubuntu trusty main' > /etc/apt/sources.list.d/avsm-official-ocaml.list && \
|
||||||
|
gpg --keyserver keyserver.ubuntu.com --recv 61707B09 && \
|
||||||
|
gpg --export --armor 61707B09 | apt-key add - && \
|
||||||
|
apt-get update && \
|
||||||
|
apt-get install -y ocaml opam && \
|
||||||
|
opam init && \
|
||||||
|
opam install oasis
|
||||||
|
|
||||||
|
# Force utf8 locale to successfully build Haskell tf-random
|
||||||
|
ENV LC_ALL C.UTF-8
|
||||||
|
|
||||||
|
ENV THRIFT_ROOT /thrift
|
||||||
|
RUN mkdir -p $THRIFT_ROOT/src
|
||||||
|
COPY Dockerfile $THRIFT_ROOT/
|
||||||
|
WORKDIR $THRIFT_ROOT/src
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Add table
Add a link
Reference in a new issue