Comments and atomics

This commit is contained in:
Mark Haines 2017-03-17 12:00:11 +00:00
parent a21477a303
commit 8a4cbdeea1
2 changed files with 23 additions and 12 deletions

View file

@ -7,6 +7,7 @@ import (
"github.com/matrix-org/dendrite/roomserver/api" "github.com/matrix-org/dendrite/roomserver/api"
"github.com/matrix-org/dendrite/roomserver/types" "github.com/matrix-org/dendrite/roomserver/types"
sarama "gopkg.in/Shopify/sarama.v1" sarama "gopkg.in/Shopify/sarama.v1"
"sync/atomic"
) )
// A ConsumerDatabase has the storage APIs needed by the consumer. // A ConsumerDatabase has the storage APIs needed by the consumer.
@ -47,10 +48,12 @@ type Consumer struct {
// If left as nil then the consumer will panic when it encounters an error // If left as nil then the consumer will panic when it encounters an error
ErrorLogger ErrorLogger ErrorLogger ErrorLogger
// If non-nil then the consumer will stop processing messages after this // If non-nil then the consumer will stop processing messages after this
// many messages and will shutdown // many messages and will shutdown. Malformed messages are included in the count.
StopProcessingAfter *int StopProcessingAfter *int64
// If not-nil then the consumer will call this to shutdown the server. // If not-nil then the consumer will call this to shutdown the server.
ShutdownCallback func(reason string) ShutdownCallback func(reason string)
// How many messages the consumer has processed.
processed int64
} }
// WriteOutputRoomEvent implements OutputRoomEventWriter // WriteOutputRoomEvent implements OutputRoomEventWriter
@ -113,14 +116,7 @@ func (c *Consumer) Start() error {
// consumePartition consumes the room events for a single partition of the kafkaesque stream. // consumePartition consumes the room events for a single partition of the kafkaesque stream.
func (c *Consumer) consumePartition(pc sarama.PartitionConsumer) { func (c *Consumer) consumePartition(pc sarama.PartitionConsumer) {
defer pc.Close() defer pc.Close()
var processed int
for message := range pc.Messages() { for message := range pc.Messages() {
if c.StopProcessingAfter != nil && processed >= *c.StopProcessingAfter {
if c.ShutdownCallback != nil {
c.ShutdownCallback(fmt.Sprintf("Stopping processing after %d messages", processed))
}
return
}
var input api.InputRoomEvent var input api.InputRoomEvent
if err := json.Unmarshal(message.Value, &input); err != nil { if err := json.Unmarshal(message.Value, &input); err != nil {
// If the message is invalid then log it and move onto the next message in the stream. // If the message is invalid then log it and move onto the next message in the stream.
@ -139,7 +135,19 @@ func (c *Consumer) consumePartition(pc sarama.PartitionConsumer) {
if err := c.DB.SetPartitionOffset(c.InputRoomEventTopic, message.Partition, message.Offset); err != nil { if err := c.DB.SetPartitionOffset(c.InputRoomEventTopic, message.Partition, message.Offset); err != nil {
c.logError(message, err) c.logError(message, err)
} }
processed++ // Update the number of processed messages using atomic addition because it is accessed from multiple goroutines.
processed := atomic.AddInt64(&c.processed, 1)
// Check if we should stop processing.
// Note that since we have multiple goroutines it's quite likely that we'll overshoot by a few messages.
// If we try to stop processing after M message and we have N goroutines then we will process somewhere
// between M and (N + M) messages because the N goroutines could all try to process what they think will be the
// last message. We could be more careful here but this is good enough for getting rough benchmarks.
if c.StopProcessingAfter != nil && processed >= int64(*c.StopProcessingAfter) {
if c.ShutdownCallback != nil {
c.ShutdownCallback(fmt.Sprintf("Stopping processing after %d messages", c.processed))
}
return
}
} }
} }

View file

@ -19,6 +19,9 @@ var (
inputRoomEventTopic = os.Getenv("TOPIC_INPUT_ROOM_EVENT") inputRoomEventTopic = os.Getenv("TOPIC_INPUT_ROOM_EVENT")
outputRoomEventTopic = os.Getenv("TOPIC_OUTPUT_ROOM_EVENT") outputRoomEventTopic = os.Getenv("TOPIC_OUTPUT_ROOM_EVENT")
bindAddr = os.Getenv("BIND_ADDRESS") bindAddr = os.Getenv("BIND_ADDRESS")
// Shuts the roomserver down after processing a given number of messages.
// This is useful for running benchmarks for seeing how quickly the server
// can process a given number of messages.
stopProcessingAfter = os.Getenv("STOP_AFTER") stopProcessingAfter = os.Getenv("STOP_AFTER")
) )
@ -47,7 +50,7 @@ func main() {
} }
if stopProcessingAfter != "" { if stopProcessingAfter != "" {
count, err := strconv.Atoi(stopProcessingAfter) count, err := strconv.ParseInt(stopProcessingAfter, 10, 64)
if err != nil { if err != nil {
panic(err) panic(err)
} }