mirror of
https://github.com/hoernschen/dendrite.git
synced 2024-12-27 23:48:27 +00:00
f05ce478f0
* Add Pushserver component with Pushers API Co-authored-by: Tommie Gannert <tommie@gannert.se> Co-authored-by: Dan Peleg <dan@globekeeper.com> * Wire Pushserver component Co-authored-by: Neil Alexander <neilalexander@users.noreply.github.com> * Add PushGatewayClient. The full event format is required for Sytest. * Add a pushrules module. * Change user API account creation to use the new pushrules module's defaults. Introduces "scope" as required by client API, and some small field tweaks to make some 61push Sytests pass. * Add push rules query/put API in Pushserver. This manipulates account data over User API, and fires sync messages for changes. Those sync messages should, according to an existing TODO in clientapi, be moved to userapi. Forks clientapi/producers/syncapi.go to pushserver/ for later extension. * Add clientapi routes for push rules to Pushserver. A cleanup would be to move more of the name-splitting logic into pushrules.go, to depollute routing.go. * Output rooms.join.unread_notifications in /sync. This is the read-side. Pushserver will be the write-side. * Implement pushserver/storage for notifications. * Use PushGatewayClient and the pushrules module in Pushserver's room consumer. * Use one goroutine per user to avoid locking up the entire server for one bad push gateway. * Split pushing by format. * Send one device per push. Sytest does not support coalescing multiple devices into one push. Matches Synapse. Either we change Sytest, or remove the group-by-url-and-format logic. * Write OutputNotificationData from push server. Sync API is already the consumer. * Implement read receipt consumers in Pushserver. Supports m.read and m.fully_read receipts. * Add clientapi route for /unstable/notifications. * Rename to UpsertPusher for clarity and handle pusher update * Fix linter errors * Ignore body.Close() error check * Fix push server internal http wiring * Add 40 newly passing 61push tests to whitelist * Add next 12 newly passing 61push tests to whitelist * Send notification data before notifying users in EDU server consumer * NATS JetStream * Goodbye sarama * Fix `NewStreamTokenFromString` * Consume on the correct topic for the roomserver * Don't panic, NAK instead * Move push notifications into the User API * Don't set null values since that apparently causes Element upsetti * Also set omitempty on conditions * Fix bug so that we don't override the push rules unnecessarily * Tweak defaults * Update defaults * More tweaks * Move `/notifications` onto `r0`/`v3` mux * User API will consume events and read/fully read markers from the sync API with stream positions, instead of consuming directly Co-authored-by: Piotr Kozimor <p1996k@gmail.com> Co-authored-by: Tommie Gannert <tommie@gannert.se> Co-authored-by: Neil Alexander <neilalexander@users.noreply.github.com>
577 lines
18 KiB
Go
577 lines
18 KiB
Go
// Copyright 2017-2018 New Vector Ltd
|
|
// Copyright 2019-2020 The Matrix.org Foundation C.I.C.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package sqlite3
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"encoding/json"
|
|
"fmt"
|
|
"sort"
|
|
|
|
"github.com/matrix-org/dendrite/internal"
|
|
"github.com/matrix-org/dendrite/roomserver/api"
|
|
"github.com/matrix-org/dendrite/syncapi/storage/tables"
|
|
"github.com/matrix-org/dendrite/syncapi/types"
|
|
|
|
"github.com/matrix-org/dendrite/internal/sqlutil"
|
|
"github.com/matrix-org/gomatrixserverlib"
|
|
log "github.com/sirupsen/logrus"
|
|
)
|
|
|
|
const outputRoomEventsSchema = `
|
|
-- Stores output room events received from the roomserver.
|
|
CREATE TABLE IF NOT EXISTS syncapi_output_room_events (
|
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
event_id TEXT NOT NULL UNIQUE,
|
|
room_id TEXT NOT NULL,
|
|
headered_event_json TEXT NOT NULL,
|
|
type TEXT NOT NULL,
|
|
sender TEXT NOT NULL,
|
|
contains_url BOOL NOT NULL,
|
|
add_state_ids TEXT, -- JSON encoded string array
|
|
remove_state_ids TEXT, -- JSON encoded string array
|
|
session_id BIGINT,
|
|
transaction_id TEXT,
|
|
exclude_from_sync BOOL NOT NULL DEFAULT FALSE
|
|
);
|
|
`
|
|
|
|
const insertEventSQL = "" +
|
|
"INSERT INTO syncapi_output_room_events (" +
|
|
"id, room_id, event_id, headered_event_json, type, sender, contains_url, add_state_ids, remove_state_ids, session_id, transaction_id, exclude_from_sync" +
|
|
") VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12) " +
|
|
"ON CONFLICT (event_id) DO UPDATE SET exclude_from_sync = (excluded.exclude_from_sync AND $13)"
|
|
|
|
const selectEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id FROM syncapi_output_room_events WHERE event_id = $1"
|
|
|
|
const selectRecentEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
const selectRecentEventsForSyncSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3 AND exclude_from_sync = FALSE"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
const selectEarlyEventsSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, session_id, exclude_from_sync, transaction_id FROM syncapi_output_room_events" +
|
|
" WHERE room_id = $1 AND id > $2 AND id <= $3"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
const selectMaxEventIDSQL = "" +
|
|
"SELECT MAX(id) FROM syncapi_output_room_events"
|
|
|
|
const updateEventJSONSQL = "" +
|
|
"UPDATE syncapi_output_room_events SET headered_event_json=$1 WHERE event_id=$2"
|
|
|
|
const selectStateInRangeSQL = "" +
|
|
"SELECT event_id, id, headered_event_json, exclude_from_sync, add_state_ids, remove_state_ids" +
|
|
" FROM syncapi_output_room_events" +
|
|
" WHERE (id > $1 AND id <= $2)" +
|
|
" AND ((add_state_ids IS NOT NULL AND add_state_ids != '') OR (remove_state_ids IS NOT NULL AND remove_state_ids != ''))"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
const deleteEventsForRoomSQL = "" +
|
|
"DELETE FROM syncapi_output_room_events WHERE room_id = $1"
|
|
|
|
const selectContextEventSQL = "" +
|
|
"SELECT id, headered_event_json FROM syncapi_output_room_events WHERE room_id = $1 AND event_id = $2"
|
|
|
|
const selectContextBeforeEventSQL = "" +
|
|
"SELECT headered_event_json FROM syncapi_output_room_events WHERE room_id = $1 AND id < $2"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
const selectContextAfterEventSQL = "" +
|
|
"SELECT id, headered_event_json FROM syncapi_output_room_events WHERE room_id = $1 AND id > $2"
|
|
|
|
// WHEN, ORDER BY and LIMIT are appended by prepareWithFilters
|
|
|
|
type outputRoomEventsStatements struct {
|
|
db *sql.DB
|
|
streamIDStatements *streamIDStatements
|
|
insertEventStmt *sql.Stmt
|
|
selectEventsStmt *sql.Stmt
|
|
selectMaxEventIDStmt *sql.Stmt
|
|
updateEventJSONStmt *sql.Stmt
|
|
deleteEventsForRoomStmt *sql.Stmt
|
|
selectContextEventStmt *sql.Stmt
|
|
selectContextBeforeEventStmt *sql.Stmt
|
|
selectContextAfterEventStmt *sql.Stmt
|
|
}
|
|
|
|
func NewSqliteEventsTable(db *sql.DB, streamID *streamIDStatements) (tables.Events, error) {
|
|
s := &outputRoomEventsStatements{
|
|
db: db,
|
|
streamIDStatements: streamID,
|
|
}
|
|
_, err := db.Exec(outputRoomEventsSchema)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return s, sqlutil.StatementList{
|
|
{&s.insertEventStmt, insertEventSQL},
|
|
{&s.selectEventsStmt, selectEventsSQL},
|
|
{&s.selectMaxEventIDStmt, selectMaxEventIDSQL},
|
|
{&s.updateEventJSONStmt, updateEventJSONSQL},
|
|
{&s.deleteEventsForRoomStmt, deleteEventsForRoomSQL},
|
|
{&s.selectContextEventStmt, selectContextEventSQL},
|
|
{&s.selectContextBeforeEventStmt, selectContextBeforeEventSQL},
|
|
{&s.selectContextAfterEventStmt, selectContextAfterEventSQL},
|
|
}.Prepare(db)
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) UpdateEventJSON(ctx context.Context, event *gomatrixserverlib.HeaderedEvent) error {
|
|
headeredJSON, err := json.Marshal(event)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
_, err = s.updateEventJSONStmt.ExecContext(ctx, headeredJSON, event.EventID())
|
|
return err
|
|
}
|
|
|
|
// selectStateInRange returns the state events between the two given PDU stream positions, exclusive of oldPos, inclusive of newPos.
|
|
// Results are bucketed based on the room ID. If the same state is overwritten multiple times between the
|
|
// two positions, only the most recent state is returned.
|
|
func (s *outputRoomEventsStatements) SelectStateInRange(
|
|
ctx context.Context, txn *sql.Tx, r types.Range,
|
|
stateFilter *gomatrixserverlib.StateFilter,
|
|
) (map[string]map[string]bool, map[string]types.StreamEvent, error) {
|
|
stmt, params, err := prepareWithFilters(
|
|
s.db, txn, selectStateInRangeSQL,
|
|
[]interface{}{
|
|
r.Low(), r.High(),
|
|
},
|
|
stateFilter.Senders, stateFilter.NotSenders,
|
|
stateFilter.Types, stateFilter.NotTypes,
|
|
nil, stateFilter.Limit, FilterOrderAsc,
|
|
)
|
|
if err != nil {
|
|
return nil, nil, fmt.Errorf("s.prepareWithFilters: %w", err)
|
|
}
|
|
|
|
rows, err := stmt.QueryContext(ctx, params...)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
defer rows.Close() // nolint: errcheck
|
|
// Fetch all the state change events for all rooms between the two positions then loop each event and:
|
|
// - Keep a cache of the event by ID (99% of state change events are for the event itself)
|
|
// - For each room ID, build up an array of event IDs which represents cumulative adds/removes
|
|
// For each room, map cumulative event IDs to events and return. This may need to a batch SELECT based on event ID
|
|
// if they aren't in the event ID cache. We don't handle state deletion yet.
|
|
eventIDToEvent := make(map[string]types.StreamEvent)
|
|
|
|
// RoomID => A set (map[string]bool) of state event IDs which are between the two positions
|
|
stateNeeded := make(map[string]map[string]bool)
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventID string
|
|
streamPos types.StreamPosition
|
|
eventBytes []byte
|
|
excludeFromSync bool
|
|
addIDsJSON string
|
|
delIDsJSON string
|
|
)
|
|
if err := rows.Scan(&eventID, &streamPos, &eventBytes, &excludeFromSync, &addIDsJSON, &delIDsJSON); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
addIDs, delIDs, err := unmarshalStateIDs(addIDsJSON, delIDsJSON)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Sanity check for deleted state and whine if we see it. We don't need to do anything
|
|
// since it'll just mark the event as not being needed.
|
|
if len(addIDs) < len(delIDs) {
|
|
log.WithFields(log.Fields{
|
|
"since": r.From,
|
|
"current": r.To,
|
|
"adds": addIDsJSON,
|
|
"dels": delIDsJSON,
|
|
}).Warn("StateBetween: ignoring deleted state")
|
|
}
|
|
|
|
// TODO: Handle redacted events
|
|
var ev gomatrixserverlib.HeaderedEvent
|
|
if err := ev.UnmarshalJSONWithEventID(eventBytes, eventID); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
needSet := stateNeeded[ev.RoomID()]
|
|
if needSet == nil { // make set if required
|
|
needSet = make(map[string]bool)
|
|
}
|
|
for _, id := range delIDs {
|
|
needSet[id] = false
|
|
}
|
|
for _, id := range addIDs {
|
|
needSet[id] = true
|
|
}
|
|
stateNeeded[ev.RoomID()] = needSet
|
|
|
|
eventIDToEvent[eventID] = types.StreamEvent{
|
|
HeaderedEvent: &ev,
|
|
StreamPosition: streamPos,
|
|
ExcludeFromSync: excludeFromSync,
|
|
}
|
|
}
|
|
|
|
return stateNeeded, eventIDToEvent, nil
|
|
}
|
|
|
|
// MaxID returns the ID of the last inserted event in this table. 'txn' is optional. If it is not supplied,
|
|
// then this function should only ever be used at startup, as it will race with inserting events if it is
|
|
// done afterwards. If there are no inserted events, 0 is returned.
|
|
func (s *outputRoomEventsStatements) SelectMaxEventID(
|
|
ctx context.Context, txn *sql.Tx,
|
|
) (id int64, err error) {
|
|
var nullableID sql.NullInt64
|
|
stmt := sqlutil.TxStmt(txn, s.selectMaxEventIDStmt)
|
|
err = stmt.QueryRowContext(ctx).Scan(&nullableID)
|
|
if nullableID.Valid {
|
|
id = nullableID.Int64
|
|
}
|
|
return
|
|
}
|
|
|
|
// InsertEvent into the output_room_events table. addState and removeState are an optional list of state event IDs. Returns the position
|
|
// of the inserted event.
|
|
func (s *outputRoomEventsStatements) InsertEvent(
|
|
ctx context.Context, txn *sql.Tx,
|
|
event *gomatrixserverlib.HeaderedEvent, addState, removeState []string,
|
|
transactionID *api.TransactionID, excludeFromSync bool,
|
|
) (types.StreamPosition, error) {
|
|
var txnID *string
|
|
var sessionID *int64
|
|
if transactionID != nil {
|
|
sessionID = &transactionID.SessionID
|
|
txnID = &transactionID.TransactionID
|
|
}
|
|
|
|
// Parse content as JSON and search for an "url" key
|
|
containsURL := false
|
|
var content map[string]interface{}
|
|
if json.Unmarshal(event.Content(), &content) != nil {
|
|
// Set containsURL to true if url is present
|
|
_, containsURL = content["url"]
|
|
}
|
|
|
|
var headeredJSON []byte
|
|
headeredJSON, err := json.Marshal(event)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
|
|
var addStateJSON, removeStateJSON []byte
|
|
if len(addState) > 0 {
|
|
addStateJSON, err = json.Marshal(addState)
|
|
}
|
|
if err != nil {
|
|
return 0, fmt.Errorf("json.Marshal(addState): %w", err)
|
|
}
|
|
if len(removeState) > 0 {
|
|
removeStateJSON, err = json.Marshal(removeState)
|
|
}
|
|
if err != nil {
|
|
return 0, fmt.Errorf("json.Marshal(removeState): %w", err)
|
|
}
|
|
|
|
streamPos, err := s.streamIDStatements.nextPDUID(ctx, txn)
|
|
if err != nil {
|
|
return 0, err
|
|
}
|
|
insertStmt := sqlutil.TxStmt(txn, s.insertEventStmt)
|
|
_, err = insertStmt.ExecContext(
|
|
ctx,
|
|
streamPos,
|
|
event.RoomID(),
|
|
event.EventID(),
|
|
headeredJSON,
|
|
event.Type(),
|
|
event.Sender(),
|
|
containsURL,
|
|
string(addStateJSON),
|
|
string(removeStateJSON),
|
|
sessionID,
|
|
txnID,
|
|
excludeFromSync,
|
|
excludeFromSync,
|
|
)
|
|
return streamPos, err
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectRecentEvents(
|
|
ctx context.Context, txn *sql.Tx,
|
|
roomID string, r types.Range, eventFilter *gomatrixserverlib.RoomEventFilter,
|
|
chronologicalOrder bool, onlySyncEvents bool,
|
|
) ([]types.StreamEvent, bool, error) {
|
|
var query string
|
|
if onlySyncEvents {
|
|
query = selectRecentEventsForSyncSQL
|
|
} else {
|
|
query = selectRecentEventsSQL
|
|
}
|
|
|
|
stmt, params, err := prepareWithFilters(
|
|
s.db, txn, query,
|
|
[]interface{}{
|
|
roomID, r.Low(), r.High(),
|
|
},
|
|
eventFilter.Senders, eventFilter.NotSenders,
|
|
eventFilter.Types, eventFilter.NotTypes,
|
|
nil, eventFilter.Limit+1, FilterOrderDesc,
|
|
)
|
|
if err != nil {
|
|
return nil, false, fmt.Errorf("s.prepareWithFilters: %w", err)
|
|
}
|
|
|
|
rows, err := stmt.QueryContext(ctx, params...)
|
|
if err != nil {
|
|
return nil, false, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectRecentEvents: rows.close() failed")
|
|
events, err := rowsToStreamEvents(rows)
|
|
if err != nil {
|
|
return nil, false, err
|
|
}
|
|
if chronologicalOrder {
|
|
// The events need to be returned from oldest to latest, which isn't
|
|
// necessary the way the SQL query returns them, so a sort is necessary to
|
|
// ensure the events are in the right order in the slice.
|
|
sort.SliceStable(events, func(i int, j int) bool {
|
|
return events[i].StreamPosition < events[j].StreamPosition
|
|
})
|
|
}
|
|
// we queried for 1 more than the limit, so if we returned one more mark limited=true
|
|
limited := false
|
|
if len(events) > eventFilter.Limit {
|
|
limited = true
|
|
// re-slice the extra (oldest) event out: in chronological order this is the first entry, else the last.
|
|
if chronologicalOrder {
|
|
events = events[1:]
|
|
} else {
|
|
events = events[:len(events)-1]
|
|
}
|
|
}
|
|
return events, limited, nil
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectEarlyEvents(
|
|
ctx context.Context, txn *sql.Tx,
|
|
roomID string, r types.Range, eventFilter *gomatrixserverlib.RoomEventFilter,
|
|
) ([]types.StreamEvent, error) {
|
|
stmt, params, err := prepareWithFilters(
|
|
s.db, txn, selectEarlyEventsSQL,
|
|
[]interface{}{
|
|
roomID, r.Low(), r.High(),
|
|
},
|
|
eventFilter.Senders, eventFilter.NotSenders,
|
|
eventFilter.Types, eventFilter.NotTypes,
|
|
nil, eventFilter.Limit, FilterOrderAsc,
|
|
)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("s.prepareWithFilters: %w", err)
|
|
}
|
|
rows, err := stmt.QueryContext(ctx, params...)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer internal.CloseAndLogIfError(ctx, rows, "selectEarlyEvents: rows.close() failed")
|
|
events, err := rowsToStreamEvents(rows)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
// The events need to be returned from oldest to latest, which isn't
|
|
// necessarily the way the SQL query returns them, so a sort is necessary to
|
|
// ensure the events are in the right order in the slice.
|
|
sort.SliceStable(events, func(i int, j int) bool {
|
|
return events[i].StreamPosition < events[j].StreamPosition
|
|
})
|
|
return events, nil
|
|
}
|
|
|
|
// selectEvents returns the events for the given event IDs. If an event is
|
|
// missing from the database, it will be omitted.
|
|
func (s *outputRoomEventsStatements) SelectEvents(
|
|
ctx context.Context, txn *sql.Tx, eventIDs []string,
|
|
) ([]types.StreamEvent, error) {
|
|
var returnEvents []types.StreamEvent
|
|
stmt := sqlutil.TxStmt(txn, s.selectEventsStmt)
|
|
for _, eventID := range eventIDs {
|
|
rows, err := stmt.QueryContext(ctx, eventID)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if streamEvents, err := rowsToStreamEvents(rows); err == nil {
|
|
returnEvents = append(returnEvents, streamEvents...)
|
|
}
|
|
internal.CloseAndLogIfError(ctx, rows, "selectEvents: rows.close() failed")
|
|
}
|
|
return returnEvents, nil
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) DeleteEventsForRoom(
|
|
ctx context.Context, txn *sql.Tx, roomID string,
|
|
) (err error) {
|
|
_, err = sqlutil.TxStmt(txn, s.deleteEventsForRoomStmt).ExecContext(ctx, roomID)
|
|
return err
|
|
}
|
|
|
|
func rowsToStreamEvents(rows *sql.Rows) ([]types.StreamEvent, error) {
|
|
var result []types.StreamEvent
|
|
for rows.Next() {
|
|
var (
|
|
eventID string
|
|
streamPos types.StreamPosition
|
|
eventBytes []byte
|
|
excludeFromSync bool
|
|
sessionID *int64
|
|
txnID *string
|
|
transactionID *api.TransactionID
|
|
)
|
|
if err := rows.Scan(&eventID, &streamPos, &eventBytes, &sessionID, &excludeFromSync, &txnID); err != nil {
|
|
return nil, err
|
|
}
|
|
// TODO: Handle redacted events
|
|
var ev gomatrixserverlib.HeaderedEvent
|
|
if err := ev.UnmarshalJSONWithEventID(eventBytes, eventID); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if sessionID != nil && txnID != nil {
|
|
transactionID = &api.TransactionID{
|
|
SessionID: *sessionID,
|
|
TransactionID: *txnID,
|
|
}
|
|
}
|
|
|
|
result = append(result, types.StreamEvent{
|
|
HeaderedEvent: &ev,
|
|
StreamPosition: streamPos,
|
|
TransactionID: transactionID,
|
|
ExcludeFromSync: excludeFromSync,
|
|
})
|
|
}
|
|
return result, nil
|
|
}
|
|
func (s *outputRoomEventsStatements) SelectContextEvent(
|
|
ctx context.Context, txn *sql.Tx, roomID, eventID string,
|
|
) (id int, evt gomatrixserverlib.HeaderedEvent, err error) {
|
|
row := sqlutil.TxStmt(txn, s.selectContextEventStmt).QueryRowContext(ctx, roomID, eventID)
|
|
var eventAsString string
|
|
if err = row.Scan(&id, &eventAsString); err != nil {
|
|
return 0, evt, err
|
|
}
|
|
|
|
if err = json.Unmarshal([]byte(eventAsString), &evt); err != nil {
|
|
return 0, evt, err
|
|
}
|
|
return id, evt, nil
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectContextBeforeEvent(
|
|
ctx context.Context, txn *sql.Tx, id int, roomID string, filter *gomatrixserverlib.RoomEventFilter,
|
|
) (evts []*gomatrixserverlib.HeaderedEvent, err error) {
|
|
stmt, params, err := prepareWithFilters(
|
|
s.db, txn, selectContextBeforeEventSQL,
|
|
[]interface{}{
|
|
roomID, id,
|
|
},
|
|
filter.Senders, filter.NotSenders,
|
|
filter.Types, filter.NotTypes,
|
|
nil, filter.Limit, FilterOrderDesc,
|
|
)
|
|
|
|
rows, err := stmt.QueryContext(ctx, params...)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer rows.Close()
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventBytes []byte
|
|
evt *gomatrixserverlib.HeaderedEvent
|
|
)
|
|
if err = rows.Scan(&eventBytes); err != nil {
|
|
return evts, err
|
|
}
|
|
if err = json.Unmarshal(eventBytes, &evt); err != nil {
|
|
return evts, err
|
|
}
|
|
evts = append(evts, evt)
|
|
}
|
|
|
|
return evts, rows.Err()
|
|
}
|
|
|
|
func (s *outputRoomEventsStatements) SelectContextAfterEvent(
|
|
ctx context.Context, txn *sql.Tx, id int, roomID string, filter *gomatrixserverlib.RoomEventFilter,
|
|
) (lastID int, evts []*gomatrixserverlib.HeaderedEvent, err error) {
|
|
stmt, params, err := prepareWithFilters(
|
|
s.db, txn, selectContextAfterEventSQL,
|
|
[]interface{}{
|
|
roomID, id,
|
|
},
|
|
filter.Senders, filter.NotSenders,
|
|
filter.Types, filter.NotTypes,
|
|
nil, filter.Limit, FilterOrderAsc,
|
|
)
|
|
|
|
rows, err := stmt.QueryContext(ctx, params...)
|
|
if err != nil {
|
|
return
|
|
}
|
|
defer rows.Close()
|
|
|
|
for rows.Next() {
|
|
var (
|
|
eventBytes []byte
|
|
evt *gomatrixserverlib.HeaderedEvent
|
|
)
|
|
if err = rows.Scan(&lastID, &eventBytes); err != nil {
|
|
return 0, evts, err
|
|
}
|
|
if err = json.Unmarshal(eventBytes, &evt); err != nil {
|
|
return 0, evts, err
|
|
}
|
|
evts = append(evts, evt)
|
|
}
|
|
return lastID, evts, rows.Err()
|
|
}
|
|
|
|
func unmarshalStateIDs(addIDsJSON, delIDsJSON string) (addIDs []string, delIDs []string, err error) {
|
|
if len(addIDsJSON) > 0 {
|
|
if err = json.Unmarshal([]byte(addIDsJSON), &addIDs); err != nil {
|
|
return
|
|
}
|
|
}
|
|
if len(delIDsJSON) > 0 {
|
|
if err = json.Unmarshal([]byte(delIDsJSON), &delIDs); err != nil {
|
|
return
|
|
}
|
|
}
|
|
return
|
|
}
|