2022-09-30 13:48:10 +02:00
|
|
|
package shared
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"database/sql"
|
|
|
|
"fmt"
|
2023-01-17 10:08:23 +01:00
|
|
|
"math"
|
2022-09-30 13:48:10 +02:00
|
|
|
|
2023-04-19 16:50:33 +02:00
|
|
|
"github.com/matrix-org/gomatrixserverlib/spec"
|
2023-01-12 10:06:03 +01:00
|
|
|
"github.com/tidwall/gjson"
|
2022-10-19 14:05:39 +02:00
|
|
|
|
2022-09-30 13:48:10 +02:00
|
|
|
"github.com/matrix-org/dendrite/internal/eventutil"
|
2023-06-07 19:14:35 +02:00
|
|
|
"github.com/matrix-org/dendrite/roomserver/api"
|
2023-04-27 13:54:20 +02:00
|
|
|
rstypes "github.com/matrix-org/dendrite/roomserver/types"
|
2023-04-04 19:16:53 +02:00
|
|
|
"github.com/matrix-org/dendrite/syncapi/synctypes"
|
2022-09-30 13:48:10 +02:00
|
|
|
"github.com/matrix-org/dendrite/syncapi/types"
|
|
|
|
userapi "github.com/matrix-org/dendrite/userapi/api"
|
|
|
|
)
|
|
|
|
|
|
|
|
type DatabaseTransaction struct {
|
|
|
|
*Database
|
2022-09-30 18:07:37 +02:00
|
|
|
ctx context.Context
|
2022-09-30 13:48:10 +02:00
|
|
|
txn *sql.Tx
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) Commit() error {
|
|
|
|
if d.txn == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
return d.txn.Commit()
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) Rollback() error {
|
|
|
|
if d.txn == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
return d.txn.Rollback()
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForPDUs(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.OutputEvents.SelectMaxEventID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.OutputEvents.SelectMaxEventID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForReceipts(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.Receipts.SelectMaxReceiptID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.Receipts.SelectMaxReceiptID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForInvites(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.Invites.SelectMaxInviteID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.Invites.SelectMaxInviteID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForSendToDeviceMessages(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.SendToDevice.SelectMaxSendToDeviceMessageID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.SendToDevice.SelectMaxSendToDeviceMessageID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForAccountData(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.AccountData.SelectMaxAccountDataID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.Invites.SelectMaxAccountDataID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForNotificationData(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.NotificationData.SelectMaxID(ctx, d.txn)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("d.NotificationData.SelectMaxID: %w", err)
|
|
|
|
}
|
|
|
|
return types.StreamPosition(id), nil
|
|
|
|
}
|
|
|
|
|
2023-04-27 13:54:20 +02:00
|
|
|
func (d *DatabaseTransaction) CurrentState(ctx context.Context, roomID string, stateFilterPart *synctypes.StateFilter, excludeEventIDs []string) ([]*rstypes.HeaderedEvent, error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
return d.CurrentRoomState.SelectCurrentState(ctx, d.txn, roomID, stateFilterPart, excludeEventIDs)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) RoomIDsWithMembership(ctx context.Context, userID string, membership string) ([]string, error) {
|
|
|
|
return d.CurrentRoomState.SelectRoomIDsWithMembership(ctx, d.txn, userID, membership)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MembershipCount(ctx context.Context, roomID, membership string, pos types.StreamPosition) (int, error) {
|
|
|
|
return d.Memberships.SelectMembershipCount(ctx, d.txn, roomID, membership, pos)
|
|
|
|
}
|
|
|
|
|
2023-01-12 10:06:03 +01:00
|
|
|
func (d *DatabaseTransaction) GetRoomSummary(ctx context.Context, roomID, userID string) (*types.Summary, error) {
|
|
|
|
summary := &types.Summary{Heroes: []string{}}
|
|
|
|
|
2023-04-19 16:50:33 +02:00
|
|
|
joinCount, err := d.CurrentRoomState.SelectMembershipCount(ctx, d.txn, roomID, spec.Join)
|
2023-01-12 10:06:03 +01:00
|
|
|
if err != nil {
|
|
|
|
return summary, err
|
|
|
|
}
|
2023-04-19 16:50:33 +02:00
|
|
|
inviteCount, err := d.CurrentRoomState.SelectMembershipCount(ctx, d.txn, roomID, spec.Invite)
|
2023-01-12 10:06:03 +01:00
|
|
|
if err != nil {
|
|
|
|
return summary, err
|
|
|
|
}
|
|
|
|
summary.InvitedMemberCount = &inviteCount
|
|
|
|
summary.JoinedMemberCount = &joinCount
|
|
|
|
|
|
|
|
// Get the room name and canonical alias, if any
|
2023-04-04 19:16:53 +02:00
|
|
|
filter := synctypes.DefaultStateFilter()
|
2023-04-19 16:50:33 +02:00
|
|
|
filterTypes := []string{spec.MRoomName, spec.MRoomCanonicalAlias}
|
2023-01-12 10:06:03 +01:00
|
|
|
filterRooms := []string{roomID}
|
|
|
|
|
|
|
|
filter.Types = &filterTypes
|
|
|
|
filter.Rooms = &filterRooms
|
|
|
|
evs, err := d.CurrentRoomState.SelectCurrentState(ctx, d.txn, roomID, &filter, nil)
|
|
|
|
if err != nil {
|
|
|
|
return summary, err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, ev := range evs {
|
|
|
|
switch ev.Type() {
|
2023-04-19 16:50:33 +02:00
|
|
|
case spec.MRoomName:
|
2023-01-12 10:06:03 +01:00
|
|
|
if gjson.GetBytes(ev.Content(), "name").Str != "" {
|
|
|
|
return summary, nil
|
|
|
|
}
|
2023-04-19 16:50:33 +02:00
|
|
|
case spec.MRoomCanonicalAlias:
|
2023-01-12 10:06:03 +01:00
|
|
|
if gjson.GetBytes(ev.Content(), "alias").Str != "" {
|
|
|
|
return summary, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If there's no room name or canonical alias, get the room heroes, excluding the user
|
2023-04-19 16:50:33 +02:00
|
|
|
heroes, err := d.CurrentRoomState.SelectRoomHeroes(ctx, d.txn, roomID, userID, []string{spec.Join, spec.Invite})
|
2023-01-12 10:06:03 +01:00
|
|
|
if err != nil {
|
|
|
|
return summary, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// "When no joined or invited members are available, this should consist of the banned and left users"
|
|
|
|
if len(heroes) == 0 {
|
2023-04-19 16:50:33 +02:00
|
|
|
heroes, err = d.CurrentRoomState.SelectRoomHeroes(ctx, d.txn, roomID, userID, []string{spec.Leave, spec.Ban})
|
2023-01-12 10:06:03 +01:00
|
|
|
if err != nil {
|
|
|
|
return summary, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
summary.Heroes = heroes
|
|
|
|
|
|
|
|
return summary, nil
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
|
2023-04-04 19:16:53 +02:00
|
|
|
func (d *DatabaseTransaction) RecentEvents(ctx context.Context, roomIDs []string, r types.Range, eventFilter *synctypes.RoomEventFilter, chronologicalOrder bool, onlySyncEvents bool) (map[string]types.RecentEvents, error) {
|
2023-02-07 14:31:23 +01:00
|
|
|
return d.OutputEvents.SelectRecentEvents(ctx, d.txn, roomIDs, r, eventFilter, chronologicalOrder, onlySyncEvents)
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) PositionInTopology(ctx context.Context, eventID string) (pos types.StreamPosition, spos types.StreamPosition, err error) {
|
|
|
|
return d.Topology.SelectPositionInTopology(ctx, d.txn, eventID)
|
|
|
|
}
|
|
|
|
|
2023-04-27 13:54:20 +02:00
|
|
|
func (d *DatabaseTransaction) InviteEventsInRange(ctx context.Context, targetUserID string, r types.Range) (map[string]*rstypes.HeaderedEvent, map[string]*rstypes.HeaderedEvent, types.StreamPosition, error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
return d.Invites.SelectInviteEventsInRange(ctx, d.txn, targetUserID, r)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) PeeksInRange(ctx context.Context, userID, deviceID string, r types.Range) (peeks []types.Peek, err error) {
|
|
|
|
return d.Peeks.SelectPeeksInRange(ctx, d.txn, userID, deviceID, r)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) RoomReceiptsAfter(ctx context.Context, roomIDs []string, streamPos types.StreamPosition) (types.StreamPosition, []types.OutputReceiptEvent, error) {
|
|
|
|
return d.Receipts.SelectRoomReceiptsAfter(ctx, d.txn, roomIDs, streamPos)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Events lookups a list of event by their event ID.
|
|
|
|
// Returns a list of events matching the requested IDs found in the database.
|
|
|
|
// If an event is not found in the database then it will be omitted from the list.
|
|
|
|
// Returns an error if there was a problem talking with the database.
|
|
|
|
// Does not include any transaction IDs in the returned events.
|
2023-04-27 13:54:20 +02:00
|
|
|
func (d *DatabaseTransaction) Events(ctx context.Context, eventIDs []string) ([]*rstypes.HeaderedEvent, error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
streamEvents, err := d.OutputEvents.SelectEvents(ctx, d.txn, eventIDs, nil, false)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// We don't include a device here as we only include transaction IDs in
|
|
|
|
// incremental syncs.
|
2023-06-07 19:14:35 +02:00
|
|
|
return d.StreamEventsToEvents(ctx, nil, streamEvents, nil), nil
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) AllJoinedUsersInRooms(ctx context.Context) (map[string][]string, error) {
|
|
|
|
return d.CurrentRoomState.SelectJoinedUsers(ctx, d.txn)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) AllJoinedUsersInRoom(ctx context.Context, roomIDs []string) (map[string][]string, error) {
|
|
|
|
return d.CurrentRoomState.SelectJoinedUsersInRoom(ctx, d.txn, roomIDs)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) AllPeekingDevicesInRooms(ctx context.Context) (map[string][]types.PeekingDevice, error) {
|
|
|
|
return d.Peeks.SelectPeekingDevices(ctx, d.txn)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) SharedUsers(ctx context.Context, userID string, otherUserIDs []string) ([]string, error) {
|
|
|
|
return d.CurrentRoomState.SelectSharedUsers(ctx, d.txn, userID, otherUserIDs)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) GetStateEvent(
|
|
|
|
ctx context.Context, roomID, evType, stateKey string,
|
2023-04-27 13:54:20 +02:00
|
|
|
) (*rstypes.HeaderedEvent, error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
return d.CurrentRoomState.SelectStateEvent(ctx, d.txn, roomID, evType, stateKey)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) GetStateEventsForRoom(
|
2023-04-04 19:16:53 +02:00
|
|
|
ctx context.Context, roomID string, stateFilter *synctypes.StateFilter,
|
2023-04-27 13:54:20 +02:00
|
|
|
) (stateEvents []*rstypes.HeaderedEvent, err error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
stateEvents, err = d.CurrentRoomState.SelectCurrentState(ctx, d.txn, roomID, stateFilter, nil)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetAccountDataInRange returns all account data for a given user inserted or
|
|
|
|
// updated between two given positions
|
|
|
|
// Returns a map following the format data[roomID] = []dataTypes
|
|
|
|
// If no data is retrieved, returns an empty map
|
|
|
|
// If there was an issue with the retrieval, returns an error
|
|
|
|
func (d *DatabaseTransaction) GetAccountDataInRange(
|
|
|
|
ctx context.Context, userID string, r types.Range,
|
2023-04-04 19:16:53 +02:00
|
|
|
accountDataFilterPart *synctypes.EventFilter,
|
2022-09-30 13:48:10 +02:00
|
|
|
) (map[string][]string, types.StreamPosition, error) {
|
|
|
|
return d.AccountData.SelectAccountDataInRange(ctx, d.txn, userID, r, accountDataFilterPart)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) GetEventsInTopologicalRange(
|
|
|
|
ctx context.Context,
|
|
|
|
from, to *types.TopologyToken,
|
|
|
|
roomID string,
|
2023-04-04 19:16:53 +02:00
|
|
|
filter *synctypes.RoomEventFilter,
|
2022-09-30 13:48:10 +02:00
|
|
|
backwardOrdering bool,
|
2023-07-13 14:18:37 +02:00
|
|
|
) (events []types.StreamEvent, start, end types.TopologyToken, err error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
var minDepth, maxDepth, maxStreamPosForMaxDepth types.StreamPosition
|
|
|
|
if backwardOrdering {
|
|
|
|
// Backward ordering means the 'from' token has a higher depth than the 'to' token
|
|
|
|
minDepth = to.Depth
|
|
|
|
maxDepth = from.Depth
|
|
|
|
// for cases where we have say 5 events with the same depth, the TopologyToken needs to
|
|
|
|
// know which of the 5 the client has seen. This is done by using the PDU position.
|
|
|
|
// Events with the same maxDepth but less than this PDU position will be returned.
|
|
|
|
maxStreamPosForMaxDepth = from.PDUPosition
|
|
|
|
} else {
|
|
|
|
// Forward ordering means the 'from' token has a lower depth than the 'to' token.
|
|
|
|
minDepth = from.Depth
|
|
|
|
maxDepth = to.Depth
|
|
|
|
}
|
|
|
|
|
|
|
|
// Select the event IDs from the defined range.
|
|
|
|
var eIDs []string
|
2023-07-13 14:18:37 +02:00
|
|
|
eIDs, start, end, err = d.Topology.SelectEventIDsInRange(
|
2022-09-30 13:48:10 +02:00
|
|
|
ctx, d.txn, roomID, minDepth, maxDepth, maxStreamPosForMaxDepth, filter.Limit, !backwardOrdering,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Retrieve the events' contents using their IDs.
|
|
|
|
events, err = d.OutputEvents.SelectEvents(ctx, d.txn, eIDs, filter, true)
|
2023-07-13 14:18:37 +02:00
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2022-09-30 13:48:10 +02:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) BackwardExtremitiesForRoom(
|
|
|
|
ctx context.Context, roomID string,
|
|
|
|
) (backwardExtremities map[string][]string, err error) {
|
|
|
|
return d.BackwardExtremities.SelectBackwardExtremitiesForRoom(ctx, d.txn, roomID)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) EventPositionInTopology(
|
|
|
|
ctx context.Context, eventID string,
|
|
|
|
) (types.TopologyToken, error) {
|
|
|
|
depth, stream, err := d.Topology.SelectPositionInTopology(ctx, d.txn, eventID)
|
|
|
|
if err != nil {
|
|
|
|
return types.TopologyToken{}, err
|
|
|
|
}
|
|
|
|
return types.TopologyToken{Depth: depth, PDUPosition: stream}, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) StreamToTopologicalPosition(
|
|
|
|
ctx context.Context, roomID string, streamPos types.StreamPosition, backwardOrdering bool,
|
|
|
|
) (types.TopologyToken, error) {
|
|
|
|
topoPos, err := d.Topology.SelectStreamToTopologicalPosition(ctx, d.txn, roomID, streamPos, backwardOrdering)
|
|
|
|
switch {
|
|
|
|
case err == sql.ErrNoRows && backwardOrdering: // no events in range, going backward
|
|
|
|
return types.TopologyToken{PDUPosition: streamPos}, nil
|
|
|
|
case err == sql.ErrNoRows && !backwardOrdering: // no events in range, going forward
|
2023-01-17 10:08:23 +01:00
|
|
|
return types.TopologyToken{Depth: math.MaxInt64, PDUPosition: math.MaxInt64}, nil
|
2022-09-30 13:48:10 +02:00
|
|
|
case err != nil: // some other error happened
|
|
|
|
return types.TopologyToken{}, fmt.Errorf("d.Topology.SelectStreamToTopologicalPosition: %w", err)
|
|
|
|
default:
|
|
|
|
return types.TopologyToken{Depth: topoPos, PDUPosition: streamPos}, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetBackwardTopologyPos retrieves the backward topology position, i.e. the position of the
|
|
|
|
// oldest event in the room's topology.
|
|
|
|
func (d *DatabaseTransaction) GetBackwardTopologyPos(
|
|
|
|
ctx context.Context,
|
2023-04-27 13:54:20 +02:00
|
|
|
events []*rstypes.HeaderedEvent,
|
2022-09-30 13:48:10 +02:00
|
|
|
) (types.TopologyToken, error) {
|
|
|
|
zeroToken := types.TopologyToken{}
|
|
|
|
if len(events) == 0 {
|
|
|
|
return zeroToken, nil
|
|
|
|
}
|
|
|
|
pos, spos, err := d.Topology.SelectPositionInTopology(ctx, d.txn, events[0].EventID())
|
|
|
|
if err != nil {
|
|
|
|
return zeroToken, err
|
|
|
|
}
|
|
|
|
tok := types.TopologyToken{Depth: pos, PDUPosition: spos}
|
|
|
|
tok.Decrement()
|
|
|
|
return tok, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetStateDeltas returns the state deltas between fromPos and toPos,
|
|
|
|
// exclusive of oldPos, inclusive of newPos, for the rooms in which
|
|
|
|
// the user has new membership events.
|
|
|
|
// A list of joined room IDs is also returned in case the caller needs it.
|
2022-10-19 14:05:39 +02:00
|
|
|
// nolint:gocyclo
|
2022-09-30 13:48:10 +02:00
|
|
|
func (d *DatabaseTransaction) GetStateDeltas(
|
|
|
|
ctx context.Context, device *userapi.Device,
|
|
|
|
r types.Range, userID string,
|
2023-06-07 19:14:35 +02:00
|
|
|
stateFilter *synctypes.StateFilter, rsAPI api.SyncRoomserverAPI,
|
2022-09-30 13:48:10 +02:00
|
|
|
) (deltas []types.StateDelta, joinedRoomsIDs []string, err error) {
|
|
|
|
// Implement membership change algorithm: https://github.com/matrix-org/synapse/blob/v0.19.3/synapse/handlers/sync.py#L821
|
|
|
|
// - Get membership list changes for this user in this sync response
|
|
|
|
// - For each room which has membership list changes:
|
|
|
|
// * Check if the room is 'newly joined' (insufficient to just check for a join event because we allow dupe joins TODO).
|
|
|
|
// If it is, then we need to send the full room state down (and 'limited' is always true).
|
|
|
|
// * Check if user is still CURRENTLY invited to the room. If so, add room to 'invited' block.
|
|
|
|
// * Check if the user is CURRENTLY (TODO) left/banned. If so, add room to 'archived' block.
|
|
|
|
// - Get all CURRENTLY joined rooms, and add them to 'joined' block.
|
|
|
|
|
|
|
|
// Look up all memberships for the user. We only care about rooms that a
|
|
|
|
// user has ever interacted with — joined to, kicked/banned from, left.
|
|
|
|
memberships, err := d.CurrentRoomState.SelectRoomIDsWithAnyMembership(ctx, d.txn, userID)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
allRoomIDs := make([]string, 0, len(memberships))
|
|
|
|
joinedRoomIDs := make([]string, 0, len(memberships))
|
|
|
|
for roomID, membership := range memberships {
|
|
|
|
allRoomIDs = append(allRoomIDs, roomID)
|
2023-04-19 16:50:33 +02:00
|
|
|
if membership == spec.Join {
|
2022-09-30 13:48:10 +02:00
|
|
|
joinedRoomIDs = append(joinedRoomIDs, roomID)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// get all the state events ever (i.e. for all available rooms) between these two positions
|
2022-10-19 14:05:39 +02:00
|
|
|
stateNeeded, eventMap, err := d.OutputEvents.SelectStateInRange(ctx, d.txn, r, nil, allRoomIDs)
|
2022-09-30 13:48:10 +02:00
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
state, err := d.fetchStateEvents(ctx, d.txn, stateNeeded, eventMap)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
2022-10-19 14:05:39 +02:00
|
|
|
// get all the state events ever (i.e. for all available rooms) between these two positions
|
2023-02-07 14:31:23 +01:00
|
|
|
stateFiltered := state
|
|
|
|
// avoid hitting the database if the result would be the same as above
|
|
|
|
if !isStatefilterEmpty(stateFilter) {
|
|
|
|
var stateNeededFiltered map[string]map[string]bool
|
|
|
|
var eventMapFiltered map[string]types.StreamEvent
|
|
|
|
stateNeededFiltered, eventMapFiltered, err = d.OutputEvents.SelectStateInRange(ctx, d.txn, r, stateFilter, allRoomIDs)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
2022-10-19 14:05:39 +02:00
|
|
|
}
|
2023-02-07 14:31:23 +01:00
|
|
|
stateFiltered, err = d.fetchStateEvents(ctx, d.txn, stateNeededFiltered, eventMapFiltered)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
2022-10-19 14:05:39 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-09-30 13:48:10 +02:00
|
|
|
// find out which rooms this user is peeking, if any.
|
|
|
|
// We do this before joins so any peeks get overwritten
|
|
|
|
peeks, err := d.Peeks.SelectPeeksInRange(ctx, d.txn, userID, device.ID, r)
|
|
|
|
if err != nil && err != sql.ErrNoRows {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// add peek blocks
|
|
|
|
for _, peek := range peeks {
|
|
|
|
if peek.New {
|
|
|
|
// send full room state down instead of a delta
|
|
|
|
var s []types.StreamEvent
|
|
|
|
s, err = d.currentStateStreamEventsForRoom(ctx, peek.RoomID, stateFilter)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
state[peek.RoomID] = s
|
|
|
|
}
|
|
|
|
if !peek.Deleted {
|
|
|
|
deltas = append(deltas, types.StateDelta{
|
2023-04-19 16:50:33 +02:00
|
|
|
Membership: spec.Peek,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, state[peek.RoomID], rsAPI),
|
2022-09-30 13:48:10 +02:00
|
|
|
RoomID: peek.RoomID,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// handle newly joined rooms and non-joined rooms
|
|
|
|
newlyJoinedRooms := make(map[string]bool, len(state))
|
|
|
|
for roomID, stateStreamEvents := range state {
|
|
|
|
for _, ev := range stateStreamEvents {
|
2022-10-11 11:58:34 +02:00
|
|
|
// Look for our membership in the state events and skip over any
|
|
|
|
// membership events that are not related to us.
|
2023-06-12 13:19:25 +02:00
|
|
|
membership, prevMembership := getMembershipFromEvent(ctx, ev.PDU, userID, rsAPI)
|
2022-10-11 11:58:34 +02:00
|
|
|
if membership == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2023-04-19 16:50:33 +02:00
|
|
|
if membership == spec.Join {
|
2022-10-11 11:58:34 +02:00
|
|
|
// If our membership is now join but the previous membership wasn't
|
|
|
|
// then this is a "join transition", so we'll insert this room.
|
|
|
|
if prevMembership != membership {
|
2022-10-19 14:05:39 +02:00
|
|
|
newlyJoinedRooms[roomID] = true
|
2022-10-11 11:58:34 +02:00
|
|
|
// Get the full room state, as we'll send that down for a newly
|
|
|
|
// joined room instead of a delta.
|
2022-09-30 13:48:10 +02:00
|
|
|
var s []types.StreamEvent
|
2022-10-11 11:58:34 +02:00
|
|
|
if s, err = d.currentStateStreamEventsForRoom(ctx, roomID, stateFilter); err != nil {
|
2022-09-30 13:48:10 +02:00
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
2022-10-11 11:58:34 +02:00
|
|
|
|
|
|
|
// Add the information for this room into the state so that
|
|
|
|
// it will get added with all of the rest of the joined rooms.
|
2022-10-19 14:05:39 +02:00
|
|
|
stateFiltered[roomID] = s
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
|
2022-10-11 11:58:34 +02:00
|
|
|
// We won't add joined rooms into the delta at this point as they
|
|
|
|
// are added later on.
|
|
|
|
continue
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
2022-10-11 11:58:34 +02:00
|
|
|
|
|
|
|
deltas = append(deltas, types.StateDelta{
|
|
|
|
Membership: membership,
|
|
|
|
MembershipPos: ev.StreamPosition,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, stateFiltered[roomID], rsAPI),
|
2022-10-11 11:58:34 +02:00
|
|
|
RoomID: roomID,
|
|
|
|
})
|
|
|
|
break
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-11 11:58:34 +02:00
|
|
|
// Finally, add in currently joined rooms, including those from the
|
|
|
|
// join transitions above.
|
2022-09-30 13:48:10 +02:00
|
|
|
for _, joinedRoomID := range joinedRoomIDs {
|
|
|
|
deltas = append(deltas, types.StateDelta{
|
2023-04-19 16:50:33 +02:00
|
|
|
Membership: spec.Join,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, stateFiltered[joinedRoomID], rsAPI),
|
2022-09-30 13:48:10 +02:00
|
|
|
RoomID: joinedRoomID,
|
|
|
|
NewlyJoined: newlyJoinedRooms[joinedRoomID],
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
return deltas, joinedRoomIDs, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// GetStateDeltasForFullStateSync is a variant of getStateDeltas used for /sync
|
|
|
|
// requests with full_state=true.
|
|
|
|
// Fetches full state for all joined rooms and uses selectStateInRange to get
|
|
|
|
// updates for other rooms.
|
|
|
|
func (d *DatabaseTransaction) GetStateDeltasForFullStateSync(
|
|
|
|
ctx context.Context, device *userapi.Device,
|
|
|
|
r types.Range, userID string,
|
2023-06-07 19:14:35 +02:00
|
|
|
stateFilter *synctypes.StateFilter, rsAPI api.SyncRoomserverAPI,
|
2022-09-30 13:48:10 +02:00
|
|
|
) ([]types.StateDelta, []string, error) {
|
|
|
|
// Look up all memberships for the user. We only care about rooms that a
|
|
|
|
// user has ever interacted with — joined to, kicked/banned from, left.
|
|
|
|
memberships, err := d.CurrentRoomState.SelectRoomIDsWithAnyMembership(ctx, d.txn, userID)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
allRoomIDs := make([]string, 0, len(memberships))
|
|
|
|
joinedRoomIDs := make([]string, 0, len(memberships))
|
|
|
|
for roomID, membership := range memberships {
|
|
|
|
allRoomIDs = append(allRoomIDs, roomID)
|
2023-04-19 16:50:33 +02:00
|
|
|
if membership == spec.Join {
|
2022-09-30 13:48:10 +02:00
|
|
|
joinedRoomIDs = append(joinedRoomIDs, roomID)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Use a reasonable initial capacity
|
|
|
|
deltas := make(map[string]types.StateDelta)
|
|
|
|
|
|
|
|
peeks, err := d.Peeks.SelectPeeksInRange(ctx, d.txn, userID, device.ID, r)
|
|
|
|
if err != nil && err != sql.ErrNoRows {
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add full states for all peeking rooms
|
|
|
|
for _, peek := range peeks {
|
|
|
|
if !peek.Deleted {
|
|
|
|
s, stateErr := d.currentStateStreamEventsForRoom(ctx, peek.RoomID, stateFilter)
|
|
|
|
if stateErr != nil {
|
|
|
|
if stateErr == sql.ErrNoRows {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
return nil, nil, stateErr
|
|
|
|
}
|
|
|
|
deltas[peek.RoomID] = types.StateDelta{
|
2023-04-19 16:50:33 +02:00
|
|
|
Membership: spec.Peek,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, s, rsAPI),
|
2022-09-30 13:48:10 +02:00
|
|
|
RoomID: peek.RoomID,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get all the state events ever between these two positions
|
|
|
|
stateNeeded, eventMap, err := d.OutputEvents.SelectStateInRange(ctx, d.txn, r, stateFilter, allRoomIDs)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
state, err := d.fetchStateEvents(ctx, d.txn, stateNeeded, eventMap)
|
|
|
|
if err != nil {
|
|
|
|
if err == sql.ErrNoRows {
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
return nil, nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
for roomID, stateStreamEvents := range state {
|
|
|
|
for _, ev := range stateStreamEvents {
|
2023-06-12 13:19:25 +02:00
|
|
|
if membership, _ := getMembershipFromEvent(ctx, ev.PDU, userID, rsAPI); membership != "" {
|
2023-04-19 16:50:33 +02:00
|
|
|
if membership != spec.Join { // We've already added full state for all joined rooms above.
|
2022-09-30 13:48:10 +02:00
|
|
|
deltas[roomID] = types.StateDelta{
|
|
|
|
Membership: membership,
|
|
|
|
MembershipPos: ev.StreamPosition,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, stateStreamEvents, rsAPI),
|
2022-09-30 13:48:10 +02:00
|
|
|
RoomID: roomID,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add full states for all joined rooms
|
|
|
|
for _, joinedRoomID := range joinedRoomIDs {
|
|
|
|
s, stateErr := d.currentStateStreamEventsForRoom(ctx, joinedRoomID, stateFilter)
|
|
|
|
if stateErr != nil {
|
|
|
|
if stateErr == sql.ErrNoRows {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
return nil, nil, stateErr
|
|
|
|
}
|
|
|
|
deltas[joinedRoomID] = types.StateDelta{
|
2023-04-19 16:50:33 +02:00
|
|
|
Membership: spec.Join,
|
2023-06-07 19:14:35 +02:00
|
|
|
StateEvents: d.StreamEventsToEvents(ctx, device, s, rsAPI),
|
2022-09-30 13:48:10 +02:00
|
|
|
RoomID: joinedRoomID,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a response array.
|
|
|
|
result := make([]types.StateDelta, len(deltas))
|
|
|
|
i := 0
|
|
|
|
for _, delta := range deltas {
|
|
|
|
result[i] = delta
|
|
|
|
i++
|
|
|
|
}
|
|
|
|
|
|
|
|
return result, joinedRoomIDs, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) currentStateStreamEventsForRoom(
|
|
|
|
ctx context.Context, roomID string,
|
2023-04-04 19:16:53 +02:00
|
|
|
stateFilter *synctypes.StateFilter,
|
2022-09-30 13:48:10 +02:00
|
|
|
) ([]types.StreamEvent, error) {
|
|
|
|
allState, err := d.CurrentRoomState.SelectCurrentState(ctx, d.txn, roomID, stateFilter, nil)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
s := make([]types.StreamEvent, len(allState))
|
|
|
|
for i := 0; i < len(s); i++ {
|
|
|
|
s[i] = types.StreamEvent{HeaderedEvent: allState[i], StreamPosition: 0}
|
|
|
|
}
|
|
|
|
return s, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) SendToDeviceUpdatesForSync(
|
|
|
|
ctx context.Context,
|
|
|
|
userID, deviceID string,
|
|
|
|
from, to types.StreamPosition,
|
|
|
|
) (types.StreamPosition, []types.SendToDeviceEvent, error) {
|
|
|
|
// First of all, get our send-to-device updates for this user.
|
|
|
|
lastPos, events, err := d.SendToDevice.SelectSendToDeviceMessages(ctx, d.txn, userID, deviceID, from, to)
|
|
|
|
if err != nil {
|
|
|
|
return from, nil, fmt.Errorf("d.SendToDevice.SelectSendToDeviceMessages: %w", err)
|
|
|
|
}
|
|
|
|
// If there's nothing to do then stop here.
|
|
|
|
if len(events) == 0 {
|
|
|
|
return to, nil, nil
|
|
|
|
}
|
|
|
|
return lastPos, events, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) GetRoomReceipts(ctx context.Context, roomIDs []string, streamPos types.StreamPosition) ([]types.OutputReceiptEvent, error) {
|
|
|
|
_, receipts, err := d.Receipts.SelectRoomReceiptsAfter(ctx, d.txn, roomIDs, streamPos)
|
|
|
|
return receipts, err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) GetUserUnreadNotificationCountsForRooms(ctx context.Context, userID string, rooms map[string]string) (map[string]*eventutil.NotificationData, error) {
|
|
|
|
roomIDs := make([]string, 0, len(rooms))
|
|
|
|
for roomID, membership := range rooms {
|
2023-04-19 16:50:33 +02:00
|
|
|
if membership != spec.Join {
|
2022-09-30 13:48:10 +02:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
roomIDs = append(roomIDs, roomID)
|
|
|
|
}
|
|
|
|
return d.NotificationData.SelectUserUnreadCountsForRooms(ctx, d.txn, userID, roomIDs)
|
|
|
|
}
|
|
|
|
|
2022-12-08 08:25:03 +01:00
|
|
|
func (d *DatabaseTransaction) GetPresences(ctx context.Context, userIDs []string) ([]*types.PresenceInternal, error) {
|
|
|
|
return d.Presence.GetPresenceForUsers(ctx, d.txn, userIDs)
|
2022-09-30 13:48:10 +02:00
|
|
|
}
|
|
|
|
|
2023-04-04 19:16:53 +02:00
|
|
|
func (d *DatabaseTransaction) PresenceAfter(ctx context.Context, after types.StreamPosition, filter synctypes.EventFilter) (map[string]*types.PresenceInternal, error) {
|
2022-09-30 13:48:10 +02:00
|
|
|
return d.Presence.GetPresenceAfter(ctx, d.txn, after, filter)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForPresence(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
return d.Presence.GetMaxPresenceID(ctx, d.txn)
|
|
|
|
}
|
2022-10-13 15:50:52 +02:00
|
|
|
|
2023-01-19 21:02:32 +01:00
|
|
|
func (d *Database) PurgeRoom(ctx context.Context, roomID string) error {
|
|
|
|
return d.Writer.Do(d.DB, nil, func(txn *sql.Tx) error {
|
|
|
|
if err := d.BackwardExtremities.PurgeBackwardExtremities(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge backward extremities: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.CurrentRoomState.DeleteRoomStateForRoom(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge current room state: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.Invites.PurgeInvites(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge invites: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.Memberships.PurgeMemberships(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge memberships: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.NotificationData.PurgeNotificationData(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge notification data: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.OutputEvents.PurgeEvents(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge events: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.Topology.PurgeEventsTopology(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge events topology: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.Peeks.PurgePeeks(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge peeks: %w", err)
|
|
|
|
}
|
|
|
|
if err := d.Receipts.PurgeReceipts(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("failed to purge receipts: %w", err)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *Database) PurgeRoomState(
|
|
|
|
ctx context.Context, roomID string,
|
|
|
|
) error {
|
|
|
|
return d.Writer.Do(d.DB, nil, func(txn *sql.Tx) error {
|
|
|
|
// If the event is a create event then we'll delete all of the existing
|
|
|
|
// data for the room. The only reason that a create event would be replayed
|
|
|
|
// to us in this way is if we're about to receive the entire room state.
|
|
|
|
if err := d.CurrentRoomState.DeleteRoomStateForRoom(ctx, txn, roomID); err != nil {
|
|
|
|
return fmt.Errorf("d.CurrentRoomState.DeleteRoomStateForRoom: %w", err)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2022-10-13 15:50:52 +02:00
|
|
|
func (d *DatabaseTransaction) MaxStreamPositionForRelations(ctx context.Context) (types.StreamPosition, error) {
|
|
|
|
id, err := d.Relations.SelectMaxRelationID(ctx, d.txn)
|
|
|
|
return types.StreamPosition(id), err
|
|
|
|
}
|
|
|
|
|
2023-04-04 19:16:53 +02:00
|
|
|
func isStatefilterEmpty(filter *synctypes.StateFilter) bool {
|
2023-02-07 14:31:23 +01:00
|
|
|
if filter == nil {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
switch {
|
|
|
|
case filter.NotTypes != nil && len(*filter.NotTypes) > 0:
|
|
|
|
return false
|
|
|
|
case filter.Types != nil && len(*filter.Types) > 0:
|
|
|
|
return false
|
|
|
|
case filter.Senders != nil && len(*filter.Senders) > 0:
|
|
|
|
return false
|
|
|
|
case filter.NotSenders != nil && len(*filter.NotSenders) > 0:
|
|
|
|
return false
|
|
|
|
case filter.NotRooms != nil && len(*filter.NotRooms) > 0:
|
|
|
|
return false
|
|
|
|
case filter.ContainsURL != nil:
|
|
|
|
return false
|
|
|
|
default:
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-13 15:50:52 +02:00
|
|
|
func (d *DatabaseTransaction) RelationsFor(ctx context.Context, roomID, eventID, relType, eventType string, from, to types.StreamPosition, backwards bool, limit int) (
|
|
|
|
events []types.StreamEvent, prevBatch, nextBatch string, err error,
|
|
|
|
) {
|
|
|
|
r := types.Range{
|
|
|
|
From: from,
|
|
|
|
To: to,
|
|
|
|
Backwards: backwards,
|
|
|
|
}
|
|
|
|
|
|
|
|
if r.Backwards && r.From == 0 {
|
|
|
|
// If we're working backwards (dir=b) and there's no ?from= specified then
|
|
|
|
// we will automatically want to work backwards from the current position,
|
|
|
|
// so find out what that is.
|
|
|
|
if r.From, err = d.MaxStreamPositionForRelations(ctx); err != nil {
|
|
|
|
return nil, "", "", fmt.Errorf("d.MaxStreamPositionForRelations: %w", err)
|
|
|
|
}
|
|
|
|
// The result normally isn't inclusive of the event *at* the ?from=
|
|
|
|
// position, so add 1 here so that we include the most recent relation.
|
|
|
|
r.From++
|
|
|
|
} else if !r.Backwards && r.To == 0 {
|
|
|
|
// If we're working forwards (dir=f) and there's no ?to= specified then
|
|
|
|
// we will automatically want to work forwards towards the current position,
|
|
|
|
// so find out what that is.
|
|
|
|
if r.To, err = d.MaxStreamPositionForRelations(ctx); err != nil {
|
|
|
|
return nil, "", "", fmt.Errorf("d.MaxStreamPositionForRelations: %w", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// First look up any relations from the database. We add one to the limit here
|
|
|
|
// so that we can tell if we're overflowing, as we will only set the "next_batch"
|
|
|
|
// in the response if we are.
|
|
|
|
relations, _, err := d.Relations.SelectRelationsInRange(ctx, d.txn, roomID, eventID, relType, eventType, r, limit+1)
|
|
|
|
if err != nil {
|
|
|
|
return nil, "", "", fmt.Errorf("d.Relations.SelectRelationsInRange: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we specified a relation type then just get those results, otherwise collate
|
|
|
|
// them from all of the returned relation types.
|
|
|
|
entries := []types.RelationEntry{}
|
|
|
|
if relType != "" {
|
|
|
|
entries = relations[relType]
|
|
|
|
} else {
|
|
|
|
for _, e := range relations {
|
|
|
|
entries = append(entries, e...)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If there were no entries returned, there were no relations, so stop at this point.
|
|
|
|
if len(entries) == 0 {
|
|
|
|
return nil, "", "", nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, let's try and work out what sensible prev_batch and next_batch values
|
|
|
|
// could be. We've requested an extra event by adding one to the limit already so
|
|
|
|
// that we can determine whether or not to provide a "next_batch", so trim off that
|
|
|
|
// event off the end if needs be.
|
|
|
|
if len(entries) > limit {
|
|
|
|
entries = entries[:len(entries)-1]
|
|
|
|
nextBatch = fmt.Sprintf("%d", entries[len(entries)-1].Position)
|
|
|
|
}
|
|
|
|
// TODO: set prevBatch? doesn't seem to affect the tests...
|
|
|
|
|
|
|
|
// Extract all of the event IDs from the relation entries so that we can pull the
|
|
|
|
// events out of the database. Then go and fetch the events.
|
|
|
|
eventIDs := make([]string, 0, len(entries))
|
|
|
|
for _, entry := range entries {
|
|
|
|
eventIDs = append(eventIDs, entry.EventID)
|
|
|
|
}
|
|
|
|
events, err = d.OutputEvents.SelectEvents(ctx, d.txn, eventIDs, nil, true)
|
|
|
|
if err != nil {
|
|
|
|
return nil, "", "", fmt.Errorf("d.OutputEvents.SelectEvents: %w", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return events, prevBatch, nextBatch, nil
|
|
|
|
}
|