mattermost-community-enterp.../channels/app/shared_channel_sync_self_referential_utils_test.go
Claude ec1f89217a Merge: Complete Mattermost Server with Community Enterprise
Full Mattermost server source with integrated Community Enterprise features.
Includes vendor directory for offline/air-gapped builds.

Structure:
- enterprise-impl/: Enterprise feature implementations
- enterprise-community/: Init files that register implementations
- enterprise/: Bridge imports (community_imports.go)
- vendor/: All dependencies for offline builds

Build (online):
  go build ./cmd/mattermost

Build (offline/air-gapped):
  go build -mod=vendor ./cmd/mattermost

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
2025-12-17 23:59:07 +09:00

310 lines
11 KiB
Go

// Copyright (c) 2015-present Mattermost, Inc. All Rights Reserved.
// See LICENSE.txt for license information.
package app
import (
"encoding/json"
"io"
"net/http"
"sync/atomic"
"testing"
"time"
"github.com/mattermost/mattermost/server/public/model"
"github.com/mattermost/mattermost/server/v8/channels/store"
"github.com/mattermost/mattermost/server/v8/platform/services/remotecluster"
"github.com/mattermost/mattermost/server/v8/platform/services/sharedchannel"
"github.com/stretchr/testify/require"
)
// writeOKResponse writes a standard OK JSON response in the format expected by remotecluster
func writeOKResponse(w http.ResponseWriter) {
response := &remotecluster.Response{
Status: "OK",
Err: "",
}
// Set empty sync response as payload
syncResp := &model.SyncResponse{}
_ = response.SetPayload(syncResp)
w.Header().Set("Content-Type", "application/json")
w.WriteHeader(http.StatusOK)
respBytes, _ := json.Marshal(response)
_, _ = w.Write(respBytes)
}
// SelfReferentialSyncHandler handles incoming sync messages for self-referential tests.
// It processes the messages, updates cursors, and returns proper responses.
type SelfReferentialSyncHandler struct {
t *testing.T
service *sharedchannel.Service
selfCluster *model.RemoteCluster
syncMessageCount *int32
SimulateUnshared bool // When true, always return ErrChannelIsNotShared for sync messages
// Callbacks for capturing sync data
OnIndividualSync func(userId string, messageNumber int32)
OnBatchSync func(userIds []string, messageNumber int32)
OnGlobalUserSync func(userIds []string, messageNumber int32)
}
// NewSelfReferentialSyncHandler creates a new handler for processing sync messages in tests
func NewSelfReferentialSyncHandler(t *testing.T, service *sharedchannel.Service, selfCluster *model.RemoteCluster) *SelfReferentialSyncHandler {
count := int32(0)
return &SelfReferentialSyncHandler{
t: t,
service: service,
selfCluster: selfCluster,
syncMessageCount: &count,
}
}
// HandleRequest processes incoming HTTP requests for the test server.
// This handler includes common remote cluster endpoints to simulate a real remote cluster:
// - /api/v4/remotecluster/msg: Main sync message endpoint
// - /api/v4/remotecluster/ping: Ping endpoint to maintain online status (prevents offline after 5 minutes)
// - /api/v4/remotecluster/confirm_invite: Invitation confirmation endpoint
func (h *SelfReferentialSyncHandler) HandleRequest(w http.ResponseWriter, r *http.Request) {
switch r.URL.Path {
case "/api/v4/remotecluster/msg":
currentCall := atomic.AddInt32(h.syncMessageCount, 1)
// Read and process the sync message
body, _ := io.ReadAll(r.Body)
// The message is wrapped in a RemoteClusterFrame
var frame model.RemoteClusterFrame
err := json.Unmarshal(body, &frame)
if err == nil {
// Simulate the remote having unshared if configured to do so
if h.SimulateUnshared && frame.Msg.Topic == "sharedchannel_sync" {
// Return HTTP error instead of JSON error response
w.WriteHeader(http.StatusBadRequest)
_, _ = w.Write([]byte("channel is no longer shared"))
return
}
// Process the message to update cursor
response := &remotecluster.Response{}
processErr := h.service.OnReceiveSyncMessageForTesting(frame.Msg, h.selfCluster, response)
if processErr != nil {
response.Status = "ERROR"
response.Err = processErr.Error()
} else {
// Success - build a proper sync response
response.Status = "OK"
response.Err = ""
var syncMsg model.SyncMsg
if unmarshalErr := json.Unmarshal(frame.Msg.Payload, &syncMsg); unmarshalErr == nil {
syncResp := &model.SyncResponse{}
// Handle global user sync
if len(syncMsg.Users) > 0 {
userIds := make([]string, 0, len(syncMsg.Users))
for userId := range syncMsg.Users {
userIds = append(userIds, userId)
syncResp.UsersSyncd = append(syncResp.UsersSyncd, userId)
}
if h.OnGlobalUserSync != nil {
h.OnGlobalUserSync(userIds, currentCall)
}
}
// Handle membership sync using unified field
if len(syncMsg.MembershipChanges) > 0 {
batch := make([]string, 0)
for _, change := range syncMsg.MembershipChanges {
if change.IsAdd {
syncResp.UsersSyncd = append(syncResp.UsersSyncd, change.UserId)
batch = append(batch, change.UserId)
}
}
// Call appropriate callback
if len(batch) > 0 {
if h.OnBatchSync != nil {
h.OnBatchSync(batch, currentCall)
}
if len(batch) == 1 && h.OnIndividualSync != nil {
h.OnIndividualSync(batch[0], currentCall)
}
}
}
_ = response.SetPayload(syncResp)
}
}
// Send the proper response
w.Header().Set("Content-Type", "application/json")
w.WriteHeader(http.StatusOK)
respBytes, _ := json.Marshal(response)
_, _ = w.Write(respBytes)
return
}
writeOKResponse(w)
case "/api/v4/remotecluster/ping":
writeOKResponse(w)
case "/api/v4/remotecluster/confirm_invite":
writeOKResponse(w)
default:
writeOKResponse(w)
}
}
// GetSyncMessageCount returns the current count of sync messages received
func (h *SelfReferentialSyncHandler) GetSyncMessageCount() int32 {
return atomic.LoadInt32(h.syncMessageCount)
}
// EnsureCleanState ensures a clean test state by removing all shared channels, remote clusters,
// and extra team/channel members. This helps prevent state pollution between tests.
func EnsureCleanState(t *testing.T, th *TestHelper, ss store.Store) {
t.Helper()
// First, wait for any pending async tasks to complete, then shutdown services
scsInterface := th.App.Srv().GetSharedChannelSyncService()
if scsInterface != nil && scsInterface.Active() {
// Cast to concrete type to access testing methods
if service, ok := scsInterface.(*sharedchannel.Service); ok {
// Wait for any pending tasks from previous tests to complete
require.Eventually(t, func() bool {
return !service.HasPendingTasksForTesting()
}, 10*time.Second, 100*time.Millisecond, "All pending sync tasks should complete before cleanup")
}
// Shutdown the shared channel service to stop any async operations
_ = scsInterface.Shutdown()
// Wait for shutdown to complete with more time
require.Eventually(t, func() bool {
return !scsInterface.Active()
}, 5*time.Second, 100*time.Millisecond, "Shared channel service should be inactive after shutdown")
}
// Clear all shared channels and remotes from previous tests
allSharedChannels, _ := ss.SharedChannel().GetAll(0, 1000, model.SharedChannelFilterOpts{})
for _, sc := range allSharedChannels {
// Delete all remotes for this channel
remotes, _ := ss.SharedChannel().GetRemotes(0, 999999, model.SharedChannelRemoteFilterOpts{ChannelId: sc.ChannelId})
for _, remote := range remotes {
_, _ = ss.SharedChannel().DeleteRemote(remote.Id)
}
// Delete the shared channel
_, _ = ss.SharedChannel().Delete(sc.ChannelId)
}
// Delete all remote clusters
allRemoteClusters, _ := ss.RemoteCluster().GetAll(0, 1000, model.RemoteClusterQueryFilter{})
for _, rc := range allRemoteClusters {
_, _ = ss.RemoteCluster().Delete(rc.RemoteId)
}
// Clear all SharedChannelUsers sync state - this is critical for test isolation
// The SharedChannelUsers table tracks per-user sync timestamps that can interfere between tests
_, _ = th.SQLStore.GetMaster().Exec("DELETE FROM SharedChannelUsers WHERE 1=1")
// Clear all SharedChannelAttachments sync state
_, _ = th.SQLStore.GetMaster().Exec("DELETE FROM SharedChannelAttachments WHERE 1=1")
// Reset sync cursors in any remaining SharedChannelRemotes (before they get deleted)
// This ensures cursors don't persist if deletion fails
_, _ = th.SQLStore.GetMaster().Exec(`UPDATE SharedChannelRemotes SET
LastPostCreateAt = 0,
LastPostCreateId = '',
LastPostUpdateAt = 0,
LastPostId = '',
LastMembersSyncAt = 0
WHERE 1=1`)
// Remove all channel members from test channels (except the basic team/channel setup)
channels, _ := ss.Channel().GetAll(th.BasicTeam.Id)
for _, channel := range channels {
// Skip direct message and group channels, and skip the default channels
if channel.Type != model.ChannelTypeDirect && channel.Type != model.ChannelTypeGroup &&
channel.Id != th.BasicChannel.Id {
members, _ := ss.Channel().GetMembers(model.ChannelMembersGetOptions{
ChannelID: channel.Id,
})
for _, member := range members {
_ = ss.Channel().RemoveMember(th.Context, channel.Id, member.UserId)
}
}
}
// Remove all users from teams except the basic test users
teams, _ := ss.Team().GetAll()
for _, team := range teams {
if team.Id == th.BasicTeam.Id {
members, _ := ss.Team().GetMembers(team.Id, 0, 10000, nil)
for _, member := range members {
// Keep only the basic test users
if member.UserId != th.BasicUser.Id && member.UserId != th.BasicUser2.Id &&
member.UserId != th.SystemAdminUser.Id {
_ = ss.Team().RemoveMember(th.Context, team.Id, member.UserId)
}
}
}
}
// Get all active users and deactivate non-basic ones
options := &model.UserGetOptions{
Page: 0,
PerPage: 200,
Active: true,
}
users, _ := ss.User().GetAllProfiles(options)
for _, user := range users {
// Keep only the basic test users active
if user.Id != th.BasicUser.Id && user.Id != th.BasicUser2.Id &&
user.Id != th.SystemAdminUser.Id {
// Deactivate the user (soft delete)
user.DeleteAt = model.GetMillis()
_, _ = ss.User().Update(th.Context, user, true)
}
}
// Verify cleanup is complete
require.Eventually(t, func() bool {
sharedChannels, _ := ss.SharedChannel().GetAll(0, 1000, model.SharedChannelFilterOpts{})
remoteClusters, _ := ss.RemoteCluster().GetAll(0, 1000, model.RemoteClusterQueryFilter{})
return len(sharedChannels) == 0 && len(remoteClusters) == 0
}, 2*time.Second, 100*time.Millisecond, "Failed to clean up shared channels and remote clusters")
// Reset batch size to default to ensure test isolation
defaultBatchSize := 20
th.App.UpdateConfig(func(cfg *model.Config) {
cfg.ConnectedWorkspacesSettings.GlobalUserSyncBatchSize = &defaultBatchSize
})
// Restart services and ensure they are running and ready
if scsInterface != nil {
// Restart the shared channel service
_ = scsInterface.Start()
if scs, ok := scsInterface.(*sharedchannel.Service); ok {
require.Eventually(t, func() bool {
return scs.Active()
}, 5*time.Second, 100*time.Millisecond, "Shared channel service should be active after restart")
}
}
rcService := th.App.Srv().GetRemoteClusterService()
if rcService != nil {
if rc, ok := rcService.(*remotecluster.Service); ok {
rc.SetActive(true)
}
require.Eventually(t, func() bool {
return rcService.Active()
}, 5*time.Second, 100*time.Millisecond, "Remote cluster service should be active")
}
}