mirror of
https://github.com/empayre/fleet.git
synced 2024-11-06 08:55:24 +00:00
e1bbcfcfda
* Generate audit logs for activities * Fix config tests * Fix TestGetConfig/IncludeServerConfig * Fix use of AddAttributes in results only * Stream activities asynchronously * Fix index and add logging * Revert change * Documentation fixes
2814 lines
86 KiB
Go
2814 lines
86 KiB
Go
package service
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"reflect"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/WatchBeam/clock"
|
|
"github.com/fleetdm/fleet/v4/server/authz"
|
|
"github.com/fleetdm/fleet/v4/server/config"
|
|
hostctx "github.com/fleetdm/fleet/v4/server/contexts/host"
|
|
fleetLogging "github.com/fleetdm/fleet/v4/server/contexts/logging"
|
|
"github.com/fleetdm/fleet/v4/server/contexts/viewer"
|
|
"github.com/fleetdm/fleet/v4/server/datastore/mysqlredis"
|
|
"github.com/fleetdm/fleet/v4/server/datastore/redis/redistest"
|
|
"github.com/fleetdm/fleet/v4/server/fleet"
|
|
"github.com/fleetdm/fleet/v4/server/live_query/live_query_mock"
|
|
"github.com/fleetdm/fleet/v4/server/mock"
|
|
mockresult "github.com/fleetdm/fleet/v4/server/mock/mockresult"
|
|
"github.com/fleetdm/fleet/v4/server/ptr"
|
|
"github.com/fleetdm/fleet/v4/server/pubsub"
|
|
"github.com/fleetdm/fleet/v4/server/service/async"
|
|
"github.com/fleetdm/fleet/v4/server/service/osquery_utils"
|
|
"github.com/fleetdm/fleet/v4/server/service/redis_policy_set"
|
|
"github.com/go-kit/kit/log"
|
|
"github.com/go-kit/kit/log/level"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestGetClientConfig(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
ds.ListPacksForHostFunc = func(ctx context.Context, hid uint) ([]*fleet.Pack, error) {
|
|
return []*fleet.Pack{}, nil
|
|
}
|
|
ds.ListScheduledQueriesInPackFunc = func(ctx context.Context, pid uint) (fleet.ScheduledQueryList, error) {
|
|
tru := true
|
|
fals := false
|
|
fortytwo := uint(42)
|
|
switch pid {
|
|
case 1:
|
|
return []*fleet.ScheduledQuery{
|
|
{Name: "time", Query: "select * from time", Interval: 30, Removed: &fals},
|
|
}, nil
|
|
case 4:
|
|
return []*fleet.ScheduledQuery{
|
|
{Name: "foobar", Query: "select 3", Interval: 20, Shard: &fortytwo},
|
|
{Name: "froobing", Query: "select 'guacamole'", Interval: 60, Snapshot: &tru},
|
|
}, nil
|
|
default:
|
|
return []*fleet.ScheduledQuery{}, nil
|
|
}
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{AgentOptions: ptr.RawMessage(json.RawMessage(`{"config":{"options":{"baz":"bar"}}}`))}, nil
|
|
}
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
return nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
if id != 1 && id != 2 {
|
|
return nil, errors.New("not found")
|
|
}
|
|
return &fleet.Host{ID: id}, nil
|
|
}
|
|
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
ctx1 := hostctx.NewContext(ctx, &fleet.Host{ID: 1})
|
|
ctx2 := hostctx.NewContext(ctx, &fleet.Host{ID: 2})
|
|
|
|
expectedOptions := map[string]interface{}{
|
|
"baz": "bar",
|
|
}
|
|
|
|
expectedConfig := map[string]interface{}{
|
|
"options": expectedOptions,
|
|
}
|
|
|
|
// No packs loaded yet
|
|
conf, err := svc.GetClientConfig(ctx1)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, expectedConfig, conf)
|
|
|
|
conf, err = svc.GetClientConfig(ctx2)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, expectedConfig, conf)
|
|
|
|
// Now add packs
|
|
ds.ListPacksForHostFunc = func(ctx context.Context, hid uint) ([]*fleet.Pack, error) {
|
|
switch hid {
|
|
case 1:
|
|
return []*fleet.Pack{
|
|
{ID: 1, Name: "pack_by_label"},
|
|
{ID: 4, Name: "pack_by_other_label"},
|
|
}, nil
|
|
|
|
case 2:
|
|
return []*fleet.Pack{
|
|
{ID: 1, Name: "pack_by_label"},
|
|
}, nil
|
|
}
|
|
return []*fleet.Pack{}, nil
|
|
}
|
|
|
|
conf, err = svc.GetClientConfig(ctx1)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, expectedOptions, conf["options"])
|
|
assert.JSONEq(t, `{
|
|
"pack_by_other_label": {
|
|
"queries": {
|
|
"foobar":{"query":"select 3","interval":20,"shard":42},
|
|
"froobing":{"query":"select 'guacamole'","interval":60,"snapshot":true}
|
|
}
|
|
},
|
|
"pack_by_label": {
|
|
"queries":{
|
|
"time":{"query":"select * from time","interval":30,"removed":false}
|
|
}
|
|
}
|
|
}`,
|
|
string(conf["packs"].(json.RawMessage)),
|
|
)
|
|
|
|
conf, err = svc.GetClientConfig(ctx2)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, expectedOptions, conf["options"])
|
|
assert.JSONEq(t, `{
|
|
"pack_by_label": {
|
|
"queries":{
|
|
"time":{"query":"select * from time","interval":30,"removed":false}
|
|
}
|
|
}
|
|
}`,
|
|
string(conf["packs"].(json.RawMessage)),
|
|
)
|
|
}
|
|
|
|
func TestAgentOptionsForHost(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
teamID := uint(1)
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{
|
|
AgentOptions: ptr.RawMessage(json.RawMessage(`{"config":{"baz":"bar"},"overrides":{"platforms":{"darwin":{"foo":"override2"}}}}`)),
|
|
}, nil
|
|
}
|
|
ds.TeamAgentOptionsFunc = func(ctx context.Context, id uint) (*json.RawMessage, error) {
|
|
return ptr.RawMessage(json.RawMessage(`{"config":{"foo":"bar"},"overrides":{"platforms":{"darwin":{"foo":"override"}}}}`)), nil
|
|
}
|
|
|
|
host := &fleet.Host{
|
|
TeamID: &teamID,
|
|
Platform: "darwin",
|
|
}
|
|
|
|
opt, err := svc.AgentOptionsForHost(ctx, host.TeamID, host.Platform)
|
|
require.NoError(t, err)
|
|
assert.JSONEq(t, `{"foo":"override"}`, string(opt))
|
|
|
|
host.Platform = "windows"
|
|
opt, err = svc.AgentOptionsForHost(ctx, host.TeamID, host.Platform)
|
|
require.NoError(t, err)
|
|
assert.JSONEq(t, `{"foo":"bar"}`, string(opt))
|
|
|
|
// Should take gobal option with no team
|
|
host.TeamID = nil
|
|
opt, err = svc.AgentOptionsForHost(ctx, host.TeamID, host.Platform)
|
|
require.NoError(t, err)
|
|
assert.JSONEq(t, `{"baz":"bar"}`, string(opt))
|
|
|
|
host.Platform = "darwin"
|
|
opt, err = svc.AgentOptionsForHost(ctx, host.TeamID, host.Platform)
|
|
require.NoError(t, err)
|
|
assert.JSONEq(t, `{"foo":"override2"}`, string(opt))
|
|
}
|
|
|
|
var allDetailQueries = osquery_utils.GetDetailQueries(config.FleetConfig{Vulnerabilities: config.VulnerabilitiesConfig{DisableWinOSVulnerabilities: true}}, &fleet.Features{EnableHostUsers: true})
|
|
|
|
func expectedDetailQueriesForPlatform(platform string) map[string]osquery_utils.DetailQuery {
|
|
queries := make(map[string]osquery_utils.DetailQuery)
|
|
for k, v := range allDetailQueries {
|
|
if v.RunsForPlatform(platform) {
|
|
queries[k] = v
|
|
}
|
|
}
|
|
return queries
|
|
}
|
|
|
|
func TestEnrollAgent(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
ds.VerifyEnrollSecretFunc = func(ctx context.Context, secret string) (*fleet.EnrollSecret, error) {
|
|
switch secret {
|
|
case "valid_secret":
|
|
return &fleet.EnrollSecret{Secret: "valid_secret", TeamID: ptr.Uint(3)}, nil
|
|
default:
|
|
return nil, errors.New("not found")
|
|
}
|
|
}
|
|
ds.EnrollHostFunc = func(ctx context.Context, osqueryHostId, nodeKey string, teamID *uint, cooldown time.Duration) (*fleet.Host, error) {
|
|
assert.Equal(t, ptr.Uint(3), teamID)
|
|
return &fleet.Host{
|
|
OsqueryHostID: osqueryHostId, NodeKey: nodeKey,
|
|
}, nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
nodeKey, err := svc.EnrollAgent(ctx, "valid_secret", "host123", nil)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, nodeKey)
|
|
}
|
|
|
|
func TestEnrollAgentEnforceLimit(t *testing.T) {
|
|
runTest := func(t *testing.T, pool fleet.RedisPool) {
|
|
const maxHosts = 2
|
|
|
|
var hostIDSeq uint
|
|
ds := new(mock.Store)
|
|
ds.VerifyEnrollSecretFunc = func(ctx context.Context, secret string) (*fleet.EnrollSecret, error) {
|
|
switch secret {
|
|
case "valid_secret":
|
|
return &fleet.EnrollSecret{Secret: "valid_secret"}, nil
|
|
default:
|
|
return nil, errors.New("not found")
|
|
}
|
|
}
|
|
ds.EnrollHostFunc = func(ctx context.Context, osqueryHostId, nodeKey string, teamID *uint, cooldown time.Duration) (*fleet.Host, error) {
|
|
hostIDSeq++
|
|
return &fleet.Host{
|
|
ID: hostIDSeq, OsqueryHostID: osqueryHostId, NodeKey: nodeKey,
|
|
}, nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
return &fleet.Host{ID: id}, nil
|
|
}
|
|
ds.DeleteHostFunc = func(ctx context.Context, id uint) error {
|
|
return nil
|
|
}
|
|
|
|
redisWrapDS := mysqlredis.New(ds, pool, mysqlredis.WithEnforcedHostLimit(maxHosts))
|
|
svc, ctx := newTestService(t, redisWrapDS, nil, nil, &TestServerOpts{
|
|
EnrollHostLimiter: redisWrapDS,
|
|
License: &fleet.LicenseInfo{DeviceCount: maxHosts},
|
|
})
|
|
ctx = viewer.NewContext(ctx, viewer.Viewer{
|
|
User: &fleet.User{
|
|
ID: 0,
|
|
GlobalRole: ptr.String(fleet.RoleAdmin),
|
|
},
|
|
})
|
|
|
|
nodeKey, err := svc.EnrollAgent(ctx, "valid_secret", "host001", nil)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, nodeKey)
|
|
|
|
nodeKey, err = svc.EnrollAgent(ctx, "valid_secret", "host002", nil)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, nodeKey)
|
|
|
|
_, err = svc.EnrollAgent(ctx, "valid_secret", "host003", nil)
|
|
require.Error(t, err)
|
|
require.Contains(t, err.Error(), fmt.Sprintf("maximum number of hosts reached: %d", maxHosts))
|
|
|
|
// delete a host with id 1
|
|
err = svc.DeleteHost(ctx, 1)
|
|
require.NoError(t, err)
|
|
|
|
// now host 003 can be enrolled
|
|
nodeKey, err = svc.EnrollAgent(ctx, "valid_secret", "host003", nil)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, nodeKey)
|
|
}
|
|
|
|
t.Run("standalone", func(t *testing.T) {
|
|
pool := redistest.SetupRedis(t, "enrolled_hosts:*", false, false, false)
|
|
runTest(t, pool)
|
|
})
|
|
|
|
t.Run("cluster", func(t *testing.T) {
|
|
pool := redistest.SetupRedis(t, "enrolled_hosts:*", true, true, false)
|
|
runTest(t, pool)
|
|
})
|
|
}
|
|
|
|
func TestEnrollAgentIncorrectEnrollSecret(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
ds.VerifyEnrollSecretFunc = func(ctx context.Context, secret string) (*fleet.EnrollSecret, error) {
|
|
switch secret {
|
|
case "valid_secret":
|
|
return &fleet.EnrollSecret{Secret: "valid_secret", TeamID: ptr.Uint(3)}, nil
|
|
default:
|
|
return nil, errors.New("not found")
|
|
}
|
|
}
|
|
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
nodeKey, err := svc.EnrollAgent(ctx, "not_correct", "host123", nil)
|
|
assert.NotNil(t, err)
|
|
assert.Empty(t, nodeKey)
|
|
}
|
|
|
|
func TestEnrollAgentDetails(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
ds.VerifyEnrollSecretFunc = func(ctx context.Context, secret string) (*fleet.EnrollSecret, error) {
|
|
return &fleet.EnrollSecret{}, nil
|
|
}
|
|
ds.EnrollHostFunc = func(ctx context.Context, osqueryHostId, nodeKey string, teamID *uint, cooldown time.Duration) (*fleet.Host, error) {
|
|
return &fleet.Host{
|
|
OsqueryHostID: osqueryHostId, NodeKey: nodeKey,
|
|
}, nil
|
|
}
|
|
var gotHost *fleet.Host
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
gotHost = host
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
details := map[string](map[string]string){
|
|
"osquery_info": {"version": "2.12.0"},
|
|
"system_info": {"hostname": "zwass.local", "uuid": "froobling_uuid"},
|
|
"os_version": {
|
|
"name": "Mac OS X",
|
|
"major": "10",
|
|
"minor": "14",
|
|
"patch": "5",
|
|
"platform": "darwin",
|
|
},
|
|
"foo": {"foo": "bar"},
|
|
}
|
|
nodeKey, err := svc.EnrollAgent(ctx, "", "host123", details)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, nodeKey)
|
|
|
|
assert.Equal(t, "Mac OS X 10.14.5", gotHost.OSVersion)
|
|
assert.Equal(t, "darwin", gotHost.Platform)
|
|
assert.Equal(t, "2.12.0", gotHost.OsqueryVersion)
|
|
assert.Equal(t, "zwass.local", gotHost.Hostname)
|
|
assert.Equal(t, "froobling_uuid", gotHost.UUID)
|
|
}
|
|
|
|
func TestAuthenticateHost(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
task := async.NewTask(ds, nil, clock.C, config.OsqueryConfig{})
|
|
svc, ctx := newTestService(t, ds, nil, nil, &TestServerOpts{Task: task})
|
|
|
|
var gotKey string
|
|
host := fleet.Host{ID: 1, Hostname: "foobar"}
|
|
ds.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
gotKey = nodeKey
|
|
return &host, nil
|
|
}
|
|
var gotHostIDs []uint
|
|
ds.MarkHostsSeenFunc = func(ctx context.Context, hostIDs []uint, t time.Time) error {
|
|
gotHostIDs = hostIDs
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
_, _, err := svc.AuthenticateHost(ctx, "test")
|
|
require.NoError(t, err)
|
|
assert.Equal(t, "test", gotKey)
|
|
assert.False(t, ds.MarkHostsSeenFuncInvoked)
|
|
|
|
host = fleet.Host{ID: 7, Hostname: "foobar"}
|
|
_, _, err = svc.AuthenticateHost(ctx, "floobar")
|
|
require.NoError(t, err)
|
|
assert.Equal(t, "floobar", gotKey)
|
|
assert.False(t, ds.MarkHostsSeenFuncInvoked)
|
|
// Host checks in twice
|
|
host = fleet.Host{ID: 7, Hostname: "foobar"}
|
|
_, _, err = svc.AuthenticateHost(ctx, "floobar")
|
|
require.NoError(t, err)
|
|
assert.Equal(t, "floobar", gotKey)
|
|
assert.False(t, ds.MarkHostsSeenFuncInvoked)
|
|
|
|
err = task.FlushHostsLastSeen(ctx, time.Now())
|
|
require.NoError(t, err)
|
|
assert.True(t, ds.MarkHostsSeenFuncInvoked)
|
|
ds.MarkHostsSeenFuncInvoked = false
|
|
assert.ElementsMatch(t, []uint{1, 7}, gotHostIDs)
|
|
|
|
err = task.FlushHostsLastSeen(ctx, time.Now())
|
|
require.NoError(t, err)
|
|
assert.True(t, ds.MarkHostsSeenFuncInvoked)
|
|
require.Len(t, gotHostIDs, 0)
|
|
}
|
|
|
|
func TestAuthenticateHostFailure(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
ds.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
return nil, errors.New("not found")
|
|
}
|
|
|
|
_, _, err := svc.AuthenticateHost(ctx, "test")
|
|
require.NotNil(t, err)
|
|
}
|
|
|
|
type testJSONLogger struct {
|
|
logs []json.RawMessage
|
|
}
|
|
|
|
func (n *testJSONLogger) Write(ctx context.Context, logs []json.RawMessage) error {
|
|
n.logs = logs
|
|
return nil
|
|
}
|
|
|
|
func TestSubmitStatusLogs(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
// Hack to get at the service internals and modify the writer
|
|
serv := ((svc.(validationMiddleware)).Service).(*Service)
|
|
|
|
testLogger := &testJSONLogger{}
|
|
serv.osqueryLogWriter = &OsqueryLogger{Status: testLogger}
|
|
|
|
logs := []string{
|
|
`{"severity":"0","filename":"tls.cpp","line":"216","message":"some message","version":"1.8.2","decorations":{"host_uuid":"uuid_foobar","username":"zwass"}}`,
|
|
`{"severity":"1","filename":"buffered.cpp","line":"122","message":"warning!","version":"1.8.2","decorations":{"host_uuid":"uuid_foobar","username":"zwass"}}`,
|
|
}
|
|
logJSON := fmt.Sprintf("[%s]", strings.Join(logs, ","))
|
|
|
|
var status []json.RawMessage
|
|
err := json.Unmarshal([]byte(logJSON), &status)
|
|
require.NoError(t, err)
|
|
|
|
host := fleet.Host{}
|
|
ctx = hostctx.NewContext(ctx, &host)
|
|
err = serv.SubmitStatusLogs(ctx, status)
|
|
require.NoError(t, err)
|
|
|
|
assert.Equal(t, status, testLogger.logs)
|
|
}
|
|
|
|
func TestSubmitResultLogs(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
// Hack to get at the service internals and modify the writer
|
|
serv := ((svc.(validationMiddleware)).Service).(*Service)
|
|
|
|
testLogger := &testJSONLogger{}
|
|
serv.osqueryLogWriter = &OsqueryLogger{Result: testLogger}
|
|
|
|
logs := []string{
|
|
`{"name":"system_info","hostIdentifier":"some_uuid","calendarTime":"Fri Sep 30 17:55:15 2016 UTC","unixTime":"1475258115","decorations":{"host_uuid":"some_uuid","username":"zwass"},"columns":{"cpu_brand":"Intel(R) Core(TM) i7-4770HQ CPU @ 2.20GHz","hostname":"hostimus","physical_memory":"17179869184"},"action":"added"}`,
|
|
`{"name":"encrypted","hostIdentifier":"some_uuid","calendarTime":"Fri Sep 30 21:19:15 2016 UTC","unixTime":"1475270355","decorations":{"host_uuid":"4740D59F-699E-5B29-960B-979AAF9BBEEB","username":"zwass"},"columns":{"encrypted":"1","name":"\/dev\/disk1","type":"AES-XTS","uid":"","user_uuid":"","uuid":"some_uuid"},"action":"added"}`,
|
|
`{"snapshot":[{"hour":"20","minutes":"8"}],"action":"snapshot","name":"time","hostIdentifier":"1379f59d98f4","calendarTime":"Tue Jan 10 20:08:51 2017 UTC","unixTime":"1484078931","decorations":{"host_uuid":"EB714C9D-C1F8-A436-B6DA-3F853C5502EA"}}`,
|
|
`{"diffResults":{"removed":[{"address":"127.0.0.1","hostnames":"kl.groob.io"}],"added":""},"name":"pack\/test\/hosts","hostIdentifier":"FA01680E-98CA-5557-8F59-7716ECFEE964","calendarTime":"Sun Nov 19 00:02:08 2017 UTC","unixTime":"1511049728","epoch":"0","counter":"10","decorations":{"host_uuid":"FA01680E-98CA-5557-8F59-7716ECFEE964","hostname":"kl.groob.io"}}`,
|
|
// fleet will accept anything in the "data" field of a log request.
|
|
`{"unknown":{"foo": [] }}`,
|
|
}
|
|
logJSON := fmt.Sprintf("[%s]", strings.Join(logs, ","))
|
|
|
|
var results []json.RawMessage
|
|
err := json.Unmarshal([]byte(logJSON), &results)
|
|
require.NoError(t, err)
|
|
|
|
host := fleet.Host{}
|
|
ctx = hostctx.NewContext(ctx, &host)
|
|
err = serv.SubmitResultLogs(ctx, results)
|
|
require.NoError(t, err)
|
|
|
|
assert.Equal(t, results, testLogger.logs)
|
|
}
|
|
|
|
func verifyDiscovery(t *testing.T, queries, discovery map[string]string) {
|
|
assert.Equal(t, len(queries), len(discovery))
|
|
// discoveryUsed holds the queries where we know use the distributed discovery feature.
|
|
discoveryUsed := map[string]struct{}{
|
|
hostDetailQueryPrefix + "google_chrome_profiles": {},
|
|
hostDetailQueryPrefix + "mdm": {},
|
|
hostDetailQueryPrefix + "munki_info": {},
|
|
hostDetailQueryPrefix + "windows_update_history": {},
|
|
hostDetailQueryPrefix + "kubequery_info": {},
|
|
hostDetailQueryPrefix + "orbit_info": {},
|
|
}
|
|
for name := range queries {
|
|
require.NotEmpty(t, discovery[name])
|
|
if _, ok := discoveryUsed[name]; ok {
|
|
require.NotEqual(t, alwaysTrueQuery, discovery[name])
|
|
} else {
|
|
require.Equal(t, alwaysTrueQuery, discovery[name])
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestHostDetailQueries(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
additional := json.RawMessage(`{"foobar": "select foo", "bim": "bam"}`)
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{AdditionalQueries: &additional, EnableHostUsers: true}}, nil
|
|
}
|
|
|
|
mockClock := clock.NewMockClock()
|
|
host := fleet.Host{
|
|
ID: 1,
|
|
UpdateCreateTimestamps: fleet.UpdateCreateTimestamps{
|
|
UpdateTimestamp: fleet.UpdateTimestamp{
|
|
UpdatedAt: mockClock.Now(),
|
|
},
|
|
CreateTimestamp: fleet.CreateTimestamp{
|
|
CreatedAt: mockClock.Now(),
|
|
},
|
|
},
|
|
|
|
Platform: "darwin",
|
|
DetailUpdatedAt: mockClock.Now(),
|
|
NodeKey: "test_key",
|
|
Hostname: "test_hostname",
|
|
UUID: "test_uuid",
|
|
}
|
|
|
|
svc := &Service{
|
|
clock: mockClock,
|
|
logger: log.NewNopLogger(),
|
|
config: config.TestConfig(),
|
|
ds: ds,
|
|
jitterMu: new(sync.Mutex),
|
|
jitterH: make(map[time.Duration]*jitterHashTable),
|
|
}
|
|
|
|
queries, discovery, err := svc.detailQueriesForHost(context.Background(), &host)
|
|
require.NoError(t, err)
|
|
assert.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
|
|
// With refetch requested detail queries should be returned
|
|
host.RefetchRequested = true
|
|
queries, discovery, err = svc.detailQueriesForHost(context.Background(), &host)
|
|
require.NoError(t, err)
|
|
assert.NotEmpty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
host.RefetchRequested = false
|
|
|
|
// Advance the time
|
|
mockClock.AddTime(1*time.Hour + 1*time.Minute)
|
|
|
|
queries, discovery, err = svc.detailQueriesForHost(context.Background(), &host)
|
|
require.NoError(t, err)
|
|
// +2: additional queries: bim, foobar
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+2, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
for name := range queries {
|
|
assert.True(t,
|
|
strings.HasPrefix(name, hostDetailQueryPrefix) || strings.HasPrefix(name, hostAdditionalQueryPrefix),
|
|
)
|
|
}
|
|
assert.Equal(t, "bam", queries[hostAdditionalQueryPrefix+"bim"])
|
|
assert.Equal(t, "select foo", queries[hostAdditionalQueryPrefix+"foobar"])
|
|
}
|
|
|
|
func TestQueriesAndHostFeatures(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
team1 := fleet.Team{
|
|
ID: 1,
|
|
Config: fleet.TeamConfig{
|
|
Features: fleet.Features{
|
|
EnableHostUsers: true,
|
|
EnableSoftwareInventory: false,
|
|
},
|
|
},
|
|
}
|
|
|
|
team2 := fleet.Team{
|
|
ID: 2,
|
|
Config: fleet.TeamConfig{
|
|
Features: fleet.Features{
|
|
EnableHostUsers: false,
|
|
EnableSoftwareInventory: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
host := fleet.Host{
|
|
ID: 1,
|
|
Platform: "darwin",
|
|
NodeKey: "test_key",
|
|
Hostname: "test_hostname",
|
|
UUID: "test_uuid",
|
|
TeamID: nil,
|
|
}
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{
|
|
Features: fleet.Features{
|
|
EnableHostUsers: false,
|
|
EnableSoftwareInventory: false,
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
ds.TeamFeaturesFunc = func(ctx context.Context, id uint) (*fleet.Features, error) {
|
|
switch id {
|
|
case uint(1):
|
|
return &team1.Config.Features, nil
|
|
case uint(2):
|
|
return &team2.Config.Features, nil
|
|
default:
|
|
return nil, errors.New("team not found")
|
|
}
|
|
}
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
|
|
lq := live_query_mock.New(t)
|
|
lq.On("QueriesForHost", uint(1)).Return(map[string]string{}, nil)
|
|
lq.On("QueriesForHost", uint(2)).Return(map[string]string{}, nil)
|
|
lq.On("QueriesForHost", nil).Return(map[string]string{}, nil)
|
|
|
|
t.Run("free license", func(t *testing.T) {
|
|
license := &fleet.LicenseInfo{Tier: fleet.TierFree}
|
|
svc, ctx := newTestService(t, ds, nil, lq, &TestServerOpts{License: license})
|
|
|
|
ctx = hostctx.NewContext(ctx, &host)
|
|
queries, _, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.NotContains(t, queries, "fleet_detail_query_users")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_macos")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_linux")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_windows")
|
|
|
|
// assign team 1 to host
|
|
host.TeamID = &team1.ID
|
|
queries, _, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.NotContains(t, queries, "fleet_detail_query_users")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_macos")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_linux")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_windows")
|
|
|
|
// assign team 2 to host
|
|
host.TeamID = &team2.ID
|
|
queries, _, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.NotContains(t, queries, "fleet_detail_query_users")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_macos")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_linux")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_windows")
|
|
})
|
|
|
|
t.Run("premium license", func(t *testing.T) {
|
|
license := &fleet.LicenseInfo{Tier: fleet.TierPremium}
|
|
svc, ctx := newTestService(t, ds, nil, lq, &TestServerOpts{License: license})
|
|
|
|
host.TeamID = nil
|
|
ctx = hostctx.NewContext(ctx, &host)
|
|
queries, _, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.NotContains(t, queries, "fleet_detail_query_users")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_macos")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_linux")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_windows")
|
|
|
|
// assign team 1 to host
|
|
host.TeamID = &team1.ID
|
|
queries, _, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Contains(t, queries, "fleet_detail_query_users")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_macos")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_linux")
|
|
require.NotContains(t, queries, "fleet_detail_query_software_windows")
|
|
|
|
// assign team 2 to host
|
|
host.TeamID = &team2.ID
|
|
queries, _, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.NotContains(t, queries, "fleet_detail_query_users")
|
|
require.Contains(t, queries, "fleet_detail_query_software_macos")
|
|
})
|
|
}
|
|
|
|
func TestGetDistributedQueriesMissingHost(t *testing.T) {
|
|
svc, ctx := newTestService(t, &mock.Store{}, nil, nil)
|
|
|
|
_, _, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NotNil(t, err)
|
|
assert.Contains(t, err.Error(), "missing host")
|
|
}
|
|
|
|
func TestLabelQueries(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
lq := live_query_mock.New(t)
|
|
svc, ctx := newTestServiceWithClock(t, ds, nil, lq, mockClock)
|
|
|
|
host := &fleet.Host{
|
|
Platform: "darwin",
|
|
}
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
return host, nil
|
|
}
|
|
ds.UpdateHostFunc = func(ctx context.Context, gotHost *fleet.Host) error {
|
|
host = gotHost
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true}}, nil
|
|
}
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
|
|
lq.On("QueriesForHost", uint(0)).Return(map[string]string{}, nil)
|
|
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
// With a new host, we should get the detail queries (and accelerate
|
|
// should be turned on so that we can quickly fill labels)
|
|
queries, discovery, acc, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.NotZero(t, acc)
|
|
|
|
// Simulate the detail queries being added.
|
|
host.DetailUpdatedAt = mockClock.Now().Add(-1 * time.Minute)
|
|
host.Hostname = "zwass.local"
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{
|
|
"label1": "query1",
|
|
"label2": "query2",
|
|
"label3": "query3",
|
|
}, nil
|
|
}
|
|
|
|
// Now we should get the label queries
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Len(t, queries, 3)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
var gotHost *fleet.Host
|
|
var gotResults map[uint]*bool
|
|
var gotTime time.Time
|
|
ds.RecordLabelQueryExecutionsFunc = func(ctx context.Context, host *fleet.Host, results map[uint]*bool, t time.Time, deferred bool) error {
|
|
gotHost = host
|
|
gotResults = results
|
|
gotTime = t
|
|
return nil
|
|
}
|
|
|
|
// Record a query execution
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostLabelQueryPrefix + "1": {{"col1": "val1"}},
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
host.LabelUpdatedAt = mockClock.Now()
|
|
assert.Equal(t, host, gotHost)
|
|
assert.Equal(t, mockClock.Now(), gotTime)
|
|
require.Len(t, gotResults, 1)
|
|
assert.Equal(t, true, *gotResults[1])
|
|
|
|
mockClock.AddTime(1 * time.Second)
|
|
|
|
// Record a query execution
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostLabelQueryPrefix + "2": {{"col1": "val1"}},
|
|
hostLabelQueryPrefix + "3": {},
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
host.LabelUpdatedAt = mockClock.Now()
|
|
assert.Equal(t, host, gotHost)
|
|
assert.Equal(t, mockClock.Now(), gotTime)
|
|
require.Len(t, gotResults, 2)
|
|
assert.Equal(t, true, *gotResults[2])
|
|
assert.Equal(t, false, *gotResults[3])
|
|
|
|
// We should get no labels now.
|
|
host.LabelUpdatedAt = mockClock.Now()
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
// With refetch requested details+label queries should be returned.
|
|
host.RefetchRequested = true
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +3 for label queries
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+3, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
// Record a query execution
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostLabelQueryPrefix + "2": {{"col1": "val1"}},
|
|
hostLabelQueryPrefix + "3": {},
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
host.LabelUpdatedAt = mockClock.Now()
|
|
assert.Equal(t, host, gotHost)
|
|
assert.Equal(t, mockClock.Now(), gotTime)
|
|
require.Len(t, gotResults, 2)
|
|
assert.Equal(t, true, *gotResults[2])
|
|
assert.Equal(t, false, *gotResults[3])
|
|
|
|
// SubmitDistributedQueryResults will set RefetchRequested to false.
|
|
require.False(t, host.RefetchRequested)
|
|
|
|
// There shouldn't be any labels now.
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
}
|
|
|
|
func TestDetailQueriesWithEmptyStrings(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
mockClock := clock.NewMockClock()
|
|
lq := live_query_mock.New(t)
|
|
svc, ctx := newTestServiceWithClock(t, ds, nil, lq, mockClock)
|
|
|
|
host := &fleet.Host{
|
|
ID: 1,
|
|
Platform: "windows",
|
|
}
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true}}, nil
|
|
}
|
|
ds.LabelQueriesForHostFunc = func(context.Context, *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
if id != 1 {
|
|
return nil, errors.New("not found")
|
|
}
|
|
return host, nil
|
|
}
|
|
|
|
lq.On("QueriesForHost", host.ID).Return(map[string]string{}, nil)
|
|
|
|
// With a new host, we should get the detail queries (and accelerated
|
|
// queries)
|
|
queries, discovery, acc, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +1 due to 'windows_update_history'
|
|
if expected := expectedDetailQueriesForPlatform(host.Platform); !assert.Equal(t, len(expected)+1, len(queries)) {
|
|
// this is just to print the diff between the expected and actual query
|
|
// keys when the count assertion fails, to help debugging - they are not
|
|
// expected to match.
|
|
require.ElementsMatch(t, osqueryMapKeys(expected), distQueriesMapKeys(queries))
|
|
}
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.NotZero(t, acc)
|
|
|
|
resultJSON := `
|
|
{
|
|
"fleet_detail_query_network_interface_windows": [
|
|
{
|
|
"address": "192.168.0.1",
|
|
"mac": "5f:3d:4b:10:25:82"
|
|
}
|
|
],
|
|
"fleet_detail_query_os_version": [
|
|
{
|
|
"platform": "darwin",
|
|
"build": "15G1004",
|
|
"major": "10",
|
|
"minor": "10",
|
|
"name": "Mac OS X",
|
|
"patch": "6"
|
|
}
|
|
],
|
|
"fleet_detail_query_osquery_info": [
|
|
{
|
|
"build_distro": "10.10",
|
|
"build_platform": "darwin",
|
|
"config_hash": "3c6e4537c4d0eb71a7c6dda19d",
|
|
"config_valid": "1",
|
|
"extensions": "active",
|
|
"pid": "38113",
|
|
"start_time": "1475603155",
|
|
"version": "1.8.2",
|
|
"watcher": "38112"
|
|
}
|
|
],
|
|
"fleet_detail_query_system_info": [
|
|
{
|
|
"computer_name": "computer",
|
|
"cpu_brand": "Intel(R) Core(TM) i7-4770HQ CPU @ 2.20GHz",
|
|
"cpu_logical_cores": "8",
|
|
"cpu_physical_cores": "4",
|
|
"cpu_subtype": "Intel x86-64h Haswell",
|
|
"cpu_type": "x86_64h",
|
|
"hardware_model": "MacBookPro11,4",
|
|
"hardware_serial": "ABCDEFGH",
|
|
"hardware_vendor": "Apple Inc.",
|
|
"hardware_version": "1.0",
|
|
"hostname": "computer.local",
|
|
"physical_memory": "17179869184",
|
|
"uuid": "uuid"
|
|
}
|
|
],
|
|
"fleet_detail_query_uptime": [
|
|
{
|
|
"days": "20",
|
|
"hours": "0",
|
|
"minutes": "48",
|
|
"seconds": "13",
|
|
"total_seconds": "1730893"
|
|
}
|
|
],
|
|
"fleet_detail_query_osquery_flags": [
|
|
{
|
|
"name": "config_tls_refresh",
|
|
"value": ""
|
|
},
|
|
{
|
|
"name": "distributed_interval",
|
|
"value": ""
|
|
},
|
|
{
|
|
"name": "logger_tls_period",
|
|
"value": ""
|
|
}
|
|
],
|
|
"fleet_detail_query_orbit_info": [
|
|
{
|
|
"name": "version",
|
|
"value": "42"
|
|
},
|
|
{
|
|
"name": "device_auth_token",
|
|
"value": "foo"
|
|
}
|
|
]
|
|
}
|
|
`
|
|
|
|
var results fleet.OsqueryDistributedQueryResults
|
|
err = json.Unmarshal([]byte(resultJSON), &results)
|
|
require.NoError(t, err)
|
|
|
|
var gotHost *fleet.Host
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
gotHost = host
|
|
return nil
|
|
}
|
|
|
|
// Verify that results are ingested properly
|
|
require.NoError(t, svc.SubmitDistributedQueryResults(ctx, results, map[string]fleet.OsqueryStatus{}, map[string]string{}))
|
|
|
|
// osquery_info
|
|
assert.Equal(t, "darwin", gotHost.Platform)
|
|
assert.Equal(t, "1.8.2", gotHost.OsqueryVersion)
|
|
|
|
// system_info
|
|
assert.Equal(t, int64(17179869184), gotHost.Memory)
|
|
assert.Equal(t, "computer.local", gotHost.Hostname)
|
|
assert.Equal(t, "uuid", gotHost.UUID)
|
|
|
|
// os_version
|
|
assert.Equal(t, "Mac OS X 10.10.6", gotHost.OSVersion)
|
|
|
|
// uptime
|
|
assert.Equal(t, 1730893*time.Second, gotHost.Uptime)
|
|
|
|
// osquery_flags
|
|
assert.Equal(t, uint(0), gotHost.ConfigTLSRefresh)
|
|
assert.Equal(t, uint(0), gotHost.DistributedInterval)
|
|
assert.Equal(t, uint(0), gotHost.LoggerTLSPeriod)
|
|
|
|
host.Hostname = "computer.local"
|
|
host.DetailUpdatedAt = mockClock.Now()
|
|
mockClock.AddTime(1 * time.Minute)
|
|
|
|
// Now no detail queries should be required
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
// Advance clock and queries should exist again
|
|
mockClock.AddTime(1*time.Hour + 1*time.Minute)
|
|
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// somehow confusingly, the query response above changed the host's platform
|
|
// from windows to darwin
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(gotHost.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
}
|
|
|
|
func TestDetailQueries(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
mockClock := clock.NewMockClock()
|
|
lq := live_query_mock.New(t)
|
|
svc, ctx := newTestServiceWithClock(t, ds, nil, lq, mockClock)
|
|
|
|
host := &fleet.Host{
|
|
ID: 1,
|
|
Platform: "linux",
|
|
}
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
lq.On("QueriesForHost", host.ID).Return(map[string]string{}, nil)
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true, EnableSoftwareInventory: true}}, nil
|
|
}
|
|
ds.LabelQueriesForHostFunc = func(context.Context, *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.SetOrUpdateMDMDataFunc = func(ctx context.Context, hostID uint, isServer, enrolled bool, serverURL string, installedFromDep bool, name string) error {
|
|
require.True(t, enrolled)
|
|
require.False(t, installedFromDep)
|
|
require.Equal(t, "hi.com", serverURL)
|
|
return nil
|
|
}
|
|
ds.SetOrUpdateMunkiInfoFunc = func(ctx context.Context, hostID uint, version string, errs, warns []string) error {
|
|
require.Equal(t, "3.4.5", version)
|
|
return nil
|
|
}
|
|
ds.SetOrUpdateHostOrbitInfoFunc = func(ctx context.Context, hostID uint, version string) error {
|
|
require.Equal(t, "42", version)
|
|
return nil
|
|
}
|
|
ds.SetOrUpdateDeviceAuthTokenFunc = func(ctx context.Context, hostID uint, authToken string) error {
|
|
require.Equal(t, uint(1), hostID)
|
|
require.Equal(t, "foo", authToken)
|
|
return nil
|
|
}
|
|
ds.SetOrUpdateHostDisksSpaceFunc = func(ctx context.Context, hostID uint, gigsAvailable, percentAvailable float64) error {
|
|
require.Equal(t, 277.0, gigsAvailable)
|
|
require.Equal(t, 56.0, percentAvailable)
|
|
return nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
if id != 1 {
|
|
return nil, errors.New("not found")
|
|
}
|
|
return host, nil
|
|
}
|
|
|
|
// With a new host, we should get the detail queries (and accelerated
|
|
// queries)
|
|
queries, discovery, acc, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +1 for software inventory
|
|
if expected := expectedDetailQueriesForPlatform(host.Platform); !assert.Equal(t, len(expected)+1, len(queries)) {
|
|
// this is just to print the diff between the expected and actual query
|
|
// keys when the count assertion fails, to help debugging - they are not
|
|
// expected to match.
|
|
require.ElementsMatch(t, osqueryMapKeys(expected), distQueriesMapKeys(queries))
|
|
}
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.NotZero(t, acc)
|
|
|
|
resultJSON := `
|
|
{
|
|
"fleet_detail_query_network_interface": [
|
|
{
|
|
"address": "192.168.0.1",
|
|
"broadcast": "192.168.0.255",
|
|
"ibytes": "1601207629",
|
|
"ierrors": "314179",
|
|
"interface": "en0",
|
|
"ipackets": "25698094",
|
|
"last_change": "1474233476",
|
|
"mac": "5f:3d:4b:10:25:82",
|
|
"mask": "255.255.255.0",
|
|
"metric": "1",
|
|
"mtu": "1453",
|
|
"obytes": "2607283152",
|
|
"oerrors": "101010",
|
|
"opackets": "12264603",
|
|
"point_to_point": "",
|
|
"type": "6"
|
|
}
|
|
],
|
|
"fleet_detail_query_os_version": [
|
|
{
|
|
"platform": "darwin",
|
|
"build": "15G1004",
|
|
"major": "10",
|
|
"minor": "10",
|
|
"name": "Mac OS X",
|
|
"patch": "6"
|
|
}
|
|
],
|
|
"fleet_detail_query_osquery_info": [
|
|
{
|
|
"build_distro": "10.10",
|
|
"build_platform": "darwin",
|
|
"config_hash": "3c6e4537c4d0eb71a7c6dda19d",
|
|
"config_valid": "1",
|
|
"extensions": "active",
|
|
"pid": "38113",
|
|
"start_time": "1475603155",
|
|
"version": "1.8.2",
|
|
"watcher": "38112"
|
|
}
|
|
],
|
|
"fleet_detail_query_system_info": [
|
|
{
|
|
"computer_name": "computer",
|
|
"cpu_brand": "Intel(R) Core(TM) i7-4770HQ CPU @ 2.20GHz",
|
|
"cpu_logical_cores": "8",
|
|
"cpu_physical_cores": "4",
|
|
"cpu_subtype": "Intel x86-64h Haswell",
|
|
"cpu_type": "x86_64h",
|
|
"hardware_model": "MacBookPro11,4",
|
|
"hardware_serial": "ABCDEFGH",
|
|
"hardware_vendor": "Apple Inc.",
|
|
"hardware_version": "1.0",
|
|
"hostname": "computer.local",
|
|
"physical_memory": "17179869184",
|
|
"uuid": "uuid"
|
|
}
|
|
],
|
|
"fleet_detail_query_uptime": [
|
|
{
|
|
"days": "20",
|
|
"hours": "0",
|
|
"minutes": "48",
|
|
"seconds": "13",
|
|
"total_seconds": "1730893"
|
|
}
|
|
],
|
|
"fleet_detail_query_osquery_flags": [
|
|
{
|
|
"name":"config_tls_refresh",
|
|
"value":"10"
|
|
},
|
|
{
|
|
"name":"config_refresh",
|
|
"value":"9"
|
|
},
|
|
{
|
|
"name":"distributed_interval",
|
|
"value":"5"
|
|
},
|
|
{
|
|
"name":"logger_tls_period",
|
|
"value":"60"
|
|
}
|
|
],
|
|
"fleet_detail_query_users": [
|
|
{
|
|
"uid": "1234",
|
|
"username": "user1",
|
|
"type": "sometype",
|
|
"groupname": "somegroup",
|
|
"shell": "someloginshell"
|
|
},
|
|
{
|
|
"uid": "5678",
|
|
"username": "user2",
|
|
"type": "sometype",
|
|
"groupname": "somegroup"
|
|
}
|
|
],
|
|
"fleet_detail_query_software_macos": [
|
|
{
|
|
"name": "app1",
|
|
"version": "1.0.0",
|
|
"source": "source1"
|
|
},
|
|
{
|
|
"name": "app2",
|
|
"version": "1.0.0",
|
|
"source": "source2",
|
|
"bundle_identifier": "somebundle"
|
|
}
|
|
],
|
|
"fleet_detail_query_disk_space_unix": [
|
|
{
|
|
"percent_disk_space_available": "56",
|
|
"gigs_disk_space_available": "277.0"
|
|
}
|
|
],
|
|
"fleet_detail_query_mdm": [
|
|
{
|
|
"enrolled": "true",
|
|
"server_url": "hi.com",
|
|
"installed_from_dep": "false"
|
|
}
|
|
],
|
|
"fleet_detail_query_munki_info": [
|
|
{
|
|
"version": "3.4.5"
|
|
}
|
|
],
|
|
"fleet_detail_query_orbit_info": [
|
|
{
|
|
"version": "42"
|
|
}
|
|
]
|
|
}
|
|
`
|
|
|
|
var results fleet.OsqueryDistributedQueryResults
|
|
err = json.Unmarshal([]byte(resultJSON), &results)
|
|
require.NoError(t, err)
|
|
|
|
var gotHost *fleet.Host
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
gotHost = host
|
|
return nil
|
|
}
|
|
var gotUsers []fleet.HostUser
|
|
ds.SaveHostUsersFunc = func(ctx context.Context, hostID uint, users []fleet.HostUser) error {
|
|
if hostID != 1 {
|
|
return errors.New("not found")
|
|
}
|
|
gotUsers = users
|
|
return nil
|
|
}
|
|
var gotSoftware []fleet.Software
|
|
ds.UpdateHostSoftwareFunc = func(ctx context.Context, hostID uint, software []fleet.Software) error {
|
|
if hostID != 1 {
|
|
return errors.New("not found")
|
|
}
|
|
gotSoftware = software
|
|
return nil
|
|
}
|
|
|
|
// Verify that results are ingested properly
|
|
require.NoError(t, svc.SubmitDistributedQueryResults(ctx, results, map[string]fleet.OsqueryStatus{}, map[string]string{}))
|
|
require.NotNil(t, gotHost)
|
|
|
|
require.True(t, ds.SetOrUpdateMDMDataFuncInvoked)
|
|
require.True(t, ds.SetOrUpdateMunkiInfoFuncInvoked)
|
|
require.True(t, ds.SetOrUpdateHostDisksSpaceFuncInvoked)
|
|
|
|
// osquery_info
|
|
assert.Equal(t, "darwin", gotHost.Platform)
|
|
assert.Equal(t, "1.8.2", gotHost.OsqueryVersion)
|
|
|
|
// system_info
|
|
assert.Equal(t, int64(17179869184), gotHost.Memory)
|
|
assert.Equal(t, "computer.local", gotHost.Hostname)
|
|
assert.Equal(t, "uuid", gotHost.UUID)
|
|
|
|
// os_version
|
|
assert.Equal(t, "Mac OS X 10.10.6", gotHost.OSVersion)
|
|
|
|
// uptime
|
|
assert.Equal(t, 1730893*time.Second, gotHost.Uptime)
|
|
|
|
// osquery_flags
|
|
assert.Equal(t, uint(10), gotHost.ConfigTLSRefresh)
|
|
assert.Equal(t, uint(5), gotHost.DistributedInterval)
|
|
assert.Equal(t, uint(60), gotHost.LoggerTLSPeriod)
|
|
|
|
// users
|
|
require.Len(t, gotUsers, 2)
|
|
assert.Equal(t, fleet.HostUser{
|
|
Uid: 1234,
|
|
Username: "user1",
|
|
Type: "sometype",
|
|
GroupName: "somegroup",
|
|
Shell: "someloginshell",
|
|
}, gotUsers[0])
|
|
assert.Equal(t, fleet.HostUser{
|
|
Uid: 5678,
|
|
Username: "user2",
|
|
Type: "sometype",
|
|
GroupName: "somegroup",
|
|
Shell: "",
|
|
}, gotUsers[1])
|
|
|
|
// software
|
|
require.Len(t, gotSoftware, 2)
|
|
assert.Equal(t, []fleet.Software{
|
|
{
|
|
Name: "app1",
|
|
Version: "1.0.0",
|
|
Source: "source1",
|
|
},
|
|
{
|
|
Name: "app2",
|
|
Version: "1.0.0",
|
|
BundleIdentifier: "somebundle",
|
|
Source: "source2",
|
|
},
|
|
}, gotSoftware)
|
|
|
|
host.Hostname = "computer.local"
|
|
host.Platform = "darwin"
|
|
host.DetailUpdatedAt = mockClock.Now()
|
|
mockClock.AddTime(1 * time.Minute)
|
|
|
|
// Now no detail queries should be required
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Empty(t, queries)
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
|
|
// Advance clock and queries should exist again
|
|
mockClock.AddTime(1*time.Hour + 1*time.Minute)
|
|
|
|
queries, discovery, acc, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +1 software inventory
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+1, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
assert.Zero(t, acc)
|
|
}
|
|
|
|
func TestNewDistributedQueryCampaign(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
rs := &mockresult.QueryResultStore{
|
|
HealthCheckFunc: func() error {
|
|
return nil
|
|
},
|
|
}
|
|
lq := live_query_mock.New(t)
|
|
mockClock := clock.NewMockClock()
|
|
svc, ctx := newTestServiceWithClock(t, ds, rs, lq, mockClock)
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
var gotQuery *fleet.Query
|
|
ds.NewQueryFunc = func(ctx context.Context, query *fleet.Query, opts ...fleet.OptionalArg) (*fleet.Query, error) {
|
|
gotQuery = query
|
|
query.ID = 42
|
|
return query, nil
|
|
}
|
|
var gotCampaign *fleet.DistributedQueryCampaign
|
|
ds.NewDistributedQueryCampaignFunc = func(ctx context.Context, camp *fleet.DistributedQueryCampaign) (*fleet.DistributedQueryCampaign, error) {
|
|
gotCampaign = camp
|
|
camp.ID = 21
|
|
return camp, nil
|
|
}
|
|
var gotTargets []*fleet.DistributedQueryCampaignTarget
|
|
ds.NewDistributedQueryCampaignTargetFunc = func(ctx context.Context, target *fleet.DistributedQueryCampaignTarget) (*fleet.DistributedQueryCampaignTarget, error) {
|
|
gotTargets = append(gotTargets, target)
|
|
return target, nil
|
|
}
|
|
|
|
ds.CountHostsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets, now time.Time) (fleet.TargetMetrics, error) {
|
|
return fleet.TargetMetrics{}, nil
|
|
}
|
|
ds.HostIDsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets) ([]uint, error) {
|
|
return []uint{1, 3, 5}, nil
|
|
}
|
|
lq.On("RunQuery", "21", "select year, month, day, hour, minutes, seconds from time", []uint{1, 3, 5}).Return(nil)
|
|
viewerCtx := viewer.NewContext(ctx, viewer.Viewer{
|
|
User: &fleet.User{
|
|
ID: 0,
|
|
GlobalRole: ptr.String(fleet.RoleAdmin),
|
|
},
|
|
})
|
|
q := "select year, month, day, hour, minutes, seconds from time"
|
|
ds.NewActivityFunc = func(ctx context.Context, user *fleet.User, activity fleet.ActivityDetails) error {
|
|
return nil
|
|
}
|
|
campaign, err := svc.NewDistributedQueryCampaign(viewerCtx, q, nil, fleet.HostTargets{HostIDs: []uint{2}, LabelIDs: []uint{1}})
|
|
require.NoError(t, err)
|
|
assert.Equal(t, gotQuery.ID, gotCampaign.QueryID)
|
|
assert.True(t, ds.NewActivityFuncInvoked)
|
|
assert.Equal(t, []*fleet.DistributedQueryCampaignTarget{
|
|
{
|
|
Type: fleet.TargetHost,
|
|
DistributedQueryCampaignID: campaign.ID,
|
|
TargetID: 2,
|
|
},
|
|
{
|
|
Type: fleet.TargetLabel,
|
|
DistributedQueryCampaignID: campaign.ID,
|
|
TargetID: 1,
|
|
},
|
|
}, gotTargets,
|
|
)
|
|
}
|
|
|
|
func TestDistributedQueryResults(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc, ctx := newTestServiceWithClock(t, ds, rs, lq, mockClock)
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{ID: 42}
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
host := &fleet.Host{
|
|
ID: 1,
|
|
Platform: "windows",
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
if id != 1 {
|
|
return nil, errors.New("not found")
|
|
}
|
|
return host, nil
|
|
}
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
if host.ID != 1 {
|
|
return errors.New("not found")
|
|
}
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true}}, nil
|
|
}
|
|
|
|
hostCtx := hostctx.NewContext(ctx, host)
|
|
|
|
lq.On("QueriesForHost", uint(1)).Return(
|
|
map[string]string{
|
|
strconv.Itoa(int(campaign.ID)): "select * from time",
|
|
},
|
|
nil,
|
|
)
|
|
lq.On("QueryCompletedByHost", strconv.Itoa(int(campaign.ID)), host.ID).Return(nil)
|
|
|
|
// Now we should get the active distributed query
|
|
queries, discovery, acc, err := svc.GetDistributedQueries(hostCtx)
|
|
require.NoError(t, err)
|
|
// +1 for the distributed query for campaign ID 42, +1 for windows update history
|
|
if expected := expectedDetailQueriesForPlatform(host.Platform); !assert.Equal(t, len(expected)+2, len(queries)) {
|
|
// this is just to print the diff between the expected and actual query
|
|
// keys when the count assertion fails, to help debugging - they are not
|
|
// expected to match.
|
|
require.ElementsMatch(t, osqueryMapKeys(expected), distQueriesMapKeys(queries))
|
|
}
|
|
verifyDiscovery(t, queries, discovery)
|
|
queryKey := fmt.Sprintf("%s%d", hostDistributedQueryPrefix, campaign.ID)
|
|
assert.Equal(t, "select * from time", queries[queryKey])
|
|
assert.NotZero(t, acc)
|
|
|
|
expectedRows := []map[string]string{
|
|
{
|
|
"year": "2016",
|
|
"month": "11",
|
|
"day": "11",
|
|
"hour": "6",
|
|
"minutes": "12",
|
|
"seconds": "10",
|
|
},
|
|
}
|
|
results := map[string][]map[string]string{
|
|
queryKey: expectedRows,
|
|
}
|
|
|
|
// TODO use service method
|
|
readChan, err := rs.ReadChannel(context.Background(), *campaign)
|
|
require.NoError(t, err)
|
|
|
|
// We need to listen for the result in a separate thread to prevent the
|
|
// write to the result channel from failing
|
|
var waitSetup, waitComplete sync.WaitGroup
|
|
waitSetup.Add(1)
|
|
waitComplete.Add(1)
|
|
go func() {
|
|
waitSetup.Done()
|
|
select {
|
|
case val := <-readChan:
|
|
if res, ok := val.(fleet.DistributedQueryResult); ok {
|
|
assert.Equal(t, campaign.ID, res.DistributedQueryCampaignID)
|
|
assert.Equal(t, expectedRows, res.Rows)
|
|
assert.Equal(t, host, res.Host.Host)
|
|
} else {
|
|
t.Error("Wrong result type")
|
|
}
|
|
assert.NotNil(t, val)
|
|
|
|
case <-time.After(1 * time.Second):
|
|
t.Error("No result received")
|
|
}
|
|
waitComplete.Done()
|
|
}()
|
|
|
|
waitSetup.Wait()
|
|
// Sleep a short time to ensure that the above goroutine is blocking on
|
|
// the channel read (the waitSetup.Wait() is not necessarily sufficient
|
|
// if there is a context switch immediately after waitSetup.Done() is
|
|
// called). This should be a small price to pay to prevent flakiness in
|
|
// this test.
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
err = svc.SubmitDistributedQueryResults(hostCtx, results, map[string]fleet.OsqueryStatus{}, map[string]string{})
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestIngestDistributedQueryParseIdError(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
host := fleet.Host{ID: 1}
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "bad_name", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "unable to parse campaign")
|
|
}
|
|
|
|
func TestIngestDistributedQueryOrphanedCampaignLoadError(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
ds.DistributedQueryCampaignFunc = func(ctx context.Context, id uint) (*fleet.DistributedQueryCampaign, error) {
|
|
return nil, errors.New("missing campaign")
|
|
}
|
|
|
|
lq.On("StopQuery", "42").Return(nil)
|
|
|
|
host := fleet.Host{ID: 1}
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "loading orphaned campaign")
|
|
}
|
|
|
|
func TestIngestDistributedQueryOrphanedCampaignWaitListener(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{
|
|
ID: 42,
|
|
UpdateCreateTimestamps: fleet.UpdateCreateTimestamps{
|
|
CreateTimestamp: fleet.CreateTimestamp{
|
|
CreatedAt: mockClock.Now().Add(-1 * time.Second),
|
|
},
|
|
},
|
|
}
|
|
|
|
ds.DistributedQueryCampaignFunc = func(ctx context.Context, id uint) (*fleet.DistributedQueryCampaign, error) {
|
|
return campaign, nil
|
|
}
|
|
|
|
host := fleet.Host{ID: 1}
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "campaign waiting for listener")
|
|
}
|
|
|
|
func TestIngestDistributedQueryOrphanedCloseError(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{
|
|
ID: 42,
|
|
UpdateCreateTimestamps: fleet.UpdateCreateTimestamps{
|
|
CreateTimestamp: fleet.CreateTimestamp{
|
|
CreatedAt: mockClock.Now().Add(-2 * time.Minute),
|
|
},
|
|
},
|
|
}
|
|
|
|
ds.DistributedQueryCampaignFunc = func(ctx context.Context, id uint) (*fleet.DistributedQueryCampaign, error) {
|
|
return campaign, nil
|
|
}
|
|
ds.SaveDistributedQueryCampaignFunc = func(ctx context.Context, campaign *fleet.DistributedQueryCampaign) error {
|
|
return errors.New("failed save")
|
|
}
|
|
|
|
host := fleet.Host{ID: 1}
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "closing orphaned campaign")
|
|
}
|
|
|
|
func TestIngestDistributedQueryOrphanedStopError(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{
|
|
ID: 42,
|
|
UpdateCreateTimestamps: fleet.UpdateCreateTimestamps{
|
|
CreateTimestamp: fleet.CreateTimestamp{
|
|
CreatedAt: mockClock.Now().Add(-2 * time.Minute),
|
|
},
|
|
},
|
|
}
|
|
|
|
ds.DistributedQueryCampaignFunc = func(ctx context.Context, id uint) (*fleet.DistributedQueryCampaign, error) {
|
|
return campaign, nil
|
|
}
|
|
ds.SaveDistributedQueryCampaignFunc = func(ctx context.Context, campaign *fleet.DistributedQueryCampaign) error {
|
|
return nil
|
|
}
|
|
lq.On("StopQuery", strconv.Itoa(int(campaign.ID))).Return(errors.New("failed"))
|
|
|
|
host := fleet.Host{ID: 1}
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "stopping orphaned campaign")
|
|
}
|
|
|
|
func TestIngestDistributedQueryOrphanedStop(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{
|
|
ID: 42,
|
|
UpdateCreateTimestamps: fleet.UpdateCreateTimestamps{
|
|
CreateTimestamp: fleet.CreateTimestamp{
|
|
CreatedAt: mockClock.Now().Add(-2 * time.Minute),
|
|
},
|
|
},
|
|
}
|
|
|
|
ds.DistributedQueryCampaignFunc = func(ctx context.Context, id uint) (*fleet.DistributedQueryCampaign, error) {
|
|
return campaign, nil
|
|
}
|
|
ds.SaveDistributedQueryCampaignFunc = func(ctx context.Context, campaign *fleet.DistributedQueryCampaign) error {
|
|
return nil
|
|
}
|
|
lq.On("StopQuery", strconv.Itoa(int(campaign.ID))).Return(nil)
|
|
|
|
host := fleet.Host{ID: 1}
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "campaign stopped")
|
|
lq.AssertExpectations(t)
|
|
}
|
|
|
|
func TestIngestDistributedQueryRecordCompletionError(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{ID: 42}
|
|
host := fleet.Host{ID: 1}
|
|
|
|
lq.On("QueryCompletedByHost", strconv.Itoa(int(campaign.ID)), host.ID).Return(errors.New("fail"))
|
|
|
|
go func() {
|
|
ch, err := rs.ReadChannel(context.Background(), *campaign)
|
|
require.NoError(t, err)
|
|
<-ch
|
|
}()
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.Error(t, err)
|
|
assert.Contains(t, err.Error(), "record query completion")
|
|
lq.AssertExpectations(t)
|
|
}
|
|
|
|
func TestIngestDistributedQuery(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
rs := pubsub.NewInmemQueryResults()
|
|
lq := live_query_mock.New(t)
|
|
svc := &Service{
|
|
ds: ds,
|
|
resultStore: rs,
|
|
liveQueryStore: lq,
|
|
logger: log.NewNopLogger(),
|
|
clock: mockClock,
|
|
}
|
|
|
|
campaign := &fleet.DistributedQueryCampaign{ID: 42}
|
|
host := fleet.Host{ID: 1}
|
|
|
|
lq.On("QueryCompletedByHost", strconv.Itoa(int(campaign.ID)), host.ID).Return(nil)
|
|
|
|
go func() {
|
|
ch, err := rs.ReadChannel(context.Background(), *campaign)
|
|
require.NoError(t, err)
|
|
<-ch
|
|
}()
|
|
time.Sleep(10 * time.Millisecond)
|
|
|
|
err := svc.ingestDistributedQuery(context.Background(), host, "fleet_distributed_query_42", []map[string]string{}, false, "")
|
|
require.NoError(t, err)
|
|
lq.AssertExpectations(t)
|
|
}
|
|
|
|
func TestUpdateHostIntervals(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
ds.ListPacksForHostFunc = func(ctx context.Context, hid uint) ([]*fleet.Pack, error) {
|
|
return []*fleet.Pack{}, nil
|
|
}
|
|
|
|
testCases := []struct {
|
|
name string
|
|
initIntervals fleet.HostOsqueryIntervals
|
|
finalIntervals fleet.HostOsqueryIntervals
|
|
configOptions json.RawMessage
|
|
updateIntervalsCalled bool
|
|
}{
|
|
{
|
|
"Both updated",
|
|
fleet.HostOsqueryIntervals{
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
json.RawMessage(`{"options": {
|
|
"distributed_interval": 11,
|
|
"logger_tls_period": 33,
|
|
"logger_plugin": "tls"
|
|
}}`),
|
|
true,
|
|
},
|
|
{
|
|
"Only logger_tls_period updated",
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
json.RawMessage(`{"options": {
|
|
"distributed_interval": 11,
|
|
"logger_tls_period": 33
|
|
}}`),
|
|
true,
|
|
},
|
|
{
|
|
"Only distributed_interval updated",
|
|
fleet.HostOsqueryIntervals{
|
|
ConfigTLSRefresh: 60,
|
|
LoggerTLSPeriod: 33,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
json.RawMessage(`{"options": {
|
|
"distributed_interval": 11,
|
|
"logger_tls_period": 33
|
|
}}`),
|
|
true,
|
|
},
|
|
{
|
|
"Fleet not managing distributed_interval",
|
|
fleet.HostOsqueryIntervals{
|
|
ConfigTLSRefresh: 60,
|
|
DistributedInterval: 11,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
json.RawMessage(`{"options":{
|
|
"logger_tls_period": 33
|
|
}}`),
|
|
true,
|
|
},
|
|
{
|
|
"config_refresh should also cause an update",
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 42,
|
|
},
|
|
json.RawMessage(`{"options":{
|
|
"distributed_interval": 11,
|
|
"logger_tls_period": 33,
|
|
"config_refresh": 42
|
|
}}`),
|
|
true,
|
|
},
|
|
{
|
|
"update intervals should not be called with no changes",
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
fleet.HostOsqueryIntervals{
|
|
DistributedInterval: 11,
|
|
LoggerTLSPeriod: 33,
|
|
ConfigTLSRefresh: 60,
|
|
},
|
|
json.RawMessage(`{"options":{
|
|
"distributed_interval": 11,
|
|
"logger_tls_period": 33
|
|
}}`),
|
|
false,
|
|
},
|
|
}
|
|
|
|
for _, tt := range testCases {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
ctx := hostctx.NewContext(ctx, &fleet.Host{
|
|
ID: 1,
|
|
NodeKey: "123456",
|
|
DistributedInterval: tt.initIntervals.DistributedInterval,
|
|
ConfigTLSRefresh: tt.initIntervals.ConfigTLSRefresh,
|
|
LoggerTLSPeriod: tt.initIntervals.LoggerTLSPeriod,
|
|
})
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{AgentOptions: ptr.RawMessage(json.RawMessage(`{"config":` + string(tt.configOptions) + `}`))}, nil
|
|
}
|
|
|
|
updateIntervalsCalled := false
|
|
ds.UpdateHostOsqueryIntervalsFunc = func(ctx context.Context, hostID uint, intervals fleet.HostOsqueryIntervals) error {
|
|
if hostID != 1 {
|
|
return errors.New("not found")
|
|
}
|
|
updateIntervalsCalled = true
|
|
assert.Equal(t, tt.finalIntervals, intervals)
|
|
return nil
|
|
}
|
|
|
|
_, err := svc.GetClientConfig(ctx)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, tt.updateIntervalsCalled, updateIntervalsCalled)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestAuthenticationErrors(t *testing.T) {
|
|
ms := new(mock.Store)
|
|
ms.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
return nil, nil
|
|
}
|
|
|
|
svc, ctx := newTestService(t, ms, nil, nil)
|
|
|
|
_, _, err := svc.AuthenticateHost(ctx, "")
|
|
require.Error(t, err)
|
|
require.True(t, err.(osqueryError).NodeInvalid())
|
|
|
|
ms.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
return &fleet.Host{ID: 1}, nil
|
|
}
|
|
ms.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
_, _, err = svc.AuthenticateHost(ctx, "foo")
|
|
require.NoError(t, err)
|
|
|
|
// return not found error
|
|
ms.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
return nil, notFoundError{}
|
|
}
|
|
|
|
_, _, err = svc.AuthenticateHost(ctx, "foo")
|
|
require.Error(t, err)
|
|
require.True(t, err.(osqueryError).NodeInvalid())
|
|
|
|
// return other error
|
|
ms.LoadHostByNodeKeyFunc = func(ctx context.Context, nodeKey string) (*fleet.Host, error) {
|
|
return nil, errors.New("foo")
|
|
}
|
|
|
|
_, _, err = svc.AuthenticateHost(ctx, "foo")
|
|
require.NotNil(t, err)
|
|
require.False(t, err.(osqueryError).NodeInvalid())
|
|
}
|
|
|
|
func TestGetHostIdentifier(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
details := map[string](map[string]string){
|
|
"osquery_info": map[string]string{
|
|
"uuid": "foouuid",
|
|
"instance_id": "fooinstance",
|
|
},
|
|
"system_info": map[string]string{
|
|
"hostname": "foohost",
|
|
},
|
|
}
|
|
|
|
emptyDetails := map[string](map[string]string){
|
|
"osquery_info": map[string]string{
|
|
"uuid": "",
|
|
"instance_id": "",
|
|
},
|
|
"system_info": map[string]string{
|
|
"hostname": "",
|
|
},
|
|
}
|
|
|
|
testCases := []struct {
|
|
identifierOption string
|
|
providedIdentifier string
|
|
details map[string](map[string]string)
|
|
expected string
|
|
shouldPanic bool
|
|
}{
|
|
// Panix
|
|
{identifierOption: "bad", shouldPanic: true},
|
|
{identifierOption: "", shouldPanic: true},
|
|
|
|
// Missing details
|
|
{identifierOption: "instance", providedIdentifier: "foobar", expected: "foobar"},
|
|
{identifierOption: "uuid", providedIdentifier: "foobar", expected: "foobar"},
|
|
{identifierOption: "hostname", providedIdentifier: "foobar", expected: "foobar"},
|
|
{identifierOption: "provided", providedIdentifier: "foobar", expected: "foobar"},
|
|
|
|
// Empty details
|
|
{identifierOption: "instance", providedIdentifier: "foobar", details: emptyDetails, expected: "foobar"},
|
|
{identifierOption: "uuid", providedIdentifier: "foobar", details: emptyDetails, expected: "foobar"},
|
|
{identifierOption: "hostname", providedIdentifier: "foobar", details: emptyDetails, expected: "foobar"},
|
|
{identifierOption: "provided", providedIdentifier: "foobar", details: emptyDetails, expected: "foobar"},
|
|
|
|
// Successes
|
|
{identifierOption: "instance", providedIdentifier: "foobar", details: details, expected: "fooinstance"},
|
|
{identifierOption: "uuid", providedIdentifier: "foobar", details: details, expected: "foouuid"},
|
|
{identifierOption: "hostname", providedIdentifier: "foobar", details: details, expected: "foohost"},
|
|
{identifierOption: "provided", providedIdentifier: "foobar", details: details, expected: "foobar"},
|
|
}
|
|
logger := log.NewNopLogger()
|
|
|
|
for _, tt := range testCases {
|
|
t.Run("", func(t *testing.T) {
|
|
if tt.shouldPanic {
|
|
assert.Panics(
|
|
t,
|
|
func() { getHostIdentifier(logger, tt.identifierOption, tt.providedIdentifier, tt.details) },
|
|
)
|
|
return
|
|
}
|
|
|
|
assert.Equal(
|
|
t,
|
|
tt.expected,
|
|
getHostIdentifier(logger, tt.identifierOption, tt.providedIdentifier, tt.details),
|
|
)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestDistributedQueriesLogsManyErrors(t *testing.T) {
|
|
buf := new(bytes.Buffer)
|
|
logger := log.NewJSONLogger(buf)
|
|
logger = level.NewFilter(logger, level.AllowDebug())
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
host := &fleet.Host{
|
|
ID: 1,
|
|
Platform: "darwin",
|
|
}
|
|
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
return authz.CheckMissingWithResponse(nil)
|
|
}
|
|
ds.RecordLabelQueryExecutionsFunc = func(ctx context.Context, host *fleet.Host, results map[uint]*bool, t time.Time, deferred bool) error {
|
|
return errors.New("something went wrong")
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
ds.SaveHostAdditionalFunc = func(ctx context.Context, hostID uint, additional *json.RawMessage) error {
|
|
return errors.New("something went wrong")
|
|
}
|
|
|
|
lCtx := &fleetLogging.LoggingContext{}
|
|
ctx = fleetLogging.NewContext(ctx, lCtx)
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
err := svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostDetailQueryPrefix + "network_interface_unix": {{"col1": "val1"}}, // we need one detail query that updates hosts.
|
|
hostLabelQueryPrefix + "1": {{"col1": "val1"}},
|
|
hostAdditionalQueryPrefix + "1": {{"col1": "val1"}},
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
|
|
lCtx.Log(ctx, logger)
|
|
|
|
logs := buf.String()
|
|
parts := strings.Split(strings.TrimSpace(logs), "\n")
|
|
require.Len(t, parts, 1)
|
|
|
|
var logData map[string]interface{}
|
|
err = json.Unmarshal([]byte(parts[0]), &logData)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, "something went wrong || something went wrong", logData["err"])
|
|
assert.Equal(t, "Missing authorization check", logData["internal"])
|
|
}
|
|
|
|
func TestDistributedQueriesReloadsHostIfDetailsAreIn(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
svc, ctx := newTestService(t, ds, nil, nil)
|
|
|
|
host := &fleet.Host{
|
|
ID: 42,
|
|
Platform: "darwin",
|
|
}
|
|
|
|
ds.UpdateHostFunc = func(ctx context.Context, host *fleet.Host) error {
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
err := svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostDetailQueryPrefix + "network_interface_unix": {{"col1": "val1"}},
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
assert.True(t, ds.UpdateHostFuncInvoked)
|
|
}
|
|
|
|
func TestObserversCanOnlyRunDistributedCampaigns(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
rs := &mockresult.QueryResultStore{
|
|
HealthCheckFunc: func() error {
|
|
return nil
|
|
},
|
|
}
|
|
lq := live_query_mock.New(t)
|
|
mockClock := clock.NewMockClock()
|
|
svc, ctx := newTestServiceWithClock(t, ds, rs, lq, mockClock)
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
ds.NewDistributedQueryCampaignFunc = func(ctx context.Context, camp *fleet.DistributedQueryCampaign) (*fleet.DistributedQueryCampaign, error) {
|
|
return camp, nil
|
|
}
|
|
ds.QueryFunc = func(ctx context.Context, id uint) (*fleet.Query, error) {
|
|
return &fleet.Query{
|
|
ID: 42,
|
|
Name: "query",
|
|
Query: "select 1;",
|
|
ObserverCanRun: false,
|
|
}, nil
|
|
}
|
|
viewerCtx := viewer.NewContext(ctx, viewer.Viewer{
|
|
User: &fleet.User{ID: 0, GlobalRole: ptr.String(fleet.RoleObserver)},
|
|
})
|
|
|
|
q := "select year, month, day, hour, minutes, seconds from time"
|
|
ds.NewActivityFunc = func(ctx context.Context, user *fleet.User, activity fleet.ActivityDetails) error {
|
|
return nil
|
|
}
|
|
_, err := svc.NewDistributedQueryCampaign(viewerCtx, q, nil, fleet.HostTargets{HostIDs: []uint{2}, LabelIDs: []uint{1}})
|
|
require.Error(t, err)
|
|
|
|
_, err = svc.NewDistributedQueryCampaign(viewerCtx, "", ptr.Uint(42), fleet.HostTargets{HostIDs: []uint{2}, LabelIDs: []uint{1}})
|
|
require.Error(t, err)
|
|
|
|
ds.QueryFunc = func(ctx context.Context, id uint) (*fleet.Query, error) {
|
|
return &fleet.Query{
|
|
ID: 42,
|
|
Name: "query",
|
|
Query: "select 1;",
|
|
ObserverCanRun: true,
|
|
}, nil
|
|
}
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.NewDistributedQueryCampaignFunc = func(ctx context.Context, camp *fleet.DistributedQueryCampaign) (*fleet.DistributedQueryCampaign, error) {
|
|
camp.ID = 21
|
|
return camp, nil
|
|
}
|
|
ds.NewDistributedQueryCampaignTargetFunc = func(ctx context.Context, target *fleet.DistributedQueryCampaignTarget) (*fleet.DistributedQueryCampaignTarget, error) {
|
|
return target, nil
|
|
}
|
|
ds.CountHostsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets, now time.Time) (fleet.TargetMetrics, error) {
|
|
return fleet.TargetMetrics{}, nil
|
|
}
|
|
ds.HostIDsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets) ([]uint, error) {
|
|
return []uint{1, 3, 5}, nil
|
|
}
|
|
ds.NewActivityFunc = func(ctx context.Context, user *fleet.User, activity fleet.ActivityDetails) error {
|
|
return nil
|
|
}
|
|
lq.On("RunQuery", "21", "select 1;", []uint{1, 3, 5}).Return(nil)
|
|
_, err = svc.NewDistributedQueryCampaign(viewerCtx, "", ptr.Uint(42), fleet.HostTargets{HostIDs: []uint{2}, LabelIDs: []uint{1}})
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestTeamMaintainerCanRunNewDistributedCampaigns(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
rs := &mockresult.QueryResultStore{
|
|
HealthCheckFunc: func() error {
|
|
return nil
|
|
},
|
|
}
|
|
lq := live_query_mock.New(t)
|
|
mockClock := clock.NewMockClock()
|
|
svc, ctx := newTestServiceWithClock(t, ds, rs, lq, mockClock)
|
|
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{}, nil
|
|
}
|
|
|
|
ds.NewDistributedQueryCampaignFunc = func(ctx context.Context, camp *fleet.DistributedQueryCampaign) (*fleet.DistributedQueryCampaign, error) {
|
|
return camp, nil
|
|
}
|
|
ds.QueryFunc = func(ctx context.Context, id uint) (*fleet.Query, error) {
|
|
return &fleet.Query{
|
|
ID: 42,
|
|
AuthorID: ptr.Uint(99),
|
|
Name: "query",
|
|
Query: "select 1;",
|
|
ObserverCanRun: false,
|
|
}, nil
|
|
}
|
|
viewerCtx := viewer.NewContext(ctx, viewer.Viewer{
|
|
User: &fleet.User{ID: 99, Teams: []fleet.UserTeam{{Team: fleet.Team{ID: 123}, Role: fleet.RoleMaintainer}}},
|
|
})
|
|
|
|
q := "select year, month, day, hour, minutes, seconds from time"
|
|
ds.NewActivityFunc = func(ctx context.Context, user *fleet.User, activity fleet.ActivityDetails) error {
|
|
return nil
|
|
}
|
|
// var gotQuery *fleet.Query
|
|
ds.NewQueryFunc = func(ctx context.Context, query *fleet.Query, opts ...fleet.OptionalArg) (*fleet.Query, error) {
|
|
// gotQuery = query
|
|
query.ID = 42
|
|
return query, nil
|
|
}
|
|
ds.NewDistributedQueryCampaignTargetFunc = func(ctx context.Context, target *fleet.DistributedQueryCampaignTarget) (*fleet.DistributedQueryCampaignTarget, error) {
|
|
return target, nil
|
|
}
|
|
ds.CountHostsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets, now time.Time) (fleet.TargetMetrics, error) {
|
|
return fleet.TargetMetrics{}, nil
|
|
}
|
|
ds.HostIDsInTargetsFunc = func(ctx context.Context, filter fleet.TeamFilter, targets fleet.HostTargets) ([]uint, error) {
|
|
return []uint{1, 3, 5}, nil
|
|
}
|
|
ds.NewActivityFunc = func(ctx context.Context, user *fleet.User, activity fleet.ActivityDetails) error {
|
|
return nil
|
|
}
|
|
lq.On("RunQuery", "0", "select year, month, day, hour, minutes, seconds from time", []uint{1, 3, 5}).Return(nil)
|
|
_, err := svc.NewDistributedQueryCampaign(viewerCtx, q, nil, fleet.HostTargets{HostIDs: []uint{2}, LabelIDs: []uint{1}, TeamIDs: []uint{123}})
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestPolicyQueries(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
lq := live_query_mock.New(t)
|
|
svc, ctx := newTestServiceWithClock(t, ds, nil, lq, mockClock)
|
|
|
|
host := &fleet.Host{
|
|
Platform: "darwin",
|
|
}
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
return host, nil
|
|
}
|
|
ds.UpdateHostFunc = func(ctx context.Context, gotHost *fleet.Host) error {
|
|
host = gotHost
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true}}, nil
|
|
}
|
|
|
|
lq.On("QueriesForHost", uint(0)).Return(map[string]string{}, nil)
|
|
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{"1": "select 1", "2": "select 42;"}, nil
|
|
}
|
|
recordedResults := make(map[uint]*bool)
|
|
ds.RecordPolicyQueryExecutionsFunc = func(ctx context.Context, gotHost *fleet.Host, results map[uint]*bool, updated time.Time, deferred bool) error {
|
|
recordedResults = results
|
|
host = gotHost
|
|
return nil
|
|
}
|
|
ds.FlippingPoliciesForHostFunc = func(ctx context.Context, hostID uint, incomingResults map[uint]*bool) (newFailing []uint, newPassing []uint, err error) {
|
|
return nil, nil, nil
|
|
}
|
|
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
queries, discovery, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +2 policy queries
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+2, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
|
|
checkPolicyResults := func(queries map[string]string) {
|
|
hasPolicy1, hasPolicy2 := false, false
|
|
for name := range queries {
|
|
if strings.HasPrefix(name, hostPolicyQueryPrefix) {
|
|
if name[len(hostPolicyQueryPrefix):] == "1" {
|
|
hasPolicy1 = true
|
|
}
|
|
if name[len(hostPolicyQueryPrefix):] == "2" {
|
|
hasPolicy2 = true
|
|
}
|
|
}
|
|
}
|
|
assert.True(t, hasPolicy1)
|
|
assert.True(t, hasPolicy2)
|
|
}
|
|
|
|
checkPolicyResults(queries)
|
|
|
|
// Record a query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {{"col1": "val1"}},
|
|
hostPolicyQueryPrefix + "2": {},
|
|
},
|
|
map[string]fleet.OsqueryStatus{
|
|
hostPolicyQueryPrefix + "2": 1,
|
|
},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.Len(t, recordedResults, 2)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.True(t, *recordedResults[1])
|
|
result, ok := recordedResults[2]
|
|
require.True(t, ok)
|
|
require.Nil(t, result)
|
|
|
|
noPolicyResults := func(queries map[string]string) {
|
|
hasAnyPolicy := false
|
|
for name := range queries {
|
|
if strings.HasPrefix(name, hostPolicyQueryPrefix) {
|
|
hasAnyPolicy = true
|
|
break
|
|
}
|
|
}
|
|
assert.False(t, hasAnyPolicy)
|
|
}
|
|
|
|
// After the first time we get policies and update the host, then there shouldn't be any policies.
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
noPolicyResults(queries)
|
|
|
|
// Let's move time forward, there should be policies now.
|
|
mockClock.AddTime(2 * time.Hour)
|
|
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +2 policy queries
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+2, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
checkPolicyResults(queries)
|
|
|
|
// Record another query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {{"col1": "val1"}},
|
|
hostPolicyQueryPrefix + "2": {},
|
|
},
|
|
map[string]fleet.OsqueryStatus{
|
|
hostPolicyQueryPrefix + "2": 1,
|
|
},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.Len(t, recordedResults, 2)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.True(t, *recordedResults[1])
|
|
result, ok = recordedResults[2]
|
|
require.True(t, ok)
|
|
require.Nil(t, result)
|
|
|
|
// There shouldn't be any policies now.
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
noPolicyResults(queries)
|
|
|
|
// With refetch requested policy queries should be returned.
|
|
host.RefetchRequested = true
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +2 policy queries
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+2, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
checkPolicyResults(queries)
|
|
|
|
// Record another query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {{"col1": "val1"}},
|
|
hostPolicyQueryPrefix + "2": {},
|
|
},
|
|
map[string]fleet.OsqueryStatus{
|
|
hostPolicyQueryPrefix + "2": 1,
|
|
},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.True(t, *recordedResults[1])
|
|
result, ok = recordedResults[2]
|
|
require.True(t, ok)
|
|
require.Nil(t, result)
|
|
|
|
// SubmitDistributedQueryResults will set RefetchRequested to false.
|
|
require.False(t, host.RefetchRequested)
|
|
|
|
// There shouldn't be any policies now.
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
noPolicyResults(queries)
|
|
}
|
|
|
|
func TestPolicyWebhooks(t *testing.T) {
|
|
mockClock := clock.NewMockClock()
|
|
ds := new(mock.Store)
|
|
lq := live_query_mock.New(t)
|
|
pool := redistest.SetupRedis(t, t.Name(), false, false, false)
|
|
failingPolicySet := redis_policy_set.NewFailingTest(t, pool)
|
|
testConfig := config.TestConfig()
|
|
svc, ctx := newTestServiceWithConfig(t, ds, testConfig, nil, lq, &TestServerOpts{
|
|
FailingPolicySet: failingPolicySet,
|
|
Clock: mockClock,
|
|
})
|
|
|
|
host := &fleet.Host{
|
|
ID: 5,
|
|
Platform: "darwin",
|
|
Hostname: "test.hostname",
|
|
}
|
|
|
|
lq.On("QueriesForHost", uint(5)).Return(map[string]string{}, nil)
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
return host, nil
|
|
}
|
|
ds.UpdateHostFunc = func(ctx context.Context, gotHost *fleet.Host) error {
|
|
host = gotHost
|
|
return nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{
|
|
Features: fleet.Features{
|
|
EnableHostUsers: true,
|
|
},
|
|
WebhookSettings: fleet.WebhookSettings{
|
|
FailingPoliciesWebhook: fleet.FailingPoliciesWebhookSettings{
|
|
Enable: true,
|
|
PolicyIDs: []uint{1, 2, 3},
|
|
},
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{
|
|
"1": "select 1;", // passing policy
|
|
"2": "select * from unexistent_table;", // policy that fails to execute (e.g. missing table)
|
|
"3": "select 1 where 1 = 0;", // failing policy
|
|
}, nil
|
|
}
|
|
recordedResults := make(map[uint]*bool)
|
|
ds.RecordPolicyQueryExecutionsFunc = func(ctx context.Context, gotHost *fleet.Host, results map[uint]*bool, updated time.Time, deferred bool) error {
|
|
recordedResults = results
|
|
host = gotHost
|
|
return nil
|
|
}
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
queries, discovery, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +3 for policies
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+3, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
|
|
checkPolicyResults := func(queries map[string]string) {
|
|
hasPolicy1, hasPolicy2, hasPolicy3 := false, false, false
|
|
for name := range queries {
|
|
if strings.HasPrefix(name, hostPolicyQueryPrefix) {
|
|
switch name[len(hostPolicyQueryPrefix):] {
|
|
case "1":
|
|
hasPolicy1 = true
|
|
case "2":
|
|
hasPolicy2 = true
|
|
case "3":
|
|
hasPolicy3 = true
|
|
}
|
|
}
|
|
}
|
|
assert.True(t, hasPolicy1)
|
|
assert.True(t, hasPolicy2)
|
|
assert.True(t, hasPolicy3)
|
|
}
|
|
|
|
checkPolicyResults(queries)
|
|
|
|
ds.FlippingPoliciesForHostFunc = func(ctx context.Context, hostID uint, incomingResults map[uint]*bool) (newFailing []uint, newPassing []uint, err error) {
|
|
return []uint{3}, nil, nil
|
|
}
|
|
|
|
// Record a query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {{"col1": "val1"}}, // succeeds
|
|
hostPolicyQueryPrefix + "2": {}, // didn't execute
|
|
hostPolicyQueryPrefix + "3": {}, // fails
|
|
},
|
|
map[string]fleet.OsqueryStatus{
|
|
hostPolicyQueryPrefix + "2": 1, // didn't execute
|
|
},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.Len(t, recordedResults, 3)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.True(t, *recordedResults[1])
|
|
result, ok := recordedResults[2]
|
|
require.True(t, ok)
|
|
require.Nil(t, result)
|
|
require.NotNil(t, recordedResults[3])
|
|
require.False(t, *recordedResults[3])
|
|
|
|
cmpSets := func(expSets map[uint][]fleet.PolicySetHost) error {
|
|
actualSets, err := failingPolicySet.ListSets()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
var expSets_ []uint
|
|
for expSet := range expSets {
|
|
expSets_ = append(expSets_, expSet)
|
|
}
|
|
sort.Slice(expSets_, func(i, j int) bool {
|
|
return expSets_[i] < expSets_[j]
|
|
})
|
|
sort.Slice(actualSets, func(i, j int) bool {
|
|
return actualSets[i] < actualSets[j]
|
|
})
|
|
if !reflect.DeepEqual(actualSets, expSets_) {
|
|
return fmt.Errorf("sets mismatch: %+v vs %+v", actualSets, expSets_)
|
|
}
|
|
for expID, expHosts := range expSets {
|
|
actualHosts, err := failingPolicySet.ListHosts(expID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
sort.Slice(actualHosts, func(i, j int) bool {
|
|
return actualHosts[i].ID < actualHosts[j].ID
|
|
})
|
|
sort.Slice(expHosts, func(i, j int) bool {
|
|
return expHosts[i].ID < expHosts[j].ID
|
|
})
|
|
if !reflect.DeepEqual(actualHosts, expHosts) {
|
|
return fmt.Errorf("hosts mismatch %d: %+v vs %+v", expID, actualHosts, expHosts)
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
assert.Eventually(t, func() bool {
|
|
err = cmpSets(map[uint][]fleet.PolicySetHost{
|
|
3: {{
|
|
ID: host.ID,
|
|
Hostname: host.Hostname,
|
|
}},
|
|
})
|
|
return err == nil
|
|
}, 1*time.Minute, 250*time.Millisecond)
|
|
require.NoError(t, err)
|
|
|
|
noPolicyResults := func(queries map[string]string) {
|
|
hasAnyPolicy := false
|
|
for name := range queries {
|
|
if strings.HasPrefix(name, hostPolicyQueryPrefix) {
|
|
hasAnyPolicy = true
|
|
break
|
|
}
|
|
}
|
|
assert.False(t, hasAnyPolicy)
|
|
}
|
|
|
|
// After the first time we get policies and update the host, then there shouldn't be any policies.
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
noPolicyResults(queries)
|
|
|
|
// Let's move time forward, there should be policies now.
|
|
mockClock.AddTime(2 * time.Hour)
|
|
|
|
queries, discovery, _, err = svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
// +3 for policies
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform))+3, len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
checkPolicyResults(queries)
|
|
|
|
ds.FlippingPoliciesForHostFunc = func(ctx context.Context, hostID uint, incomingResults map[uint]*bool) (newFailing []uint, newPassing []uint, err error) {
|
|
return []uint{1}, []uint{3}, nil
|
|
}
|
|
|
|
// Record another query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {}, // 1 now fails
|
|
hostPolicyQueryPrefix + "2": {}, // didn't execute
|
|
hostPolicyQueryPrefix + "3": {{"col1": "val1"}}, // 1 now succeeds
|
|
},
|
|
map[string]fleet.OsqueryStatus{
|
|
hostPolicyQueryPrefix + "2": 1, // didn't execute
|
|
},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.Len(t, recordedResults, 3)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.False(t, *recordedResults[1])
|
|
result, ok = recordedResults[2]
|
|
require.True(t, ok)
|
|
require.Nil(t, result)
|
|
require.NotNil(t, recordedResults[3])
|
|
require.True(t, *recordedResults[3])
|
|
|
|
assert.Eventually(t, func() bool {
|
|
err = cmpSets(map[uint][]fleet.PolicySetHost{
|
|
1: {{
|
|
ID: host.ID,
|
|
Hostname: host.Hostname,
|
|
}},
|
|
3: {},
|
|
})
|
|
return err == nil
|
|
}, 1*time.Minute, 250*time.Millisecond)
|
|
require.NoError(t, err)
|
|
|
|
// Simulate webhook trigger by removing the hosts.
|
|
err = failingPolicySet.RemoveHosts(1, []fleet.PolicySetHost{{
|
|
ID: host.ID,
|
|
Hostname: host.Hostname,
|
|
}})
|
|
require.NoError(t, err)
|
|
|
|
ds.FlippingPoliciesForHostFunc = func(ctx context.Context, hostID uint, incomingResults map[uint]*bool) (newFailing []uint, newPassing []uint, err error) {
|
|
return []uint{}, []uint{2}, nil
|
|
}
|
|
|
|
// Record another query execution.
|
|
err = svc.SubmitDistributedQueryResults(
|
|
ctx,
|
|
map[string][]map[string]string{
|
|
hostPolicyQueryPrefix + "1": {}, // continues to fail
|
|
hostPolicyQueryPrefix + "2": {{"col1": "val1"}}, // now passes
|
|
hostPolicyQueryPrefix + "3": {{"col1": "val1"}}, // continues to succeed
|
|
},
|
|
map[string]fleet.OsqueryStatus{},
|
|
map[string]string{},
|
|
)
|
|
require.NoError(t, err)
|
|
require.Len(t, recordedResults, 3)
|
|
require.NotNil(t, recordedResults[1])
|
|
require.False(t, *recordedResults[1])
|
|
require.NotNil(t, recordedResults[2])
|
|
require.True(t, *recordedResults[2])
|
|
require.NotNil(t, recordedResults[3])
|
|
require.True(t, *recordedResults[3])
|
|
|
|
assert.Eventually(t, func() bool {
|
|
err = cmpSets(map[uint][]fleet.PolicySetHost{
|
|
1: {},
|
|
3: {},
|
|
})
|
|
return err == nil
|
|
}, 1*time.Minute, 250*time.Millisecond)
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
// If the live query store (Redis) is down we still (see #3503)
|
|
// want hosts to get queries and continue to check in.
|
|
func TestLiveQueriesFailing(t *testing.T) {
|
|
ds := new(mock.Store)
|
|
lq := live_query_mock.New(t)
|
|
cfg := config.TestConfig()
|
|
buf := new(bytes.Buffer)
|
|
logger := log.NewLogfmtLogger(buf)
|
|
svc, ctx := newTestServiceWithConfig(t, ds, cfg, nil, lq, &TestServerOpts{
|
|
Logger: logger,
|
|
})
|
|
|
|
hostID := uint(1)
|
|
host := &fleet.Host{
|
|
ID: hostID,
|
|
Platform: "darwin",
|
|
}
|
|
lq.On("QueriesForHost", hostID).Return(
|
|
map[string]string{},
|
|
errors.New("failed to get queries for host"),
|
|
)
|
|
|
|
ds.LabelQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
ds.HostLiteFunc = func(ctx context.Context, id uint) (*fleet.Host, error) {
|
|
return host, nil
|
|
}
|
|
ds.AppConfigFunc = func(ctx context.Context) (*fleet.AppConfig, error) {
|
|
return &fleet.AppConfig{Features: fleet.Features{EnableHostUsers: true}}, nil
|
|
}
|
|
ds.PolicyQueriesForHostFunc = func(ctx context.Context, host *fleet.Host) (map[string]string, error) {
|
|
return map[string]string{}, nil
|
|
}
|
|
|
|
ctx = hostctx.NewContext(ctx, host)
|
|
|
|
queries, discovery, _, err := svc.GetDistributedQueries(ctx)
|
|
require.NoError(t, err)
|
|
require.Equal(t, len(expectedDetailQueriesForPlatform(host.Platform)), len(queries), distQueriesMapKeys(queries))
|
|
verifyDiscovery(t, queries, discovery)
|
|
|
|
logs, err := io.ReadAll(buf)
|
|
require.NoError(t, err)
|
|
require.Contains(t, string(logs), "level=error")
|
|
require.Contains(t, string(logs), "failed to get queries for host")
|
|
}
|
|
|
|
func distQueriesMapKeys(m map[string]string) []string {
|
|
keys := make([]string, 0, len(m))
|
|
for k := range m {
|
|
keys = append(keys, strings.TrimPrefix(k, "fleet_detail_query_"))
|
|
}
|
|
sort.Strings(keys)
|
|
return keys
|
|
}
|
|
|
|
func osqueryMapKeys(m map[string]osquery_utils.DetailQuery) []string {
|
|
keys := make([]string, 0, len(m))
|
|
for k := range m {
|
|
keys = append(keys, k)
|
|
}
|
|
sort.Strings(keys)
|
|
return keys
|
|
}
|