Skip to content

Commit

Permalink
workload: add indexes workload
Browse files Browse the repository at this point in the history
Closes #34912.

This adds a new workload called `indexes`. Unlike `kv`, the workload is
specifically designed to stress secondary indexes. Its schema looks
like:

```
CREATE TABLE indexes (
    key     UUID  NOT NULL PRIMARY KEY,
    col0    INT   NOT NULL,
    col1    INT   NOT NULL,
    col2    INT   NOT NULL,
    col3    INT   NOT NULL,
    col4    INT   NOT NULL,
    col5    INT   NOT NULL,
    col6    INT   NOT NULL,
    col7    INT   NOT NULL,
    col8    INT   NOT NULL,
    col9    INT   NOT NULL,
    payload BYTES NOT NULL
)
```

and it includes knobs to specify how many of the "col" columns should be
indexed and whether they should be indexed using a UNIQUE specifier. The
workload also includes a knob to specify the size of the `payload`
column.

Release note: None
  • Loading branch information
nvanbenschoten committed Feb 25, 2019
1 parent 873bf74 commit 74e3029
Show file tree
Hide file tree
Showing 3 changed files with 195 additions and 1 deletion.
1 change: 1 addition & 0 deletions pkg/ccl/workloadccl/allccl/all.go
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@ import (
_ "github.com/cockroachdb/cockroach/pkg/ccl/workloadccl/roachmartccl"
_ "github.com/cockroachdb/cockroach/pkg/workload/bank"
_ "github.com/cockroachdb/cockroach/pkg/workload/examples"
_ "github.com/cockroachdb/cockroach/pkg/workload/indexes"
_ "github.com/cockroachdb/cockroach/pkg/workload/interleavedpartitioned"
_ "github.com/cockroachdb/cockroach/pkg/workload/jsonload"
_ "github.com/cockroachdb/cockroach/pkg/workload/kv"
Expand Down
193 changes: 193 additions & 0 deletions pkg/workload/indexes/indexes.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,193 @@
// Copyright 2019 The Cockroach Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
// implied. See the License for the specific language governing
// permissions and limitations under the License.

package indexes

import (
"context"
gosql "database/sql"
"fmt"
"math"
"math/rand"
"strings"

"github.com/cockroachdb/cockroach/pkg/util/timeutil"
"github.com/cockroachdb/cockroach/pkg/util/uint128"
"github.com/cockroachdb/cockroach/pkg/util/uuid"
"github.com/cockroachdb/cockroach/pkg/workload"
"github.com/pkg/errors"
"github.com/spf13/pflag"
)

const (
schemaBase = `(
key UUID NOT NULL PRIMARY KEY,
col0 INT NOT NULL,
col1 INT NOT NULL,
col2 INT NOT NULL,
col3 INT NOT NULL,
col4 INT NOT NULL,
col5 INT NOT NULL,
col6 INT NOT NULL,
col7 INT NOT NULL,
col8 INT NOT NULL,
col9 INT NOT NULL,
payload BYTES NOT NULL`
)

type indexes struct {
flags workload.Flags
connFlags *workload.ConnFlags

seed int64
idxs int
unique bool
payload int
}

func init() {
workload.Register(indexesMeta)
}

var indexesMeta = workload.Meta{
Name: `indexes`,
Description: `Indexes writes to a table with a variable number of secondary indexes`,
Version: `1.0.0`,
New: func() workload.Generator {
g := &indexes{}
g.flags.FlagSet = pflag.NewFlagSet(`indexes`, pflag.ContinueOnError)
g.flags.Int64Var(&g.seed, `seed`, 1, `Key hash seed.`)
g.flags.IntVar(&g.idxs, `secondary-indexes`, 1, `Number of indexes to add to the table.`)
g.flags.BoolVar(&g.unique, `unique-indexes`, false, `Use UNIQUE secondary indexes.`)
g.flags.IntVar(&g.payload, `payload`, 64, `Size of the unindexed payload column.`)
g.connFlags = workload.NewConnFlags(&g.flags)
return g
},
}

// Meta implements the Generator interface.
func (*indexes) Meta() workload.Meta { return indexesMeta }

// Flags implements the Flagser interface.
func (w *indexes) Flags() workload.Flags { return w.flags }

// Hooks implements the Hookser interface.
func (w *indexes) Hooks() workload.Hooks {
return workload.Hooks{
Validate: func() error {
if w.idxs < 0 || w.idxs > 10 {
return errors.Errorf(`--secondary-indexes must be in range [0, 10]`)
}
if w.payload < 1 {
return errors.Errorf(`--payload size must be equal to or greater than 1`)
}
return nil
},
PostLoad: func(sqlDB *gosql.DB) error {
// Split at the beginning of each index so that as long as the
// table has a single index, all writes will be multi-range.
for i := 0; i < w.idxs; i++ {
split := fmt.Sprintf(`ALTER INDEX idx%d SPLIT AT VALUES (%d)`, i, math.MinInt64)
if _, err := sqlDB.Exec(split); err != nil {
return err
}
}
return nil
},
}
}

// Tables implements the Generator interface.
func (w *indexes) Tables() []workload.Table {
// Construct the schema with all indexes.
var unique string
if w.unique {
unique = "UNIQUE "
}
var b strings.Builder
b.WriteString(schemaBase)
for i := 0; i < w.idxs; i++ {
fmt.Fprintf(&b, ",\n\t\t%sINDEX idx%d (col%d)", unique, i, i)
}
b.WriteString("\n)")

return []workload.Table{{
Name: `indexes`,
Schema: b.String(),
}}
}

// Ops implements the Opser interface.
func (w *indexes) Ops(urls []string, reg *workload.HistogramRegistry) (workload.QueryLoad, error) {
ctx := context.Background()
sqlDatabase, err := workload.SanitizeUrls(w, w.connFlags.DBOverride, urls)
if err != nil {
return workload.QueryLoad{}, err
}
mcp, err := workload.NewMultiConnPool(w.connFlags.Concurrency+1, urls...)
if err != nil {
return workload.QueryLoad{}, err
}

ql := workload.QueryLoad{SQLDatabase: sqlDatabase}
const stmt = `UPSERT INTO indexes VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12)`
for i := 0; i < w.connFlags.Concurrency; i++ {
op := &indexesOp{
config: w,
hists: reg.GetHandle(),
rand: rand.New(rand.NewSource(int64((i + 1)) * w.seed)),
buf: make([]byte, w.payload),
}
op.stmt = op.sr.Define(stmt)
if err := op.sr.Init(ctx, "indexes", mcp, w.connFlags); err != nil {
return workload.QueryLoad{}, err
}
ql.WorkerFns = append(ql.WorkerFns, op.run)
}
return ql, nil
}

type indexesOp struct {
config *indexes
hists *workload.Histograms
rand *rand.Rand
sr workload.SQLRunner
stmt workload.StmtHandle
buf []byte
}

func (o *indexesOp) run(ctx context.Context) error {
keyHi, keyLo := o.rand.Uint64(), o.rand.Uint64()
_, _ = o.rand.Read(o.buf[:])
args := []interface{}{
uuid.FromUint128(uint128.FromInts(keyHi, keyLo)).String(), // key
int64(keyLo + 0), // col0
int64(keyLo + 1), // col1
int64(keyLo + 2), // col2
int64(keyLo + 3), // col3
int64(keyLo + 4), // col4
int64(keyLo + 5), // col5
int64(keyLo + 6), // col6
int64(keyLo + 7), // col7
int64(keyLo + 8), // col8
int64(keyLo + 9), // col9
o.buf[:], // payload
}

start := timeutil.Now()
_, err := o.stmt.Exec(ctx, args...)
elapsed := timeutil.Since(start)
o.hists.Get(`write`).Record(elapsed)
return err
}
2 changes: 1 addition & 1 deletion pkg/workload/kv/kv.go
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ func init() {
var kvMeta = workload.Meta{
Name: `kv`,
Description: `
KV reads and writes to keys spread (by default, uniformly at random) across
KV reads and writes to keys spread (by default, uniformly at random) across
the cluster.
--concurrency workers alternate between doing selects and upserts (according
to a --read-percent ratio). Each select/upsert reads/writes a batch of --batch
Expand Down

0 comments on commit 74e3029

Please sign in to comment.