mirror of
https://github.com/postgresml/pgcat.git
synced 2026-03-23 01:16:30 +00:00
Allow configuring routing decision when no shard is selected (#578)
The TL;DR for the change is that we allow QueryRouter to set the active shard to None. This signals to the Pool::get method that we have no shard selected. The get method follows a no_shard_specified_behavior config to know how to route the query. Original PR description Ruby-pg library makes a startup query to SET client_encoding to ... if Encoding.default_internal value is set (Code). This query is troublesome because we cannot possibly attach a routing comment to it. PgCat, by default, will route that query to the default shard. Everything is fine until shard 0 has issues, Clients will all be attempting to send this query to shard0 which increases the connection latency significantly for all clients, even those not interested in shard0 This PR introduces no_shard_specified_behavior that defines the behavior in case we have routing-by-comment enabled but we get a query without a comment. The allowed behaviors are random: Picks a shard at random random_healthy: Picks a shard at random favoring shards with the least number of recent connection/checkout errors shard_<number>: e.g. shard_0, shard_4, etc. picks a specific shard, everytime In order to achieve this, this PR introduces an error_count on the Address Object that tracks the number of errors since the last checkout and uses that metric to sort shards by error count before making a routing decision. I didn't want to use address stats to avoid introducing a routing dependency on internal stats (We might do that in the future but I prefer to avoid this for the time being. I also made changes to the test environment to replace Ruby's TOML reader library, It appears to be abandoned and does not support mixed arrays (which we use in the config toml), and it also does not play nicely with single-quoted regular expressions. I opted for using yj which is a CLI tool that can convert from toml to JSON and back. So I refactor the tests to use that library.
This commit is contained in:
committed by
GitHub
parent
33db0dffa8
commit
0b01d70b55
@@ -1,5 +1,7 @@
|
||||
FROM rust:bullseye
|
||||
|
||||
COPY --from=sclevine/yj /bin/yj /bin/yj
|
||||
RUN /bin/yj -h
|
||||
RUN apt-get update && apt-get install llvm-11 psmisc postgresql-contrib postgresql-client ruby ruby-dev libpq-dev python3 python3-pip lcov curl sudo iproute2 -y
|
||||
RUN cargo install cargo-binutils rustfilt
|
||||
RUN rustup component add llvm-tools-preview
|
||||
|
||||
@@ -185,7 +185,7 @@ describe "Auth Query" do
|
||||
},
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
context 'and with cleartext passwords set' do
|
||||
it 'it uses local passwords' do
|
||||
|
||||
@@ -33,18 +33,18 @@ module Helpers
|
||||
"0" => {
|
||||
"database" => "shard0",
|
||||
"servers" => [
|
||||
["localhost", primary.port.to_s, "primary"],
|
||||
["localhost", replica.port.to_s, "replica"],
|
||||
["localhost", primary.port.to_i, "primary"],
|
||||
["localhost", replica.port.to_i, "replica"],
|
||||
]
|
||||
},
|
||||
},
|
||||
"users" => { "0" => user.merge(config_user) }
|
||||
}
|
||||
}
|
||||
pgcat_cfg["general"]["port"] = pgcat.port
|
||||
pgcat_cfg["general"]["port"] = pgcat.port.to_i
|
||||
pgcat.update_config(pgcat_cfg)
|
||||
pgcat.start
|
||||
|
||||
|
||||
pgcat.wait_until_ready(
|
||||
pgcat.connection_string(
|
||||
"sharded_db",
|
||||
@@ -92,13 +92,13 @@ module Helpers
|
||||
"0" => {
|
||||
"database" => database,
|
||||
"servers" => [
|
||||
["localhost", primary.port.to_s, "primary"],
|
||||
["localhost", replica.port.to_s, "replica"],
|
||||
["localhost", primary.port.to_i, "primary"],
|
||||
["localhost", replica.port.to_i, "replica"],
|
||||
]
|
||||
},
|
||||
},
|
||||
"users" => { "0" => user.merge(config_user) }
|
||||
}
|
||||
}
|
||||
end
|
||||
# Main proxy configs
|
||||
pgcat_cfg["pools"] = {
|
||||
@@ -109,7 +109,7 @@ module Helpers
|
||||
pgcat_cfg["general"]["port"] = pgcat.port
|
||||
pgcat.update_config(pgcat_cfg.deep_merge(extra_conf))
|
||||
pgcat.start
|
||||
|
||||
|
||||
pgcat.wait_until_ready(pgcat.connection_string("sharded_db0", pg_user['username'], pg_user['password']))
|
||||
|
||||
OpenStruct.new.tap do |struct|
|
||||
|
||||
@@ -7,10 +7,24 @@ class PgInstance
|
||||
attr_reader :password
|
||||
attr_reader :database_name
|
||||
|
||||
def self.mass_takedown(databases)
|
||||
raise StandardError "block missing" unless block_given?
|
||||
|
||||
databases.each do |database|
|
||||
database.toxiproxy.toxic(:limit_data, bytes: 1).toxics.each(&:save)
|
||||
end
|
||||
sleep 0.1
|
||||
yield
|
||||
ensure
|
||||
databases.each do |database|
|
||||
database.toxiproxy.toxics.each(&:destroy)
|
||||
end
|
||||
end
|
||||
|
||||
def initialize(port, username, password, database_name)
|
||||
@original_port = port
|
||||
@original_port = port.to_i
|
||||
@toxiproxy_port = 10000 + port.to_i
|
||||
@port = @toxiproxy_port
|
||||
@port = @toxiproxy_port.to_i
|
||||
|
||||
@username = username
|
||||
@password = password
|
||||
@@ -48,9 +62,9 @@ class PgInstance
|
||||
|
||||
def take_down
|
||||
if block_given?
|
||||
Toxiproxy[@toxiproxy_name].toxic(:limit_data, bytes: 5).apply { yield }
|
||||
Toxiproxy[@toxiproxy_name].toxic(:limit_data, bytes: 1).apply { yield }
|
||||
else
|
||||
Toxiproxy[@toxiproxy_name].toxic(:limit_data, bytes: 5).toxics.each(&:save)
|
||||
Toxiproxy[@toxiproxy_name].toxic(:limit_data, bytes: 1).toxics.each(&:save)
|
||||
end
|
||||
end
|
||||
|
||||
@@ -89,6 +103,6 @@ class PgInstance
|
||||
end
|
||||
|
||||
def count_select_1_plus_2
|
||||
with_connection { |c| c.async_exec("SELECT SUM(calls) FROM pg_stat_statements WHERE query = 'SELECT $1 + $2'")[0]["sum"].to_i }
|
||||
with_connection { |c| c.async_exec("SELECT SUM(calls) FROM pg_stat_statements WHERE query LIKE '%SELECT $1 + $2%'")[0]["sum"].to_i }
|
||||
end
|
||||
end
|
||||
|
||||
@@ -38,9 +38,9 @@ module Helpers
|
||||
"automatic_sharding_key" => "data.id",
|
||||
"sharding_function" => "pg_bigint_hash",
|
||||
"shards" => {
|
||||
"0" => { "database" => "shard0", "servers" => [["localhost", primary0.port.to_s, "primary"]] },
|
||||
"1" => { "database" => "shard1", "servers" => [["localhost", primary1.port.to_s, "primary"]] },
|
||||
"2" => { "database" => "shard2", "servers" => [["localhost", primary2.port.to_s, "primary"]] },
|
||||
"0" => { "database" => "shard0", "servers" => [["localhost", primary0.port.to_i, "primary"]] },
|
||||
"1" => { "database" => "shard1", "servers" => [["localhost", primary1.port.to_i, "primary"]] },
|
||||
"2" => { "database" => "shard2", "servers" => [["localhost", primary2.port.to_i, "primary"]] },
|
||||
},
|
||||
"users" => { "0" => user },
|
||||
"plugins" => {
|
||||
@@ -100,7 +100,7 @@ module Helpers
|
||||
"0" => {
|
||||
"database" => "shard0",
|
||||
"servers" => [
|
||||
["localhost", primary.port.to_s, "primary"]
|
||||
["localhost", primary.port.to_i, "primary"]
|
||||
]
|
||||
},
|
||||
},
|
||||
@@ -146,10 +146,10 @@ module Helpers
|
||||
"0" => {
|
||||
"database" => "shard0",
|
||||
"servers" => [
|
||||
["localhost", primary.port.to_s, "primary"],
|
||||
["localhost", replica0.port.to_s, "replica"],
|
||||
["localhost", replica1.port.to_s, "replica"],
|
||||
["localhost", replica2.port.to_s, "replica"]
|
||||
["localhost", primary.port.to_i, "primary"],
|
||||
["localhost", replica0.port.to_i, "replica"],
|
||||
["localhost", replica1.port.to_i, "replica"],
|
||||
["localhost", replica2.port.to_i, "replica"]
|
||||
]
|
||||
},
|
||||
},
|
||||
|
||||
@@ -1,8 +1,10 @@
|
||||
require 'pg'
|
||||
require 'toml'
|
||||
require 'json'
|
||||
require 'tempfile'
|
||||
require 'fileutils'
|
||||
require 'securerandom'
|
||||
|
||||
class ConfigReloadFailed < StandardError; end
|
||||
class PgcatProcess
|
||||
attr_reader :port
|
||||
attr_reader :pid
|
||||
@@ -18,7 +20,7 @@ class PgcatProcess
|
||||
end
|
||||
|
||||
def initialize(log_level)
|
||||
@env = {"RUST_LOG" => log_level}
|
||||
@env = {}
|
||||
@port = rand(20000..32760)
|
||||
@log_level = log_level
|
||||
@log_filename = "/tmp/pgcat_log_#{SecureRandom.urlsafe_base64}.log"
|
||||
@@ -30,7 +32,7 @@ class PgcatProcess
|
||||
'../../target/debug/pgcat'
|
||||
end
|
||||
|
||||
@command = "#{command_path} #{@config_filename}"
|
||||
@command = "#{command_path} #{@config_filename} --log-level #{@log_level}"
|
||||
|
||||
FileUtils.cp("../../pgcat.toml", @config_filename)
|
||||
cfg = current_config
|
||||
@@ -46,22 +48,34 @@ class PgcatProcess
|
||||
|
||||
def update_config(config_hash)
|
||||
@original_config = current_config
|
||||
output_to_write = TOML::Generator.new(config_hash).body
|
||||
output_to_write = output_to_write.gsub(/,\s*["|'](\d+)["|']\s*,/, ',\1,')
|
||||
output_to_write = output_to_write.gsub(/,\s*["|'](\d+)["|']\s*\]/, ',\1]')
|
||||
File.write(@config_filename, output_to_write)
|
||||
Tempfile.create('json_out', '/tmp') do |f|
|
||||
f.write(config_hash.to_json)
|
||||
f.flush
|
||||
`cat #{f.path} | yj -jt > #{@config_filename}`
|
||||
end
|
||||
end
|
||||
|
||||
def current_config
|
||||
loadable_string = File.read(@config_filename)
|
||||
loadable_string = loadable_string.gsub(/,\s*(\d+)\s*,/, ', "\1",')
|
||||
loadable_string = loadable_string.gsub(/,\s*(\d+)\s*\]/, ', "\1"]')
|
||||
TOML.load(loadable_string)
|
||||
JSON.parse(`cat #{@config_filename} | yj -tj`)
|
||||
end
|
||||
|
||||
def raw_config_file
|
||||
File.read(@config_filename)
|
||||
end
|
||||
|
||||
def reload_config
|
||||
`kill -s HUP #{@pid}`
|
||||
sleep 0.5
|
||||
conn = PG.connect(admin_connection_string)
|
||||
|
||||
conn.async_exec("RELOAD")
|
||||
rescue PG::ConnectionBad => e
|
||||
errors = logs.split("Reloading config").last
|
||||
errors = errors.gsub(/\e\[([;\d]+)?m/, '') # Remove color codes
|
||||
errors = errors.
|
||||
split("\n").select{|line| line.include?("ERROR") }.
|
||||
map { |line| line.split("pgcat::config: ").last }
|
||||
raise ConfigReloadFailed, errors.join("\n")
|
||||
ensure
|
||||
conn&.close
|
||||
end
|
||||
|
||||
def start
|
||||
@@ -116,11 +130,11 @@ class PgcatProcess
|
||||
cfg = current_config
|
||||
user_idx, user_obj = cfg["pools"][pool_name]["users"].detect { |k, user| user["username"] == username }
|
||||
connection_string = "postgresql://#{username}:#{password || user_obj["password"]}@0.0.0.0:#{@port}/#{pool_name}"
|
||||
|
||||
|
||||
# Add the additional parameters to the connection string
|
||||
parameter_string = parameters.map { |key, value| "#{key}=#{value}" }.join("&")
|
||||
connection_string += "?#{parameter_string}" unless parameter_string.empty?
|
||||
|
||||
|
||||
connection_string
|
||||
end
|
||||
|
||||
|
||||
@@ -11,9 +11,9 @@ describe "Query Mirroing" do
|
||||
before do
|
||||
new_configs = processes.pgcat.current_config
|
||||
new_configs["pools"]["sharded_db"]["shards"]["0"]["mirrors"] = [
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
]
|
||||
processes.pgcat.update_config(new_configs)
|
||||
processes.pgcat.reload_config
|
||||
@@ -31,7 +31,8 @@ describe "Query Mirroing" do
|
||||
runs.times { conn.async_exec("SELECT 1 + 2") }
|
||||
sleep 0.5
|
||||
expect(processes.all_databases.first.count_select_1_plus_2).to eq(runs)
|
||||
expect(mirror_pg.count_select_1_plus_2).to eq(runs * 3)
|
||||
# Allow some slack in mirroring successes
|
||||
expect(mirror_pg.count_select_1_plus_2).to be > ((runs - 5) * 3)
|
||||
end
|
||||
|
||||
context "when main server connection is closed" do
|
||||
@@ -42,9 +43,9 @@ describe "Query Mirroing" do
|
||||
new_configs = processes.pgcat.current_config
|
||||
new_configs["pools"]["sharded_db"]["idle_timeout"] = 5000 + i
|
||||
new_configs["pools"]["sharded_db"]["shards"]["0"]["mirrors"] = [
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_s, "0"],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
[mirror_host, mirror_pg.port.to_i, 0],
|
||||
]
|
||||
processes.pgcat.update_config(new_configs)
|
||||
processes.pgcat.reload_config
|
||||
|
||||
@@ -252,7 +252,7 @@ describe "Miscellaneous" do
|
||||
end
|
||||
|
||||
expect(processes.primary.count_query("RESET ROLE")).to eq(10)
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
context "transaction mode" do
|
||||
@@ -317,7 +317,7 @@ describe "Miscellaneous" do
|
||||
conn.async_exec("SET statement_timeout to 1500")
|
||||
expect(conn.async_exec("SHOW statement_timeout")[0]["statement_timeout"]).to eq(orignal_statement_timeout)
|
||||
end
|
||||
|
||||
|
||||
end
|
||||
|
||||
context "transaction mode with transactions" do
|
||||
@@ -354,7 +354,6 @@ describe "Miscellaneous" do
|
||||
conn.async_exec("SET statement_timeout TO 1000")
|
||||
conn.close
|
||||
|
||||
puts processes.pgcat.logs
|
||||
expect(processes.primary.count_query("RESET ALL")).to eq(0)
|
||||
end
|
||||
|
||||
@@ -365,7 +364,6 @@ describe "Miscellaneous" do
|
||||
|
||||
conn.close
|
||||
|
||||
puts processes.pgcat.logs
|
||||
expect(processes.primary.count_query("RESET ALL")).to eq(0)
|
||||
end
|
||||
end
|
||||
@@ -376,10 +374,9 @@ describe "Miscellaneous" do
|
||||
before do
|
||||
current_configs = processes.pgcat.current_config
|
||||
correct_idle_client_transaction_timeout = current_configs["general"]["idle_client_in_transaction_timeout"]
|
||||
puts(current_configs["general"]["idle_client_in_transaction_timeout"])
|
||||
|
||||
|
||||
current_configs["general"]["idle_client_in_transaction_timeout"] = 0
|
||||
|
||||
|
||||
processes.pgcat.update_config(current_configs) # with timeout 0
|
||||
processes.pgcat.reload_config
|
||||
end
|
||||
@@ -397,9 +394,9 @@ describe "Miscellaneous" do
|
||||
context "idle transaction timeout set to 500ms" do
|
||||
before do
|
||||
current_configs = processes.pgcat.current_config
|
||||
correct_idle_client_transaction_timeout = current_configs["general"]["idle_client_in_transaction_timeout"]
|
||||
correct_idle_client_transaction_timeout = current_configs["general"]["idle_client_in_transaction_timeout"]
|
||||
current_configs["general"]["idle_client_in_transaction_timeout"] = 500
|
||||
|
||||
|
||||
processes.pgcat.update_config(current_configs) # with timeout 500
|
||||
processes.pgcat.reload_config
|
||||
end
|
||||
@@ -418,7 +415,7 @@ describe "Miscellaneous" do
|
||||
conn.async_exec("BEGIN")
|
||||
conn.async_exec("SELECT 1")
|
||||
sleep(1) # above 500ms
|
||||
expect{ conn.async_exec("COMMIT") }.to raise_error(PG::SystemError, /idle transaction timeout/)
|
||||
expect{ conn.async_exec("COMMIT") }.to raise_error(PG::SystemError, /idle transaction timeout/)
|
||||
conn.async_exec("SELECT 1") # should be able to send another query
|
||||
conn.close
|
||||
end
|
||||
|
||||
@@ -7,11 +7,11 @@ describe "Sharding" do
|
||||
|
||||
before do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
|
||||
# Setup the sharding data
|
||||
3.times do |i|
|
||||
conn.exec("SET SHARD TO '#{i}'")
|
||||
conn.exec("DELETE FROM data WHERE id > 0")
|
||||
|
||||
conn.exec("DELETE FROM data WHERE id > 0") rescue nil
|
||||
end
|
||||
|
||||
18.times do |i|
|
||||
@@ -19,10 +19,11 @@ describe "Sharding" do
|
||||
conn.exec("SET SHARDING KEY TO '#{i}'")
|
||||
conn.exec("INSERT INTO data (id, value) VALUES (#{i}, 'value_#{i}')")
|
||||
end
|
||||
|
||||
conn.close
|
||||
end
|
||||
|
||||
after do
|
||||
|
||||
processes.all_databases.map(&:reset)
|
||||
processes.pgcat.shutdown
|
||||
end
|
||||
@@ -48,4 +49,148 @@ describe "Sharding" do
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
describe "no_shard_specified_behavior config" do
|
||||
context "when default shard number is invalid" do
|
||||
it "prevents config reload" do
|
||||
admin_conn = PG::connect(processes.pgcat.admin_connection_string)
|
||||
|
||||
current_configs = processes.pgcat.current_config
|
||||
current_configs["pools"]["sharded_db"]["default_shard"] = "shard_99"
|
||||
|
||||
processes.pgcat.update_config(current_configs)
|
||||
|
||||
expect { processes.pgcat.reload_config }.to raise_error(ConfigReloadFailed, /Invalid shard 99/)
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
describe "comment-based routing" do
|
||||
context "when no configs are set" do
|
||||
it "routes queries with a shard_id comment to the default shard" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
10.times { conn.async_exec("/* shard_id: 2 */ SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([10, 0, 0])
|
||||
end
|
||||
|
||||
it "does not honor no_shard_specified_behavior directives" do
|
||||
end
|
||||
end
|
||||
|
||||
[
|
||||
["shard_id_regex", "/\\* the_shard_id: (\\d+) \\*/", "/* the_shard_id: 1 */"],
|
||||
["sharding_key_regex", "/\\* the_sharding_key: (\\d+) \\*/", "/* the_sharding_key: 3 */"],
|
||||
].each do |config_name, config_value, comment_to_use|
|
||||
context "when #{config_name} config is set" do
|
||||
let(:no_shard_specified_behavior) { nil }
|
||||
|
||||
before do
|
||||
admin_conn = PG::connect(processes.pgcat.admin_connection_string)
|
||||
|
||||
current_configs = processes.pgcat.current_config
|
||||
current_configs["pools"]["sharded_db"][config_name] = config_value
|
||||
if no_shard_specified_behavior
|
||||
current_configs["pools"]["sharded_db"]["default_shard"] = no_shard_specified_behavior
|
||||
else
|
||||
current_configs["pools"]["sharded_db"].delete("default_shard")
|
||||
end
|
||||
|
||||
processes.pgcat.update_config(current_configs)
|
||||
processes.pgcat.reload_config
|
||||
end
|
||||
|
||||
it "routes queries with a shard_id comment to the correct shard" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("#{comment_to_use} SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([0, 25, 0])
|
||||
end
|
||||
|
||||
context "when no_shard_specified_behavior config is set to random" do
|
||||
let(:no_shard_specified_behavior) { "random" }
|
||||
|
||||
context "with no shard comment" do
|
||||
it "sends queries to random shard" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2).all?(&:positive?)).to be true
|
||||
end
|
||||
end
|
||||
|
||||
context "with a shard comment" do
|
||||
it "honors the comment" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("#{comment_to_use} SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([0, 25, 0])
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
context "when no_shard_specified_behavior config is set to random_healthy" do
|
||||
let(:no_shard_specified_behavior) { "random_healthy" }
|
||||
|
||||
context "with no shard comment" do
|
||||
it "sends queries to random healthy shard" do
|
||||
|
||||
good_databases = [processes.all_databases[0], processes.all_databases[2]]
|
||||
bad_database = processes.all_databases[1]
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
250.times { conn.async_exec("SELECT 99") }
|
||||
bad_database.take_down do
|
||||
250.times do
|
||||
conn.async_exec("SELECT 99")
|
||||
rescue PG::ConnectionBad => e
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
end
|
||||
end
|
||||
|
||||
# Routes traffic away from bad shard
|
||||
25.times { conn.async_exec("SELECT 1 + 2") }
|
||||
expect(good_databases.map(&:count_select_1_plus_2).all?(&:positive?)).to be true
|
||||
expect(bad_database.count_select_1_plus_2).to eq(0)
|
||||
|
||||
# Routes traffic to the bad shard if the shard_id is specified
|
||||
25.times { conn.async_exec("#{comment_to_use} SELECT 1 + 2") }
|
||||
bad_database = processes.all_databases[1]
|
||||
expect(bad_database.count_select_1_plus_2).to eq(25)
|
||||
end
|
||||
end
|
||||
|
||||
context "with a shard comment" do
|
||||
it "honors the comment" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("#{comment_to_use} SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([0, 25, 0])
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
context "when no_shard_specified_behavior config is set to shard_x" do
|
||||
let(:no_shard_specified_behavior) { "shard_2" }
|
||||
|
||||
context "with no shard comment" do
|
||||
it "sends queries to the specified shard" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([0, 0, 25])
|
||||
end
|
||||
end
|
||||
|
||||
context "with a shard comment" do
|
||||
it "honors the comment" do
|
||||
conn = PG.connect(processes.pgcat.connection_string("sharded_db", "sharding_user"))
|
||||
25.times { conn.async_exec("#{comment_to_use} SELECT 1 + 2") }
|
||||
|
||||
expect(processes.all_databases.map(&:count_select_1_plus_2)).to eq([0, 25, 0])
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
end
|
||||
|
||||
Reference in New Issue
Block a user