Skip to content
Snippets Groups Projects
Commit ed7c1f2a authored by Sergey Yakubov's avatar Sergey Yakubov
Browse files

discover mongodb via discovery service, not dns

parent fd907cf6
No related branches found
No related tags found
No related merge requests found
Showing
with 60 additions and 60 deletions
...@@ -53,6 +53,13 @@ using asapo::ReceiverConfig; ...@@ -53,6 +53,13 @@ using asapo::ReceiverConfig;
namespace { namespace {
TEST(DbWriterHandler, Constructor) {
RequestHandlerDbWrite handler{""};
ASSERT_THAT(dynamic_cast<asapo::HttpClient*>(handler.http_client__.get()), Ne(nullptr));
}
class DbWriterHandlerTests : public Test { class DbWriterHandlerTests : public Test {
public: public:
std::string expected_collection_name = asapo::kDBDataCollectionName; std::string expected_collection_name = asapo::kDBDataCollectionName;
...@@ -81,7 +88,7 @@ class DbWriterHandlerTests : public Test { ...@@ -81,7 +88,7 @@ class DbWriterHandlerTests : public Test {
handler.db_client__ = std::unique_ptr<asapo::Database> {&mock_db}; handler.db_client__ = std::unique_ptr<asapo::Database> {&mock_db};
handler.log__ = &mock_logger; handler.log__ = &mock_logger;
mock_request.reset(new NiceMock<MockRequest> {request_header, 1, ""}); mock_request.reset(new NiceMock<MockRequest> {request_header, 1, ""});
config.broker_db_uri = "127.0.0.1:27017"; config.database_uri = "127.0.0.1:27017";
config.source_host = expected_hostname; config.source_host = expected_hostname;
config.dataserver.listen_port = expected_port; config.dataserver.listen_port = expected_port;
SetReceiverConfig(config, "none"); SetReceiverConfig(config, "none");
...@@ -127,7 +134,7 @@ void DbWriterHandlerTests::ExpectRequestParams(asapo::Opcode op_code, const std: ...@@ -127,7 +134,7 @@ void DbWriterHandlerTests::ExpectRequestParams(asapo::Opcode op_code, const std:
std::string db_name = expected_beamtime_id; std::string db_name = expected_beamtime_id;
db_name += "_" + stream; db_name += "_" + stream;
EXPECT_CALL(mock_db, Connect_t(config.broker_db_uri, db_name, expected_collection_name)). EXPECT_CALL(mock_db, Connect_t(config.database_uri, db_name, expected_collection_name)).
WillOnce(testing::Return(nullptr)); WillOnce(testing::Return(nullptr));
EXPECT_CALL(*mock_request, GetDataSize()) EXPECT_CALL(*mock_request, GetDataSize())
...@@ -180,7 +187,7 @@ TEST_F(DbWriterHandlerTests, CallsInsert) { ...@@ -180,7 +187,7 @@ TEST_F(DbWriterHandlerTests, CallsInsert) {
WillOnce(testing::Return(nullptr)); WillOnce(testing::Return(nullptr));
EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("insert record"), EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("insert record"),
HasSubstr(config.broker_db_uri), HasSubstr(config.database_uri),
HasSubstr(expected_beamtime_id), HasSubstr(expected_beamtime_id),
HasSubstr(expected_stream), HasSubstr(expected_stream),
HasSubstr(expected_collection_name) HasSubstr(expected_collection_name)
...@@ -201,7 +208,7 @@ TEST_F(DbWriterHandlerTests, CallsInsertSubset) { ...@@ -201,7 +208,7 @@ TEST_F(DbWriterHandlerTests, CallsInsertSubset) {
WillOnce(testing::Return(nullptr)); WillOnce(testing::Return(nullptr));
EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("insert record"), EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("insert record"),
HasSubstr(config.broker_db_uri), HasSubstr(config.database_uri),
HasSubstr(expected_beamtime_id), HasSubstr(expected_beamtime_id),
HasSubstr(expected_collection_name) HasSubstr(expected_collection_name)
) )
......
...@@ -65,8 +65,8 @@ class SenderFluentdTests : public Test { ...@@ -65,8 +65,8 @@ class SenderFluentdTests : public Test {
statistics.tags.push_back(std::make_pair("name1", "value1")); statistics.tags.push_back(std::make_pair("name1", "value1"));
statistics.tags.push_back(std::make_pair("name2", "value2")); statistics.tags.push_back(std::make_pair("name2", "value2"));
config.monitor_db_uri = "test_uri"; config.performance_db_uri = "test_uri";
config.monitor_db_name = "test_name"; config.performance_db_name = "test_name";
SetReceiverConfig(config, "none"); SetReceiverConfig(config, "none");
sender.statistics_log__.reset(&mock_logger); sender.statistics_log__.reset(&mock_logger);
......
...@@ -66,8 +66,8 @@ class SenderInfluxDbTests : public Test { ...@@ -66,8 +66,8 @@ class SenderInfluxDbTests : public Test {
statistics.tags.push_back(std::make_pair("name1", "value1")); statistics.tags.push_back(std::make_pair("name1", "value1"));
statistics.tags.push_back(std::make_pair("name2", "value2")); statistics.tags.push_back(std::make_pair("name2", "value2"));
config.monitor_db_uri = "test_uri"; config.performance_db_uri = "test_uri";
config.monitor_db_name = "test_name"; config.performance_db_name = "test_name";
SetReceiverConfig(config, "none"); SetReceiverConfig(config, "none");
sender.httpclient__.reset(&mock_http_client); sender.httpclient__.reset(&mock_http_client);
...@@ -88,7 +88,7 @@ TEST_F(SenderInfluxDbTests, SendStatisticsCallsPost) { ...@@ -88,7 +88,7 @@ TEST_F(SenderInfluxDbTests, SendStatisticsCallsPost) {
Return("") Return("")
)); ));
EXPECT_CALL(mock_logger, Error(AllOf(HasSubstr("sending statistics"), HasSubstr(config.monitor_db_uri)))); EXPECT_CALL(mock_logger, Error(AllOf(HasSubstr("sending statistics"), HasSubstr(config.performance_db_uri))));
sender.SendStatistics(statistics); sender.SendStatistics(statistics);
...@@ -113,8 +113,8 @@ TEST_F(SenderInfluxDbTests, LogDebugSendStatistics) { ...@@ -113,8 +113,8 @@ TEST_F(SenderInfluxDbTests, LogDebugSendStatistics) {
)); ));
EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("sending statistics"), EXPECT_CALL(mock_logger, Debug(AllOf(HasSubstr("sending statistics"),
HasSubstr(config.monitor_db_uri), HasSubstr(config.performance_db_uri),
HasSubstr(config.monitor_db_name) HasSubstr(config.performance_db_name)
) )
) )
); );
......
...@@ -55,7 +55,7 @@ Cleanup() { ...@@ -55,7 +55,7 @@ Cleanup() {
influx -execute "create database ${monitor_database_name}" influx -execute "create database ${monitor_database_name}"
sed -i 's/27017/27016/g' receiver.json.tpl sed -i 's/27017/27016/g' receiver.json.tpl
sed -i 's/27017/27016/g' broker.json.tpl sed -i 's/27017/27016/g' discovery.json.tpl
sed -i 's/info/debug/g' broker.json.tpl sed -i 's/info/debug/g' broker.json.tpl
start_mongo start_mongo
......
...@@ -54,7 +54,7 @@ wait_mongo ...@@ -54,7 +54,7 @@ wait_mongo
# create db before worker starts reading it. todo: git rid of it # create db before worker starts reading it. todo: git rid of it
echo "db.${beamtime_id}_detector.insert({dummy:1})" | mongo --port 27016 ${beamtime_id}_detector echo "db.${beamtime_id}_detector.insert({dummy:1})" | mongo --port 27016 ${beamtime_id}_detector
sed -i 's/27017/27016/g' receiver.json.tpl sed -i 's/27017/27016/g' discovery.json.tpl
nomad run authorizer.nmd nomad run authorizer.nmd
......
{ {
"BrokerDbAddress":"127.0.0.1:27017", "DatabaseServer":"127.0.0.1:27017",
"MonitorDbAddress": "localhost:8086", "PerformanceDbServer": "localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"port":5005, "port":5005,
"LogLevel":"info", "LogLevel":"info",
"SecretFile":"auth_secret.key" "SecretFile":"auth_secret.key"
......
{ {
"BrokerDbAddress":"127.0.0.1:27017", "DatabaseServer":"auto",
"MonitorDbAddress": "localhost:8086", "DiscoveryServer": "localhost:8400/discovery",
"MonitorDbName": "db_test", "PerformanceDbServer": "localhost:8086",
"PerformanceDbName": "db_test",
"port":{{ env "NOMAD_PORT_broker" }}, "port":{{ env "NOMAD_PORT_broker" }},
"LogLevel":"info", "LogLevel":"info",
"SecretFile":"auth_secret.key" "SecretFile":"auth_secret.key"
......
{
"Mode": "consul",
"Receiver": {
"MaxConnections": 32
},
"Port": {{ env "NOMAD_PORT_discovery" }},
"LogLevel":"debug"
}
...@@ -3,6 +3,9 @@ ...@@ -3,6 +3,9 @@
"Receiver": { "Receiver": {
"MaxConnections": 32 "MaxConnections": 32
}, },
"Mongo": {
"StaticEndpoint": "127.0.0.1:27017"
},
"Port": {{ env "NOMAD_PORT_discovery" }}, "Port": {{ env "NOMAD_PORT_discovery" }},
"LogLevel":"debug" "LogLevel":"debug"
} }
......
...@@ -22,15 +22,9 @@ http { ...@@ -22,15 +22,9 @@ http {
set $fluentd_endpoint fluentd.service.asapo; set $fluentd_endpoint fluentd.service.asapo;
set $kibana_endpoint kibana.service.asapo; set $kibana_endpoint kibana.service.asapo;
set $grafana_endpoint grafana.service.asapo; set $grafana_endpoint grafana.service.asapo;
set $mongo_endpoint mongo.service.asapo;
set $influxdb_endpoint influxdb.service.asapo; set $influxdb_endpoint influxdb.service.asapo;
set $elasticsearch_endpoint elasticsearch.service.asapo; set $elasticsearch_endpoint elasticsearch.service.asapo;
location /mongo/ {
rewrite ^/mongo(/.*) $1 break;
proxy_pass http://$mongo_endpoint:27017$uri$is_args$args;
}
location /influxdb/ { location /influxdb/ {
rewrite ^/influxdb(/.*) $1 break; rewrite ^/influxdb(/.*) $1 break;
proxy_pass http://$influxdb_endpoint:8086$uri$is_args$args; proxy_pass http://$influxdb_endpoint:8086$uri$is_args$args;
......
{ {
"MonitorDbAddress":"localhost:8086", "PerformanceDbServer":"localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"auto",
"DiscoveryServer": "localhost:8400/discovery",
"DataServer": { "DataServer": {
"NThreads": 2, "NThreads": 2,
"ListenPort": {{ env "NOMAD_PORT_recv_ds" }} "ListenPort": {{ env "NOMAD_PORT_recv_ds" }}
......
{ {
"MonitorDbAddress":"localhost:8086", "PerformanceDbServer":"localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"auto",
"DiscoveryServer": "localhost:8400/discovery",
"AuthorizationServer": "localhost:8400/authorizer", "AuthorizationServer": "localhost:8400/authorizer",
"AuthorizationInterval": 10000, "AuthorizationInterval": 10000,
"ListenPort": {{ env "NOMAD_PORT_recv" }}, "ListenPort": {{ env "NOMAD_PORT_recv" }},
......
{ {
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"localhost:27017",
"MonitorDbAddress": "localhost:8086", "PerformanceDbServer": "localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"port":5005, "port":5005,
"LogLevel":"info", "LogLevel":"info",
"SecretFile":"auth_secret.key" "SecretFile":"auth_secret.key"
......
...@@ -22,7 +22,7 @@ worker_dir=~/broker_test ...@@ -22,7 +22,7 @@ worker_dir=~/broker_test
service_dir=~/broker_test service_dir=~/broker_test
cat settings.json | jq ".MonitorDbAddress = \"${monitor_node}:${monitor_port}\"" > settings_tmp.json cat settings.json | jq ".PerformanceDbServer = \"${monitor_node}:${monitor_port}\"" > settings_tmp.json
cat discovery.json | jq ".Broker.StaticEndpoint = \"${service_node}:5005\"" > discovery_tmp.json cat discovery.json | jq ".Broker.StaticEndpoint = \"${service_node}:5005\"" > discovery_tmp.json
......
{ {
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"localhost:27017",
"MonitorDbAddress": "localhost:8086", "PerformanceDbServer": "localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"port":5005, "port":5005,
"LogLevel":"info", "LogLevel":"info",
"SecretFile":"auth_secret.key" "SecretFile":"auth_secret.key"
......
{ {
"MonitorDbAddress":"localhost:8086", "PerformanceDbServer":"localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"localhost:27017",
"AuthorizationServer": "localhost:5007", "AuthorizationServer": "localhost:5007",
"DiscoveryServer": "localhost:8400/discovery",
"AuthorizationInterval": 10000, "AuthorizationInterval": 10000,
"ListenPort":4200, "ListenPort":4200,
"DataServer": { "DataServer": {
......
...@@ -43,7 +43,7 @@ ssh ${receiver_node} mkdir -p ${receiver_dir}/files/${beamline}/${beamtime_id} ...@@ -43,7 +43,7 @@ ssh ${receiver_node} mkdir -p ${receiver_dir}/files/${beamline}/${beamtime_id}
scp ../../../cmake-build-release/receiver/receiver ${receiver_node}:${receiver_dir} scp ../../../cmake-build-release/receiver/receiver ${receiver_node}:${receiver_dir}
cat receiver.json | cat receiver.json |
jq "to_entries | jq "to_entries |
map(if .key == \"MonitorDbAddress\" map(if .key == \"PerformanceDbServer\"
then . + {value:\"${monitor_node}:${monitor_port}\"} then . + {value:\"${monitor_node}:${monitor_port}\"}
elif .key == \"ListenPort\" elif .key == \"ListenPort\"
then . + {value:${receiver_port}} then . + {value:${receiver_port}}
...@@ -96,7 +96,7 @@ broker_dir=~/fullchain_tests ...@@ -96,7 +96,7 @@ broker_dir=~/fullchain_tests
ssh ${broker_node} mkdir -p ${broker_dir}/logs ssh ${broker_node} mkdir -p ${broker_dir}/logs
cat broker.json | cat broker.json |
jq "to_entries | jq "to_entries |
map(if .key == \"MonitorDbAddress\" map(if .key == \"PerformanceDbServer\"
then . + {value:\"${monitor_node}:${monitor_port}\"} then . + {value:\"${monitor_node}:${monitor_port}\"}
else . else .
end end
......
{ {
"MonitorDbAddress":"localhost:8086", "PerformanceDbServer":"localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"localhost:27017",
"DiscoveryServer": "localhost:8400/discovery",
"AuthorizationServer": "localhost:5007", "AuthorizationServer": "localhost:5007",
"AuthorizationInterval": 10000, "AuthorizationInterval": 10000,
"ListenPort":4200, "ListenPort":4200,
......
...@@ -54,7 +54,7 @@ scp ../../../cmake-build-release/examples/producer/dummy-data-producer/dummy-dat ...@@ -54,7 +54,7 @@ scp ../../../cmake-build-release/examples/producer/dummy-data-producer/dummy-dat
function do_work { function do_work {
cat receiver.json | cat receiver.json |
jq "to_entries | jq "to_entries |
map(if .key == \"MonitorDbAddress\" map(if .key == \"PerformanceDbServer\"
then . + {value:\"${monitor_node}:${monitor_port}\"} then . + {value:\"${monitor_node}:${monitor_port}\"}
elif .key == \"ListenPort\" elif .key == \"ListenPort\"
then . + {value:${receiver_port}} then . + {value:${receiver_port}}
......
{ {
"MonitorDbAddress":"localhost:8086", "PerformanceDbServer":"localhost:8086",
"MonitorDbName": "db_test", "PerformanceDbName": "db_test",
"BrokerDbAddress":"localhost:27017", "DatabaseServer":"localhost:27017",
"DiscoveryServer": "localhost:8400/discovery",
"DataServer": { "DataServer": {
"NThreads": 2, "NThreads": 2,
"ListenPort": {{ env "NOMAD_PORT_recv_ds" }} "ListenPort": {{ env "NOMAD_PORT_recv_ds" }}
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment