Skip to content

Commit

Permalink
chore: update CAS ipfs ports as well
Browse files Browse the repository at this point in the history
also renamed some constants
  • Loading branch information
dav1do committed Jul 2, 2024
1 parent 3eff430 commit 4680223
Show file tree
Hide file tree
Showing 11 changed files with 30 additions and 30 deletions.
8 changes: 4 additions & 4 deletions operator/src/network/cas.rs
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ use crate::{
network::{
ceramic::NetworkConfig,
controller::{
CAS_APP, CAS_IPFS_APP, CAS_IPFS_SERVICE_NAME, CAS_POSTGRES_APP,
CAS_APP, CAS_IPFS_APP, CAS_IPFS_SERVICE_NAME, CAS_IPFS_SERVICE_PORT, CAS_POSTGRES_APP,
CAS_POSTGRES_SECRET_NAME, CAS_POSTGRES_SERVICE_NAME, CAS_SERVICE_NAME,
DEFAULT_METRICS_PORT, GANACHE_APP, GANACHE_SERVICE_NAME, LOCALSTACK_APP,
LOCALSTACK_SERVICE_NAME, NETWORK_DEV_MODE_RESOURCES,
Expand Down Expand Up @@ -426,7 +426,7 @@ pub fn cas_stateful_set_spec(
},
EnvVar {
name: "IPFS_API_URL".to_owned(),
value: Some(format!("http://{CAS_IPFS_SERVICE_NAME}:5001")),
value: Some(format!("http://{CAS_IPFS_SERVICE_NAME}:{CAS_IPFS_SERVICE_PORT}")),
..Default::default()
},
EnvVar {
Expand Down Expand Up @@ -621,9 +621,9 @@ pub fn cas_ipfs_service_spec() -> ServiceSpec {
ServiceSpec {
ports: Some(vec![ServicePort {
name: Some("cas-ipfs".to_owned()),
port: 5001,
port: CAS_IPFS_SERVICE_PORT,
protocol: Some("TCP".to_owned()),
target_port: Some(IntOrString::Int(5001)),
target_port: Some(IntOrString::Int(CAS_IPFS_SERVICE_PORT)),
..Default::default()
}]),
selector: selector_labels(CAS_IPFS_APP),
Expand Down
12 changes: 6 additions & 6 deletions operator/src/network/ceramic.rs
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ use crate::{
network::{
controller::{
CAS_SERVICE_NAME, CERAMIC_APP, CERAMIC_POSTGRES_SECRET_NAME, CERAMIC_SERVICE_API_PORT,
CERAMIC_SERVICE_IPFS_PORT, DEFAULT_METRICS_PORT, GANACHE_SERVICE_NAME,
CERAMIC_ONE_IPFS_PORT, DEFAULT_METRICS_PORT, GANACHE_SERVICE_NAME,
INIT_CONFIG_MAP_NAME, NETWORK_DEV_MODE_RESOURCES, NODE_INSPECTION_PORT,
},
datadog::DataDogConfig,
Expand All @@ -34,7 +34,7 @@ use crate::{
};

use super::{
controller::CERAMIC_SERVICE_SWARM_PORT, debug_mode_security_context,
controller::CERAMIC_ONE_SWARM_PORT, debug_mode_security_context,
storage::PersistentStorageConfig,
};

Expand Down Expand Up @@ -118,13 +118,13 @@ pub fn service_spec() -> ServiceSpec {
..Default::default()
},
ServicePort {
port: CERAMIC_SERVICE_IPFS_PORT,
port: CERAMIC_ONE_IPFS_PORT,
name: Some("ipfs".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
},
ServicePort {
port: CERAMIC_SERVICE_SWARM_PORT,
port: CERAMIC_ONE_SWARM_PORT,
name: Some("swarm-tcp".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
Expand Down Expand Up @@ -239,7 +239,7 @@ impl CeramicInfo {
/// Determine the IPFS RPC address of a Ceramic peer
pub fn ipfs_rpc_addr(&self, ns: &str, peer: i32) -> String {
format!(
"http://{}-{peer}.{}.{ns}.svc.cluster.local:{CERAMIC_SERVICE_IPFS_PORT}",
"http://{}-{peer}.{}.{ns}.svc.cluster.local:{CERAMIC_ONE_IPFS_PORT}",
self.stateful_set, self.service
)
}
Expand Down Expand Up @@ -363,7 +363,7 @@ pub fn stateful_set_spec(ns: &str, bundle: &CeramicBundle<'_>) -> StatefulSetSpe
},
EnvVar {
name: "CERAMIC_IPFS_HOST".to_owned(),
value: Some(format!("http://localhost:{CERAMIC_SERVICE_IPFS_PORT}")),
value: Some(format!("http://localhost:{CERAMIC_ONE_IPFS_PORT}")),
..Default::default()
},
EnvVar {
Expand Down
8 changes: 4 additions & 4 deletions operator/src/network/controller.rs
Original file line number Diff line number Diff line change
Expand Up @@ -71,8 +71,8 @@ use crate::{
/// network.
pub const PEERS_CONFIG_MAP_NAME: &str = "keramik-peers";

pub const CERAMIC_SERVICE_SWARM_PORT: i32 = 4101;
pub const CERAMIC_SERVICE_IPFS_PORT: i32 = 5101;
pub const CERAMIC_ONE_SWARM_PORT: i32 = 4101;
pub const CERAMIC_ONE_IPFS_PORT: i32 = 5101;
pub const CERAMIC_SERVICE_API_PORT: i32 = 7007;
pub const CERAMIC_POSTGRES_SECRET_NAME: &str = "ceramic-postgres-auth";

Expand All @@ -81,7 +81,7 @@ pub const ADMIN_SECRET_NAME: &str = "ceramic-admin";

pub const CAS_SERVICE_NAME: &str = "cas";
pub const CAS_IPFS_SERVICE_NAME: &str = "cas-ipfs";
pub const CAS_SERVICE_IPFS_PORT: i32 = 5001;
pub const CAS_IPFS_SERVICE_PORT: i32 = 5101;
pub const CAS_POSTGRES_SERVICE_NAME: &str = "cas-postgres";
pub const CAS_POSTGRES_SECRET_NAME: &str = "postgres-auth";
pub const GANACHE_SERVICE_NAME: &str = "ganache";
Expand Down Expand Up @@ -891,7 +891,7 @@ async fn update_peer_status(
let network_config: NetworkConfig = network.spec().into();
if network_config.network_type == NetworkType::Local {
// CAS IPFS peer
let ipfs_rpc_addr = format!("http://{CAS_IPFS_SERVICE_NAME}-0.{CAS_IPFS_SERVICE_NAME}.{ns}.svc.cluster.local:{CAS_SERVICE_IPFS_PORT}");
let ipfs_rpc_addr = format!("http://{CAS_IPFS_SERVICE_NAME}-0.{CAS_IPFS_SERVICE_NAME}.{ns}.svc.cluster.local:{CAS_IPFS_SERVICE_PORT}");
match cx.rpc_client.peer_info(&ipfs_rpc_addr).await {
Ok(info) => {
status.peers.push(Peer::Ipfs(info));
Expand Down
16 changes: 8 additions & 8 deletions operator/src/network/ipfs.rs
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ use crate::{
};

use super::{
controller::{CERAMIC_SERVICE_IPFS_PORT, CERAMIC_SERVICE_SWARM_PORT},
controller::{CERAMIC_ONE_IPFS_PORT, CERAMIC_ONE_SWARM_PORT},
debug_mode_security_context,
};

Expand Down Expand Up @@ -174,7 +174,7 @@ impl RustIpfsConfig {
},
EnvVar {
name: "CERAMIC_ONE_BIND_ADDRESS".to_owned(),
value: Some(format!("0.0.0.0:{CERAMIC_SERVICE_IPFS_PORT}")),
value: Some(format!("0.0.0.0:{CERAMIC_ONE_IPFS_PORT}")),
..Default::default()
},
EnvVar {
Expand All @@ -184,7 +184,7 @@ impl RustIpfsConfig {
},
EnvVar {
name: "CERAMIC_ONE_SWARM_ADDRESSES".to_owned(),
value: Some(format!("/ip4/0.0.0.0/tcp/{CERAMIC_SERVICE_SWARM_PORT}")),
value: Some(format!("/ip4/0.0.0.0/tcp/{CERAMIC_ONE_SWARM_PORT}")),
..Default::default()
},
EnvVar {
Expand Down Expand Up @@ -227,13 +227,13 @@ impl RustIpfsConfig {
// Construct the set of ports
let mut ports = vec![
ContainerPort {
container_port: CERAMIC_SERVICE_SWARM_PORT,
container_port: CERAMIC_ONE_SWARM_PORT,
name: Some("swarm-tcp".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
},
ContainerPort {
container_port: CERAMIC_SERVICE_IPFS_PORT,
container_port: CERAMIC_ONE_IPFS_PORT,
name: Some("rpc".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
Expand Down Expand Up @@ -363,7 +363,7 @@ ipfs config --json Addresses.Gateway '[]'
# Enable pubsub
ipfs config --json PubSub.Enabled true
# Only listen on specific tcp address as nothing else is exposed
ipfs config --json Addresses.Swarm '["/ip4/0.0.0.0/tcp/4001"]'
ipfs config --json Addresses.Swarm '["/ip4/0.0.0.0/tcp/4101"]'
# Set explicit resource manager limits as Kubo computes them based off
# the k8s node resources and not the pods limits.
ipfs config Swarm.ResourceMgr.MaxMemory '400 MB'
Expand Down Expand Up @@ -417,13 +417,13 @@ ipfs config --json Swarm.ResourceMgr.MaxFileDescriptors 500000
name: IPFS_CONTAINER_NAME.to_owned(),
ports: Some(vec![
ContainerPort {
container_port: CERAMIC_SERVICE_SWARM_PORT,
container_port: CERAMIC_ONE_SWARM_PORT,
name: Some("swarm-tcp".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
},
ContainerPort {
container_port: CERAMIC_SERVICE_IPFS_PORT,
container_port: CERAMIC_ONE_IPFS_PORT,
name: Some("rpc".to_owned()),
protocol: Some("TCP".to_owned()),
..Default::default()
Expand Down
2 changes: 1 addition & 1 deletion operator/src/network/testdata/ceramic_weighted_peers
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Request {
body: {
"apiVersion": "v1",
"data": {
"peers.json": "[{\"ceramic\":{\"peerId\":\"peer_id_0\",\"ipfsRpcAddr\":\"http://ceramic-0-0.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-0.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_1\",\"ipfsRpcAddr\":\"http://ceramic-0-1.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-1.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_2\",\"ipfsRpcAddr\":\"http://ceramic-0-2.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-2.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_3\",\"ipfsRpcAddr\":\"http://ceramic-0-3.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-3.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_4\",\"ipfsRpcAddr\":\"http://ceramic-0-4.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-4.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_5\",\"ipfsRpcAddr\":\"http://ceramic-0-5.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-5.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_6\",\"ipfsRpcAddr\":\"http://ceramic-0-6.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-6.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_7\",\"ipfsRpcAddr\":\"http://ceramic-0-7.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-7.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_8\",\"ipfsRpcAddr\":\"http://ceramic-0-8.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-8.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_9\",\"ipfsRpcAddr\":\"http://ceramic-0-9.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-9.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_10\",\"ipfsRpcAddr\":\"http://ceramic-1-0.ceramic-1.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-1-0.ceramic-1.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_11\",\"ipfsRpcAddr\":\"http://ceramic-1-1.ceramic-1.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-1-1.ceramic-1.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_12\",\"ipfsRpcAddr\":\"http://ceramic-2-0.ceramic-2.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-2-0.ceramic-2.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_13\",\"ipfsRpcAddr\":\"http://ceramic-3-0.ceramic-3.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-3-0.ceramic-3.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_14\",\"ipfsRpcAddr\":\"http://ceramic-4-0.ceramic-4.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-4-0.ceramic-4.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_15\",\"ipfsRpcAddr\":\"http://ceramic-5-0.ceramic-5.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-5-0.ceramic-5.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_16\",\"ipfsRpcAddr\":\"http://ceramic-6-0.ceramic-6.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-6-0.ceramic-6.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_17\",\"ipfsRpcAddr\":\"http://ceramic-7-0.ceramic-7.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-7-0.ceramic-7.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_18\",\"ipfsRpcAddr\":\"http://ceramic-8-0.ceramic-8.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-8-0.ceramic-8.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_19\",\"ipfsRpcAddr\":\"http://ceramic-9-0.ceramic-9.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-9-0.ceramic-9.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ipfs\":{\"peerId\":\"cas_peer_id\",\"ipfsRpcAddr\":\"http://cas-ipfs-0.cas-ipfs.keramik-test.svc.cluster.local:5001\",\"p2pAddrs\":[]}}]"
"peers.json": "[{\"ceramic\":{\"peerId\":\"peer_id_0\",\"ipfsRpcAddr\":\"http://ceramic-0-0.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-0.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_1\",\"ipfsRpcAddr\":\"http://ceramic-0-1.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-1.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_2\",\"ipfsRpcAddr\":\"http://ceramic-0-2.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-2.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_3\",\"ipfsRpcAddr\":\"http://ceramic-0-3.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-3.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_4\",\"ipfsRpcAddr\":\"http://ceramic-0-4.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-4.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_5\",\"ipfsRpcAddr\":\"http://ceramic-0-5.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-5.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_6\",\"ipfsRpcAddr\":\"http://ceramic-0-6.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-6.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_7\",\"ipfsRpcAddr\":\"http://ceramic-0-7.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-7.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_8\",\"ipfsRpcAddr\":\"http://ceramic-0-8.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-8.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_9\",\"ipfsRpcAddr\":\"http://ceramic-0-9.ceramic-0.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-0-9.ceramic-0.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_10\",\"ipfsRpcAddr\":\"http://ceramic-1-0.ceramic-1.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-1-0.ceramic-1.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_11\",\"ipfsRpcAddr\":\"http://ceramic-1-1.ceramic-1.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-1-1.ceramic-1.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_12\",\"ipfsRpcAddr\":\"http://ceramic-2-0.ceramic-2.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-2-0.ceramic-2.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_13\",\"ipfsRpcAddr\":\"http://ceramic-3-0.ceramic-3.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-3-0.ceramic-3.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_14\",\"ipfsRpcAddr\":\"http://ceramic-4-0.ceramic-4.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-4-0.ceramic-4.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_15\",\"ipfsRpcAddr\":\"http://ceramic-5-0.ceramic-5.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-5-0.ceramic-5.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_16\",\"ipfsRpcAddr\":\"http://ceramic-6-0.ceramic-6.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-6-0.ceramic-6.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_17\",\"ipfsRpcAddr\":\"http://ceramic-7-0.ceramic-7.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-7-0.ceramic-7.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_18\",\"ipfsRpcAddr\":\"http://ceramic-8-0.ceramic-8.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-8-0.ceramic-8.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ceramic\":{\"peerId\":\"peer_id_19\",\"ipfsRpcAddr\":\"http://ceramic-9-0.ceramic-9.keramik-test.svc.cluster.local:5101\",\"ceramicAddr\":\"http://ceramic-9-0.ceramic-9.keramik-test.svc.cluster.local:7007\",\"p2pAddrs\":[]}},{\"ipfs\":{\"peerId\":\"cas_peer_id\",\"ipfsRpcAddr\":\"http://cas-ipfs-0.cas-ipfs.keramik-test.svc.cluster.local:5101\",\"p2pAddrs\":[]}}]"
},
"kind": "ConfigMap",
"metadata": {
Expand Down
2 changes: 1 addition & 1 deletion operator/src/network/testdata/ceramics_weighted_status
Original file line number Diff line number Diff line change
Expand Up @@ -172,7 +172,7 @@ Request {
},
{
"ipfs": {
"ipfsRpcAddr": "http://cas-ipfs-0.cas-ipfs.keramik-test.svc.cluster.local:5001",
"ipfsRpcAddr": "http://cas-ipfs-0.cas-ipfs.keramik-test.svc.cluster.local:5101",
"p2pAddrs": [],
"peerId": "cas_peer_id"
}
Expand Down
4 changes: 2 additions & 2 deletions operator/src/network/testdata/default_stubs/cas_ipfs_service
Original file line number Diff line number Diff line change
Expand Up @@ -20,9 +20,9 @@ Request {
"ports": [
{
"name": "cas-ipfs",
"port": 5001,
"port": 5101,
"protocol": "TCP",
"targetPort": 5001
"targetPort": 5101
}
],
"selector": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -249,7 +249,7 @@ Request {
},
{
"name": "IPFS_API_URL",
"value": "http://cas-ipfs:5001"
"value": "http://cas-ipfs:5101"
},
{
"name": "IPFS_API_TIMEOUT",
Expand Down
2 changes: 1 addition & 1 deletion operator/src/network/testdata/go_ipfs_configmap
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Request {
body: {
"apiVersion": "v1",
"data": {
"001-config.sh": "#!/bin/sh\nset -ex\n# Do not bootstrap against public nodes\nipfs bootstrap rm all\n# Do not sticky peer with ceramic specific peers\n# We want an isolated network\nipfs config --json Peering.Peers '[]'\n# Disable the gateway\nipfs config --json Addresses.Gateway '[]'\n# Enable pubsub\nipfs config --json PubSub.Enabled true\n# Only listen on specific tcp address as nothing else is exposed\nipfs config --json Addresses.Swarm '[\"/ip4/0.0.0.0/tcp/4001\"]'\n# Set explicit resource manager limits as Kubo computes them based off\n# the k8s node resources and not the pods limits.\nipfs config Swarm.ResourceMgr.MaxMemory '400 MB'\nipfs config --json Swarm.ResourceMgr.MaxFileDescriptors 500000\n"
"001-config.sh": "#!/bin/sh\nset -ex\n# Do not bootstrap against public nodes\nipfs bootstrap rm all\n# Do not sticky peer with ceramic specific peers\n# We want an isolated network\nipfs config --json Peering.Peers '[]'\n# Disable the gateway\nipfs config --json Addresses.Gateway '[]'\n# Enable pubsub\nipfs config --json PubSub.Enabled true\n# Only listen on specific tcp address as nothing else is exposed\nipfs config --json Addresses.Swarm '[\"/ip4/0.0.0.0/tcp/4101\"]'\n# Set explicit resource manager limits as Kubo computes them based off\n# the k8s node resources and not the pods limits.\nipfs config Swarm.ResourceMgr.MaxMemory '400 MB'\nipfs config --json Swarm.ResourceMgr.MaxFileDescriptors 500000\n"
},
"kind": "ConfigMap",
"metadata": {
Expand Down
Loading

0 comments on commit 4680223

Please sign in to comment.