- Installing the Migration Toolkit for Containers in a restricted network environment
Installing the Migration Toolkit for Containers in a restricted network environment
- Compatibility guidelines
- Installing the Migration Toolkit for Containers Operator on OKD 4.9
- Installing the legacy Migration Toolkit for Containers Operator on OKD 4.2 to 4.5
- Configuring proxies
- Configuring a replication repository
You can install the Migration Toolkit for Containers (MTC) on OKD 4 in a restricted network environment by performing the following procedures:
Create a mirrored Operator catalog.
This process creates a
mapping.txt
file, which contains the mapping between theregistry.redhat.io
image and your mirror registry image. Themapping.txt
file is required for installing the legacy Migration Toolkit for Containers Operator on an OKD 4.2 to 4.5 source cluster.Install the Migration Toolkit for Containers Operator on the OKD 4.9 target cluster by using Operator Lifecycle Manager.
By default, the MTC web console and the
Migration Controller
pod run on the target cluster. You can configure theMigration Controller
custom resource manifest to run the MTC web console and theMigration Controller
pod on a remote cluster.Install the Migration Toolkit for Containers Operator on the source cluster:
OKD 4.6 or later: Install the Migration Toolkit for Containers Operator by using Operator Lifecycle Manager.
OKD 4.2 to 4.5: Install the legacy Migration Toolkit for Containers Operator from the command line interface.
To install MTC on OKD 3, see Installing the legacy Migration Toolkit for Containers Operator on OKD 3.
Configure object storage to use as a replication repository.
Compatibility guidelines
You must install the Migration Toolkit for Containers (MTC) Operator that is compatible with your OKD version.
You cannot install MTC 1.6 on OKD 4.5, or earlier versions, because the custom resource definition API versions are incompatible.
You can migrate workloads from a source cluster with MTC 1.5.3 to a target cluster with MTC 1.6 as long as the MigrationController
custom resource and the MTC web console are running on the target cluster.
OKD version | MTC version | Migration Toolkit for Containers Operator |
---|---|---|
4.5 and earlier | 1.5.3 | Legacy Migration Toolkit for Containers Operator, installed manually with the |
4.6 and later | Latest 1.6.x z-stream release | Migration Toolkit for Containers Operator, installed with Operator Lifecycle Manager. |
Installing the Migration Toolkit for Containers Operator on OKD 4.9
You install the Migration Toolkit for Containers Operator on OKD 4.9 by using the Operator Lifecycle Manager.
Prerequisites
You must be logged in as a user with
cluster-admin
privileges on all clusters.You must create an Operator catalog from a mirror image in a local registry.
Procedure
In the OKD web console, click Operators → OperatorHub.
Use the Filter by keyword field to find the Migration Toolkit for Containers Operator.
Select the Migration Toolkit for Containers Operator and click Install.
Click Install.
On the Installed Operators page, the Migration Toolkit for Containers Operator appears in the openshift-migration project with the status Succeeded.
Click Migration Toolkit for Containers Operator.
Under Provided APIs, locate the Migration Controller tile, and click Create Instance.
Click Create.
Click Workloads → Pods to verify that the MTC pods are running.
Installing the legacy Migration Toolkit for Containers Operator on OKD 4.2 to 4.5
You can install the legacy Migration Toolkit for Containers Operator manually on OKD versions 4.2 to 4.5.
Prerequisites
You must be logged in as a user with
cluster-admin
privileges on all clusters.You must have access to
registry.redhat.io
.You must have
podman
installed.You must have a Linux workstation with network access in order to download files from
registry.redhat.io
.You must create a mirror image of the Operator catalog.
You must install the Migration Toolkit for Containers Operator from the mirrored Operator catalog on OKD 4.9.
Procedure
Log in to
registry.redhat.io
with your Red Hat Customer Portal credentials:$ sudo podman login registry.redhat.io
Download the
operator.yml
file:$ sudo podman cp $(sudo podman create \
registry.redhat.io/rhmtc/openshift-migration-legacy-rhel8-operator:v1.5.3):/operator.yml ./
Download the
controller.yml
file:$ sudo podman cp $(sudo podman create \
registry.redhat.io/rhmtc/openshift-migration-legacy-rhel8-operator:v1.5.3):/controller.yml ./
Obtain the Operator image mapping by running the following command:
$ grep openshift-migration-legacy-rhel8-operator ./mapping.txt | grep rhmtc
The
mapping.txt
file was created when you mirrored the Operator catalog. The output shows the mapping between theregistry.redhat.io
image and your mirror registry image.Example output
registry.redhat.io/rhmtc/openshift-migration-legacy-rhel8-operator@sha256:468a6126f73b1ee12085ca53a312d1f96ef5a2ca03442bcb63724af5e2614e8a=<registry.apps.example.com>/rhmtc/openshift-migration-legacy-rhel8-operator
Update the
image
values for theansible
andoperator
containers and theREGISTRY
value in theoperator.yml
file:containers:
- name: ansible
image: <registry.apps.example.com>/rhmtc/openshift-migration-legacy-rhel8-operator@sha256:<468a6126f73b1ee12085ca53a312d1f96ef5a2ca03442bcb63724af5e2614e8a> (1)
...
- name: operator
image: <registry.apps.example.com>/rhmtc/openshift-migration-legacy-rhel8-operator@sha256:<468a6126f73b1ee12085ca53a312d1f96ef5a2ca03442bcb63724af5e2614e8a> (1)
...
env:
- name: REGISTRY
value: <registry.apps.example.com> (2)
1 Specify your mirror registry and the sha256
value of the Operator image.2 Specify your mirror registry. Log in to your OKD 3 cluster.
Create the Migration Toolkit for Containers Operator object:
$ oc create -f operator.yml
Example output
namespace/openshift-migration created
rolebinding.rbac.authorization.k8s.io/system:deployers created
serviceaccount/migration-operator created
customresourcedefinition.apiextensions.k8s.io/migrationcontrollers.migration.openshift.io created
role.rbac.authorization.k8s.io/migration-operator created
rolebinding.rbac.authorization.k8s.io/migration-operator created
clusterrolebinding.rbac.authorization.k8s.io/migration-operator created
deployment.apps/migration-operator created
Error from server (AlreadyExists): error when creating "./operator.yml":
rolebindings.rbac.authorization.k8s.io "system:image-builders" already exists (1)
Error from server (AlreadyExists): error when creating "./operator.yml":
rolebindings.rbac.authorization.k8s.io "system:image-pullers" already exists
1 You can ignore Error from server (AlreadyExists)
messages. They are caused by the Migration Toolkit for Containers Operator creating resources for earlier versions of OKD 3 that are provided in later releases.Create the
MigrationController
object:$ oc create -f controller.yml
Verify that the MTC pods are running:
$ oc get pods -n openshift-migration
Configuring proxies
For OKD 4.1 and earlier versions, you must configure proxies in the MigrationController
custom resource (CR) manifest after you install the Migration Toolkit for Containers Operator because these versions do not support a cluster-wide proxy
object.
For OKD 4.2 to 4.9, the Migration Toolkit for Containers (MTC) inherits the cluster-wide proxy settings. You can change the proxy parameters if you want to override the cluster-wide proxy settings.
You must configure the proxies to allow the SPDY protocol and to forward the Upgrade HTTP
header to the API server. Otherwise, an Upgrade request required
error is displayed. The MigrationController
CR uses SPDY to run commands within remote pods. The Upgrade HTTP
header is required in order to open a websocket connection with the API server.
Direct volume migration
If you are performing a direct volume migration (DVM) from a source cluster behind a proxy, you must configure an Stunnel proxy. Stunnel creates a transparent tunnel between the source and target clusters for the TCP connection without changing the certificates.
DVM supports only one proxy. The source cluster cannot access the route of the target cluster if the target cluster is also behind a proxy.
Prerequisites
- You must be logged in as a user with
cluster-admin
privileges on all clusters.
Procedure
Get the
MigrationController
CR manifest:$ oc get migrationcontroller <migration_controller> -n openshift-migration
Update the proxy parameters:
apiVersion: migration.openshift.io/v1alpha1
kind: MigrationController
metadata:
name: <migration_controller>
namespace: openshift-migration
...
spec:
stunnel_tcp_proxy: http://<username>:<password>@<ip>:<port> (1)
httpProxy: http://<username>:<password>@<ip>:<port> (2)
httpsProxy: http://<username>:<password>@<ip>:<port> (3)
noProxy: example.com (4)
1 Stunnel proxy URL for direct volume migration. 2 Proxy URL for creating HTTP connections outside the cluster. The URL scheme must be http
.3 Proxy URL for creating HTTPS connections outside the cluster. If this is not specified, then httpProxy
is used for both HTTP and HTTPS connections.4 Comma-separated list of destination domain names, domains, IP addresses, or other network CIDRs to exclude proxying. Preface a domain with
.
to match subdomains only. For example,.y.com
matchesx.y.com
, but noty.com
. Use*
to bypass proxy for all destinations. If you scale up workers that are not included in the network defined by thenetworking.machineNetwork[].cidr
field from the installation configuration, you must add them to this list to prevent connection issues.This field is ignored if neither the
httpProxy
nor thehttpsProxy
field is set.Save the manifest as
migration-controller.yaml
.Apply the updated manifest:
$ oc replace -f migration-controller.yaml -n openshift-migration
For more information, see Configuring the cluster-wide proxy.
Configuring a replication repository
The Multicloud Object Gateway is the only supported option for a restricted network environment.
MTC supports the file system and snapshot data copy methods for migrating data from the source cluster to the target cluster. You can select a method that is suited for your environment and is supported by your storage provider.
Prerequisites
All clusters must have uninterrupted network access to the replication repository.
If you use a proxy server with an internally hosted replication repository, you must ensure that the proxy allows access to the replication repository.
Configuring the Multicloud Object Gateway
You can configure the Multicloud Object Gateway (MCG) as a replication repository for the Migration Toolkit for Containers (MTC). MCG is a component of OpenShift Container Storage.
Prerequisites
Ensure that you have downloaded the pull secret from the Red Hat OpenShift Cluster Manager site as shown in Obtaining the installation program in the installation documentation for your platform.
If you have the pull secret, add the
redhat-operators
catalog to the OperatorHub custom resource (CR) as shown in Configuring OKD to use Red Hat Operators.
Procedure
Deploy OpenShift Container Storage by using the appropriate OpenShift Container Storage deployment guide.
Obtain the S3 endpoint,
AWS_ACCESS_KEY_ID
, andAWS_SECRET_ACCESS_KEY
by running the describe command on theNooBaa
custom resource.These values are required in order to add MCG as a replication repository to the MTC web console.
Additional resources for configuring a replication repository
Disconnected environment in the Red Hat OpenShift Container Storage documentation.
Adding a replication repository to the MTC web console :installing-mtc-restricted!: