[Amazon FSx for NetApp ONTAP] FlexCloneで書き込み可能なクローンボリュームを作ってみた
検証用に一時的に使用するボリュームをサクッと用意したい
こんにちは、のんピ(@non____97)です。
皆さんは検証用に一時的に使用するボリュームをサクッと用意したいなと思ったことはありますか? 私はあります。
既に存在しているデータを使って書き込みを伴う検証を行う場合、影響範囲を広げないために対象データをコピーして行うことが多いと思います。しかし、コピー対象のデータサイズが大きい場合は、そのコピーにかかる時間が非常にもったいないです。特に定期的に本番環境のデータにリフレッシュしたい場合はコピー時間は気になるでしょう。
そのようなお悩みはONTAPのFlexCloneが解決できそうです。
FlexCloneとは対象ボリュームのSnapshotを利用して、書き込み可能なクローンボリュームを作成する機能です。クローンボリューム上でデータを変更しても、クローン元のボリュームには影響を与えません。Autoraのクローンと似たような使い勝手と思っていただくと良いと思います。
FlexCloneを使用することで、テスト環境や開発環境用はもちろん、ユーザー毎のトレーニング環境のボリュームを高速に準備することができます。
また、障害やセキュリティインシデントなどでボリューム内のデータを何か調査が必要になった場合においても、インシデント発生前のSnapshotからクローンボリュームを作成して業務は継続させた上で、オリジナルのボリュームでフォレンジック調査を継続することも可能です。
実際に触ってみたので紹介します。
いきなりまとめ
- FlexCloneとはクローン元ボリュームのSnapshotを利用して高速にクローンボリュームを作成する機能
- クローンボリュームを作成するタイミングで新規にSnapshotを作成することも、既存のSnapshotを選択することも可能
- 実際のデータのコピーは行われないので即座にクローンボリュームが作成される
- クローンへ新たに書き込まれたデータ分だけのサイズが消費される
- クローンボリュームのクローンボリュームも作成可能
- ボリューム単位クローンだけでなく、ボリューム内のファイルやLUNのクローンも作成可能
- クローンボリューム作成後にクローン元ボリュームを削除したい場合は、依存関係を切ることで対応可能
- FlexCloneをネストさせている場合、中間のクローンボリュームをスプリットさせることはできない
FlexCloneとは
FlexCloneは迅速かつ効率的にクローンボリュームを作成する機能です。もう少し裏側を紹介します。
FlexCloneではクローン元ボリュームのSnapshotを利用し、クローンボリュームとデータブロックを共有します。たとえ1TBのデータが保存されているボリュームのクローンを作成しても、クローンボリュームに変更を加えなければ、消費する物理ストレージサイズは1TBです。クローンボリュームに上で更新分だけが消費されます。
抜粋 : NetApp FlexCloneテクノロジでソフトウェア開発を高速化
抜粋 : NetApp Tech Community ONLINE vol.46
実際のデータのコピーは発生しないので、数秒でクローンボリュームを作成できます。重複したデータブロックを保持しないのでコスト削減にもつながりますね。
やってみた
テストファイルの書き込み
実際にやってみます。
テストファイル作成前のボリューム、aggregate、Storage Efficiencyの状態は以下のとおりです。物理ストレージ消費量の変化をわかりやすくするため、Storage Efficiencyは無効にし、Tiering Policyはnoneにしています。
::> set diag Warning: These diagnostic commands are for use by NetApp personnel only. Do you want to continue? {y|n}: y ::*> version NetApp Release 9.13.1P7D3: Wed Feb 14 13:11:46 UTC 2024 ::*> volume efficiency show -volume vol1* -fields changelog-usage, changelog-size, logical-data-size, state, progress, last-op-size, last-op-begin, last-op-end vserver volume state progress last-op-begin last-op-end last-op-size changelog-usage changelog-size logical-data-size ------- ------ -------- ----------------- ------------------------ ------------------------ ------------ --------------- -------------- ----------------- svm vol1 Disabled Idle for 00:12:28 Sun Apr 07 01:38:49 2024 Sun Apr 07 01:38:49 2024 0B 0% 0B 308KB ::*> volume efficiency show -volume vol1* -fields state, policy, storage-efficiency-mode, inline-compression, inline-dedupe, compression, data-compaction, auto-adaptive-compression-savings, auto-adaptive-compression-existing-volume, using-auto-adaptive-compression vserver volume state policy compression inline-compression storage-efficiency-mode inline-dedupe data-compaction auto-adaptive-compression-savings using-auto-adaptive-compression auto-adaptive-compression-existing-volume ------- ------ -------- ------ ----------- ------------------ ----------------------- ------------- --------------- --------------------------------- ------------------------------- ----------------------------------------- svm vol1 Disabled auto false false efficient false false true false false ::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percentlogical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ----- ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ ------------------------------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 121.6GB 128GB 121.6GB 308KB 0% 0B 0% 0B 308KB 0% 308KB 0%- 308KB none 0B 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 136KB Total Physical Used: 284KB Total Storage Efficiency Ratio: 1.00:1 Total Data Reduction Logical Used Without Snapshots: 136KB Total Data Reduction Physical Used Without Snapshots: 284KB Total Data Reduction Efficiency Ratio Without Snapshots: 1.00:1 Total Data Reduction Logical Used without snapshots and flexclones: 136KB Total Data Reduction Physical Used without snapshots and flexclones: 284KB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 616KB Total Physical Used in FabricPool Performance Tier: 4.80MB Total FabricPool Performance Tier Storage Efficiency Ratio: 1.00:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 616KB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.80MB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 136KB Physical Space Used for All Volumes: 136KB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 284KB Physical Space Used by the Aggregate: 284KB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 0B Physical Size Used by Snapshot Copies: 0B Snapshot Volume Data Reduction Ratio: 1.00:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 1.00:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 1 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 860.6GB 861.8GB 1.12GB 46.98MB 0% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 1.12GB 0% Aggregate Metadata 4.14MB 0% Snapshot Reserve 45.36GB 5% Total Used 46.48GB 5% Total Physical Used 46.98MB 0% Total Provisioned Space 129GB 14% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed.
1GiBのテストファイルを書き込みます。
$ sudo mount -t nfs svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 /mnt/fsxn/vol1 $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 320K 122G 1% /mnt/fsxn/vol1 $ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1/random_pattern_binary_block_1GiB bs=1M count=1024 1024+0 records in 1024+0 records out 1073741824 bytes (1.1 GB, 1.0 GiB) copied, 6.32884 s, 170 MB/s $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 1.1G 121G 1% /mnt/fsxn/vol1
テストファイル書き込み後のボリューム、aggregateの状態は以下のとおりです。論理的にも物理的にも1GiB消費していることが分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 120.6GB 128GB 121.6GB 1.00GB 0% 0B 0% 0B 1.00GB 1% 1.00GB 1% - 1.00GB none 0B 0% ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 1.00GB Total Physical Used: 1.00GB Total Storage Efficiency Ratio: 1.00:1 Total Data Reduction Logical Used Without Snapshots: 1.00GB Total Data Reduction Physical Used Without Snapshots: 1.00GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.00:1 Total Data Reduction Logical Used without snapshots and flexclones: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones: 1.00GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 1.00GB Total Physical Used in FabricPool Performance Tier: 1.01GB Total FabricPool Performance Tier Storage Efficiency Ratio: 1.00:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.01GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 1.00GB Physical Space Used for All Volumes: 1.00GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 1.00GB Physical Space Used by the Aggregate: 1.00GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 0B Physical Size Used by Snapshot Copies: 0B Snapshot Volume Data Reduction Ratio: 1.00:1 Logical Size Used by FlexClone Volumes: 0B Physical Sized Used by FlexClone Volumes: 0B FlexClone Volume Data Reduction Ratio: 1.00:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 1.00:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 1 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 859.6GB 861.8GB 2.13GB 1.06GB 0% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 2.12GB 0% Aggregate Metadata 5.43MB 0% Snapshot Reserve 45.36GB 5% Total Used 47.48GB 5% Total Physical Used 1.06GB 0% Total Provisioned Space 129GB 14% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed.
FlexCloneボリュームの作成
それではFlexCloneボリュームを作成します。
作成方法は以下NetApp公式ドキュメントが参考になります。
::*> snapshot show -volume vol1 There are no entries matching your query. ::*> volume clone create sharing-by-split show split ::*> volume clone show This table is currently empty. :*> volume clone create -parent-volume vol1 -flexclone vol1_clone -junction-path /vol1_clone [Job 44] Job succeeded: Successful ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 1.00GB Split Estimate: 1.01GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 0% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock ::*> ::*> volume clone sharing-by-split show This table is currently empty.
5秒ほどで作成完了しました。
今回は同じSVMにクローンボリュームを作成しましたがvolume clone createのコマンドリファレンスを確認すると、-parent-vserver
との指定が可能なのでクローン元ボリュームは異なるSVMでも問題なさそうです。
また、Snapshot一覧を確認すると、Snapshotが取得されていることを確認できました。
::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 164KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 220KB 0% 0% 2 entries were displayed.
FlexCloneボリューム作成後のボリューム、aggregateの状態は以下のとおりです。物理的な消費量は1.06GBから1.10GBとほとんど変わっていないことが分かります。また、クローンボリュームvol_clone
の物理データ消費量は1.27MBであることも分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 120.6GB 128GB 121.6GB 1.00GB 0% 0B 0% 0B 1.00GB 1% 1.00GB 1% - 1.00GB none 0B 0% svm vol1_clone 128GB 120.6GB 128GB 121.6GB 1.00GB 0% 0B 0% 0B 1.27MB 0% 1.00GB 1% - 1.00GB none 0B 0% 2 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 6.03GB Total Physical Used: 1.00GB Total Storage Efficiency Ratio: 6.00:1 Total Data Reduction Logical Used Without Snapshots: 2.01GB Total Data Reduction Physical Used Without Snapshots: 1.00GB Total Data Reduction Efficiency Ratio Without Snapshots: 2.00:1 Total Data Reduction Logical Used without snapshots and flexclones: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones: 1.00GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 6.03GB Total Physical Used in FabricPool Performance Tier: 1.01GB Total FabricPool Performance Tier Storage Efficiency Ratio: 5.94:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.01GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 2.01GB Physical Space Used for All Volumes: 2.01GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 1.00GB Physical Space Used by the Aggregate: 1.00GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 4.02GB Physical Size Used by Snapshot Copies: 852KB Snapshot Volume Data Reduction Ratio: 4945.92:1 Logical Size Used by FlexClone Volumes: 1.00GB Physical Sized Used by FlexClone Volumes: 1.27MB FlexClone Volume Data Reduction Ratio: 810.30:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 2447.64:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 2 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 859.6GB 861.8GB 2.17GB 1.10GB 0% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 2.17GB 0% Aggregate Metadata 7.48MB 0% Snapshot Reserve 45.36GB 5% Total Used 47.53GB 5% Total Physical Used 1.10GB 0% Total Provisioned Space 257GB 28% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed.
FlexCloneボリュームへの書き込み
続いて、FlexCloneボリュームへの書き込みをします。
$ sudo mkdir -p /mnt/fsxn/vol1_clone $ sudo mount -t nfs svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone /mnt/fsxn/vol1_clone$ df -hT -t nfs4Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 1.1G 121G 1% /mnt/fsxn/vol1 svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone nfs4 122G 1.1G 121G 1% /mnt/fsxn/vol1_clone $ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1_clone/random_pattern_binary_block_2GiB bs=1M count=20482048+0 records in 2048+0 records out 2147483648 bytes (2.1 GB, 2.0 GiB) copied, 13.7145 s, 157 MB/s $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 1.1G 121G 1% /mnt/fsxn/vol1
FlexCloneボリュームへの書き込み後のボリューム、aggregateの状態は以下のとおりです。物理的な消費量は1.10GBから3.12GBと、書き込んだ2GiB分だけ増加していることが分かります。また、クローンボリュームvol_clone
の物理データ消費量は1.27MBから3.01に増加していることも分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 120.6GB 128GB 121.6GB 1.00GB 0% 0B 0% 0B 1.00GB 1% 1.00GB 1% - 1.00GB none 0B 0% svm vol1_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 2.01GB 2% 3.01GB 2% - 3.01GB none 0B 0% 2 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 8.03GB Total Physical Used: 3.01GB Total Storage Efficiency Ratio: 2.67:1 Total Data Reduction Logical Used Without Snapshots: 4.02GB Total Data Reduction Physical Used Without Snapshots: 3.01GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.33:1 Total Data Reduction Logical Used without snapshots and flexclones: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones: 1GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 8.04GB Total Physical Used in FabricPool Performance Tier: 3.03GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.65:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.00GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 1.02GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 4.02GB Physical Space Used for All Volumes: 4.02GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 3.01GB Physical Space Used by the Aggregate: 3.01GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 4.02GB Physical Size Used by Snapshot Copies: 900KB Snapshot Volume Data Reduction Ratio: 4682.14:1 Logical Size Used by FlexClone Volumes: 3.01GB Physical Sized Used by FlexClone Volumes: 2.01GB FlexClone Volume Data Reduction Ratio: 1.50:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 3.49:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 2 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 857.6GB 861.8GB 4.19GB 3.12GB 0% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 4.18GB 0% Aggregate Metadata 13.55MB 0% Snapshot Reserve 45.36GB 5% Total Used 49.54GB 5% Total Physical Used 3.12GB 0% Total Provisioned Space 257GB 28% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 1.00GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock
FlexClone元のボリュームへの書き込み
FlexClone元のボリュームへの書き込みも試します。
$ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1/random_pattern_binary_block_3GiB bs=1M count=30723072+0 records in 3072+0 records out 3221225472 bytes (3.2 GB, 3.0 GiB) copied, 20.8923 s, 154 MB/s $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 4.1G 118G 4% /mnt/fsxn/vol1
FlexCloneボリュームへの書き込み後のボリューム、aggregateの状態は以下のとおりです。FlexCloneボリュームの物理ストレージ消費量は変動ないことが分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 117.6GB 128GB 121.6GB 4.02GB 3% 0B 0% 0B 4.02GB 3% 4.02GB 3% - 4.02GB none 0B 0% svm vol1_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 2.01GB 2% 3.01GB 2% - 3.01GB none 0B 0% 2 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 11.05GB Total Physical Used: 6.02GB Total Storage Efficiency Ratio: 1.83:1 Total Data Reduction Logical Used Without Snapshots: 7.03GB Total Data Reduction Physical Used Without Snapshots: 6.02GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.17:1 Total Data Reduction Logical Used without snapshots and flexclones: 4.01GB Total Data Reduction Physical Used without snapshots and flexclones: 4.01GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 11.05GB Total Physical Used in FabricPool Performance Tier: 6.05GB Total FabricPool Performance Tier Storage Efficiency Ratio: 1.83:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.02GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.04GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 7.03GB Physical Space Used for All Volumes: 7.03GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 6.02GB Physical Space Used by the Aggregate: 6.02GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 4.02GB Physical Size Used by Snapshot Copies: 928KB Snapshot Volume Data Reduction Ratio: 4540.87:1 Logical Size Used by FlexClone Volumes: 3.01GB Physical Sized Used by FlexClone Volumes: 2.01GB FlexClone Volume Data Reduction Ratio: 1.50:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 3.49:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 2 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 854.6GB 861.8GB 7.21GB 6.16GB 1% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 7.19GB 1% Aggregate Metadata 15.78MB 0% Snapshot Reserve 45.36GB 5% Total Used 52.56GB 6% Total Physical Used 6.16GB 1% Total Provisioned Space 257GB 28% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 1.00GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock
FlexCloneのネスト
次にFlexCloneのネストをしてみます。
FlexCloneボリュームからさらにFlexCloneボリュームを作成します。
::*> volume clone create -parent-volume vol1_clone -flexclone vol1_clone_clone -junction-path /vol1_clone_clone [Job 48] Job succeeded: Successful
FlexCloneボリュームのネスト後のボリューム、aggregateの状態は以下のとおりです。問題なくFlexCloneボリュームが作成されています。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 117.6GB 128GB 121.6GB 4.02GB 3% 0B 0% 0B 4.02GB 3% 4.02GB 3% - 4.02GB none 0B 0% svm vol1_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 2.01GB 2% 3.01GB 2% - 3.01GB none 0B 0% svm vol1_clone_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 1.13MB 0% 3.01GB 2% - 3.01GB none 0B 0% 3 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 20.08GB Total Physical Used: 6.02GB Total Storage Efficiency Ratio: 3.33:1 Total Data Reduction Logical Used Without Snapshots: 10.04GB Total Data Reduction Physical Used Without Snapshots: 6.02GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.67:1 Total Data Reduction Logical Used without snapshots and flexclones: 4.01GB Total Data Reduction Physical Used without snapshots and flexclones: 4.01GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 20.09GB Total Physical Used in FabricPool Performance Tier: 6.05GB Total FabricPool Performance Tier Storage Efficiency Ratio: 3.32:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.02GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.04GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 10.04GB Physical Space Used for All Volumes: 10.04GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 6.02GB Physical Space Used by the Aggregate: 6.02GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 10.04GB Physical Size Used by Snapshot Copies: 1.41MB Snapshot Volume Data Reduction Ratio: 7274.14:1 Logical Size Used by FlexClone Volumes: 6.03GB Physical Sized Used by FlexClone Volumes: 2.01GB FlexClone Volume Data Reduction Ratio: 2.99:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 7.98:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 3 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 854.5GB 861.8GB 7.25GB 6.17GB 1% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 7.23GB 1% Aggregate Metadata 16.35MB 0% Snapshot Reserve 45.36GB 5% Total Used 52.61GB 6% Total Physical Used 6.17GB 1% Total Provisioned Space 385GB 42% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW vol1_clone_clone svm vol1_clone clone_vol1_clone_clone.2024-04-07_021551.0 online RW 2 entries were displayed. ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 1.00GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock Vserver Name: svm FlexClone Volume: vol1_clone_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1_clone FlexClone Parent Snapshot: clone_vol1_clone_clone.2024-04-07_021551.0 FlexClone Volume State: online Junction Path: /vol1_clone_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1028 FlexClone Master Data Set ID: 2163179382 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 3.02GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 3.01GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - Press <space> to page down, <return> for next line, or 'q' to quit... 2 entries were displayed. ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 168KB 0% 0% hourly.2024-04-07_0205 208KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 224KB 0% 0% hourly.2024-04-07_0205 200KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 172KB 0% 0% vol1_clone_clone clone_vol1_clone_clone.2024-04-07_021551.0 316KB 0% 0% 6 entries were displayed.
ネストしたFlexCloneボリュームにデータを追加書き込みします。
$ sudo mkdir -p /mnt/fsxn/vol1_clone_clone $ sudo mount -t nfs svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone_clone /mnt/fsxn/vol1_clone_clone $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 4.1G 118G 4% /mnt/fsxn/vol1 svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone nfs4 122G 3.1G 119G 3% /mnt/fsxn/vol1_clone svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone_clone nfs4 122G 3.1G 119G 3% /mnt/fsxn/vol1_clone_clone $ sudo dd if=/dev/urandom of=/mnt/fsxn/vol1_clone_clone/random_pattern_binary_block_5GiB bs=1M count=5120 5120+0 records in 5120+0 records out 5368709120 bytes (5.4 GB, 5.0 GiB) copied, 35.1638 s, 153 MB/s $ df -hT -t nfs4 Filesystem Type Size Used Avail Use% Mounted on svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1 nfs4 122G 4.1G 118G 4% /mnt/fsxn/vol1 svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone nfs4 122G 3.1G 119G 3% /mnt/fsxn/vol1_clone svm-0365ba78d7ad91348.fs-009351b227391d1f1.fsx.us-east-1.amazonaws.com:/vol1_clone_clone nfs4 122G 8.1G 114G 7% /mnt/fsxn/vol1_clone_clone $ ls -lR /mnt/fsxn/vol1* /mnt/fsxn/vol1: total 4210836 -rw-r--r--. 1 root root 1073741824 Apr 7 01:53 random_pattern_binary_block_1GiB -rw-r--r--. 1 root root 3221225472 Apr 7 02:12 random_pattern_binary_block_3GiB /mnt/fsxn/vol1_clone: total 3158132 -rw-r--r--. 1 root root 1073741824 Apr 7 01:53 random_pattern_binary_block_1GiB -rw-r--r--. 1 root root 2147483648 Apr 7 02:09 random_pattern_binary_block_2GiB /mnt/fsxn/vol1_clone_clone: total 8421664 -rw-r--r--. 1 root root 1073741824 Apr 7 01:53 random_pattern_binary_block_1GiB -rw-r--r--. 1 root root 2147483648 Apr 7 02:09 random_pattern_binary_block_2GiB -rw-r--r--. 1 root root 5368709120 Apr 7 02:19 random_pattern_binary_block_5GiB
ボリューム、aggregateの状態は以下のとおりです。書き込まれたデータサイズ分だけ物理消費量が増加していることが分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 117.6GB 128GB 121.6GB 4.02GB 3% 0B 0% 0B 4.02GB 3% 4.02GB 3% - 4.02GB none 0B 0% svm vol1_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 2.01GB 2% 3.01GB 2% - 3.01GB none 0B 0% svm vol1_clone_clone 128GB 113.6GB 128GB 121.6GB 8.03GB 6% 0B 0% 0B 5.03GB 4% 8.03GB 7% - 8.03GB none 0B 0% 3 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 25.10GB Total Physical Used: 11.04GB Total Storage Efficiency Ratio: 2.27:1 Total Data Reduction Logical Used Without Snapshots: 15.06GB Total Data Reduction Physical Used Without Snapshots: 11.04GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.36:1 Total Data Reduction Logical Used without snapshots and flexclones: 4.01GB Total Data Reduction Physical Used without snapshots and flexclones: 4.00GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 25.11GB Total Physical Used in FabricPool Performance Tier: 11.09GB Total FabricPool Performance Tier Storage Efficiency Ratio: 2.26:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.02GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 4.05GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 15.06GB Physical Space Used for All Volumes: 15.06GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 11.04GB Physical Space Used by the Aggregate: 11.04GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 10.04GB Physical Size Used by Snapshot Copies: 1.46MB Snapshot Volume Data Reduction Ratio: 7059.62:1 Logical Size Used by FlexClone Volumes: 11.05GB Physical Sized Used by FlexClone Volumes: 7.04GB FlexClone Volume Data Reduction Ratio: 1.57:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 3.00:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 3 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 849.5GB 861.8GB 12.29GB 11.23GB 1% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 12.26GB 1% Aggregate Metadata 30.16MB 0% Snapshot Reserve 45.36GB 5% Total Used 57.64GB 6% Total Physical Used 11.23GB 1% Total Provisioned Space 385GB 42% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW vol1_clone_clone svm vol1_clone clone_vol1_clone_clone.2024-04-07_021551.0 online RW 2 entries were displayed. ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 1.00GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock Vserver Name: svm FlexClone Volume: vol1_clone_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1_clone FlexClone Parent Snapshot: clone_vol1_clone_clone.2024-04-07_021551.0 FlexClone Volume State: online Junction Path: /vol1_clone_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1028 FlexClone Master Data Set ID: 2163179382 FlexClone Size: 128GB Used Size: 8.03GB Split Estimate: 3.01GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 3.01GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 6% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - Press <space> to page down, <return> for next line, or 'q' to quit... 2 entries were displayed. ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 168KB 0% 0% hourly.2024-04-07_0205 208KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 224KB 0% 0% hourly.2024-04-07_0205 200KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 184KB 0% 0% vol1_clone_clone clone_vol1_clone_clone.2024-04-07_021551.0 348KB 0% 0% 6 entries were displayed.
FlexCloneのスプリット
FlexClone元ボリュームを削除する際にはFlexCloneを分割(スプリット)させる必要があります。
試しにvol1_clone
というFlexCloneボリュームとvol1
をスプリットさせます。
スプリットをすると物理ストレージ消費量が増加することが見込まれます。volume clone split estimateを実行して、どのぐらい増加するのか、空きは十分かチェックしておきましょう。
::*> volume clone split estimate Split Vserver FlexClone Estimate --------- ------------- ---------- svm vol1_clone 1.00GB vol1_clone_clone 3.01GB 2 entries were displayed.
注意事項は以下のとおりです。
- FlexCloneボリュームの新しいSnapshotコピーは、スプリット処理中は作成できません。
- データ保護関係に属しているFlexCloneボリュームや負荷共有ミラーに属しているFlexCloneボリュームは、親ボリュームからスプリットすることはできません。
- スプリットの実行中にFlexCloneボリュームをオフラインにすると、スプリット処理が中断されます。FlexCloneボリュームをオンラインに戻すと、スプリット処理が再開されます。
- スプリットの実行後、親FlexVolボリュームとクローンの両方で、それぞれのボリュームギャランティに基づいたスペースの完全な割り当てが必要になります。
- FlexCloneボリュームを親ボリュームからスプリットしたあとは、この2つを再び結合することはできません。
- ONTAP 9.4 以降では、 AFF システム上のボリュームのギャランティが none である場合、 FlexClone ボリュームのスプリット処理では物理ブロックが共有され、データはコピーされません。そのため、ONTAP 9.4以降では、AFFシステムのFlexCloneボリュームのスプリットは、他のFASシステムのFlexCloneスプリット処理よりも高速です。AFF システムでの FlexClone スプリット処理の向上には、次の利点があります。
- 親からクローンをスプリットしたあともストレージ効率が維持されます。
- 既存の Snapshot コピーは削除されません。
- 処理時間が短縮されます。
- FlexClone ボリュームをクローン階層の任意のポイントからスプリットできます。
実際に試してみます。
::*> volume clone split start -flexclone vol1_clone Warning: Are you sure you want to split clone volume vol1_clone in Vserver svm ? {y|n}: y [Job 49] Job is queued: Split vol1_clone. ::*> volume clone split show This table is currently empty. ::*> volume show -volume vol1* -fields clone-volume vserver volume clone-volume ------- ------ ------------ svm vol1 false svm vol1_clone true svm vol1_clone_clone true 3 entries were displayed. 1::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW vol1_clone_clone svm vol1_clone clone_vol1_clone_clone.2024-04-07_021551.0 online RW 2 entries were displayed. ::*> security audit log show -fields timestamp, node, application, vserver, username, input, state, message -application ssh -state Error|Success timestamp node application vserver username input state message -------------------------- ------------------------- ----------- ---------------------- -------- ------------- ----- --------------------------------------------- "Sun Apr 07 01:50:34 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin Login Attempt Error Your privileges has changed since last login. "Sun Apr 07 01:50:34 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin Logging in Success - "Sun Apr 07 01:51:10 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin Question: Warning: These diagnostic command... : y Success - "Sun Apr 07 01:51:10 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin set diag Success - "Sun Apr 07 02:03:21 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin volume clone create -parent-volume vol1 -flexclone vol1_clone -junction-path /vol1_clone Success - "Sun Apr 07 02:15:56 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin volume clone create -parent-volume vol1_clone -flexclone vol1_clone_clone -junction-path /vol1_clone_clone Success - "Sun Apr 07 02:28:07 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin Question: Are you sure you want to split cl... : y Success - "Sun Apr 07 02:28:07 2024" FsxId009351b227391d1f1-01 ssh FsxId009351b227391d1f1 fsxadmin volume clone split start -flexclone vol1_clone Success - 8 entries were displayed.
正常に実行が受け付けられましたが、いくら待てどもスプリットが行われません。もしかすると、FlexCloneをネストさせている場合、中間のクローンボリュームをスプリットさせることはできないのでしょうか。
vol1_clone_clone
とvol1_clone
をスプリットさせます。
::*> volume clone split start -flexclone vol1_clone_clone Warning: Are you sure you want to split clone volume vol1_clone_clone in Vserver svm ? {y|n}: y [Job 51] Job is queued: Split vol1_clone_clone. ::*> volume clone split show Blocks ------------------ Vserver FlexClone Scanned Updated % Complete --------- ------------- -------- -------- -------- svm vol1_clone_clone 201127 198786 9 ::*> volume clone split show -instance Vserver Name: svm FlexClone Volume: vol1_clone_clone Percentage Complete: 21 Blocks Scanned: 458687 Blocks Updated: 456346 ::*> volume clone split show Blocks ------------------ Vserver FlexClone Scanned Updated % Complete --------- ------------- -------- -------- -------- svm vol1_clone_clone 651995 649654 30 ::*> volume clone split show Blocks ------------------ Vserver FlexClone Scanned Updated % Complete --------- ------------- -------- -------- -------- svm vol1_clone_clone 2106795 786661 99 ::*> volume clone split show This table is currently empty. ::*> volume show -volume vol1* -fields clone-volume vserver volume clone-volume ------- ------ ------------ svm vol1 false svm vol1_clone true svm vol1_clone_clone false 3 entries were displayed.
今度はスプリットが行われました。やはり、FlexCloneをネストさせている場合、中間のクローンボリュームをスプリットさせることはできないようです。
スプリット後のボリューム、aggregate、Snapshotの状態は以下のとおりです。vol_clone_clone
の物理データ消費量は5.03GBから8.03とvolume clone split estimate
の結果分増加していることが分かります。
::*> volume show -volume vol1* -fields available, filesystem-size, total, used, percent-used, performance-tier-inactive-user-data, performance-tier-inactive-user-data-percent, size, dedupe-space-saved, dedupe-space-saved-percent, dedupe-space-shared,logical-used, logical-used-percent,logical-used-by-afs, logical-available, physical-used, physical-used-percent, tiering-policy vserver volume size available filesystem-size total used percent-used dedupe-space-saved dedupe-space-saved-percent dedupe-space-shared physical-used physical-used-percent logical-used logical-used-percent logical-available logical-used-by-afs tiering-policy performance-tier-inactive-user-data performance-tier-inactive-user-data-percent ------- ------ ----- --------- --------------- ------- ------ ------------ ------------------ -------------------------- ------------------- ------------- --------------------- ------------ -------------------- ----------------- ------------------- -------------- ----------------------------------- ------------------------------------------- svm vol1 128GB 117.6GB 128GB 121.6GB 4.02GB 3% 0B 0% 0B 4.02GB 3% 4.02GB 3% - 4.02GB none 0B 0% svm vol1_clone 128GB 118.6GB 128GB 121.6GB 3.01GB 2% 0B 0% 0B 2.01GB 2% 3.01GB 2% - 3.01GB none 0B 0% svm vol1_clone_clone 128GB 113.6GB 128GB 121.6GB 8.03GB 6% 0B 0% 0B 8.03GB 6% 8.03GB 7% - 8.03GB none 0B 0% 3 entries were displayed. ::*> aggr show-efficiency -instance Name of the Aggregate: aggr1 Node where Aggregate Resides: FsxId009351b227391d1f1-01 Logical Size Used by Volumes, Clones, Snapshot Copies in the Aggregate: 22.09GB Total Physical Used: 14.06GB Total Storage Efficiency Ratio: 1.57:1 Total Data Reduction Logical Used Without Snapshots: 15.06GB Total Data Reduction Physical Used Without Snapshots: 14.05GB Total Data Reduction Efficiency Ratio Without Snapshots: 1.07:1 Total Data Reduction Logical Used without snapshots and flexclones: 12.05GB Total Data Reduction Physical Used without snapshots and flexclones: 12.04GB Total Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Total Logical Size Used by Volumes, Clones, Snapshot Copies in the FabricPool Performance Tier: 22.10GB Total Physical Used in FabricPool Performance Tier: 14.11GB Total FabricPool Performance Tier Storage Efficiency Ratio: 1.57:1 Total Data Reduction Logical Used without snapshots and flexclones in the FabricPool Performance Tier: 12.05GB Total Data Reduction Physical Used without snapshots and flexclones in the FabricPool Performance Tier: 12.10GB Total FabricPool Performance Tier Data Reduction Efficiency Ratio without snapshots and flexclones: 1.00:1 Logical Space Used for All Volumes: 15.06GB Physical Space Used for All Volumes: 15.06GB Space Saved by Volume Deduplication: 0B Space Saved by Volume Deduplication and pattern detection: 0B Volume Deduplication Savings ratio: 1.00:1 Space Saved by Volume Compression: 0B Volume Compression Savings ratio: 1.00:1 Space Saved by Inline Zero Pattern Detection: 0B Volume Data Reduction SE Ratio: 1.00:1 Logical Space Used by the Aggregate: 14.06GB Physical Space Used by the Aggregate: 14.06GB Space Saved by Aggregate Data Reduction: 0B Aggregate Data Reduction SE Ratio: 1.00:1 Logical Size Used by Snapshot Copies: 7.03GB Physical Size Used by Snapshot Copies: 1.12MB Snapshot Volume Data Reduction Ratio: 6445.31:1 Logical Size Used by FlexClone Volumes: 3.01GB Physical Sized Used by FlexClone Volumes: 2.01GB FlexClone Volume Data Reduction Ratio: 1.50:1 Snapshot And FlexClone Volume Data Reduction SE Ratio: 4.99:1 Number of Volumes Offline: 0 Number of SIS Disabled Volumes: 1 Number of SIS Change Log Disabled Volumes: 3 ::*> aggr show -fields availsize, usedsize, size, physical-used, physical-used-percent, data-compaction-space-saved, data-compaction-space-saved-percent, data-compacted-count, composite-capacity-tier-used, sis-space-saved, sis-space-saved-percent, sis-shared-count, inactive-data-reporting-start-timestamp aggregate availsize size usedsize physical-used physical-used-percent data-compaction-space-saved data-compaction-space-saved-percent data-compacted-count composite-capacity-tier-used sis-space-saved sis-space-saved-percent sis-shared-count inactive-data-reporting-start-timestamp --------- --------- ------- -------- ------------- --------------------- --------------------------- ----------------------------------- -------------------- ---------------------------- --------------- ----------------------- ---------------- --------------------------------------- aggr1 846.4GB 861.8GB 15.38GB 14.33GB 2% 0B 0% 0B 0B 0B 0% 0B - ::*> aggr show-space Aggregate : aggr1 Performance Tier Feature Used Used% -------------------------------- ---------- ------ Volume Footprints 15.36GB 2% Aggregate Metadata 25.44MB 0% Snapshot Reserve 45.36GB 5% Total Used 60.74GB 7% Total Physical Used 14.33GB 2% Total Provisioned Space 385GB 42% Aggregate : aggr1 Object Store: FSxFabricpoolObjectStore Feature Used Used% -------------------------------- ---------- ------ Logical Used 0B - Logical Referenced Capacity 0B - Logical Unreferenced Capacity 0B - Total Physical Used 0B - 2 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW ::*> volume clone show -instance Vserver Name: svm FlexClone Volume: vol1_clone FlexClone Type: RW FlexClone Parent Vserver: svm FlexClone Parent Volume: vol1 FlexClone Parent Snapshot: clone_vol1_clone.2024-04-07_020316.0 FlexClone Volume State: online Junction Path: /vol1_clone Junction Active: true Space Guarantee Style: none Space Guarantee In Effect: true FlexClone Aggregate: aggr1 FlexClone Data Set ID: 1027 FlexClone Master Data Set ID: 2163179381 FlexClone Size: 128GB Used Size: 3.01GB Split Estimate: 1GB Blocks Scanned: - Blocks Updated: - Comment: QoS Policy Group Name: - QoS Adaptive Policy Group Name: - Caching Policy Name: - Parent volume type: READ_WRITE Inherited Physical Used from Base Snapshot: 1.00GB Inherited Savings from Base Snapshot: 0B FlexClone Used Percentage: 2% Vserver DR Protection: - Percentage Complete: - Volume-Level UID: - Volume-Level GID: - UUID of the FlexGroup: - FlexGroup Master Data Set ID: - FlexGroup Index: - Maximum size of a FlexGroup Constituent: - Constituent Volume Role: - Is Active FlexGroup Constituent: true Is Constituent Volume: false Is Volume a FlexGroup: false Extended Volume Style: flexvol Type of Workflow: auto SnapLock Type: non-snaplock ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 168KB 0% 0% hourly.2024-04-07_0205 208KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 224KB 0% 0% hourly.2024-04-07_0205 200KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 184KB 0% 0% 5 entries were displayed.
SnapMirrorとFlexCloneの相性確認
転送元がFlexClone元ボリュームとするSnapMirrorの実行
NetApp公式ドキュメントを眺めていると、「SnapMirrorの転送元ボリュームからFlexCloneを作成すると、正常にSnapMirrorの転送がされないケースがある」と記載がありました。
既存の Volume SnapMirror 関係にあるソースボリュームまたはデスティネーションボリュームから FlexClone ボリュームを作成できます。ただし、これを行うと、以降に行う SnapMirror のレプリケーション処理が正常に完了しないことがあります。
FlexClone ボリュームを作成すると、 SnapMirror によって使用される Snapshot コピーがロックされる可能性があるため、レプリケーションが機能しないことがあります。この場合、 FlexClone ボリュームが削除されるか、親ボリュームからスプリットされるまで、 SnapMirror はデスティネーションボリュームへのレプリケーションを停止します。この問題には、次の 2 つの方法で対処できます。
- FlexClone ボリュームが一時的に必要で、 SnapMirror レプリケーションが一時的に停止されても構わない場合は、 FlexClone ボリュームを作成し、可能となった時点で削除するか親からスプリットします。
- FlexClone ボリュームが削除されるか親からスプリットされた時点で、 SnapMirror レプリケーションが正常に続行されます。
- SnapMirror レプリケーションの一時的な停止を許容できない場合は、 SnapMirror ソースボリュームで Snapshot コピーを作成し、その Snapshot コピーを使用して FlexClone ボリュームを作成します。( FlexClone ボリュームをデスティネーションボリュームから作成している場合、 Snapshot コピーが SnapMirror デスティネーションボリュームにレプリケートされるまで待機する必要があります)。
この方法で SnapMirror ソースボリューム内に Snapshot コピーを作成すると、 SnapMirror によって使用されている Snapshot コピーをロックすることなくクローンを作成できます。
SnapMirror のソースボリュームまたはデスティネーションボリュームから FlexClone ボリュームを作成する際の考慮事項
転送元をFlexClone元ボリュームとするSnapMirrorを試してみます。
::*> snapmirror protect -path-list svm:vol1 -destination-vserver svm -policy MirrorAllSnapshots -auto-initialize true [Job 52] Job is queued: snapmirror protect for list of source endpoints beginning with "svm:vol1". ::*> snapmirror show This table is currently empty. ::*> job show 52 Owning Job ID Name Vserver Node State ------ -------------------- ---------- -------------- ---------- 52 SnapMirror protect svm FsxId009351b227391d1f1-02 Failure Description: snapmirror protect for list of source endpoints beginning with "svm:vol1" ::*> job show 52 -instance Job ID: 52 Owning Vserver: svm Name: SnapMirror protect Description: snapmirror protect for list of source endpoints beginning with "svm:vol1" Priority: Medium Node: FsxId009351b227391d1f1-02 Affinity: Cluster Schedule: @now Queue Time: 04/07 05:57:23 Start Time: 04/07 05:57:26 End Time: 04/07 05:57:26 Drop-dead Time: - Restarted?: false State: Failure Status Code: 1 Completion String: Out of 1 endpoints, 0 endpoints have been protected and protection has failed for 1 endpoints. List of paths that failed protection and reasons for failure: svm:vol1 : Failed to identify aggregate on which to create volume "vol1_dst". (No eligible aggregate to place the volume. Make sure that a non-root, non-taken-over, non-snaplock, non-composite aggregate with enough free space exists at the destination cluster. If aggr-list is configured for the Vserver make sure that an eligible aggregate is present in that list. Verify that the number of volumes is less than the supported maximum on the nodes in the destination cluster.) Job Type: SnapMirror-ProtectV1 Job Category: SnapMirror UUID: b3ece0ff-f4a3-11ee-8e23-0fe9d475982c Execution Progress: Complete: Out of 1 endpoints, 0 endpoints have been protected and protection has failed for 1 endpoints. List of paths that failed protection and reasons for failure: svm:vol1 : Failed to identify aggregate on which to create volume "vol1_dst". (No eligible aggregate to place the volume. Make sure that a non-root, non-taken-over, non-snaplock, non-composite aggregate with enough free space exists at the destination cluster. If aggr-list is configured for the Vserver make sure that an eligible aggregate is present in that list. Verify that the number of volumes is less than the supported maximum on the nodes in the destination cluster.) [1] User Name: fsxadmin Process: mgwd Restart Is or Was Delayed?: false Restart Is Delayed by Module: -
snapmirror protect
によるSnapMirror relastionshipの作成に失敗しました。aggregateの特定ができないとのことでしたが、他の検証では同様のコマンドで成功したので謎です。
手動でボリュームの作成、SnapMirror relastionshipの作成、initializeを行います。
::*> volume create -vserver svm -volume vol1_dst -aggregate aggr1 -state online -type DP -size 4GB -tiering-policy none -autosize-mode grow [Job 54] Job succeeded: Successful ::*> snapmirror create -source-path svm:vol1 -destination-path svm:vol1_dst -policy MirrorAllSnapshots Operation succeeded: snapmirror create for the relationship with destination "svm:vol1_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Uninitialized Idle - true - ::*> snapmirror initialize -destination-path svm:vol1_dst Operation is queued: snapmirror initialize of destination "svm:vol1_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Uninitialized Transferring 0B true 04/07 06:09:55 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Uninitialized Finalizing 1.02GB true 04/07 06:10:05 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Transferring 0B true 04/07 06:10:13 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Transferring 1.90GB true 04/07 06:10:21 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Finalizing 1.90GB true 04/07 06:10:21 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Finalizing 3.07GB true 04/07 06:10:36 ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Idle - true -
転送が完了しました。特にエラーにはなっていません。
FlexCloneやSnapshotの状態も確認します。特にエラーはなっていませんね。
::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_060955 140KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 384KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_060955 172KB 0% 0% 6 entries were displayed.
差分同期もしてみます。
::*> snapmirror update -destination-path svm:vol1_dst Operation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapmirror show Progress Source Destination Mirror Relationship Total Last Path Type Path State Status Progress Healthy Updated ----------- ---- ------------ ------- -------------- --------- ------- -------- svm:vol1 XDP svm:vol1_dst Snapmirrored Idle - true - ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061310 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 384KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_060955 320KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061310 156KB 0% 0% 7 entries were displayed.
特に問題なく、転送が完了します。
その後、繰り返し差分同期をしましたが、いずれも成功しています。
::*> snapmirror show-history Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:14:14 4/7/2024 06:14:16 success svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:13:10 4/7/2024 06:13:12 success svm:vol1_dst svm:vol1 scheduled-update 4/7/2024 06:09:55 4/7/2024 06:10:46 success svm:vol1_dst svm:vol1 initialize 4/7/2024 06:09:55 4/7/2024 06:10:13 success svm:vol1_dst svm:vol1 create 4/7/2024 06:09:18 4/7/2024 06:09:18 success 5 entries were displayed.
そのため、転送元がFlexClone元ボリュームとするSnapMirrorを実行するのは問題ないようです。
SnapMirror relastionship作成後にFlexCloneをした場合
先ほどのパターンはFlexCloneでクローンボリュームを作成した後に、クローン元ボリュームが転送元になるSnapMirrorを実行するパターンでした。
それでは、SnapMirror relastionship作成後にFlexCloneをした場合、SnapMirrorの転送は正常に行えるでしょうか。
vol1
をクローン元とするFlexCloneボリュームを作成します。
::*> volume clone create -parent-volume vol1 -flexclone vol1_clone2 -junction-path /vol1_clone2 [Job 56] Job succeeded: Successful ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061414 148KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 140KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 996KB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061310 280KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061414 156KB 0% 0% 9 entries were displayed. ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW vol1_clone2 svm vol1 clone_vol1_clone2.2024-04-07_061613.0 online RW 2 entries were displayed.
SnapMirrorの差分転送をします。
::*> snapmirror update -destination-path svm:vol1_dst Operation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061414 232KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 156KB 0% 0% 10 entries were displayed. ::*> snapmirror show-history Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:17:06 4/7/2024 06:17:10 success . . (以下略) . .
問題なく転送できましたね。
SnapMirror relastionship作成後にFlexCloneをした場合 (sm_createdのSnapshotを指定)
それでは、FlexCloneで使用するSnapshotをSnapMirrorによって作成されたもの(sm_created)を指定した場合はどうでしょうか。sm_created
のSnapshotがロックされてしまい、正常に転送できないのでしょうか。
vol1
のsm_created
をクローン元とするFlexCloneボリュームを作成します。
::*> volume clone create -parent-volume vol1 -flexclone vol1_clone3 -junction-path /vol1_clone3 -parent-snapshot snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 [Job 57] Job succeeded: Successful ::*> volume clone show Parent Parent Parent Vserver FlexClone Vserver Volume Snapshot State Type ------- ------------- ------- ------------- -------------------- --------- ---- svm vol1_clone svm vol1 clone_vol1_clone.2024-04-07_020316.0 online RW vol1_clone2 svm vol1 clone_vol1_clone2.2024-04-07_061613.0 online RW vol1_clone3 svm vol1 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 online RW 3 entries were displayed. ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_clone3 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 1.09MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061414 232KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 156KB 0% 0% 11 entries were displayed.
問題なくFlexCloneボリュームを作成できました。
SnapMirrorの差分転送を行います。
::*> snapmirror update -destination-path svm:vol1_dstOperation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061924 140KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_clone3 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 1.12MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061924 156KB 0% 0% 12 entries were displayed. ::*> snapmirror show-history -max-rows-per-relationship 1 Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:19:24 4/7/2024 06:19:26 success
特にエラーにはなっていません。
差分転送を繰り返してFlexCloneボリューム作成時に使用したsnapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706
がどうなるのか確認します。
::*> snapmirror update -destination-path svm:vol1_dst Operation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062014 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_clone3 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 1.12MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061924 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062014 156KB 0% 0% 13 entries were displayed. ::*> snapmirror show-history -max-rows-per-relationship 1 Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:20:14 4/7/2024 06:20:16 success
::*> snapmirror update -destination-path svm:vol1_dst Operation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapmirror show-history -max-rows-per-relationship 1 Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:21:17 4/7/2024 06:21:19 success ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062117 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_clone3 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 1.12MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062014 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062117 156KB 0% 0% 13 entries were displayed.
::*> snapmirror update -destination-path svm:vol1_dst Operation is queued: snapmirror update of destination "svm:vol1_dst". ::*> snapmirror show-history -max-rows-per-relationship 1 Destination Source Start End Path Path Operation Time Time Result ----------- ----------- --------- ----------- ----------- ------- svm:vol1_dst svm:vol1 manual-update 4/7/2024 06:21:56 4/7/2024 06:21:58 success ::*> snapshot show -volume vol1* ---Blocks--- Vserver Volume Snapshot Size Total% Used% -------- -------- ------------------------------------- -------- ------ ----- svm vol1 clone_vol1_clone.2024-04-07_020316.0 200KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 144KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 148KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062156 148KB 0% 0% vol1_clone clone_vol1_clone.2024-04-07_020316.0 252KB 0% 0% clone_vol1_clone_clone.2024-04-07_021551.0 188KB 0% 0% vol1_clone2 clone_vol1_clone2.2024-04-07_061613.0 1.09MB 0% 0% vol1_clone3 snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 1.12MB 0% 0% vol1_dst clone_vol1_clone.2024-04-07_020316.0 388KB 0% 0% clone_vol1_clone2.2024-04-07_061613.0 232KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062117 236KB 0% 0% snapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_062156 156KB 0% 0% 13 entries were displayed.
FlexCloneボリューム作成時に使用したsnapmirror.646592b9-f47f-11ee-8e23-0fe9d475982c_2163179383.2024-04-07_061706
は削除されずに残っていますね。ロックされずに転送に失敗するということも発生していません。
私が検証をしている中ではSnapMirrorの転送が失敗するシチュエーションは確認できませんでした。
FlexCloneを有効活用して開発スピードアップやコスト削減に繋げよう
NetApp ONTAPのFlexCloneを紹介しました。
FlexCloneを有効活用することで、開発スピードアップやコスト削減に繋がりそうです。積極的に使っていきたいですね。
この記事が誰かの助けになれば幸いです。
以上、AWS事業本部 コンサルティング部の のんピ(@non____97)でした!