Quantcast
Channel: ONTAP Discussions topics
Viewing all 4954 articles
Browse latest View live

7 mode to cluster mode transition questions

$
0
0

HI 

preparation is underway to migrate 7 mode volumes to cluser mode and have come across few questions releated to a warning reported by a precheck and cutover releated questions.  appreciate your expert opinion here. sorry for lenghy questions. appreciate your responses.

1) Is there a command to disconnect NFS client access to a 7-mode volume being migrated from the storage side during storage cutover? CIFS access to a 7-mode volume being migrated can be terminated from storage side using “cifs terminate –v vol_name”, is there a way to terminate NFS access to a volume from the storage side to prevent additional writes from being performed while final copy is in progress during storage cutover phase?

 

2) 7Mtt precheck reported below warning (warning no 90702)

Access to non-Unicode directories on the transitioned volumes by NFSv4 or CIFS clients might fail in 8.3.0 or later versions of Clustered Data ONTAP.

Identify the non-Unicode directories and fix the issues immediately after the transition is complete. For more information about how to handle non-Unicode directories, refer to the 7-Mode Transition Tool Release Notes.

  Is above warning ignorable. What is the criticality to address above warning for successful transition with regard to client access after the cutover?

 

3) I plan to use new identify for CIFS on the cluster (different CIFS server NETBIOS name in AD domain than the one 7 mode using currently) . Per 7MTT 3.3.1 user guide, new CIFS identify can be used having configured the CIFS server on the target SVMwith a new CIFS identity.  I hope not but wanted to double check, will there be any impact with regard to client access using new CIFS identify on the cluster than the one that 7 mode using after the cutover.

4) In our current 7-mode configuration, we have two controllers in HA. As part of our migration project, we plan to migrate volumes on one controller now and volume on partner controller will be migrated later in the year. We have a plan to migrate the IP address of the controller being migrated manually and configure it on cluster after 7MTT cutover task completed so clients can continue to use same IP after the storage cutover to cluster. Volumes being migrated will be taken offline as part of the 7mtt storage cutover, if no IP selected for transition, then IP on 7 mode controller will not be brought offline and it need to be brought down manually correct ?. When I bring down IP address of a 7-mode controller manually after the storage cutover, complete to reconfigure cluster LIF to use same IP, will 7-mode Ontap trigger a takeover? because negotiated failover is enabled on data interface of the 7 mode controllers, if it bring down a NFO enabled 7 mode interface , will Ontap trigger takeover of the 7 mode node where interface being brought down. (OR) taking down interface manually does not account for NFO and takeover will not be initiated.


File Auditing using ELK stack

$
0
0

Currently we are using Netwrix for CIFS auditing of our CIFS SVMs. We are looking to make the move to using the ELK stack if we can. Is it somehow possible to send the CIFS audit files to ELK, or accomplish this some other way?

 

Thanks.

 

Can QoS for IOPS be applied or assigned to a Volume that is being accessed by the client ?

$
0
0

Hi All,

 

Can QoS be appied (For max IOPS) dynamically ? While the clients are accessing the volume be it CIFS or NFS ?

 

Do I need to stop the CIFS service or NFS service first,  and then apply the QoS IOPS Policy, and reestart the Protocol service ?

 

In a simple simulator test on ONTAP 9.4 : I was copying a large file, and it was consuming most of the IOPS available to it. When I assigned QoS policy for [100 IOPS}, it did not ahered to it.

 

However, once the copy was over, I re-assigned the Policy (IOPS 100)  and started another copy and this time I could see constant IOPS of 110. I did not have to stop the CIFS service, as this is a SIMULATOR and I knew it's just one activity, so it was simple.

 

Does anyone has more insight ?

 

Thanks,

-Ash

Failed Ontap cluster update

$
0
0

Trying to run a cluster update from 9.4P3 to 9.4P5 via System Manager on a newly installed FAS2720.    Get the attached error when upgrade process is trying to download the image to the nodes.    We have verified all steps outlined in the error including re-downloading the 9.4P5 image from NetApp support site and still get the error.   Verified the checksum on the downloaded image and it looks good.   Tried a direct update of the image via HTTP from the web server on my laptop using "system node image update" and get the same error.   I know that the image on my  laptop is good as I upgraded another customer's system with it 2 days earlier.

CPU Concurrency of 1+ CPU on AFF-A300 with ONTAP 9.1P13

$
0
0

Hi,

 

According to this KB all these 'Domains' has CPU Concurrency of 1+ CPUs:

https://kb.netapp.com/app/answers/answer_view/a_id/1001217

 

Host os : 1+
exempt : 1+
SM_Exempt : 1+ [No SnapMirror configured yet]
WAFL_XCleaner : 1+
WAFL_EX : 1+
ssan_exempt : 1+ [No FCP/iSCSI processing, Pure NFS environment]
XO_Ex : 1+
raid_exempt : 1+
storage : 1+

nwk_exempt: 1+ [IP & NFS Processing]


My question:
Given a situation, where all the above domains except (SM_Exempt & ssan_exempt) are running parallely and are able to process information using single CPU. So, total '7' CPU are being used at a given time.


For  a NetApp FILER such as  AFF-A300: 16 Cores [16-7=9]

My question : If need be, will 'nwk_exempt' domain be able to maximum it's processing using all the available CPU cores ? Or there is a limit to, it can grab ?, given they are indeed available.

 

 

Dawson : Do you mind attending to this query,  you have been very helpful in the past on this subject.

 

Thanks,
-Ash

Netapp E-Series(E-2700) can be accessed through OnTap(manageontap.jar) ?

$
0
0

1. I have NetApp E-Series(E-2700) how can i access it through manageontap.jar or NMSDK ?

 

Thanks and Regards,

Divya Jyoti Das

discrepancy btw volume size and transferred size by SnapMirror

$
0
0

I am calculating the snapmirroring speed.

 

Total Transferred suze us 25.7GB by "snapmirror show" command. 
However, it shows me total used size is 40GB by "df -g" command on the destination after initialization and completed with snapmirroring. 

So, why they have showed me in different sizes, and there are 14.3GB differences?

 

How much data really have been transferred?

 

Thanks!

NetApp CIFS auting security options are disabled

$
0
0

Hi Team,

 

I am using latest data ontap version in cluster mode. I have enabled CIFS auditing on NetApp and while configuring Auditing through  Windows Explorer I found options disabled in security tab. Please check attached file and suggest solution.

I am using domain admin account which is also have built in Administrator privileges.

 

Regards

Prajyot Katakdound

prajyot.katakdound.wg@hitachi-systems.com

 


vifmgr.lifs.noredundancy on IFGRP

$
0
0

Running Multiple FAS 8200s

All running On-Tap 9.3 p10

 

Hi All

 

Im seing the below on my filers (in the Event Logs) which im 95% certain is a red herring.

Would like a little clarity or options before i dump it to support so any help greatly appreaciated....

 

My network is air gapped so logs etc may me difficult.

The "Intercluster Replication" IFGRP comprises of 2 uplinks per filer, carrying the replication vLAN which i have confirmed from the command line.

 

The replication IFGRP is assigned to the Broadcast Domain "Replication" when checked the correct IFGROUP is shown inside the bradcast domain. So all looks good to me..

 

However (and im gussing here) that On-Tap cannot determine that IFGRP shown in the Broadcast Domain "Replication" acually consists of multiple physical ports and is therefore beleives that there is actually no redundancey here, hece the error log...?

(Bearing in mind there is some very very heavy design for my estates, and even assiging just an extra physical port will take months of change controll as the designs are updated the documentation valadidated the work instructions ammended then followed with annother round of testing etc etc etc etc...

 

Is there a quicker way of surpressing just these SPECIFIC events for the Replication Broadcast Domain, or otherwise stopping them without too much modification..?

 

[Log Snippet below.]  

 

Severity:- Alert

Source:- vifmgr

Event:- vifmgr.lifs.noredundancy: No redundancy in the failover configuration for 1 lif assigned to node "MY-FILER". LIFs "INTERFACE NAME" (Intercluster replication)

 

Description: This message occurs when one or more logical interfaces (LIFs) are configured
to use a failover policy that implies failover to one or more ports but have no failover
targets beyond their home ports. If any affected home port or home node is offline or
unavailable, the corresponding LIFs will be operationally down and unable to serve data.

 

Cheers all

Any help much appreaciated.

SVM_DR Snapshots not deleted on source

$
0
0

Hello,

I've a question regarding the Snapshot (sm_created) used by SnapMirror for replication. I've probably misunderstood somehting but I cannot figure out what.

 

I have a cascaded replication: Cluster A  >  Cluster B (SVM_DR)  >  Cluster C (SnapVault - long term of Weekly)

 

The source volume (on Cluster A) has a snapshot policy set to "default", unchanged:

‌‌‌

Vserver: cluster1
Number of IsPolicy Name Schedules Enabled Comment
------------------------ --------- ------- ----------------------------------
default 4 true Default policy with hourly, daily & weekly schedules.
Schedule Count Prefix SnapMirror Label
---------------------- ----- ---------------------- -------------------
hourly     6      hourly        -
daily      2      daily         daily
weekly     2      weekly        weekly
5min      12      5min          5min

 

So no "sm_created" label.  At this time, the SVM_DR replication ran for a couple of days and on my source volume (cluster 1) I still have the following snapshots:

 

snap list -vserver volume snapshot snapmirror-label
------------------------------------------------------------------ ----------------
vs1 cascade_src1 vserverdr.0.d3c6b107-390d-11e9-a1d2-005056b013db.2019-02-25_190000 sm_created

[...] Ouput truncated

vs1 cascade_src1 vserverdr.2.d3c6b107-390d-11e9-a1d2-005056b013db.2019-02-27_122500 sm_created
vs1 cascade_src1 vserverdr.0.d3c6b107-390d-11e9-a1d2-005056b013db.2019-02-27_123000 sm_created
12 entries were displayed.

 

12 Snapshots are kept, the retention count for 5min snapshot is 12 and the Snapmirror schedule for the SVM_DR is indeed 5min. 

If the retention count is defined by the snamirror label, why is is this using 5min instead of "sm_created" which is the real label ?

 

Do you have any best practices or recommandation to use in this kind of situation ?

 

Thanks for your help !

 

 

NetApp CIFS Auting Security Options Are Disabled

$
0
0

Hi Team,

I am using latest data ontap os version of cmode. I am using CIFS shares and enabled Auditing from NetApp end. While confifuring  operations to be audited I found securiry options are disabled. I have checked I am using domain Admin account with BUILT in Administrator privileges. Please check attached file for reference and provide solution.

 

Vserver migrate feature

$
0
0

Hello

 

We want to migrate some vservers on our 4node metrocluster to even out load and volume/aggregate usage but when doing a
"vserver migrate start -check-only true -vserver ......." we get this an error stating that "Vserver migrate feature must be enabled".

Question is, where can we find this feature and if its licensed or not?

Is it part of some package or other license?

 

Any help woud be much appreciated

 

Regards

Marco

LUN MOVE Sanity Check

$
0
0

I have two (2) LUNs in a single volume.  This prevents me from doing a Guest File Restore via SnapCenter.

If I were to move one of the LUNs to a new volume would this be at all disruptive if the LUN is currently presented to clustered ESXi hosts (6.5 U2) and contains several VMDKs which are attached as Basic Disks to Windows 2012 Servers?

 

Documentation doesn't specifically call out any risk of disruption and I've done similar operations via volume move without disruption.  But rather than assume anything I was hoping the community could provide some real-world practical advice from the LUN perspective.

 

Thank you in advance for said advice.

 

how is the IOPS calculated in OCUM ? The output in the OCUM does not match up with the CLI output.

$
0
0

Below is the CLI & OCUM output performed against the OCUM server “r00uy0n0c”. Please review and let me know your findings.

 

OCUM output:

 

Status

Volume

Style

Latency

IOPS

MBps

Free Capacity

Total Capacity

Cluster

Node

SVM

Aggregate

Ok

HON_v2

FlexVol

N/A

N/A

N/A

327.2181702

400

xxxxx

yyyy

zzz

aggr1_cdotsg230b_SATA

 

Output from command  provided:

 

yyyy::> qos statistics volume latency show -volume HON_v2 -vserver zzz -iterations 5

Workload            ID    Latency    Network    Cluster       Data       Disk        QoS      NVRAM

--------------- ------ ---------- ---------- ---------- ---------- ---------- ---------- ----------

-total-              -   207.00us    82.00us     1.00us    72.00us    52.00us        0ms        0ms

-total-              -   165.00us    85.00us        0ms    78.00us     2.00us        0ms        0ms

-total-              -   155.00us    83.00us        0ms    69.00us     3.00us        0ms        0ms

-total-              -   356.00us    91.00us     1.00us   201.00us    63.00us        0ms        0ms

-total-              -   333.00us    83.00us        0ms   117.00us   133.00us        0ms        0ms

 

Original command output:

 

yyyy::> qos statistics volume performance show -volume HON_v2  -vserver zzz -iterations 5

Workload            ID     IOPS       Throughput    Latency

--------------- ------ -------- ---------------- ----------

-total-              -      134        79.99KB/s   254.00us

-total-              -      126        67.33KB/s   185.00us

-total-              -      103        59.21KB/s   230.00us

-total-              -      128        52.89KB/s   339.00us

-total-              -      111        63.73KB/s   150.00us

 

New TR Released:TR-4650: NetApp ONTAP and Splunk Enterprise

$
0
0

This document presents the performance and reliability validation test results for NetApp ONTAP in a Splunk Enterprise environment. It also includes storage efficiency test results for Splunk indexer data.

For more info, please click here

 


ONTAP 9.4 SSH Public Key Access Broken ( key type ssh-rsa not in PubkeyAcceptedKeyTypes)

$
0
0

One of my powers users as role based restricted access to the FAS using a ssh-rsa 2048 public key only.   This previously worked OK we started at ONTAP 9.1, then 9.2 and  until recently was  on 9.3P4 all working OK for about 2 years.

 

The Problem:

The user can nolonger access the FAS using the pubkey.  I suspect but I  cannot be certain this broke when we updated to

9.4.P3 in December 2018.  The error is:  key type ssh-rsa not in PubkeyAcceptedKeyTypes .   I also tried a new key ssh-ed25519 both have the same error.  See below:

--------------

00000018.001cc78e 0dcc3fa7 Sat Mar 02 2019 12:04:13 +00:00 [auth_sshd:info:8218] userauth_pubkey: key type ssh-rsa not in PubkeyAcceptedKeyTypes [preauth]
00000018.001cc78f 0dcc3fa7 Sat Mar 02 2019 12:04:13 +00:00 [auth_sshd:info:8218] userauth_pubkey: key type ssh-ed25519 not in PubkeyAcceptedKeyTypes

--------------

The ssh keys are good, I checked the fingerprint at both end and tested to other servers (Linux / AIX) both worked with the keys OK.   Also SSH password based access to the FAS works fine.  The MFA second authentication method is set to none. 

 

Question:

1) As anybody seen this before.  I am struggling to get any good hits googling using the error message for ONTAP.

    Linux hits indicate sshd_config can be updated to allow key types removed at later SSH 7.x levels. For example to

   allow ssh-dss which was removed from the defaults at openssh 7.x.

2) I cannot see any means of querying or modifying the ONTAP (FAS) settings for PubkeyAcceptedKeyTypes.

 

I am able to log a support ticket via the NETAPP Partner IBM who provide our L1/L2 support before it esculates to NETAPP directly via IBM if they cannot resolve it.  However I want to ask in the community first and potentially build a stronger testcase to demonstrate the problem. 

【Ontap select】 NodeCreateFailed, ClusterDeployFailed

$
0
0

With ONTAP Select Deploy 2.10.1(evaluation license) , I got following errors when creating a 2node-Cluster with following condition.

 

<Host Information>

Hypervisor: ESXi 6.5 U2 (evaluation License) ×2

VMs on host#1: VCSA6.5, ONTAP select Deploy, (!!ontap select)

VMs on hsot#2: (!!ontap select)

!! 2node-cluster will be created here

 

<Error messages>

-NodeCreateFailed

 Node"[ClusterName]" create failed. Reason: Faild to write file([DatastorePass]): 401  Unauthorized. Manual deletion of this  node from its host may be required.

-Cluster Deploy Faild

Node StartErr: Node "[ClusterName]" failed to start : (Invalid/missing Cdrom device)

 

How to solve this issues ?

I would appreciate it very much if you could advise me anything.

Ontap 9.3 Nblade.nfsV4PoolExhaust File-ID Reverse Search

$
0
0

Hi all,

 

we had a big outtage on the weekend,

an application farm took down the NFS Service on a shared Metrocluster for 3rd party customers.

 

EMS Message:
Error message observed : 3/2/2019 18:39:06 <Cluster> EMERGENCY Nblade.nfsV4PoolExhaust: NFS Store Pool for OpenState exhausted.
Output of nfsv4_diag:

<Cluster>::*> statistics show -object nfsv4_diag -counter *storePool_* -raw
 
Object: nfsv4_diag
Instance: nfs4_diag
Start-time: 3/2/2019 19:28:27
End-time: 3/2/2019 19:28:27
Scope: <Clusternode-01>
 
    Counter                                                     Value
    -------------------------------- --------------------------------
    storePool_ByteLockAlloc                                       885
    storePool_ByteLockMax                                      512010
    storePool_ClientAlloc                                          51
    storePool_ClientMax                                        102401
    storePool_CopyStateAlloc                                        0
    storePool_CopyStateMax                                      10250
    storePool_DelegAlloc                                            0
    storePool_DelegMax                                         512010
    storePool_DelegStateAlloc                                       0
    storePool_DelegStateMax                                    512001
    storePool_LayoutAlloc                                           0
    storePool_LayoutMax                                        512010
    storePool_LayoutStateAlloc                                      0
    storePool_LayoutStateMax                                   512001
    storePool_LockStateAlloc                                      885
    storePool_LockStateMax                                     512008
    storePool_OpenAlloc                                        512001
    storePool_OpenMax                                          512010
    storePool_OpenStateAlloc                                   512001
    storePool_OpenStateMax                                     512001
 
Output of lock.txt from command:
vserver locks nfsv4 show -inst
showed me the causer who filled up the   storePool_OpenAlloc and storePool_OpenStateAlloc , for example:
Vserver: <poor_SVM>
Logical Interface: <LIF_name>
State ID: 00000001-0018ff8b5b004d1d00300000
Lock UUID: 664c7c33-66cf-4667-ab2b-968aaed86740
Client ID: 0x8bff18000000000f
Client Verifier: 0x5c3abd71172bc43b
Client Name: Linux NFSv4.1 <Application Servier>
Lease Remaining: 26
Lease State: active
Lease Count: 507848
Bad Client Count: 0
Owner Name: 6f70656e2069643a000000270000000000000057804820010000000000000000000000
Owner Type: open
Owner Mode: normal
Last Owner Sequence: 0xc96d82a
Lock Type: share-level
Shared Lock Mode: 0x40006
Bytelock Starting offset: -
Number Of Bytes Locked: -
Bytelock Is Exclusive: -
Volume MSID Major: 8000042c
Volume MSID Minor: 0
File ID: d4de65
Generation ID: 3d97186e
State Is Revoked: false
Delegation Type: -
Delegation Mode: -
Subsumed Open Count: -
Reclaim Count: 0
 
My question is, I can use the volume file show-filehandle if I have the past, the application developer need the reverese way,
I have the file-ID and need the path inc. file name,
could please anyone assist?
 
 
cluster1::> volume file show-filehandle -vserver vs0 -path /vol/vol1/file1.txt
             Vserver                    Path
             ----------------------     ---------------------------
             vs0                         /vol/vol1/file1.txt

             flags   snapid  fileid     generation  fsid        msid          dsid
             ------- ------  ---------  ----------  ----------  ------------  ------------
             0x0     0       0x60       0x206b6     0x402       0x80000402    0x402
 
Thanks in advance,
 
Henrik

NetApp CIFS Auting Security tab some of options disabled

$
0
0
Hi Team, We have enabled CIFS auditing on NetApp cluster mode with one of SVM . When I checked with CIFS share to specify operation to be monitored we found options are disabled please check attached file for details. Please help on same

Planning DR with Flex clone

$
0
0

Hi,

We have 2 node cluster in production as well as in DR site at 2 diffrent locations. Quterly we perform DR test activity with a bunch of volumes. However, we dont disturb the existing snapmirror setup between production and DR site. Rather, we create new snapmirror relation between existing DR volumes (say DR1, eg. vol_test_DR1) with newly created DR volumes (created for DR purpose only..say DR2, eg. vol_test_DR2. Means, DR site = DR1 volumes + DR2 volumes...We create new volumes in DR site, create snapmirror relation of them with existing DR volumes and destroy the volumes once the DR test activity is over.)  Now, as the environment is getting older, aggregates are getting used day by day and customer doesnt have a plan to invest in new disks right now.

 

We have only 20TB of space left in all the aggregates in DR site, wherein we have to create DR2 volumes for 50TB data. We are planning for Flex clone in this case. As this is a point in time copy, we would have to create all of them at the night before our activity, which is tough.

 

Here, I need the sugestion if there is any alternate way to use the flexclone in smother way or if we can include all the volumes in a script and run flex clone, what will the template of that script?

 

Regards,

Joy Mukhopadhyay.

Viewing all 4954 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>