Quantcast
Viewing all 4519 articles
Browse latest View live

S2D quorum question

Hi All

I´d truly appreciate your help to clarify the quorum operative in an S2D environment. I´ve read a bunch about Cluster and pool quorum, I know how these two quorum works, but I dont clearly see which one applies for Storage spaces Direct.

Quorum scheme should apply , so every disk should have a vote, and the owner should have another.

But I see a failover cluster supporing this S2d cluster, and this leevrages a File Witness (for example)  

Which one takes effect?

Best Regards


Can we configure Windows Failover Clustering without Central Storage(iSCSI)

Hi,

I am completely new to Windows Server Failover Clustering subject, i have few doubts please clarify me.

I setup a Virtual lab with a DC+iSCSI installed and 2 File Server Nodes, configured Failover Clustering at both the nodes and verified the working.

When i shutdown the 1st Node dynamically the 2nd Node takes place, here my question is cant we create a Failover Cluster for the Existing File Server share folder??? instead have to configure SAN(iSCSI) for storage here both the nodes HDD space goes unused when all the data gets save in that SAN storage.

The same for SQL Failover too, when we have SQL DB server running in a server when i install Failover Clustering again on the same server for data storage have to select SAN, here the failover clustering only helps in keeping the Server Roles up and running, what will happen when Storage SAN device goes down??

I can understand SAN(iSCSI) is required such that both the Nodes get access to the storage device only then when one goes down 2nd one act behalf of that and takes the request to SAN server.

Any help please.


Mohammed...

Windows Server 2016 and High Availability Remote Desktop Services with thin clients

Hello,

I need to find a way on how to configure two redundant Window Server 2016 servers, in a way that for the thin client it is transparent which of them is really active to grant the Remote Desktop Service session. The thin client would not know which of the two servers he would be physically connected to. If this server fails, the other server will continue to grant the session to the thin client without this noticing any interruption. 

In principle I will not run any virtual machine. The software that I use will be installed in the real machines. 


Can you please suggest how this could be done?


Thanks in advance.

VMs located on one of CSV volumes stopped migrating on one of cluster nodes

We have a 3 node cluster Windows 2016 with many VMs on 3 CSV volumes. At one moment (I'm not sure when) VMs located on first CSV volume stoped to migrate (live and quick) to fist node (only to first node). 1st volume is still visible from 1st node. Cluster validation didn't show any problem.
In event log Microsoft-Windows-Hyper-V-VMMS/Admin on 1st node:
EventID:16300 
Cannot load a virtual machine configuration: The system cannot find file specified. (0x80070002) (Virtual machine ID ....)
EventID:21002
'VM name' Failed to create Planned Virtual Machine at migration destination:The system cannot find file specified. (0x80070002) (Virtual machine ID ....)

Any ideas how to fix this problem?

I would appreciate any help.  

Thanks.

Drain role Failed

We have three Node N-1, N-2, N-3. I drain role from N-2 and there10 VM's Moved out of 14. 4 VM are not moving getting error . Tried to move manually but the error same. Please assists All the Node in are WIN-2012 R2

Error Message : "operation did not complete on resource virtual machine live migration"

SAN HPE SV3200 iScsiIPrt errors crashing VMs and after cascade also Fail Over Server Nodes?

We have a three node W2012 R2 Fail over cluster that has been running spotless for years with the HPE P4300 SAN but after adding the HPE Storevirtual SV3200 as a new SAN we are having iScsiPrt errors that HPE Support cannot fix, crashing VMs and also two of the three fail over nodes. 

At first everything seemed to work, but after adding additional disks on the SAN a SAN controller crashed. That has been replaced under warranty but now when moving our servers and especially SQL 2008 Servers to the SAN, problems start to occur. The VHDX volumes of the SQL servers are thin provisioned.  

Live moving of the storage worked fine for none SQL servers. For some SQL servers the servers frooze and operation was halted, so we needed to perform an offline move. Then during high disk IO and especially during backups W2012 R2 FOC started to behave erratic eventually crashing VMs and in one instance rebooting two fail over nodes, as a result of a flood of iScsciPrt errors in the eventlog:

System iScsiPrt event ID 27 error Initiator could not find a match for the initiator task tag in the received PDU. Dump data contains the entire iSCSI header.
System iScsiPrt event 129 warning The description for Event ID 129 from source iScsiPrt cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.

If the event originated on another computer, the display information had to be saved with the event.

The following information was included with the event:

\Device\RaidPort4

the message resource is present but the message is not found in the string/message table

System iScsiPrt event ID 39 error Initiator sent a task management command to reset the target. The target name is given in the dump data.
System iScsiPrt event ID 9 error Target did not respond in time for a SCSI request. The CDB is given in the dump data.
System iScsiPrt event 129 warning The description for Event ID 129 from source iScsiPrt cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.

If the event originated on another computer, the display information had to be saved with the event.

The following information was included with the event:

\Device\RaidPort4

the message resource is present but the message is not found in the string/message table
System iScsiPrt event ID 27 error Initiator could not find a match for the initiator task tag in the received PDU. Dump data contains the entire iSCSI header.
System FailOverClustering event id 5121 Information Cluster Shared Volume 'Volume4' ('NEMCL01_CSV04') is no longer directly accessible from this cluster node. I/O access will be redirected to the storage device over the network to the node that owns the volume. If this results in degraded performance, please troubleshoot this node's connectivity to the storage device and I/O will resume to a healthy state once connectivity to the storage device is reestablished.

After a 2 hour period of these events the FailOver Cluster services started to give errors, VMs failed and finally 2 nodes of our 3 node failover cluster rebooted because of a crash.

Sofar HPE has not been able to fix this. The SV3200 logs has occasional ISCSI controller errors but the error logging in the SVMC is minimal. 

HPE support blamed using a VIP and using Sites (a label). Both are supported according to the HPE product documentation. This has been removed and ISCSI initiator has been set to the Eth0 bond IP adresses directly. As problems persist they blamed that we are using the Lefthand DSM MPIO driver on the initiator connections to the SV3200 which is not the case. Standard MS DSM. Yes the Lefthand driver is on the system for our old SAN but not configured for the SV3200 initiator sessions, which is round robin with supset.     

We  are currently facing a legal warranty standoff.

Any pointers  or other comparable experiences with the HPE Storevirtual SV3200 SAN?

TIA,

Fred

Switch Embedded Teaming

I am creating a new storage spaces direct cluster on server 2019 to replace an existing 2016 cluster.  I am using dual port Mellanox x-5 cards after adding the physical NICs to the SET switch I create the virtual NICs for hyper-v after adding the vnics the server reports their connection speed as 10 gigabits which is the speed of the physical links but on my 2016 storage spaces direct cluster the virtual NICs report a connection speed of 20 gigabits per second which would be the aggregate speed of the two 10 gigabit links.  What I am trying to find out is what the virtual NICs are supposed to report their connection speed as the single physical links or the total of all the links in the virtual switch set.  I can’t seems to find an answer of what the correct answer is in my searching.  I want to make sure I am not missing a set somewhere.  Can anyone shed some light on what should the virtual network cards report their speed as?

Unable to migrate machines within Failover Cluster Manager

We have 4 physical servers which are all clustered. When I try to migrate a VM from one host to another I get the following error:

event id 1069 - Microsoft windows failover clustering
the error code was 0x775 the reference account is currently locked out and may not be logged onto.

The virtual machine then either goes directly into a failed state or tries to move to another server before going into a failed state. The only way to get it back online is to migrate it back to the original host.

I have noticed that the hosts do not all have access to the all of the volumes within cluster storage and it is only when I migrate the disk to a different host am I then able to access it from that host. If I try to access a volume hosted on host 1 from host 1 I can access it fine but if I try to access that same volume from host 2 I get an error stating that the account is locked out.

This has only stated causing problems in the last week and no permissions have changed to any of the volumes nor have any updates been applied to the system.

Thanks


NLB Cluser add host | Error: The RPC is unavailable on the specified computer

Hi,

I have facing problem adding host in NLB Cluster. it shows error: The RPC is unavailable on the specified computer.

1. account is administrator.

2. File and printer sharing is enabled.

3. DNS record resolving and ping each other.

4. 2 NIC card.

5. cluster method "Multicast".

Please give suggestion, how to solve the problem.

Image may be NSFW.
Clik here to view.

Failover cluster quorum problem

Hi all. I´v gato a Win server 2016 failover cluser with a File Witness reporting no error or warning on "test failover cluster Wizard"

However, when testing one node failure, Clsuter Stopped (And S2D running upon this also with this error:

event 1564,  File Share Witness Resource failed to arbitrate for the share \\xxx\xxx. Pleasse ensure its accesible.

I´ve checked and I can access with no proble to this share from both hosts. REplacing the File Witness solves the problem, But i Got no trust in this quorum anymore. Is there a way to check Witness access from a faioover cluster?  somehow like a "ping"

Regards 

CPU Usage: 50% or above on all Server 2016 Hyper-V Clustered nodes with 3 VM's

Hello,

We have a 3-node Server 2016 cluster, based on Dell R730 servers.All Hosts show a CPU Usage of 50% or above. TaskManager shows the correct value (around 0%)

Only 3 small test-vm's are running on those hosts. Has anyone seen this behavior?

We would like resolve this.

Image may be NSFW.
Clik here to view.

Regards,

Jan


Jan

Howto get a clustered MSDCT in Azure on Windows Server 2019 with S2D to function

I've been pulling my hair for the last few days trying to get a FCI working with SQL Server and MSDTC, no matter what I try I can't get the MSDTC role to get online.

I've followed the instructions here: https://www.ryanjadams.com/2018/07/sql-server-failover-cluster-instance-azure-msdtc/

Current setup:
- Domain joined servers (ADDS)
- Two CSV (one for SQL, one for DTC) running on S2D (2 data disks per server = 4 disks in total)
- Windows Server 2019 Datacenter with latest updates applied
- SQL Server 2016 SP2 and latest CU
- Standard IP's (one NIC per server) and a Standard load balancer for the VM's
- Cloud witness for quorum
- Static ip addresses for all servers and FCI (cluster and roles)
- Pre-staged DNS records allowing authenticated users to update the records
- Running install/configuration with domain admin account
- Created the FC using the ManagementPointNetworkType = Singleton
- Allowed DTC, COM+ Network Access, Network Discovery, File and Printer Sharing plus the SQL port in Firewall
- Configured DTC to allow all options and tested different auth levels

All other settings are made according to the video Ryan posted but as soon as I set the static ip for the DTC role the resource fails and I can't get it started.

Anyone know what I'm missing? Has anyone else successfully configured a similar setup in Azure?

regards
/andreas

Hyper-V Cluster 2016

Hello,

I'm installing new 2 node Hyper-V Cluster 2016 all up and running 

1. I have few VMs running under Hyper-V manager.

2. Trying to make the VMs highly available.

3. I'm able to perform the following task

Under the Hyper-V Cluster Manger I select Roles, Configure Role, Select Virtual Michine, I select the VM and it runs without any problem the VM is now High Availability.

Now The problem is when I do a second VM  

I repeat the same steps and it doesn't created an new VM it creates a group that I can't manage 

can't do live migrate I'm attaching 2 images 

fist image showing first VM and second showing to VMs on my don't know if this is a 2016 settingImage may be NSFW.
Clik here to view.
Image may be NSFW.
Clik here to view.

S2D verification

quick overview. 

I have had a 4 node cluster working on 2016 for a couple years as a learning project. during that time my friends also played with it adding around 40 vms. the whole thing slowed to a crawl with even clicking on a start menu taking around 2 minutes to respond. I decided at add some SSD's as it was all 15K disks before but was told if I enabled s2d with no cache there is no supported method for adding cache.

I destroyed the disk and pool and upgraded all 4 nodes to 2019. disabled and re-enabled S2d and built a pool. tried to add a disk but the Gui is useless. power-shell'd a small disk then expand to proper size but I cant tell if I am properly caching to SSDs or not. is there any way to know? my SSDs are showing as journaling which is a good sign but I don't want to move any VMs back till I know my storage is good.

setup:

4 x dell r710s with dual e5-2697's

1 146gb 15k OS drive

5 146gb 15K storage drives

2 200GB industrail SSDs for cache

future setup

I have 2 more nodes with same hardrive/ssd setup but I want to make  nodes work before adding 2 more.


VMs Fail Randomly on 2012 Cluster

Our 2-node Server 2012 Hyper-V cluster is having an issue where VMs seem to randomly fail for no apparent reason.  We are using Dell R900s with a MD3200i SAN and we have separate networks for iSCSI and Heartbeat.  On one of the nodes we get an error 1069 "Cluster Resource 'Virtual Machine VM200X32' of type 'Virtual Machine' in clustered role 'VM200X32' failed."  Below I have listed the relevant cluster log events from around the time it fails.  The cluster has passed validation and is running 50+ test VMs just fine, it's only a few of them that seem to be having this issue.

Just wondering if anyone else might have some input on what the problem could be.

Cluster event log:

2013/08/25-16:41:18.448 WARN  [RHS] Resource Virtual Machine VM200X32 IsAlive has indicated failure.
2013/08/25-16:41:18.463 INFO  [RCM] HandleMonitorReply: FAILURENOTIFICATION for 'Virtual Machine VM200X32', gen(0) result 1/0.
2013/08/25-16:41:18.463 INFO  [RCM] Res Virtual Machine VM200X32: Online -> ProcessingFailure( StateUnknown )
2013/08/25-16:41:18.463 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) Online-->ProcessingFailure.
2013/08/25-16:41:18.463 INFO  [RCM] rcm::RcmGroup::UpdateStateIfChanged: (VM200X32, Online --> Pending)
2013/08/25-16:41:18.463 ERR   [RCM] rcm::RcmResource::HandleFailure: (Virtual Machine VM200X32)
2013/08/25-16:41:18.463 INFO  [RCM] resource Virtual Machine VM200X32: failure count: 0, restartAction: 0 persistentState: 1.
2013/08/25-16:41:18.463 INFO  [RCM] Will queue immediate restart (500 milliseconds) of Virtual Machine VM200X32 after terminate is complete.
2013/08/25-16:41:18.463 INFO  [RCM] Res Virtual Machine VM200X32: ProcessingFailure -> WaitingToTerminate( DelayRestartingResource )
2013/08/25-16:41:18.463 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) ProcessingFailure-->[WaitingToTerminate to DelayRestartingResource].
2013/08/25-16:41:18.463 INFO  [RCM] Res Virtual Machine VM200X32: [WaitingToTerminate to DelayRestartingResource] -> Terminating( DelayRestartingResource )
2013/08/25-16:41:18.463 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) [WaitingToTerminate to DelayRestartingResource]-->[Terminating to DelayRestartingResource].
2013/08/25-16:41:18.463 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: Current state 'Online', event 'Terminate'
2013/08/25-16:41:18.463 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: State change 'Online' -> 'Terminated'
2013/08/25-16:41:18.463 INFO  [RCM] ignored non-local state Pending for group VM200X32
2013/08/25-16:41:18.479 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine Configuration VM200X32', gen(0) result 0/0.
2013/08/25-16:41:18.479 INFO  [RCM] Virtual Machine Configuration VM200X32: Flags 1 added to StatusInformation. New StatusInformation 1
2013/08/25-16:41:18.479 INFO  [RCM] VM200X32: Added Flags 1 to StatusInformation. New StatusInformation 1
2013/08/25-16:41:18.479 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.275 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine Configuration VM200X32', gen(0) result 0/0.
2013/08/25-16:41:19.275 INFO  [RCM] Virtual Machine Configuration VM200X32: Flags 1 removed from StatusInformation. New StatusInformation 0
2013/08/25-16:41:19.275 INFO  [RCM] VM200X32: Removed Flags 1 from StatusInformation. New StatusInformation 0
2013/08/25-16:41:19.275 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.275 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: Current state 'Terminated', event 'VmStopped'
2013/08/25-16:41:19.306 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.836 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.836 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.836 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: State change 'Terminated' -> 'Offline'
2013/08/25-16:41:19.836 INFO  [RCM] HandleMonitorReply: TERMINATERESOURCE for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:19.836 INFO  [RCM] Res Virtual Machine VM200X32: [Terminating to DelayRestartingResource] -> DelayRestartingResource( StateUnknown )
2013/08/25-16:41:19.836 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) [Terminating to DelayRestartingResource]-->DelayRestartingResource.
2013/08/25-16:41:19.836 WARN  [RCM] Queueing immediate delay restart of resource Virtual Machine VM200X32 in 500 ms.
2013/08/25-16:41:20.351 INFO  [RCM] Delay-restarting Virtual Machine VM200X32 and any waiting dependents.
2013/08/25-16:41:20.351 INFO  [RCM-rbtr] giving default token to group VM200X32
2013/08/25-16:41:20.351 INFO  [RCM-rbtr] giving default token to group VM200X32
2013/08/25-16:41:20.351 INFO  [RCM] Res Virtual Machine VM200X32: DelayRestartingResource -> OnlineCallIssued( StateUnknown )
2013/08/25-16:41:20.351 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) DelayRestartingResource-->OnlineCallIssued.
2013/08/25-16:41:20.351 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: Current state 'Offline', event 'Online'
2013/08/25-16:41:20.351 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: State change 'Offline' -> 'OnlinePending'
2013/08/25-16:41:20.351 INFO  [RCM] HandleMonitorReply: ONLINERESOURCE for 'Virtual Machine VM200X32', gen(1) result 997/0.
2013/08/25-16:41:20.351 INFO  [RCM] Res Virtual Machine VM200X32: OnlineCallIssued -> OnlinePending( StateUnknown )
2013/08/25-16:41:20.351 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) OnlineCallIssued-->OnlinePending.
2013/08/25-16:41:20.351 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:20.351 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine Configuration VM200X32', gen(0) result 0/0.
2013/08/25-16:41:20.351 INFO  [RCM] Virtual Machine Configuration VM200X32: Flags 1 added to StatusInformation. New StatusInformation 1
2013/08/25-16:41:20.351 INFO  [RCM] VM200X32: Added Flags 1 to StatusInformation. New StatusInformation 1
2013/08/25-16:41:20.367 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:20.694 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:21.911 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine Configuration VM200X32', gen(0) result 0/0.
2013/08/25-16:41:21.911 INFO  [RCM] Virtual Machine Configuration VM200X32: Flags 1 removed from StatusInformation. New StatusInformation 0
2013/08/25-16:41:21.911 INFO  [RCM] VM200X32: Removed Flags 1 from StatusInformation. New StatusInformation 0
2013/08/25-16:41:21.911 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:21.911 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: Current state 'OnlinePending', event 'VmRunning'
2013/08/25-16:41:21.942 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:21.942 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: 'Virtual Machine VM200X32' successfully started the virtual machine.
2013/08/25-16:41:21.958 INFO  [RES] Virtual Machine <Virtual Machine VM200X32>: State change 'OnlinePending' -> 'Online'
2013/08/25-16:41:21.958 INFO  [RHS] Resource Virtual Machine VM200X32 has come online. RHS is about to report status change to RCM
2013/08/25-16:41:21.958 INFO  [RCM] HandleMonitorReply: ONLINERESOURCE for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:21.958 INFO  [RCM] Res Virtual Machine VM200X32: OnlinePending -> Online( StateUnknown )
2013/08/25-16:41:21.958 INFO  [RCM] TransitionToState(Virtual Machine VM200X32) OnlinePending-->Online.
2013/08/25-16:41:21.958 INFO  [RCM] rcm::RcmGroup::UpdateStateIfChanged: (VM200X32, Pending --> Online)
2013/08/25-16:41:21.958 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VM200X32', gen(1) result 0/0.
2013/08/25-16:41:21.958 INFO  [RCM] ignored non-local state Online for group VM200X32


ADMIN $ share on the failover cluster does not exist

Colleagues, help!
I have a working CLUSTER cluster of two nodes. On both nodes is windows server 2012R2. I am trying to install the Data Protection Manager 2012R2 agent to protect my cluster. In doing so, I get the error "The agent operation failed because the ADMIN $ share on CLUSTER does not exist".
I checked this situation and this is what happened:
\\ node01 \ admin $ - available
\\ node02 \ admin $ - available
\\ CLUSTER \ admin $ - not available
How can I open access to the administrative shared folder on the cluster?
Thanks!

CSV Autopause - Single client contification start.

HI,

I've just got a warning from my cluster that one of my CSVs was stopped. But I just dont get what was going on.

From the Failoverclustering-CsvFs protocol I get this message:

"Volume {44179469-89e8-4971-b9ff-057c4579c647} is autopaused. Status 0xC00000C4. Source: Single client contification start."

What does that even mean? Single Client contification?

Best Regards

Daniel

AlwaysOn WSFC 2016 CNO was moved

Hi All,
I have SQL Server 2016 (EE) with AlwaysOn configured in my environment. During a recent maintenance window I was updating IP Addresses/Names of resources in my WSFC and as a result the CNO was moved from under the Cluster Core Resources to being listed as a resource under AG Name as if it was the listener. This caused this one particular AG to then use the CNO as if it were the listener and made the AG failover to the Synchronous replica as the primary. Not sure how to get the CNO back listed under the Cluster Core Resources and removed from AG w/o breaking my WSFC along with the other AG's. CNO is circled in red below. Any help would be greatly appreciated

Listed under the Roles is the name of each AG. The bottom of screen provides more details on each AG. However this one particular AG has 2 resources listed under the Server Name. 1 is the AG Listener and the other resource is the (CNO) cluster named object. Not sure how the cno got included as AG Server Name Resource but i need to remove CNO from AG and get it back listed under Cluster Core resources

Image may be NSFW.
Clik here to view.
Listed under Roles is the name of each AG. The bottom half of screen provides additional details for each AG Role. This particular role has 2 resources listed under Server Name 1 is the AG Listener and the other is the (CNO) Cluster Named Object


KrisT_SQL


2012 R2 Scale-Out File Server Performance Issue

I'm implementing a product called AppLayering by Citrix in a VMware environment. It creates a unique .vhd for each piece of software you install and want to deploy to end users. We created a Scale-Out File Server for the share so that we could have 100% up time from crashes and updates/reboots. The end user machines mount the .vhds at login; usually anywhere from 5-15 of these .vhds which range from 1Gb to 12GB in size.


Now that I'm increasing the amount of machines accessing this share, sometimes I experience a very long delay, as much as 6 minutes, before the layers are mounted. They usually mount within seconds. However, it's not consistently worse the more machines that are logged in, rarely it's still instant, but it does seem to get worse in general the more machines are mounting these layers.


The only performance settings I've tried to tinker with is the MaxThreadsPerQueue from 20 to 64. This reg entry was not in the registry by default, I had to make it myself, so I'm not sure if that means anything. Also not sure if 64 is even a good number to change it to either, just shooting in the dark here, any help would be much appreciated!


Darin

clussvc causing CPU cores to hit 100% constantly

Hello,

We are seeing a really strange situation. In a 4 node cluster HCI setup, newest available Microsoft certified hardware, running fully patched (2019-10) Windows Server 2019 DC with S2D enabled, there is constantly a CPU being maxed out at 100%

According to Nagios alerts I have setup, this happens every hour or so:

Service: CPU-7-Node-0 / CRITICAL: CPU-7-Node-0 = 100 and seeing the performance monitor on all cores, there is always a core being at 99% and falling back down.

I thought initially this was due to a problem VM running or storage running on the node, but even if the VMs are fully off the node, disks are owned by other nodes and Infrastructure File Server role, but the CPU is still being hit. I do not see any alerts in the event viewer about any issues with the cluster and it fully validates with Success on every applicable test.

Restarting theclussvc service on the node, alleviates the problem for about 12 hours, but the issue comes back. I am using latest BIOS/drivers.

Thank you in advance.


Viewing all 4519 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>