Resources

Setting Up Ansys Remote Solve Manager (RSM) on Windows for Fluids/Structures Tools: A Step-by-Step Guide

Written by Samuel Lopez | Jan 8, 2025 12:59:37 AM

Introduction to Remote Solve Manager (RSM)

Ansys Remote Solve Manager (RSM) is a versatile tool that enhances computational workflows by enabling users to submit simulation jobs to remote clusters or high-performance computing (HPC) systems. Whether managing large simulations or optimizing resource usage, RSM simplifies job scheduling, execution, and resource allocation.

This guide provides step-by-step instructions for configuring RSM on a Windows machine using the Ansys Remote Cluster (ARC) job scheduler. While this example focuses on ARC, RSM also supports third-party job schedulers like Platform LSF, PBS Professional, Slurm, Altair Grid Engine, and Windows HPC. For details on compatibility with other schedulers, refer to the Ansys support documentation.

These instructions specifically apply to RSM for Fluids/Structures tools. A slightly different setup is required for Electronics tools. For those steps, please refer to our dedicated blog here.

Scenario Overview

This tutorial demonstrates configuring a multi-node ARC cluster for RSM and setting up a custom job scheduling system. In this example:

  • Master Node (MASTER): Handles job submission and management.

  • Compute Nodes (HOST1, HOST2): These machines will generally be the main work machines which execute the simulations. In this example, we will assume HOST1 is a smaller machine with limited compute resources, while HOST2 is a larger, high-capacity machine.

Step 1: Using the ARC Configuration Manager to Create Your Cluster

Setting Up the MASTER Machine (Master Node)

  1. Log In: Access the MASTER machine with administrator credentials.

  2. Launch ARC Configuration:
    • Navigate to Start > Ansys 202# R# > ARC Configuration 202# R# or run the executable located at [RSM_Install_Location]\ARC\bin\arcConfigConsole.exe.



  3. Define Cluster Usage:
     
    • Option 1: Submit, schedule, and run jobs on the same machine (local queue).

    • Option 2: Submit jobs to a master node on the network, which may also act as an execution node alongside additional compute nodes.

  4. Start Master Service: Start the Master Node Service via the configuration interface. You can allow the MASTER machine to execute jobs for additional computational capacity or restrict it to job distribution only. For this example, the MASTER node is also configured to run jobs.

Adding Execution Nodes

With the master node designated, the next step is to add the other two machines to the cluster as execution nodes. These nodes will handle the actual job processing. The master node can also participate as an execution node if the corresponding option was selected earlier. Once configured, each node will appear in the dialog box. By selecting a node from the list, you can adjust its settings, such as limiting the maximum number of cores it can use for job execution. This helps prevent overloading and ensures at least 1-2 CPUs remain available for the operating system's background tasks. Similarly, when additional execution nodes are added, they will be listed below the master node and can be customized to suit the cluster’s needs.

For this example, we will choose the following settings:

  1. Designate Compute Nodes: Right-click Execution Nodes, select “Add an execution node,” and input the names of the execution nodes (in this case, HOST1 and HOST2).

  2. Configure Resource Limits:
    • For both MASTER and HOST1, restrict the maximum number of cores available for jobs to prevent overloading.
    • For HOST2, remove restrictions to fully utilize its capacity.

  3. Enable Node Services: Start the Execution Node Service on each compute node.

Creating a Custom Queue

A job queue serves as a virtual pipeline where computational tasks (jobs) are organized, prioritized, and executed using the resources available within a cluster. Creating a custom queue allows you to optimize resource allocation for specific requirements—for example, reserving high-capacity nodes for memory-intensive simulations or capping the number of concurrent jobs on a machine to prevent overloading. Custom queues provide greater flexibility and control than default or local queues, making them a valuable tool for efficiently managing resources and ensuring critical tasks are handled according to predefined priorities. As an illustration, we will configure a custom queue accessible exclusively to HOST2, specifically for handling larger, memory-intensive jobs. For smaller clusters, the default and local queues are often sufficient. To set up a custom queue, follow these steps.

  • Add Queue: Right-click Queues, select “Add a queue,” and name it high_mem.

Configure Queue Settings:
 
  • State:
    • Enabled: The queue is active and can accept jobs.
    • Disabled: The queue will reject any incoming jobs.
    • Suspended: The queue can accept jobs, but they will remain in the queue without being processed.
  • Priority: Define the priority level for jobs in this queue relative to other queues. Jobs in higher-priority queues will be executed first. Priority levels include: 
    • High
    • Above Normal
    • Normal
    • Below Normal
    • Low

Active Time Range: Check the "Set Active Time Range" box to ensure the queue remains active at all times.

Job Limit: Specify the maximum number of jobs that can run simultaneously in this queue. In this example, we’ll set the limit to 100 jobs.



Assign Execute Nodes: Assign specific nodes to this queue by navigating to the Access Control settings. For this example, select HOST2 as the designated execution node, as shown below.

Save Changes: Apply the configuration. Optionally, enable credential caching for secure job submissions.

Step 2: Configuring Client Machines to Submit to the Cluster

To enable RSM client machines to submit jobs to the ARC submit host (MASTER), a configuration must be set up on the client side of the configuration to establish communication between the client and the submit host, define the file transfer method, and map RSM queues to the corresponding ARC cluster queues. To do this, we can follow the steps below.

  1. Launch RSM Configuration: Open Start > Ansys 202# R# > RSM Configuration 202# R#.

  2. Add an HPC Resource:
    • Assign a name (e.g., ARC), choose ARC as the HPC type, and specify the MASTER machine name. It should be noted that, here, the HPC type should be set to ARC as shown below. This denotes that we are using the Ansys default job scheduler we just created. If you have a 3rd party job scheduler, you can select it from the available list of schedulers listed here.




  3. Set File Management:
    • Use RSM Internal File Transfer Mechanism, and define a staging directory. This determines how files get from the RSM cluster to a location where they can be readily manipulated or read. In this example, we will choose the option for “RSM Internal File Transfer Mechanism”, which requires the specification of a staging directory. The staging directory is the location where all the job files are written to during/after the solve. We use the option for internal file transfer mechanism when the remote HPC staging directory is not visible to client machines. RSM uses TCP sockets to stream files from the client machine to the submit host machine. In this case, you must specify the path to the directory where the job files will be staged. This directory must be visible to all HPC nodes.



  4. Import Queues: Import local, default, and custom queues (e.g., high_mem).

  5. Test Queues: Verify functionality by submitting test jobs and optionally renaming queues for clarity.

Job Diagnostics and Troubleshooting

When submitting test jobs through the RSM Configuration tool, the status bar provides real-time updates on submission and execution progress. Upon completion, whether the job is successful or fails, a detailed diagnostics report is available. This report includes critical information for troubleshooting and ensures the configuration is correctly set up.

Step 3: Sharing the Configuration with Users

Option 1: Centralized Sharing

  1. Create a Shared Directory: Example path: C:\SharedRSMConfig\ANSYS\v2##\RSM.
  2. Copy Files: Transfer ARC.rsmcc and queues.rsmq to the shared directory.
  3. Distribute Path: Share the directory path with users for seamless integration.

Option 2: Local Copy Distribution

  1. Distribute Configuration Files: Provide users with the required files to copy into their local directories.
    • Windows Path: %APPDATA%\ANSYS\v2##\RSM.

Following these steps ensures a robust and efficient setup of Ansys RSM, enabling users to maximize simulation performance and streamline computational workflows.

 

Step 4: How to Configure in Mechanical/Fluent

Now that we have our RSM configuration setup, we can review how we can submit jobs to our new cluster in Mechanical or Fluent.

Mechanical Setup

In Ansys Mechanical, you can set up your job as usual on your local machine. However, when it's time to submit the solve, an adjustment is needed to direct Mechanical to use RSM. To configure this, open the Mechanical interface and navigate to File → Solve Process Settings. A window will appear similar to the one shown below. By default, the option "My Computer" will be selected. To enable RSM, switch this setting to My Computer, Background.

From here, you can configure the RSM setup. Set the Solution Execution Target to RSM, which specifies where the job will be submitted. Next, choose the appropriate queue for the job. While the default is "local," you can select a different queue if multiple options are available. Because the RSM Configuration Tool was set up in previous steps, the fields for HPC Configuration, HPC Type, and HPC Queue will automatically populate based on the saved configuration.

Finally, provide a job name (default is "Mechanical") and select the license to use for solving. Once everything is set, click OK, specify the number of CPUs to use for the solve, and proceed with solving the model!

 

Fluent Setup

Submitting jobs in Fluent requires a slightly different workflow. If you are using Workbench Fluent, you can use the process below. For standalone Fluent, the options can be specified from the launcher.

Access Solution Properties:

  • In Workbench, right-click the Solution cell and select Properties.

Configure RSM Submission Settings:

  • Under Solution Process settings, set the Update Option to Submit to Remote Solve Manager (RSM).

Select an RSM Queue:

  • Choose the appropriate RSM queue for your job.

Set Execution Mode:

  • Configure the Execution Mode as either Serial or Parallel based on your requirements.

Specify Parallel Processes (if applicable):

  • For a parallel solution, set the Number of Processes.
  • Ensure the number of processes is equal to or less than the available cores in the cluster and that each process handles at least 100,000 nodes or elements.

Submit the Job:

  • Update the project to submit the job to the RSM.
  • During the solution process, the Solution cell will move to Pending mode, reconnecting automatically once the run is complete.

Manage the Update Process:

  • You can interrupt or abort an update at any time by right-clicking the Solution cell.
  • This option is available for foreground, background, and RSM updates.