Veritas Data Insight Administrator's Guide
- Section I. Getting started
- Introduction to Veritas Data Insight administration
- Configuring Data Insight global settings
- About scanning and event monitoring
- About filtering certain accounts, IP addresses, and paths
- About archiving data
- About Data Insight integration with Symantec Data Loss Prevention (DLP)
- Configuring advanced analytics
- About open shares
- About user risk score
- About bulk assignment of custodians
- Configuring Metadata Framework
- Section II. Configuring Data Insight
- Configuring Data Insight product users
- Configuring Data Insight product servers
- About node templates
- About automated alerts for patches and upgrades
- Configuring saved credentials
- Configuring directory service domains
- Adding a directory service domain to Data Insight
- Configuring containers
- Section III. Configuring native file systems in Data Insight
- Configuring NetApp 7-mode file server monitoring
- Configuring clustered NetApp file server monitoring
- About configuring secure communication between Data Insight and cluster-mode NetApp devices
- Configuring EMC Celerra or VNX monitoring
- Configuring EMC Isilon monitoring
- Configuring EMC Unity VSA file servers
- Configuring Hitachi NAS file server monitoring
- Configuring Windows File Server monitoring
- Configuring Veritas File System (VxFS) file server monitoring
- Configuring monitoring of a generic device
- Managing file servers
- Adding filers
- Adding shares
- Renaming storage devices
- Configuring NetApp 7-mode file server monitoring
- Section IV. Configuring SharePoint data sources
- Configuring monitoring of SharePoint web applications
- About the Data Insight web service for SharePoint
- Adding web applications
- Adding site collections
- Configuring monitoring of SharePoint Online accounts
- About SharePoint Online account monitoring
- Adding site collections to SharePoint Online accounts
- Configuring monitoring of SharePoint web applications
- Section V. Configuring cloud data sources
- Configuring monitoring of Box accounts
- Configuring OneDrive account monitoring
- Managing cloud sources
- Section VI. Configuring Object Storage Sources
- Section VII. Health and monitoring
- Section VIII. Alerts and policies
- Section IX. Remediation
- Configuring remediation settings
- Section X. Reference
- Appendix A. Data Insight best practices
- Appendix B. Migrating Data Insight components
- Appendix C. Backing up and restoring data
- Appendix D. Data Insight health checks
- About Data Insight health checks
- About Data Insight health checks
- Appendix E. Command File Reference
- Appendix F. Data Insight jobs
- Appendix G. Troubleshooting
- Troubleshooting FPolicy issues on NetApp devices
Using the metadata framework for classification and remediation
To apply tags to devices, files, folders, and shares, you must create a CSV file with the metadata key value pairs. You may either create the CSV file manually or use a third-party tool or script to generate the CSV with tagging information for paths or devices.
Caution:
Make sure you provide valid file paths or devices in CSV file. Incorrect devices or paths with incorrect device or share name will be discarded.
To apply metadata tags
- Create a CSV file with the tagging information. You can create more than one CSV file with tagging information for paths.
To assign tags to the files, ensure that the CSV file name starts with
File_
(for example,File_tags.csv
). Enter paths for different files with the tag name and tag values.To assign tags to the devices, ensure that the CSV file name starts with
Device_
(for example,Device_tags.csv
). Enter device with the tag name and tag values.To assign tags to the share/folder level, ensure that the CSV file name starts with Folder_ (for example, Folder_tags.csv). Enter share paths with the tag name and tag values.
Note:
i18n and special characters are not supported in tag names.
- Save the CSV files in the
data/console/tags
folder in the Data Insight installation directory on the Management Server. - A scheduled job TagsConsumerJob parses the CSV file and creates a
Tags
database for each share. The job imports the tags for the paths into Data Insight. The job runs once in a day by default.If the job is executed manually using the configcli command, the job forcefully consumes all the CSV files under
Tags
folder.Whenever the job runs, it checks if the modified time of any of the CSV files under the
Tags
folder is greater than the time of the previous execution of job. If the job finds any such CSV, it processes all the CSV files underTags
folder. If the CSV file(s) have not been modified after the job was last executed, the job does not take any action.The job does not accept any tag name that starts with mx_ because they are reserved for Data Insight internal tags usage. Whenever the job processes the CSV, Data Insight deletes all existing tags (except tags starting with "mx_") from all files and folders and attaches new tags.
Note:
If a path is tagged in two different CSV files with the same tag name but with a different value, then the value in the last CSV file that is processed is applied. For device tags, multiple CSV is supported and all tags are applied. For device level tags, you can associate multiple tag values with single tag.
- To replace existing tags, update the CSV with new tags. The scheduled job replaces existing tags with the new tags. If any paths are discarded during the last run of the job, these are logged in a database
$DATADIR/console/generictags_scan_status_5.0.db
that stores the discarded paths during the last run of the job.If any devices are discarded during the last run of the job, these are logged in a database
$DATADIR/console/devicetags_scan_status_5.0.db.
that stores the discarded device during the last run of the job. - Create a DQL report to retrieve the tags from the database.
Here are a few example queries that you can use:
To fetch all paths in your storage environment along with the tags (my_tag) assigned to them.
FROM path GET name, TAG my_tag
To get all paths owned by user Joe Camel tagged with the needs_assesment tag.
FROM owner GET TAG owner.path.needs_assessment, owner.path.name IF user.name="joe_camel"
To fetch the device in your storage environment along with the tags (my_tag) assigned to the device.
FROM device GET name, Tag device.my_tag
To fetch the share in your storage environment along with the tags (my_tag) assigned to the device.
FROM msu GET name, Tag msu.my_tag
Note:
The DQL report output does not return any tag if the content does not match any predefined classification tag.
- To verify the names of tags that are stored for a share, run the idxreader command on the indexer node.
idxreader - i $MATRIX_DATA_DIR/indexer/default/99/99 - gettags all
- To remove all the tags, delete the CSV file but if it is the last CSV file in the Tag folder, remove the content from the CSV file. Do not delete the last CSV in the folder.
The CSV file with the metadata tags should be in the following format:
Device name/File/folder path, tag name, tag value
Examples:
For File tags, \\filer\share\foo,tname,tvalue
For Device tags, BOX,tname,tvalue
Where, tname refers to the name of the tag, and tvalue refers to the tag value.
Note:
For file and folder level tags, multiple values for the same tags are not supported.
If the path or the tag name contains a comma, enclose the text in double quotes (","). For example, if the folder name is foo, bar, you can add the path in the CSV as follows:
"\\filer\share\foo,bar",t_name,t_value
For shares, the path should be present in the CSV file containing folder paths. Following are examples of share level paths:
CIFS/DFS | \\filer\share |
SharePoint | Device\filer\folder\file |
SharePoint On-prem | URL of the site collection |
NFS | <export path> For example, /data/finance/docs |
Box | \\Box\<box name in Data Insight> |
OneDrive | \\OneDriveAccountName\OneDriveUserAccount |