Sei sulla pagina 1di 180

IBM StoredIQ Platform

Version 7.5.0.1

Data Server Administration Guide



SC27-5692-00
IBM StoredIQ Platform
Version 7.5.0.1

Data Server Administration Guide



SC27-5692-00
Note
Before using this information and the product it supports, read the information in Notices on page 161.

This edition applies to Version 7.5.0.1 of (product number 5725-M86) and to all subsequent releases and
modifications until otherwise indicated in new editions.
Copyright IBM Corporation 2001, 2013.
US Government Users Restricted Rights Use, duplication or disclosure restricted by GSA ADP Schedule Contract
with IBM Corp.
Contents
How to send your comments . . . . . v Configuring audit settings . . . . . . . . 30
Configuring hash settings . . . . . . . . 30
Contacting IBM . . . . . . . . . . . vii Configuring desktop settings . . . . . . . 31
Downloading the IBM Desktop Data Collector
installer from the application . . . . . . . 31
Contacting IBM StoredIQ customer Upgrading the IBM Desktop Data Collector agent 31
support. . . . . . . . . . . . . . . ix Using the Encrypted File System recovery agent 32

Understanding IBM StoredIQ Platform Creating volumes and data sources . . 33


distributed architecture . . . . . . . . 1 Volume indexing . . . . . . . . . . . . 33
What is IBM StoredIQ Administrator? . . . . . . 2 Configuring server platforms . . . . . . . . 33
What is IBM StoredIQ Data Workbench? . . . . . 3 Windows Share (CIFS). . . . . . . . . . 34
Supporting IBM StoredIQ Data Workbench . . . 4 NFS . . . . . . . . . . . . . . . . 34
What is IBM StoredIQ eDiscovery? . . . . . . . 4 Exchange servers . . . . . . . . . . . 34
What is IBM StoredIQ Data Script? . . . . . . . 5 Enabling integrated Windows authentication
What Is IBM StoredIQ Policy Manager. . . . . . 5 on Exchange servers . . . . . . . . . 35
Improving performance for IIS 6.0 and
Understanding the user interface . . . . 7 Exchange 2003 . . . . . . . . . . . 35
Navigating within IBM StoredIQ Platform . . . . 7 SharePoint . . . . . . . . . . . . . . 35
Administration . . . . . . . . . . . . 7 Privileges required by user account . . . . 36
Folders . . . . . . . . . . . . . . . 9 Alternate-access mappings . . . . . . . 36
Audit . . . . . . . . . . . . . . . . 9 Documentum . . . . . . . . . . . . . 37
Web interface icons and buttons . . . . . . . . 9 Enterprise Vault . . . . . . . . . . . . 37
Discovery Accelerator . . . . . . . . . . 38
Discovery Accelerator permissions. . . . . 38
Performing IBM StoredIQ Platform Configuring security settings for Enterprise
administration . . . . . . . . . . . 13 Vault servers . . . . . . . . . . . . 39
Checking IBM StoredIQ Platform's status . . . . 13 Configuring Enterprise Vault . . . . . . 39
Restarting and rebooting IBM StoredIQ Platform . . 15 NewsGator . . . . . . . . . . . . . 40
Configuring retention servers . . . . . . . . 40
Performing system configurations . . . 17 Creating Centera pools . . . . . . . . . 40
Configuring IBM StoredIQ Platform . . . . . . 17 Configuring the Dell DX Object Storage Platform 41
Configuring system configuration options . . . . 18 Configuring Enterprise Vault (retention volumes) 42
Configuring DA Gateway settings . . . . . . 18 Enabling remote DCOM . . . . . . . . 42
Configuring network settings . . . . . . . 18 Allowing DCOM traffic through the Windows
Configuring mail settings . . . . . . . . . 20 firewall . . . . . . . . . . . . . . 43
Configuring SNMP settings . . . . . . . . 20 Configuring IBM Information Archive retention
Configuring notifications from IBM StoredIQ servers . . . . . . . . . . . . . . . 43
Platform . . . . . . . . . . . . . . 21 NetApp SnapLock . . . . . . . . . . . 44
Setting the system time and date . . . . . . 22 Hitachi HCAP configuration requirements . . . 45
Setting system backup configurations. . . . . 22 Configuring FileNet . . . . . . . . . . 45
Managing users . . . . . . . . . . . . 23 Configuring Chatter messages . . . . . . . 45
Logging in and out of the system . . . . . 23 Creating volumes . . . . . . . . . . . . 46
Changing the administrative account . . . . 23 Creating primary volumes . . . . . . . . 46
Creating users . . . . . . . . . . . 23 Configuring Exchange 2007 Client Access
Editing users . . . . . . . . . . . . 24 Server support . . . . . . . . . . . 53
Deleting a user's account . . . . . . . . 24 Adding Domino as a primary volume . . . 53
Locking a user's account . . . . . . . . 24 Special note: adding SharePoint volumes . . 54
Unlocking a user's account . . . . . . . 24 Configuring primary volumes using
Importing cncrypted NSF files from Lotus Notes 25 Enterprise Vault . . . . . . . . . . . 55
Configuring application configuration options . . . 25 About Editing Volume Definitions . . . . . 55
Configuring harvester settings . . . . . . . 26 Creating retention volumes . . . . . . . . 55
Optical character recognition processing . . . 27 Adding a retention volume . . . . . . . 56
Configuring full-text index settings . . . . . 28 Configuring Enterprise Vault retention
Specifying data object types . . . . . . . . 30 volumes . . . . . . . . . . . . . 59

Copyright IBM Corp. 2001, 2013 iii


Creating discovery export volumes . . . . . 59 Moving a folder . . . . . . . . . . . . 86
Creating system volumes . . . . . . . . . 61 Renaming a folder . . . . . . . . . . . 86
Exporting and importing volume data . . . . 62 Copying items to different folders . . . . . . 86
Volume data export to a system volume . . . 62 Saving items into different folders . . . . . . 86
Volume data import to a system volume. . . 63 Filtering items within the folder view . . . . 87
Deleting volumes . . . . . . . . . . . . 64
Policy limitations for volume types . . . . . . 65 Using audits and logs . . . . . . . . 89
Understanding harvest audits . . . . . . . . 89
Harvesting data . . . . . . . . . . . 69 Viewing harvest audits . . . . . . . . . 91
Understanding Harvests . . . . . . . . . . 69 Downloading harvest list details . . . . . . 92
Harvesting properties and libraries . . . . . . 70 Understanding import audits . . . . . . . . 92
Performing a lightweight harvest . . . . . . . 70 Viewing volume import audit details . . . . . 93
Lightweight harvest parameters . . . . . . 70 Understanding event logs . . . . . . . . . 93
Determining volume configuration settings . . 70 Working with event logs . . . . . . . . . 93
Determining harvester configuration settings 71 Viewing event logs . . . . . . . . . . 93
Determining full-text settings . . . . . . 72 Subscribing to an event . . . . . . . . 94
Determining hash settings . . . . . . . 72 Clearing the current event log . . . . . . 94
Downloading an event log . . . . . . . 94
Configuring jobs . . . . . . . . . . 73 Understanding policy audits. . . . . . . . . 94
Types of IBM StoredIQ Platform jobs . . . . . . 73 Viewing policy audit details . . . . . . . . 96
Working with jobs . . . . . . . . . . . . 74 Viewing a policy audit by name . . . . . 97
Creating a job . . . . . . . . . . . . 74 Viewing a policy audit by volume . . . . . 97
Creating a job to discover retention volumes . . 75 Viewing a policy audit by time . . . . . . 97
Editing a job . . . . . . . . . . . . . 75 Viewing a policy audit by discovery export. . 97
Starting a job . . . . . . . . . . . . . 76 Understanding the search audit feature . . . . 98
Saving a job . . . . . . . . . . . . . 76 Saving results from an audit. . . . . . . . 99
Running a predefined job. . . . . . . . . 77 Policy audit messages . . . . . . . . . . 99
Deleting a job . . . . . . . . . . . . 77
Monitoring processing . . . . . . . . . . . 77 Appendix A. Supported file types . . . 101
Deleting a volume cache . . . . . . . . . 79 Supported file types by name . . . . . . . . 101
Determining if a harvest is stuck . . . . . . 79 Supported file types by category . . . . . . . 111
SharePoint attributes . . . . . . . . . . . 126
Utilizing desktop collection . . . . . . 81
Desktop collection processes . . . . . . . . . 81 Appendix B. Supported server
IBM Desktop Data Collector client installation . . 81 platforms and protocols . . . . . . . 129
IBM Desktop Data Collector installation methods 82
Installing the IBM Desktop Data Collector in Appendix C. Event log messages . . . 133
stealth mode . . . . . . . . . . . . . 83
ERROR event log messages. . . . . . . . . 133
Configuring IBM Desktop Data Collector collection 83
INFO event log messages . . . . . . . . . 143
Using the delete policy with the IBM Desktop
WARN event log messages . . . . . . . . . 152
Data Collector: special notes . . . . . . . . . 83

Notices . . . . . . . . . . . . . . 161
Using Folders. . . . . . . . . . . . 85
Understanding folder types . . . . . . . . . 85
Creating a folder . . . . . . . . . . . 85 Index . . . . . . . . . . . . . . . 165
Deleting a folder . . . . . . . . . . . 85

iv Administration Guide
How to send your comments
Your feedback helps IBM to provide quality information. Please share any
comments that you have about this information or other documentation that IBM
Software Development ships with its products.

You can use any of the following methods to provide comments:


v Add comments by using the Comments pane at the bottom of every page in the
information center.
v Send your comments by clicking the Feedback link at the bottom of any topic in
the information center.
v Send your comments by using the online readers' comment form at
http://www.ibm.com/software/data/rcf/.
v Send your comments by e-mail to comments@us.ibm.com. Include the name of
the product, the version number of the product, and the name and publication
number of the information (if applicable). If you are commenting on specific
text, please include the location of the text (for example, a title, a table number,
or a page number).

Copyright IBM Corp. 2001, 2013 v


vi Administration Guide
Contacting IBM
To contact IBM customer service in the United States or Canada, call
1-800-IBM-SERV (1-800-426-7378).

To learn about available service options, call one of the following numbers:
v In the United States: 1-888-426-4343
v In Canada: 1-800-465-9600

For more information about how to contact IBM, see the Contact IBM web site at
http://www.ibm.com/contact/us/.

Copyright IBM Corp. 2001, 2013 vii


viii Administration Guide
Contacting IBM StoredIQ customer support
For IBM StoredIQ technical support or to learn about available service options,
contact IBM StoredIQ customer support at this phone number:
v 1-866-227-2068

To e-mail IBM StoredIQ customer support, use this email address:


v storediqsupport@us.ibm.com

For more information about how to contact IBM, see the Contact IBM web site at
http://www.ibm.com/contact/us/.

Copyright IBM Corp. 2001, 2013 ix


x Administration Guide
Understanding IBM StoredIQ Platform distributed architecture
Visualizing IBM StoredIQ Platform distributed architecture, its functions, and how
the different components work together is key to understanding how the system
works.

v IBM StoredIQ Platform Data ServerIBM StoredIQ Platform Data Server is an


invaluable tool in helping you to truly understand your company's data
landscape. Using IBM StoredIQ Platform Data Server, all unstructured data in a
company's enterprise network is indexed. Indexing of unstructured data allows
you to gain information about it. By indexing this data, you gain information
about unstructured data such as file size, file data types, file owners.
v IBM StoredIQ AdministratorIBM StoredIQ Administrator monitors and
manages the IBM StoredIQ Administrator distributed infrastructure at a
customer site. IBM StoredIQ Administrator sits between the IBM StoredIQ
Platform interface and the applications and facilitates the transfer and
communication of information. The IBM StoredIQ Platform Administrator
understands and manages IBM StoredIQ Platform concepts such as volumes,
indexes, harvests, and configurations. At the same time, the Administrator is
also concerned with application concerns such as infoset lifecycle and creation,
data server and volume configuration, and action and target set management. To
this end, the Administrator is divided into two sections, platform and
application, so Administrators have a clear place to go in order to accomplish a
task.
v IBM StoredIQ Data WorkbenchIBM StoredIQ Data Workbench is the tool that
allows you to visualize this indexed data, helping you to identify potential
red-flag issues, to know how much data you have on different types of servers,
and to alert people about potentially interesting or useful data. IBM StoredIQ
Data Workbench helps you to ensure that your company's data is an asset, not a
liability.
v IBM StoredIQ eDiscoveryIBM StoredIQ eDiscovery is integral to the
eDiscovery process, helping you to discover, identify, refine, preserve, and collect
data that is relevant to legal matters. By using IBM StoredIQ eDiscovery, you can
perform eDiscovery work more efficiently while ensuring that you've captured
the proper data.
v IBM StoredIQ Data ScriptIBM StoredIQ Data Script enables automated
execution within IBM StoredIQ Platform, allowing you to script, automate, and
monitor processes that would normally be a manual process run within IBM
StoredIQ Data Workbench. IBM StoredIQ Data Script focuses on repeatable,
understood, and approved processes for the purposes of culling and refining
data in an approved manner.

Copyright IBM Corp. 2001, 2013 1


v IBM StoredIQ Policy ManagerIBM StoredIQ Policy Manager allows you to
act on your data in an automatic fashion at scale, executing policies that affect
data objects without requiring review.

What is IBM StoredIQ Administrator?


IBM StoredIQ Administrator helps you to manage global assets common to the
distributed infrastructure behind IBM StoredIQ Platform applications.

IBM StoredIQ Administrator provides at-a-glance understanding of the different


issues that can crop up in the IBM StoredIQ Platform environment. These views
are unique to the IBM StoredIQ Administrator application as they provide an
overview of how the system is running, allow access to various pieces of
information that are being shared across applications or allow for the management
of resources in a centralized manner.

The Administrator is the person responsible for managing the IBM StoredIQ
Platform. This individual has strong understanding of data sources, indexes, data
servers, jobs, infosets, and actions. This list provides an overview as to how IBM
StoredIQ Administrator works:
v Viewing data servers and volumesUsing IBM StoredIQ Administrator, the
Administrator can identify what data servers are deployed, their location, what
data is being managed, and the status of each data server in the system. Volume
management is a central component of IBM StoredIQ Platform. IBM StoredIQ
Administrator also allows the Administrator to see what volumes are currently
under management, which data server is responsible for that volume, the state
of the volume after indexing, and the amount and size of information that is
contained by each volume. Administrators can also add volumes to data servers
through this interface.
v Creating system infosetsSystem infosets that use only specific indexed
volumes can be created and managed within IBM StoredIQ Administrator.
Although infosets are a core component of IBM StoredIQ Data Workbench,
system infosets are created as a shortcut for users in IBM StoredIQ
Administrator.
v Managing usersThe user management area allows Administrators to create
users and manage users' access to the various IBM StoredIQ applications.
v Configuring and managing actionsAn action is any process that is taken
upon the data that is represented by the indexes. Actions are run by data servers
on indexed data objects. Any errors or warnings that are generated as a result of
an action are recorded as exceptions in IBM StoredIQ Data Workbench.

Note: Actions can be created within IBM StoredIQ Administrator and then made
available to otherIBM StoredIQ applications such as IBM StoredIQ Data
Workbench.
v Managing target setsProvides an interface that allows the user to set the
wanted targets for specific actions that require a destination volume for their
actions.
v Managing conceptsThis feature allows you to relate business concepts to
indexed data.
Related concepts:
Viewing data servers and adding volumes
Creating system infosets
Managing users

2 Administration Guide
Configuring and managing actions
Managing target sets
Managing concepts

What is IBM StoredIQ Data Workbench?


This section provides a high-level explanation of what IBM StoredIQ Data
Workbench is and its potential uses.

You have a big data problem: SharePoint sites, wikis, email, files, blogs, discussion
threads, and attachments. Your company's information is its most valuableand
potentially most dangerousasset.

Big data is a pervasive problem, not a one-time occurrence. Most companies


willingly admit that big data is problematic, but quite often, they don't even know
what problems they have. This is one instance where out of sight definitely does
not mean out of mind. Big data is all about the unknown, but the unknown cannot
be off limits. Plausible deniability is a risky bet. IBM StoredIQ Data Workbench can
help you learn about your data, to help you make educated decisions with your
most valuable asset, and can help you turn your company's most dangerous risk
into its most valuable asset.

IBM StoredIQ Data Workbench is a data visualization and management tool that
helps you to actively manage your company's data. IBM StoredIQ Data Workbench
helps you to determine how much data you have, where it resides, who owns it,
when it was last utilized, and so on. Then, once you have a clear understanding of
your company's data landscape, IBM StoredIQ Data Workbench helps you take
control of and make informed decisions about your data and act on that
knowledge by copying, copying to retention, or performing a discovery export.
That once-risky data is now a legitimate company asset.

Here are just some of the ways you could use IBM StoredIQ Data Workbench.
v Let's say that you need to find all company email sent from or received by
Eileen Sideways (esideways@thecompany.com). You can use IBM StoredIQ Data
Workbench to find all email and then copy that data to a predefined repository.
You could also use IBM StoredIQ Data Workbench to find all of the
esideways@thecompany.com email that occurred between specific dates and then
make that email available for review.
v As an administrator, you'd like to rid your networks and storage of unused data.
You can use IBM StoredIQ Data Workbench to find all files that have not been
modified in more than five years.
v You would like to find all image files created in 2007. Not only can IBM
StoredIQ Data Workbench find all image files created in 2007, it can also tell you
how much space they occupy on your network.
v A user needs to understand how data regarding Windows is being retained.
Using IBM StoredIQ Data Workbench, you can provide that user with a visual
overview of the number of objects retained and a breakdown of files per data
source. Additionally, you can apply overlays to show the user if those files
contain forbidden information such as credit-card numbers or Social Security
numbers.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1

Understanding IBM StoredIQ Platform distributed architecture 3


Supporting IBM StoredIQ Data Workbench
IBM StoredIQ Platform is an invaluable tool in helping you to truly understand
your company's data landscape. Using IBM StoredIQ Platform data server, all
unstructured data in a company's enterprise network is indexed. Indexing of
unstructured data allows you to gain information about it. By indexing this data,
you gain information about unstructured data such as file size, file data types, file
owners, and so on.

IBM StoredIQ Data Workbench is the tool that allows you to visualize this indexed
data, helping you to identify potential red-flag issues, to know how much data you
have on different servers, and to alert people about potentially interesting or useful
data. IBM StoredIQ Data Workbench helps you to ensure that your company's data
is an asset, not a liability.

What is IBM StoredIQ eDiscovery?


This section provides a high-level explanation of what IBM StoredIQ eDiscovery is,
by whom it will be used, prerequistes, and its potential uses.

IBM StoredIQ eDiscovery is an end-user application that helps legal users during
the initial phases of the eDiscovery process. By helping you to capture needed
electronic data, you can use IBM StoredIQ eDiscovery to communicate easily with
other users regarding the status and review process of work being done. IBM
StoredIQ eDiscovery does not drive the eDiscovery process, but instead helps legal
users to control and communicate those processes more effectively.

IBM StoredIQ eDiscovery helps to address the left side of the Electronic Discovery
Reference Model, as shown below.

Here are just some of the ways you could use IBM StoredIQ eDiscovery.
v Let's say that you need to find all electronic information regarding an upcoming
personal-injury lawsuit. You can use IBM StoredIQ eDiscovery to create a matter
for the suit, and then create boxes to contain email and reports regarding the
case.
v As a legal user, you'd like to see the status of all currently active matters. You
can use the Matter Dashboard to visualize the different matters' states.

4 Administration Guide
v As a data expert, you know that you need to respond to users in a timely
fashion. Using IBM StoredIQ eDiscovery, people can email you directly from the
application regarding a matter's questions.

Proactive eDiscovery consists of identifying and cataloging data or types of data


that may eventually be responsive to legal matters, collecting and retaining the
data, then producing that data in a way that matches accepted/required legal
practices.

Within IBM StoredIQ eDiscovery, the user is a non-technical end user who needs to
obtain certain pieces of data in order to complete a legal process. The IBM
StoredIQ eDiscovery user has the knowledge of the information or parameters for
what kinds of data they want, but does not always have knowledge of how to
obtain that data.

Before using IBM StoredIQ eDiscovery, ensure that prerequisites are met:
v The IBM StoredIQ Platform must be deployed, configured, and ready for use.
v IBM StoredIQ eDiscovery is dependent upon other IBM StoredIQ applications
such as IBM StoredIQ Data Workbench and IBM StoredIQ Administrator.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1

What is IBM StoredIQ Data Script?


IBM StoredIQ Data Script enables automated execution, allowing you to script,
automate, and monitor processes that would normally be a manual process run
within IBM StoredIQ Data Workbench. IBM StoredIQ Data Script focuses on
repeatable, understood, and approved processes for the purposes of culling and
refining data in an approved manner.

Through the IBM StoredIQ Data Script interface, you can monitor each of the steps
as they are executed and then view any defined outputs, such as reports, infosets,
or exports that are generated as a result. By running IBM StoredIQ Data
Workbench workflows, the user can reapply processes that have been reviewed
and approved to apply to a wide variety of different data problems.
Related concepts:
Understanding IBM StoredIQ Platform distributed architecture on page 1

What Is IBM StoredIQ Policy Manager


IBM StoredIQ Policy Manager enables policy execution at scale.

IBM StoredIQ Policy Manager allows users to run mature policies and processes at
scale across a wider range of data. What makes IBM StoredIQ Policy Manager so
powerful is that it lets users define and execute systemwide policies, focusing on
the execution of the process rather than understanding or reviewing affected data
objects. Additionally, IBM StoredIQ Policy Manager's reports lets you record what
actions were performed, when they were performed, and what data was affected
by the policy's execution. IBM StoredIQ Policy Manager is an extremely powerful
tool in managing your data effectively.

Understanding IBM StoredIQ Platform distributed architecture 5


6 Administration Guide
Understanding the user interface
This section provides an introduction to the IBM StoredIQ Platform Web interface.
It outlines the features within each tab and provides references to sections where
you can find additional information on each topic.

Navigating within IBM StoredIQ Platform


Primary tabs and subtabs found within the user interface allow you to access
data-server functionality.

Primary Tabs

IBM StoredIQ Platform users perform most tasks using the Web interface. The
menu bar at the top of the interface contains three primary tabs that are described
in this table.

This table describes the primary tabs.

Name Description

Administration Allows Administrators to perform various


configurations on these subtabs: Dashboard,
Data Sources, and Configuration.
Folders Create folders and jobs; run jobs.

Audit Examine a comprehensive history of all


harvests, executed policies, imports, and
event logs.

Related concepts:
Web interface icons and buttons on page 9

Administration
This topic provides descriptions of the Administration tab as well as its subtabs:
Dashboard, Data Sources, and Configuration.

Dashboard

The Dashboard subtab provides an overview of the systems current, ongoing, and
previous processes as well as its current status. This table describes
Administrator-level features and descriptions.

Page refresh Choose from 30-, 60-, or 90-second intervals to refresh the page.

Todays job schedule View a list of jobs scheduled for that day with links to the jobs
summary.

Copyright IBM Corp. 2001, 2013 7


System summary View a summary of system details, including system data objects,
contained data objects, volumes, and the dates of the last
completed harvest.

Jobs in progress View details of each job step as it is running, including estimated
time to completion, average speed, total system and contained
objects encountered, harvest exceptions, and binary processing
information.

Harvest statistics Review the performance over the last hour for all harvests.

Event log Review the last 500 events or download the entire event log for
the current date or previous dates.

Appliance status Provides a status view of the appliance. Reboot or restart the
appliance through the about appliance link. View cache details for
volumes and discovery exports.

Data Sources

The Data sources subtab is where Administrators define servers and volumes.
These can be places that are indexed or copied to. A variety of server types and
volumes can be configured for use in managing data. Administrators can add
Enterprise Vault sites, Centera pools, Dell DX Storage Clusters, Celerra, NetApp,
and FileNet servers through the Specify servers area. Volumes are configured and
imported in the Specify volumes section.

Configuration

The Administrator configures system and application settings for IBM StoredIQ
Platform through the Configuration subtab.

Settings Description

System
DA Gateway settings Configure the DA Gateway host or IP address.

Network settings Configure the private and public network interfaces.

Mail server settings Configure what mail server to use and how often to
send email.

SNMP settings Configure Simple Network Management Protocol


(SNMP) servers and communities.

System time and date Set the system time and date on the appliance.

Manage users Add, remove and edit users.

Lotus Notes user administration Add a Lotus Notes User.

Application

8 Administration Guide
Settings Description

Harvester settings Set basic parameters and limits, data object extensions
and directories to skip, and reasons to run binary
processing.

Full-text settings Set full-text search limits for length of word and
numbers and edit stop words. See Configuring full-text
index settings on page 28.

Data object types Set the object types that appear in the disk use by data
object type report.

Audit settings Configure how long and how many audits will be kept.

Hash settings Configure whether to compute a hash when harvesting


and which kind of hash.

Desktop settings Configure the desktop collection service.

Related concepts:
Creating volumes and data sources on page 33
Navigating within IBM StoredIQ Platform on page 7

Folders
The Folders tab is where users create and manage application objects.
Related concepts:
Navigating within IBM StoredIQ Platform on page 7
Using Folders on page 85

Audit
The audit feature allows you to review all actions taken using the data server.

Administrators can review harvests and examine the results of actions.


Related concepts:
Using audits and logs on page 89

Web interface icons and buttons


This section describe the icons and buttons used throughout IBM StoredIQ
Platform Web interface.

Understanding the user interface 9


IBM StoredIQ Platform icons and buttons

These icons and buttons are seen within the IBM StoredIQ Platform dashboard.

Icon or Button Name Description

User The User account icon accesses your user


account account, provides information regarding
version and system times, and logs you
out of the system. For more information,
see Logging In and Out of the System.

Inbox The inbox link allows you to access PDF


audit reports.

Help Clicking the Help icon loads IBM


StoredIQ technical documentation in a
separate browser window. By default, the
technical documentation is loaded as
HTML help.

Folders icons and buttons

The table describes the different icons and buttons seen within the Folders tab.

Icon or Button Name Description

New button The New button enables you


to add jobs and folders.

Action button The Action button allows


you to take action on
Workspace objects, including
the ability to move, and
delete jobs and folders.
Jobs tasks such as harvesting
Job are either a step or a series
of steps. For more
information, see Types of
IBM StoredIQ Platform Jobs.

Folder Folders are a container object


that can be accessed and
used by Administrators. For
more information, see Using
Folders.

10 Administration Guide
Icon or Button Name Description

Folder Up Folders are a container object


that can be accessed and
used by Administrators. By
default, you view the
contents of the Workspace
folder; however, by clicking
this icon, you move to the
parent folder in the structure.
For more information, see
Using Folders.

Audit icons and buttons

There are no specialized icons or buttons used on the Audit tab.


Related concepts:
Understanding the user interface on page 7

Understanding the user interface 11


12 Administration Guide
Performing IBM StoredIQ Platform administration
This section provides procedural information regarding IBM StoredIQ Platform
administration.

Checking IBM StoredIQ Platform's status


This topic provides procedural information regarding how to check IBM StoredIQ
Platform's status.

To check the status of IBM StoredIQ Platform:


1. Go to Administration > Dashboard > Appliance status.
2. Click About appliance to open the Appliance details page. The Appliance
details page shows the following information:
v Node
v Harvester processes
v Status
v Software version
v View details link
3. Click the View details link for the controller. This table defines appliance
details data and describes the data provided for the node.

Option Description
View appliance details Shows software version and details of
harvester processes running on the controller
for the appliance.
Application services Shows a list of all services and current
status, including:
v Servicethe name of each service on the
appliance component
v PIDthe process ID associated with each
service
v Current memory (MB)the memory
being used by each service
v Total memory (MB)total memory being
used by each service and all child services
v CPU percentagethe percentage of CPU
usage for each service. This value is zero
when a service is idle.
v Statusthe current status of each service.
Status messages include Running,
Stopped, Error, Initializing, Unknown.

Copyright IBM Corp. 2001, 2013 13


Option Description
System services
Shows a list of basic system information
details and memory usage statistics. System
information includes:
v System timecurrent time on the
appliance component
v GMT Offsetthe amount of variance
between the system time and GMT.
v Time upthe period of time the appliance
component has been running since the
last reboot, in days, hours, minutes and
seconds
v System processesthe total number of
processes running on the node
v Number of CPUsthe number of CPUs in
use on the component
v Load Average (1 Minute)the average
load for system processes during a
one-minute interval
v Load Average (5 Minutes)the average
load for system processes during a
five-minute interval
v Load Average (10 Minutes)the average
load for system processes during a
ten-minute interval

Memory details include:


v Totaltotal physical memory on the
appliance component
v In usehow much physical memory is in
use
v Freehow much physical memory is free
v Cachedamount of memory allocated to
disk cache
v Bufferedthe amount of physical memory
used for file buffers
v Swap totalthe total amount of swap
space available (in use plus free)
v System services Swap in usethe total
amount of swap space being used
v Swap freethe total amount of swap
space free
v Database Connections
v Configured
v Active
v Idle
v Network interfaces
v Up or down status for each interface

14 Administration Guide
Option Description
Storage
Storage information for a controller includes:
v Volume
v Total Space
v Used space
v Percentage
Controller and compute node status
Indicator lights show component status:
v GreenRunning
v YellowThe node is functional but is in
the process of rebuilding; performance
may be degraded during this time. Note:
The rebuild will progress faster if the
system is not in use.
v Rednot running

Expand the node to obtain details of the


appliance component by clicking the image.

Related concepts:
Performing IBM StoredIQ Platform administration on page 13

Restarting and rebooting IBM StoredIQ Platform


This topic provides procedural information regarding how to restart and reboot the
IBM StoredIQ Platform.

Note: The Web application is temporarily unavailable if you restart or reboot.

To restart services or reboot IBM StoredIQ Platform:


1. On Administration > Dashboard > Appliance status. On the Appliance status
page, you have two options:
v Click the Controller link.
v Click About Appliance.
2. The Restart services and Reboot buttons appear at the bottom of the window.
These buttons are available on the View details page and on each of the tabs.
Click either of these options:
v Restart services to restart all system services running on the node.
v Reboot to reboot the IBM StoredIQ Platform components.
The Web application is temporarily unavailable if you restart or reboot.
Related concepts:
Performing IBM StoredIQ Platform administration on page 13

Performing IBM StoredIQ Platform administration 15


16 Administration Guide
Performing system configurations
This section provides procedural information regarding system configurations.
Related reference:
Appendix B, Supported server platforms and protocols, on page 129

Configuring IBM StoredIQ Platform


This topic lists what an Administrator can configure from the Application and
Network areas when configuring IBM StoredIQ Platform.

The Configuration subtab (Administration > Configuration) is divided into


System and Application sections.

System Section

In the System section, an Administrator can:


v Configure the DA gateway. For more information, see Configuring DA Gateway
settings.
v View and modify network settings, including hostname, IP address, NIS domain
membership, and use. See Configuring network settings.
v View and modify settings to enable the generation of email notification
messages. See Configuring mail settings.
v Configure SNMP servers and communities. See Configuring SNMP settings.
v Manage notifications for system and application events. See Configuring
notifications from IBM StoredIQ Platform on page 21.
v View and modify date and time settings for the IBM StoredIQ Platform. See
Setting the system time and date.
v Setting backup configurations. See Setting system backup configurations on
page 22.
v Manage users. See Managing Users.
v Upload Lotus Notes user ids so that encrypted NSF files can be imported into
IBM StoredIQ Platform. See Importing Encrypted NSF Files from Lotus Notes.

Application Section

In the Application section, an Administrator can:


v Specify directory patterns to exclude during harvests. See Configuring
Harvester Settings.
v Specify options for full-text indexing. See Configuring Full-text Index Settings.
v View, add, and edit known data object types. See Specifying Data Object Types.
v View and edit settings for policy audit expiration and removal. See Configuring
Audit Settings.
v Specify options for computing hash settings when harvesting. See Configuring
Hash Settings.
v Specify options to configure the desktop collection service. See Configuring
desktop settings on page 31.

Copyright IBM Corp. 2001, 2013 17


Configuring system configuration options
This topic lists all System configuration options.
Related concepts:
Configuring application configuration options on page 25
Configuring IBM StoredIQ Platform on page 17
Performing system configurations on page 17

Configuring DA Gateway settings


This topic provides procedural information regarding how to configure the DA
Gateway.

To configure DA Gateway settings:


1. Go to Administration > Configuration > System > DA Gateway Settings.
2. In the Host text box, enter the gateway host. For example, enter
mgmt.example.com or 192.168.10.10.
3. Click OK. Services must then be restarted.
Related concepts:
Configuring system configuration options
Related tasks:
Restarting and rebooting IBM StoredIQ Platform on page 15

Configuring network settings


This topic provides procedural information regarding how to configure various
network settings.
1. Go to Administration > Configuration > System > Network settings.
2. Click Controller Settings. Set or modify the following options.

Option Description
Primary Network Interface v IP typeSet to stic or dynamic. If set to
dynamic, the IP address, Netmask, and
Default Gateway fields are disabled.
v IP addressEnter the IP address if
specifying the address manually.
v NetmaskEnter the network mask of the
IP address.
v Default gatewayEnter the IP address of
the default gateway.
v HostnameEnter the fully qualified
domain name assigned to the appliance.
v Ethernet speedSelect the Ethernet speed.
v Available portsIndicate the available
ports.
v Separate network for file/email
serversSpecify the additional subnet for
accessing file/email servers. Select this
check box if you are using the Web
application from one subnet and
harvesting from another subnet.

18 Administration Guide
A system restart is required for any primary network interface changes to take
effect. See Restarting and Rebooting the Appliance.

Option Description
DNS Settings v DNS search orderEnter the DNS search
order for multiple DNS servers.
v Nameserver 1Set the IP address of the
primary DNS server for name resolution
(required).
v Nameserver 2 Set the IP address of the
secondary DNS server for name resolution
(optional).
v Nameserver 3Set the IP address of the
tertiary DNS server for name resolution
(optional).

DNS settings take effect after they have been saved. Changes to the server's IP
address take effect immediately. Because the server has a new IP address, you
must reflect this new address in the browser address line before continuing.
3. Click OK.
4. Click Server name resolution. Set the following for the data server:
v Windows Share (CIFS) file server name resolutionThese settings take
effect upon saving.

Option Description
Windows Share File Server Name v LMHOSTSEnter the IP hostname
Resolution format.
v WINS ServerEnter the name of the
WINS server.

v NIS (for NFS)These settings take effect upon saving.

Option Description
NIS (for NFS) v Use NISSelect this box to enable NIS to
perform UID/GID to friendly name
resolution in an NFS environment.
v NIS DomainSpecify the NIS domain.
v Broadcast for server on local
networkSelect this box if the NIS
domain server is located on the local
network and can be discovered by
broadcasting. This option does not work if
the NIS domain server is on another
subnet.
v Specify NIS serverIf not using
broadcast, specify the IP address of the
NIS domain server here.

v Active DirectoryThese settings take effect upon saving.

Option Description
Active Directory Active Directory serverEnter the name of
the Active Directory server.

v Doc broker settings (for Documentum)

Performing system configurations 19


Option Description
Doc Broker Settings (for Documentum) v Enter Host name for doc broker.
v Documentum global registry:
Registry name
User
Password

5. Click OK.
Related concepts:
Configuring system configuration options on page 18

Configuring mail settings


This topic provides procedural information regarding how to configure mail
settings.

To configure mail settings:


1. Go to Administration > Configuration > System > Mail Server settings.
2. In Mail server, enter the name of the SMTP mail server.
3. In From address, enter a valid sender address. Some mail servers reject email if
the sender is invalid. A sender address also simplifies the process of filtering
email notifications based on the sender's email.
4. Click OK to save changes.
Related concepts:
Configuring system configuration options on page 18

Configuring SNMP settings


You can both configure the system to make Object Identifier (OID) values available
to Simple Network Management Protocol (SNMP) client applications, as well as to
receive status information or messages about system events in a designated trap.
For information about environmental circumstances monitored by IBM StoredIQ
Platform, see the table below.

To configure SNMP settings:


1. Go to Administration > Configuration > System > SNMP settings.
2. To make OID values available to SNMP client applications, in the Appliance
Public MIB area:
v Select the Enabled check box to make the MIB available (that is, to open port
161 on the controller).
v In the Community field, enter the community string that the SNMP clients
will use to connect to the SNMP server.
v To view the MIB, click Download Appliance MIB. This document provides
the MIB definition, which can be provided to an SNMP client application.
3. To capture messages containing status information in the Trap destination area:
v In the Host field, enter the common name or IP address for the host.
v In the Port field, enter the port number. Port number 162 is the default.
v In the Community field, enter the SNMP community name.
4. To modify the frequency of notifications, complete these fields in the
Environmental trap delivery area:
v Send environmental traps only every __ minutes.
20 Administration Guide
v Send environmental traps again after __ minutes.
5. Click OK. Environmental traps monitored by IBM StoredIQ Platform are
described in the table below.

Option Description
siqConsoleLogLineTrap This is a straight conversion of a console log
line into a trap. It uses these parameters:
messageSource, messageID, severity,
messageText.
siqRaidControllerTrap Sent when the RAID controller status is
anything but normal. Refer to the MIB for
status code information. It uses this
parameter: nodeNum.
siqRaidDiskTrap Sent when any attached raid disk's status is
anything but OK. It uses this parameter:
nodeNum.
siqBbuTrap Battery Backup Unit (BBU) error on the
RAID controller detected. It uses this
parameter: nodeNum.
siqCacheBitTrap Caching indicator for RAID array is off. It
uses this parameter: nodeNum.
siqNetworkTrap Network interface is not UP when it should
be. It uses this parameter: nodeNum.
siqDbConnTrap Delivered when the active Postgres
connection percentage exceeds an acceptable
threshold. It uses this parameter: nodeNum.
siqFreeMemTrap Delivered when available memory falls too
low. It uses this parameter: nodeNum.
siqSwapUseTrap Sent when swap use exceeds an acceptable
threshold. Often indicates memory leakage.
It uses this parameter: nodeNum.
siqCpuTrap Sent when CPU load averages are too high.
It uses this parameter: nodeNum.
siqTzMismatchTrap Sent when a nodes time zone offset does
not match that of the controller. It uses this
parameter: nodeNum.

Related concepts:
Configuring system configuration options on page 18

Configuring notifications from IBM StoredIQ Platform


You can configure the system to notify you using email or SNMP when certain
events occur.

For a list of events that can be configured, see Event Logs.

To configure notifications from IBM StoredIQ Platform:


1. Go to Administration > Configuration > System > Manage notifications.
2. Click Create a notification.
3. In the Event number: field, search for events by clicking Browse or by typing
the event number or sample message into the field.
4. Select the event level by clicking the ERROR, WARN, or INFO link.

Performing system configurations 21


5. Scroll through the list, and select each event by clicking on it. The selected
events appear in the create notification window. To delete an event, click the
delete icon to the right of the event.
6. In the Destination: field, select the method of notification: SNMP, or Email
address, or both. If you choose email address, enter one or more addresses in
the Email address field. If you choose SNMP, the messages will be sent to the
trap host identified in the SNMP settings window, with a trap type of
siqConsoleLogLineTrap.
7. Click OK.
8. To delete an item from the Manage notifications window, select the check box
next to the event, and then click Delete. You can also request a notification for
a specific event from the dashboards event log. Click the Subscribe link next
to any error message and a prepopulated edit notification screen containing the
event is provided.
Related concepts:
Configuring system configuration options on page 18

Setting the system time and date


This topic provides procedural information regarding how to set the system's time
and date.

Note: A system restart is required for any changes made to the system time and
date. See Restarting and Rebooting the Appliance.

To set the system time and date:


1. Go to Administration > Configuration > System > System time and date.
2. Enter the current date and time.
3. Select the appropriate time zone for your location.
4. Enable Use NTP to set system time to use an NTP server to automatically set
the system date and time for the data server. If NTP is used to set the system
time, then the time and date fields set automatically. However, you do need to
specify the time zone.
5. Enter the name or IP address of the NTP server.
6. Click OK to save changes.
Related concepts:
Configuring system configuration options on page 18

Setting system backup configurations


To prepare for disaster recovery, you can back up the configuration of a data server
to a gateway. This will back up volume definitions, discovery export records, and
data-server settings. This will not back up infosets, data maps, or indexes.

The gateway must be manually configured to support this backup. For additional
information, configure the gateway in the util menu. This can done by contacting
your service representative.

To initiate the run-configuration back up:


1. Go to Administration > Configuration > System > Backup configuration.
2. Press Start backup. To better understand this button and its functionality,
please contact your service representative.
Related concepts:

22 Administration Guide
Configuring system configuration options on page 18
Related reference:
Contacting IBM StoredIQ customer support on page ix

Managing users
When managing users, you can log in and out of the system and perform various
account administration tasks.
Related concepts:
Configuring system configuration options on page 18

Logging in and out of the system


The system comes with a default administrative account named admin, the default
password for which is admin. For security purposes, IBM StoredIQ Platform
recommends that this password be changed as soon as possible.

To log into or log out of the Web application:


1. Open a browser and in the address line, enter the URL for your system.
2. On the Login page, enter admin for the email address and admin for the
password the first time you log in.
3. Click Log In to enter the system.

Note: Database CompactorIf someone tries to log in while the appliance is


performing database maintenance, the Administrator can override the
maintenance procedure and use the system. For more information, see Types of
IBM StoredIQ Platform Jobs.

4. To log out of the application, click in the upper right-hand corner, and
then click the Log out link.
Related concepts:
Managing users

Changing the administrative account


To change the admin account:
1. On the Administration > Configuration tab, click Manage users.
2. From the list, select The Administrator account, and then select Change the
admin password.
3. Enter a new password, and then click OK to save the change.
Related concepts:
Administering Accounts
Managing users

Creating users
To create a user:
1. Go to Administration > Configuration > Manage users, and then click
Create new user.
2. In the First name field, enter the user's first name.
3. In the Last name field, enter the user's last name.
4. In the Email address field, enter the user's email address.
5. For Authentication, select Active Directory or Local.

Performing system configurations 23


In the Active Directory principal, enter the name of the Active Directory
6.
principal.
7. To receive Notification of reports by email, select either Yes or No.
8. Select the Administrator button to select the Role.
9. Select the View data objects in the viewer check box if you would like to see
data objects in the viewer.
10. Click OK to create the user.
Related concepts:
Managing users on page 23

Editing users
To edit a user:
1. Go to Administration > Configuration > Manage users, and then click the
name of the user that you would like to edit.
2. Click Edit User and edit the fields as needed.
3. Click OK to save your changes.
Related concepts:
Managing users on page 23

Deleting a user's account


To delete a user account:
1. Go to Administration > Configuration > Manage users, and then click the
name of the user that you would like to delete.
2. Click Delete, and then click OK.
Related concepts:
Administering Accounts
Managing users on page 23

Locking a user's account


To lock a user account:
1. On the Administration > Configuration page, click Manage users.
2. In the list, select the user name of the account you wish to lock.
3. Click Lock user account to disable the account. A padlock icon appears next to
the account name in the list, indicating that the account is now locked. The
user cannot log in to the account while it is locked.
Related concepts:
Administering Accounts
Managing users on page 23

Unlocking a user's account


To unlock a user account:
1. On the Administration > Configuration page, click Manage users.
2. In the list, select the user name of the account you wish to unlock.
3. Click Unlock account to re-enable the account.
Related concepts:
Administering Accounts

24 Administration Guide
Managing users on page 23

Importing cncrypted NSF files from Lotus Notes


IBM StoredIQ Platform can decrypt, import, and process encrypted NSF files from
IBM Lotus Domino v7. The feature works by comparing a list of user.id/key
pairs that have been imported into the system with the key values that lock each
encrypted container or email. When the correct match is found, the file is unlocked
using the key. Once the emails or containers have been unlocked, IBM StoredIQ
Platform analyzes and processes them in the usual fashion. This topic provides
procedural information regarding how to import encrypted NSF files from IBM
Lotus Notes.

These use cases are supported:


v Multiple unencrypted emails contained within a journaling database that has
been encrypted with a single user.id key
v Multiple unencrypted emails contained in an encrypted NSF file
v Multiple encrypted emails contained within an unencrypted NSF file
v Multiple encrypted emails, using the same or different user.id keys, contained in
an encrypted NSF file
v Encrypted emails from within a journaling database

To import encrypted NSF files from IBM Lotus Notes:


1. On the primary data server, go to Administration > Configuration > Lotus
Notes user administration.
2. Click Upload a Lotus user ID file.
a. In the dialog that appears, click Browse, and navigate to a user file.
b. In the Lotus Notes password: field, type the password that unlocks the
selected file.
c. (Optional) In the Description: field, enter a description for the file.
d. Click OK. Repeat until the keys for all encrypted items have been uploaded.
Note that once the list has been compiled, you can add new entries to it
later.
e. To delete an item from the list, from the Registered Lotus users screen,
select the check box next to a user, and then click Delete. In the
confirmation dialog that appears, click OK.

Note: After uploading user ids, you must restart services.


Related concepts:
Configuring system configuration options on page 18

Configuring application configuration options


This topic lists all Application configuration options.
Related concepts:
Configuring system configuration options on page 18
Configuring IBM StoredIQ Platform on page 17
Performing system configurations on page 17

Performing system configurations 25


Configuring harvester settings
This topic provides procedural information regarding how to configure harvester
settings.

To configure harvester settings:


1. Go to Administration > Configuration > Application > Harvester settings.
2. To configure Basic settings:
a. Harvester ProcessesSpecify the number of harvester processes according
to your solution.
b. Harvest miscellaneous email itemsSelect to harvest contacts, calendar
appointments, notes, and tasks from the Exchange server.
c. Harvest non-standard Exchange message classesSelect to harvest
message classes that do not represent standard Exchange email and
miscellaneous items.
d. Include extended characters in object namesSelect to allow extended
characters to be included in data object names during a harvest.
e. Determine whether data objects have NSRL digital signatureSelect to
check data objects for NSRL digital signatures.
f. Enable parallel grazingSelect to harvest volumes that have already been
harvested and are going to be reharvested. Parallelized grazing enables
harvests to begin where they left off when interrupted and to start at the
beginning if the harvest completes normally.
g. Enable OCR image processingSelect this option to control at a global
level whether or not Optical Character Recognition (OCR) processing is
attempted on image files. If you select this option, you must restart services.
See Restarting and Rebooting the Appliance.
3. Specify Skip Content processing.
a. Data object extensions to be skippedSpecify those file types you want
the harvest to ignore by adding data object extensions to be skipped.
4. To configure Locations to ignore, enter each directory that should be skipped.
Note that IBM StoredIQ Platform accepts only one entry per line and that
regular expressions can be used.
5. To configure Limits:
a. Maximum data object sizeSpecify the maximum data object size to be
processed during a harvest. During a harvest, iles exceeding the maximum
data object size will not be read. As a result, if full-text/content processing
is enabled for the volume, they are audited as skipped: Configured max.
object size. These objects still appear in the volume cluster along with all
file-system metadata. Since they were not read, the hash is a hash of the
file-path and size of the object, regardless of what the hash settings are for
the volume (full/partial/off).
b. Max entity values per entityFor any given entity type (date, city, address
and the like), the system records, per data object, the number of values set
in this field. The values do not need to be unique; for example, if the
maximum value is 1,000, and the harvester collects 1,000 instances of the
same date (8/15/2009) in a Word document, the system stops counting
dates. This setting applies to all user-defined expressions (keyword, regular
expression, scoped, and proximity) and all standard attributes.
c. Max entity values per data objectAcross all entity types, the total
(cumulative) number of values that will be collected from a data object

26 Administration Guide
during a harvest. A 0 in this field means unlimited. This setting applies to all
user-defined expressions (keyword, regular expression, scoped, and
proximity) and all standard attributes.
6. Configure Binary Processing.
a. Run binary processing when text processing failsSelect this option to
run binary processing. The system runs further processes against content
that failed in a harvesting. You can select options for when to perform this
extended processing and how to scan content. Binary processing does not
search image file types such as .GIF and .JPG for text extraction.
b. Failure reasons to begin binary processingSelect the check boxes of the
options that define when to perform extended processing. Binary processing
can enact when extracting text from a file fails, when the format of the file
is unknown to the system parameters, when the data object type is not
supported by the harvester scan, or when the data object format does not
contain actual text.
c. Data object extensionsSet binary processing to process all data files or
only files of entered extensions. To add extensions, enter one per line
without a period.
d. Text encodingSet options for what data to scan and extract when
performing binary processing. This extended processing can accept
extended characters and UTF-16 and UTF-32 encoded characters as text. The
system searches UTF-16 and UTF-32 by default.
e. MinimumsSet the minimum required number of located, consecutive
characters to begin processing for text extraction. For example, if you enter
4, the system begins text processing when four consecutive characters of a
particular select text encoding are found. This setting helps find and extract
helpful data from the binary processing, reducing the amount of false
positives.
7. Click OK. Changes to harvester settings do not take effect until the appliance
is rebooted or the application services are restarted.
Related concepts:
Configuring application configuration options on page 25

Optical character recognition processing


This topic provides conceptual information regarding optical character recognition
(OCR) processing.

How Optical Character Recognition Processing Works

OCR processing enables text extraction from graphic image files stored directly on
or inside archives stored on volumes where the Include content tagging and
full-text index option is selected. After content typing inside theIBM StoredIQ
Platform processing pipeline, enabling OCR processing will route the following
files types through an optical character recognition engine OCR to extract
recognizable text.
v Windows or OS/2 bitmap (BMP)
v Tag image bitmap file (TIFF) (TIFF)
v Bitmap (CompuServe) (GIF)
v Portable Network Graphics (PNG)

The text extracted from image files is processed through theIBM StoredIQ Platform
pipeline in the same manner as text extracted from other supported file types.

Performing system configurations 27


Policies that have a specific feature to write out extracted text to a separate file for
supported file types will do so for image files while OCR processing is enabled, an
example of which is discovery export policies.

How Fast Is Optical Character Recognition Processing

The OCR processing rate of image files is approximately 710 KB/sec per IBM
StoredIQ Platform harvester process.
Related concepts:
Configuring application configuration options on page 25
Related tasks:
Configuring harvester settings on page 26

Configuring full-text index settings


Prior to configuring or searching the full-text index, consider the following:
v Full-text filters containing words may not return all instances of those
wordsYou can limit full-text indexing for words based on their length. For
example, if you choose to full-text index words limited to 50 characters, then no
words greater than 50 characters are indexed.
v Full-text filters containing numbers may not return all instances of those
numbersThis can occur when number searches have been configured as
follows:
The length of numbers to full-text index have been definedIf you configure the
full-text filter to index numbers with three digits or more and try to index the
numbers 9, 99, 999, and the word stock, only the number 999 and the word
stock are indexed. The numbers 9 and 99 are not indexed.
Number indexing in data objects limited by file extensionsFor example, if you
choose to full-text index the number 999 when it appears in data objects with
the file extensions .XLS and .DOC, then a full-text filter returns only those
instances of the number 999 that exist in data objects with the file extensions
.XLS and .DOC. Although the number 999 may exist in other data objects
harvested, these data objects do not have the file extensions .XLS or .DOC.

To configure full-text index settings:


1. Go to Administration > Configuration > Application > Full-text settings.
2. To configure Limits:
a. Do not limit the length of the words that are indexedSelect this option
to have no limits on the length of words that are indexed.
b. Limit the length of words indexed to___charactersSelect this option to
limit the length of words that are indexed. Enter the maximum number of
characters at which to index words. Words with more characters than the
specified amount will not be indexed.
3. To configure Numbers:
v Do not include numbers in the full-text indexSelect this option to have
no numbers indexed.
v Do not include numbers in the full-text indexSelect this option to have
no numbers indexed.
v Include numbers in full-text index but limit them bySelect this option to
have only certain numbers indexed. Define these limits as follows:

28 Administration Guide
Number lengthOnly include numbers that are longer than ____
characters. Enter the number of characters a number must contain in order
to be indexed. The Number length feature enables you to index longer
numbers and ignore shorter numbers. By not indexing shorter numbers,
such as one- and two-character numbers that rarely mean anything, you
can focus your filter on meaningful numbers, such as account numbers,
Social Security numbers, credit card numbers, license plate numbers,
telephone numbers, and more.
ExtensionsIndex numbers based on the file extensions of the data
objects in which they appear. Select Limit numbers for all extensions to
limit numbers in all file extensions to the character limits set above in
Number length. Alternatively, select Limit numbers for these extensions
to limit the numbers selected in Numbers length only to data objects
having certain file extensions. Enter the file extensions one per line that
should have limited number indexing. Any data object with a file
extension that is not listed will have all numbers indexed.
4. Configure Stop words. Stop words are common words found in data objects
that are not indexed in the full-text index. To index stop words like the or but
would compromise full-text indexing speed and the relevancy of the results. By
default, the following words are omitted from the full-text index: a, an, and,
are, as, at, be, but, by, for, if, in, into, is, it, no, not, of, on, or, such, that,
the, their, then, there, these, they, this, to, was, will, with.
v If you want the system to find a word listed, delete the word from the list to
allow the full-text index to find it. You must reindex to find the word.
v To add a stop word, enter one word per line, without punctuation, which
includes hyphens and apostrophes.
5. Configure Punctuation or special characters to index by entering punctuation
characters to be included in the index. Note that punctuation characters should
be entered without spaces separating them. By default, only some punctuation
is indexed as a letter in a word by default. Most punctuation is turned into a
space. To make certain that a specific punctuation character is indexed, add it
to this list.
6. To configure Include word stems in index, select whether or not to stem words
that have been indexed. For example, employ is the stem word of words such as
employed, employment, employs, and so on. If you utilize stemming and search
for the term employed, IBM StoredIQ Platform will denote any found instances
of employment, employ, employee, and so on when viewing the data object.
v Do not stem words that are indexed (faster indexing)By not stemming
indexed words, data sources will be indexed faster.
v Stem words that are indexed (improved searching)By stemming indexed
words, filters will be more precise, although slower. Without stemming, a
filter for trade would need to be written as trade or trades or trading or
traded to get the same effect, and even then a user may miss an interesting
variant.

Note: If stemming is enabled, the use of double quotes will return stemmed
terms in results. To find exact words with no stemmed terms, use single quotes.
7. To Optimize wildcard suffix searches, select whether or not to optimize
searches by suffix such as *ology, *tion, *ious, *ive, or *less.
v Do not optimize searches by suffix (faster indexing)By not optimizing
searches by suffix, data sources will be indexed faster.

Performing system configurations 29


v Optimize searches by suffix (faster searching)By optimizing searches by
suffix, searches will be conducted more quickly; however, non-suffix searches
or standard wildcard searches (such as bird*) are not affected.

Note: Stemming will not be performed for search terms with wildcards ("?",
"*") in them. This is true regardless of the term being placed within single
quotes.
8. Click OK.
Related concepts:
Configuring application configuration options on page 25

Specifying data object types


On the Data object types page, you can add new data object types as well as view
and edit known data object types. These data objects appear in the Disk usage (by
data object type) report. There are currently over 400 data object types available.

To specify data object types:


1. Go to Administration > Configuration > Application > Data object types.
2. In the add data object type section, enter one or more extensions to associate
with the data object type. These entries must be separates by spaces. For
example, enter doc txt xls.
3. Enter the name of the data object type to be used with the extension or
extensions. For example, enter Microsoft Word.
4. Click Add to add the extension(s) to the list.
Related concepts:
Configuring application configuration options on page 25

Configuring audit settings


This topic provides procedural information regarding how to configure audit
settings.

To configure audit settings:


1. Go to Administration > Configuration > Application > Audit settings.
2. Specify the number of days to keep the policy audits before automatically
deleting them.
3. Specify the maximum number of policy audits to keep before automatically
deleting them.
4. Specify the file limit for drill down in policy audits.
5. Click OK to save changes.
Related concepts:
Configuring application configuration options on page 25

Configuring hash settings


The Hash Settings page allows you to configure whether to compute a hash when
harvesting and provides different types of hashes.

To configure hash settings:


1. Go to Administration > Configuration > Application > Hash settings.
2. Select Compute data object hash. Note: When hashing emails, click Choose
email fields and select which email attributes are to be used. Note that the

30 Administration Guide
email hash selections operate independently from the data object hash settings;
that is, a data object can have a binary hash or an email hash, but not both.
3. When hashing data objects, except for emails, specify the following options:
v Entire data object content (required for data object typing)
v Partial data object content
4. Click OK.
Related concepts:
Configuring application configuration options on page 25

Configuring desktop settings


When configuring desktop settings, you are enabling or disabling encryption
within IBM StoredIQ Platform. The IBM Desktop Data Collector (desktop client or
client) enables desktops as a volume type or data source, allowing them to be used
just as other types of added data sources. The client is provided as a standard MSI
file, and is installed according to the typical method (such as Microsoft Systems
Management Service (SMS)) used within your organization. The client can collect
PSTs and .ZIP files as well as other data objects and is capable of removing itself
once its work is completed.

To configure desktop settings:


1. Go to Administration > Configuration > Application > Desktop settings.
2. In the Desktop Services area, select the Encrypt all traffic to/from desktops
check box.
3. Select either the Enabled or Disabled button to enable or disable desktop
services.
4. Click Apply.

Downloading the IBM Desktop Data Collector installer from


the application
This topic provides procedural information regarding how to download the IBM
Desktop Data Collector installer from the application.

To download the IBM Desktop Data Collector installer from the application:
1. Go to Administration > Configuration > Application > Desktop settings.
2. In the Download the Desktop Agent installer area, click Download the
desktop client installer.
3. Once the download is complete, click Save File.

Once the desktop client has been installed on a desktop and then connected to and
registered with the data server, that desktop is available as a data source within
the list of primary volumes. Connectivity and the correct IP address is required.
Related tasks:
Configuring desktop settings

Upgrading the IBM Desktop Data Collector agent


This topic provides procedural information regarding how to upgrade the IBM
Desktop Data Collector agent.

Performing system configurations 31


During IBM Desktop Data Collector collection, if IBM StoredIQ Platform finds an
Encrypted File Systemencrypted file, then IBM Desktop Data Collector will install
a recovery agent certificate, allowing the client to open the encrypted file and
harvest from it.

To upgrade clients for registered workstations:


1. Go to Administration > Configuration > Application > Desktop settings.
2. In the Upgrades area, select from either Automatic upgrade options or
Available versions options.
3. For Automatic upgrade options:
v Upgrades disabledAll upgrades are disabled, meaning that none will be
applied.
v Upgrade all workstationsAll workstations will be upgraded.
4. For Available versions options:
v Select Manually publish new version, and then select that version.
v Select Automatically publish the latest version.
5. Click Apply.
Related tasks:
Configuring desktop settings on page 31

Using the Encrypted File System recovery agent


During IBM Desktop Data Collector collection, if IBM StoredIQ Platform finds an
Encrypted File Systemencrypted file, the IBM Desktop Data Collector will install a
recovery agent certificate, allowing you to open the encrypted file.

To add an encrypted file-system user:


1. Go to Administration > Configuration > Application > Desktop settings.
2. In the Encrypted file system recovery agent users area, click Add encrypted
file system user. The Upload Recovery Agent Certificate dialog box appears.
3. In the Select a .PFX file to upload: text box, click Browse to navigate to the
desired .PFX file. By default, the system takes a .PFX file.
4. Enter the .PFX password in the .PFX password: text box. This password
protects the file itself.
5. In the Username: text box, enter the username for the user, a SAM
compatible/NT4 Domain namestyle user name. For example, enter
MYCOMPANY\esideways. This is the credential of the user to whom this recovery
agent belongs.
6. In the Password: text box, enter the password for the user.
7. Optionally, enter a description in the Description: text box.
8. Click OK. The file is uploaded, and the added user is visible within the User
name column. Note that once users have been added, they can also be edited
or deleted using the Edit or Delete options.
Related tasks:
Configuring desktop settings on page 31

32 Administration Guide
Creating volumes and data sources
A volume represents a data source or destination that is available on the network to
the IBM StoredIQ Platform appliance. A volume can be a disk partition or group of
partitions that is available to network users as a single designated drive or mount
point. IBM StoredIQ Platform volumes perform the same function as partitions on
a hard drive. When you format the hard drive on your PC into drive partitions A,
B, and C, you are creating three partitions that function like three separate physical
drives. Volumes behave the same way that hard drive disk partitions behave. You
can set up three separate volumes originating from the same server or across many
servers. Note that only Administrators can define, configure, and add or remove
volumes to IBM StoredIQ Platform.
Related reference:
Supported server platforms by volume type

Volume indexing
This topic describes volume indexing and the different depths at which volumes
can be indexed.

When defining volumes, you also determine the depth at which you want the
volume to be indexed. There are three levels of analysis:
v System metadata indexThis level of analysis runs with each data collection
cycle and provides only system metadata for system data objects in its results. It
is useful as a simple inventory of what data objects are present in the volumes
you have defined and for monitoring resource constraints (such as file size) or
prohibited file types (such as .MP3s).
v System metadata plus containersIn a simple system metadata index, container
data objects (.ZIP files, PSTs, emails with attachments, and the like) are not
included. This level of analysis provides container-level metadata in addition to
the system metadata for system data objects.
v Full-text and content taggingThis option provides the full native language
analysis that yields the more sophisticated entity tags. Naturally, completing a
full-text index requires more system resources than a metadata index. Users
must carefully design their volume structure and harvests so that the maximum
benefit of IBM StoredIQ Platform sophisticated analytics are used where
necessary, but not on resources that do not require them. Parameters and
limitations on full-text indexing are set when the system is configured.
Related concepts:
Creating volumes and data sources

Configuring server platforms


Prior to configuring volumes on IBM StoredIQ Platform, you must configure the
server platforms you will use for the different volume types. Each server type has
requisite permissions and settings, which are described here.
Related concepts:
Creating volumes and data sources

Copyright IBM Corp. 2001, 2013 33


Windows Share (CIFS)
This topic describes the Windows Share (CIFS) server platform.

To harvest and run policies on volumes on Windows Share (CIFS) servers, the user
must be in the backup operator group on the Windows Share server exposing the
shares on IBM StoredIQ Platform.
Related concepts:
Configuring server platforms on page 33

NFS
This topic describes the NFS server platform.

To harvest and run policies on NFS servers, you must enable root access on the
NFS server that is connected to IBM StoredIQ Platform.
Related concepts:
Configuring server platforms on page 33

Exchange servers
This topic describes Microsoft Exchange servers and various connections and
permissions.
v Secure connectionIf you want to connect to Exchange volumes over HTTPS,
you must add port number 443 after the server name, for example,
qa03exch2000.qaw2k.local:443. In some cases, this secure connection can result
in some performance degradation due to SSL overhead. If you enter the volume
information without the 443 suffix, the default connection will be over HTTP.
v Permissions for Exchange 2003The following permissions must be set on the
Exchange server to the mailbox store or the mailboxes from which you will
harvest.
Read
Execute
Read permissions
List contents
Read properties
List object
Receive as
v Permissions for Exchange 2007 and 2010The Full Access permissions must be
set on the Exchange server to the mailbox store or the mailboxes from which
you will harvest.
v Recoverable items in Exchange 2010To harvest the recoverable items folders
in Exchange 2010, you must be logged in as either an Administrator role.
v Deleted itemsTo harvest items that have been deleted from the Exchange
server, enable Exchanges transport dumpster settings. For more information,
refer to Microsoft Exchange Server 2010 Administrator's Pocket Consultant.
Configuration information is also available online at www.microsoft.com.
v Windows AuthenticationFor all supported versions, enable Integrated
Windows Authentication on each Exchange server.
Related concepts:
NFS

34 Administration Guide
Enabling integrated Windows authentication on Exchange
servers
This topic provides procedural information regarding how to integrate Windows
authentication on Exchange servers.

To enable integrated Windows authentication on Exchange servers:


1. From Microsoft Windows, log in to the Exchange Server.
2. Go to Administrative Tools > Internet Information Services (IIS) Manager.
3. Go to Internet Information Services > Name of Exchange Server > Web Sites
> Default Web Site.
4. Right-click Default Web Site, and then click the Directory Security tab.
5. In the Authentication and access control pane, click Edit.
6. Select Properties. The Authentication Methods window appears.
7. In the Authentication access pane, select the Integrated Windows
authentication check box.
8. Click OK.
9. Restart IIS Services.
Related concepts:
Exchange servers on page 34

Improving performance for IIS 6.0 and Exchange 2003


This topic provides procedural information regarding how to improve performance
between IIS 6.0 and Exchange 2003.

To improve performance for IIS 6.0 and Exchange 2003:


1. From Microsoft Windows, log onto the Exchange Server.
2. Go to Administrative Tools > Internet Information Services (IIS) Manager.
3. Select Internet Information Services > <Name of Exchange Server> > Web
Sites > Application Pools.
4. Right-click Application Pools and select Properties.
5. On the Performance tab, locate the Web Garden section.
6. If the number of worker processes is different from the default value of 1, then
change the number of worker processes to 1.
7. Click OK.
8. Restart IIS Services.
Related concepts:
Exchange servers on page 34

SharePoint
This topic describes SharePoint servers and various connections and privileges.

Secure Connection

If you wish to connect to SharePoint volumes over HTTPS, you may use the Use
SSL checkbox, or may add the port number 443 after the server name when setting
up the volume on StoredIQ. For a example, qa01.company.com:443. In some cases,
this secure connection can result in some performance degradation due to Secure
Socket Layer (SSL) overhead. If you enter the volume information without the 443
suffix, the default connection is over HTTP.

Creating volumes and data sources 35


Privileges

To run policies on SharePoint servers, you must use credentials with Full Control
privileges. We recommend using a Site collection administrator to harvest subsites
of a site collection.
Related concepts:
Configuring server platforms on page 33

Privileges required by user account


This topic provides conceptual information regarding what SharePoint privileges
are required by user account along with IBM StoredIQ Platform recommendations.

IBM StoredIQ Platform is typically used with SharePoint for one of these instances:
to harvest and treat SharePoint as a source for policy actions or to use as a
destination for policy actions, meaning that you can write content into SharePoint
using IBM StoredIQ Platform. Consider these points:
v Attributes are not set/reset on a SharePoint harvest or if you copy from
SharePoint.
v Attributes are set only if you copy to SharePoint.

If you plan to only read from the SharePoint (harvest and source copies from),
then you must use user credentials with Read privileges on the site and on all of
the lists and data objects that you expect to process.

If you plan to use SharePoint as a destination for policies, you must use user
credentials with Contribute privileges on the site.

Additional Privileges for Social Data

If you wish to index all the social data for a user profile in SharePoint 2010, then
the user credentials must own privileges to Manage Social Data as well.

Recommended Privileges

We recommend using a Site Collection Administrator to ensure all data is


harvested from a given site or site collection.
Related concepts:
SharePoint on page 35

Alternate-access mappings
SharePoint 2007 and 2010 require the configuration of alternate-access mappings to
map IBM StoredIQ Platform requests to the correct Web sites.

Alternate-access mappings map URLs presented by IBM StoredIQ Platform to


internal URLs received by Windows SharePoint Services. An alternate-access
mapping is required between the server name and optional port defined in the
SharePoint volume definition and the internal URL of the Web application
containing the Web site being accessed is required. If SSL will be used to access the
site, ensure that the alternate-access mapping URL uses https:// as the protocol.

Refer to Microsoft SharePoint 2007 or 2010 documentation to configure


alternate-access mappings based on the public URL configured by the local
SharePoint administrator and the mapping used by the IBM StoredIQ Platform
SharePoint volume definitions.

36 Administration Guide
Example

You are accessing a SharePoint volume with the fully qualified domain name,
http://itweb.storediqexample.com, from the Intranet zone. An alternate-access
mapping for the public URL http://itweb.storediqexample.com for the Intranet
zone must be configured for the SharePoint 2007 or 2010 Web application hosting
the site to be accessed by the volume definition. If you are accessing the same
volume using SSL, the mapping added should be for the URL
https://itweb.storediqexample.com instead.

Note: When configuring SharePoint volumes using non-qualified names, you are
entering the URL for a SharePoint site collection/site that is leveraged by IBM
StoredIQ Platform in the volume definition. Consider the following:
v The URL must be valid with respect to the configured Alternate Access
Mappings configured in SharePoint.
v If the host name in the URL does not convey the fully qualified domain that
should be used to authenticate the configured user, an Active Directory server
must be specified. The specified Active Directory must be a fully qualified
domain name and will be used for authentication.
Related concepts:
SharePoint on page 35

Documentum
This topic describes requirements for Documentum servers.

To run harvests and copy from Documentum servers, you must use the
Contributor role.
Related concepts:
Configuring server platforms on page 33

Enterprise Vault
In order to configure an Enterprise Vault volume, you must first configure an
Enterprise Vault site. This topic provides procedural information regarding how to
configure an Enterprise Vault site.

To configure an Enterprise Vault site:


1. Go to Administration > Data sources > Specify servers.
2. Click Enterprise Vault sites, and then click Add new Enterprise Vault site.
3. In the Site name field, enter a unique logical site name. This name will appear
in the screens used to configure Enterprise Vault volumes.
4. In the Enterprise Vault site alias field, enter the fully qualified domain name
(FQDN) of the Enterprise Vault Server. Each server can only be added one time.
5. In the User name field, enter the login name. If the user is a domain user, then
enter the login name as domain\user. It is recommended that the Enterprise
Vault Service Account or a user with equivalent privileges be used.
6. In the Password field, enter the users password to authenticate with Active
Directory.
7. Click OK to save the site.
Related concepts:
Configuring server platforms on page 33
Related tasks:

Creating volumes and data sources 37


Configuring primary volumes using Enterprise Vault on page 55

Discovery Accelerator
This topic describes configurations that must be made to the Discovery Accelerator
server.

Prior to configuring Discovery Accelerator primary volumes, you must configure


Discovery Accelerator customer information and Enterprise Vault sites (first
customer, then site) so that certain configuration items can appear in the volume
configuration lists. Before performing these tasks, you must log in to the Discovery
Accelerator server and run the ImportExport.exe tool (located in the install folder)
to obtain the appropriate Customer IDs and customer database names.
v Discovery Accelerator Web-Services InterfaceIn order for IBM StoredIQ
Platform to interface with Enterprise Vault using the Discovery Accelerator Web
services, the following configurations must be made on the Discovery
Accelerator service by logging into the Discovery Accelerator client interface as a
Vault User or Discovery Administrator.
1. Click the Configuration tab and expand the API options on the Settings
page.
2. In the API settings group, ensure that the API Enabled setting is enabled.
3. Configure a Temporary Storage Area, as needed. Ensure that it has sufficient
free space and that any authenticated users that will define volumes against
Discovery Accelerator have Full Control permissions on this storage area.
Additionally, consider configuring the Temporary Storage Area Cleanup
Interval, depending on the size of the cluster that will be
deployedtypically, the default value of 30 minutes should be sufficient. If
greater than four nodes will be deployed in the cluster, this interval must be
reduced accordingly for more frequent cleanups to free up storage space.
v Improve performance for IIS 6.0If the Discovery Accelerator server runs over
IIS 6.0, an existing bug in IIS causes severe performance degradation when used
along with Kerberos authentication. We recommend that the hotfix described in
Microsoft Knowledge Base article 917557 (http://support.microsoft.com/kb/
917557) be applied to the server in this case.
Related concepts:
Configuring server platforms on page 33

Discovery Accelerator permissions


This topic describes Discovery Accelerator permissions and credentials.

It is recommended that the credentials used for referencing the Enterprise Vault
Site are those of the Vault User or any other administrator. IBM StoredIQ Platform
will validate that the credentials are strong enough for it to:
v Login remotely to the specified server
v Perform DCOM operations over RPC remotely

To harvest a Discovery Accelerator volume successfully, a user must have the


following privileges:
v A role defined in the Discovery Accelerator Web Application.
v Review messages permission for the case used in the volume definition.
v Folder review permissions on a case, if a folder (sometimes also called a
Research Folder) in the case is going to be harvested.

38 Administration Guide
v Permission to set all of the review marks that will be selected for the volume
definition.
Related concepts:
Discovery Accelerator on page 38

Configuring security settings for Enterprise Vault servers


This section lists the standard security settings that must be configured on the
Windows Servers hosting Enterprise Vault to allow it to interact with IBM
StoredIQ Platform. Because any number of security applications, including network
security providers, could be in use on a given target server, it is impossible to
address them all.
v Enable Remote DCOMRequired on all Enterprise Vault Servers and Discovery
Accelerator. See Enabling Remote DCOM.
v Allow DCOM traffic through Windows FirewallRequired on all Enterprise
Vault servers and Discovery Accelerator. See Allowing DCOM Traffic through
the Windows Firewall.

To configure Discovery Accelerator Customer Information:


1. Go to Administration > Data sources > Specify servers.
2. Click Discovery Accelerator customers, and then click Add new Discovery
Accelerator customer.
3. In the Customer name field, enter a unique display name of the DA customer.
This name will appear in the screens used to configure Enterprise Vault
volumes.
4. In the Discovery Accelerator server field, enter the DNS name of the physical
server running Discovery Accelerator.
5. In the Discovery Accelerator customer ID field, enter the customer ID value
obtained from Discovery Accelerator. Each unique combination of Discovery
Accelerator server name and customer ID can be used only one time.
6. In the Customer virtual directory field, enter the IIS Virtual Directory where
the Discovery Accelerator Web service is located.
7. In the Discovery Accelerator installation folder field, enter the path where
Enterprise Vault Business Accelerator was installed on the Discovery
Accelerator server. IBM StoredIQ Platform requires this path so that Discovery
Accelerator can be accessed during configuration.
8. Click OK to save the site.
Related concepts:
Discovery Accelerator on page 38

Configuring Enterprise Vault


This topic provides procedural information regarding how to configure an
Enterprise Vault site.

To configure an Enterprise Vault site:


1. Go to Administration > Data sources > Specify servers.
2. Click Enterprise Vault sites, and then click Add new Enterprise Vault site.
3. In the Site name field, enter a unique logical site name. This name will appear
in the screens used to configure Enterprise Vault volumes.
4. In the Enterprise Vault site alias field, enter the FQDN of the Enterprise Vault
Server. Each server can only be added one time.

Creating volumes and data sources 39


5. In the User name field, enter the login name. If the user is a domain user, then
enter the login name as domain\user. It is recommended that the Enterprise
Vault Service Account or a user with equivalent privileges be used.
6. In the Password field, enter the users password to authenticate with Active
Directory.
7. Click OK to save the site.
Related concepts:
Discovery Accelerator on page 38

NewsGator
This topic provides procedural information regarding required NewsGator
privileges and how they should be configured.

Privileges Required by User Account

The user account used to harvest or copy from a NewsGator volume must have
the Legal Audit permission on the NewsGator Social Platform Services running on
the SharePoint farm.

To configure this permission:


1. Login as an administrator to your SharePoint Central Administration Site.
2. Under Application Management, select Manage Service Applications.
3. In the Manage Service Applications screen, select the NewsGator Social
Platform Services row.
4. From the button ribbon that activates near the top of the page, select
Administrators.
5. Add the user account that will be used for the NewsGator harvest to the list of
Administrators, ensuring that the account has the Legal Audit permission.
Related concepts:
Configuring server platforms on page 33

Configuring retention servers


IBM StoredIQ supports various types of retention servers, which must be
configured prior to adding retention volumes due to unique requirements.
Related concepts:
Creating volumes and data sources on page 33

Creating Centera pools


When a Centera pool is created, an empty volume set is automatically created and
associated with the Centera pool. This feature enables the harvesting of unknown
Centera volumes so business policies can be applied to data objects on already
retained storage servers. The volume set cannot be edited or deleted from the
manage volume sets page.

The retention feature requires the Advanced retention feature to be enabled.


v PermissionsTo support all IBM StoredIQ Platform features, the following
effective access profile rights to a Centera pool must be enabled:
Read
(D)elete
(Q)uery

40 Administration Guide
(E)xist
(W)rite
Retention
(H)old
v Centera PoolsIf you have an integrated Centera server, you need to create a
Centera pool before you can add a Centera volume. Unlike other volumes,
Centera servers are not placed into volume sets but into Centera Pools. Once
you have created a pool, the Centera pool will be in the list of available choices
when adding a Centera volume.

To create a Centera pool:


1. Go to Administration > Data sources > Specify volumes.
2. In Centera pools, click Create new pool to open the Centera pool editor.
3. Enter a unique name for the Centera pool in the IBM StoredIQ Platform pool
profile name text box.
4. Choose how to define the pool. Select either Structured Output or Single text
field and enter a connection string.
v Specify Access and enter a profile name and Secret for the pool.
v Use .pea fileThis is a pool-entry authorization. For information about
creating this type of file, refer to the Centera Monitor and Administration Guide.
5. If you chose Structured input, click Add new connection in the Connections
section.
6. Enter the IP address for an access point on the server, and then click Add.
Multiple access points can be specified allowing for fail over in the event of a
problematic access node.
7. Click OK to save the pool. The pool is now available in the Add volumes
dialog when adding a Centera volume
Related concepts:
Centera

Configuring the Dell DX Object Storage Platform


A Dell DX Storage Cluster must be defined before a Dell DX Storage volume can
be added. Once you have defined the cluster, the cluster will be available in the list
of available options when adding a Dell DX Storage volume.

To create a Dell DX Storage Cluster:


1. Go to Administration > Data sources > Specify Servers > Dell DX Storage
Clusters.
2. In the Dell DX Storage Cluster list, select Add new Dell DX Storage Cluster.
3. Enter a unique name for the Profile name.
4. Enter the Dell DX Storage Cluster name. This should be the same name used
when configuring the DX CSN.
5. Enter a port number that will be used to communicate with the storage
cluster. The default value is 80.
6. Optionally, specify alternate addresses to be used to communicate with the
storage cluster. This should be one or more IP addresses or DNS names that
map to DX Storage Nodes or DX CSN in the cluster. Individual entries should
be separated by a comma.
7. Optionally, specify the IP address or DNS name for a DR Site DX CSN that
can be used for failover.

Creating volumes and data sources 41


8. Enter a Starting IP address for IBM StoredIQ Platform cluster. Each node in
the IBM StoredIQ Platform cluster will be assigned an IP address in the Dell
DX Storage Cluster private network. The Starting IP address should be an
unused, valid IP address within the Dell DX Storage Cluster private network.
For example, if 172.17.17.100 is entered for a five-node IBM StoredIQ
Platform cluster, 172.17.17.100, 172.17.17.101, 172.17.17.102,
172.17.17.103, and 172.17.17.104 will be used by IBM StoredIQ Platform.
9. Enter the Netmask value for the Dell DX Storage Cluster private network.
Typically, this value will be 255.255.255.0.
10. Note the value of the Switch port, which should be used on the IBM StoredIQ
Platform switch to link with the Dell DX Storage Cluster.
Related concepts:
Configuring retention servers on page 40

Configuring Enterprise Vault (retention volumes)


This section provides both conceptual and procedural information regarding
Enterprise Vault, including security settings and DCOM.

Prior to creating retention volumes using Enterprise Vault, you must configure
Discovery Accelerator customers and Enterprise Vault sites (customers first, then
sites) so that certain configuration items can appear in the retention volume
configuration lists. Before performing these tasks, you must log in to Discovery
Accelerator and run the ImportExport.exe tool to obtain the appropriate Customer
IDs and customer database names.

Note: If you use a system other than IBM StoredIQ Platform to ingest data into
Enterprise Vault but still want to use IBM StoredIQ Platform for exporting out of
Discovery Accelerator, you must define an Enterprise Vault site within IBM
StoredIQ Platform and then use that site to define a Discovery Accelerator volume.
When defining the Enterprise Vault site, DCOM configuration is a prerequisite.

Enabling remote DCOM


This topic provides procedural information regarding how to enable remote
DCOM for Enterprise Vault servers.

Note: Remote DCOM is required on all Enterprise Vault servers and Discovery
Accelerator.

You can configure DCOM settings using the DCOM Config utility (DCOMCnfg.exe)
found in Administrative Tools in Control Panel as Component Services. This utility
exposes the settings that enable certain users to connect to the computer remotely
through DCOM. Members of the Administrators group are allowed to connect
remotely to the computer by default.

If the Enterprise Vault Service Account (or the user whose credentials are used to
define the Enterprise Vault Site IBM StoredIQ Platform) does not have permissions
to connect remotely for DCOM, then perform this procedure on the target server.

To enable remote DCOM:


1. Run dcomcnfg as a user with Administrator privileges.
2. In the Component Services dialog box, expand Component Services, expand
Computers, and then right-click My Computer and click Default Properties. If
not already enabled, select the Enable Distributed COM on this
Computercheck box.

42 Administration Guide
3. Click the COM Security tab.
4. Under Launch and Activation Permissions, click Edit Limits.
5. In the Launch Permission dialog box, follow these steps if the user name does
not appear in the Groups or user names list:
v In the Launch Permission dialog box, click Add.
v In the Select Users, Computers, or Groups dialog box, add the username
and then click OK.
v In the Launch Permission dialog box, select your user and in the Allow
column under Permissions for User, select Remote Launch and select
Remote Activation, and then click OK.

Allowing DCOM traffic through the Windows firewall


This topic provides procedural information regarding how to allow DCOM traffic
through the Windows firewall.

Note: Required on all Enterprise Vault servers and Discovery Accelerator.

To allow DCOM traffic over the network on the target server, the DCOM TCP port
(135) must be open on the Firewall. This command opens this port if it is closed:
netsh firewall add portopening protocol=tcp port=135
name=DCOM_TCP135

The port can also be opened using the Firewall User Interface.

To open the Firewall Port to allow DCOM Traffic:


1. In the Control Panel, double-click Windows Firewall.
2. Click Change Settings, and then click the Exceptions tab.
3. In the Exceptions window, select the check box for DCOM to enable DCOM
traffic through the firewall. If there is no such check box:
v Click Add Port.
v In the dialog box, enter Name as DCOM and Port number as 135. Ensure that
the TCP radio button is selected and click OK.
Related concepts:
Configuring retention servers on page 40

Configuring IBM Information Archive retention servers


Before creating retention volumes, you need to configure settings on your IBM
Information Archive retention server. Files sent to retention must have a
corresponding management class to manage the retention period settings. These
classes require a defined class, policy domain, policy set, and node on the IBM
Information Archive retention server.

Use the Tivoli Storage administrative client, dsmadmc, to enter commands. You
can use the administrative client in either interactive or batch mode. Consult the
Tivoli Storage Manager Administrator's Reference for more details about using the
administrative client.

To configure an IBM Information Archive retention server:


1. Create a policy domain. For example:
define domain Example_PD desc=Example Domain archret=1
2. Create a policy set for the policy domain.
define policyset Example_PD Example_PS

Creating volumes and data sources 43


3. Create one or more management classes. These classes will be referred to as
Retention Classes in the applications user interface. Each management class
will have only one copy group assigned to it.
v
Define mgmtclass Example_PD Example_PS Example_MG_CR desc=Exam
ple Domain chronological retention management class'
v
Define mgmtclass Example_PD Example_PS Example_MG_0DAY
desc=Example Domain zero day retention
v
Define mgmtclass Example_PD Example_PS Example_MG_10DAY
desc=Example Domain 10 day retention
4. Create an archive copy group for each management class defined above. The
copy group must include retinit=creation. For example:
v
Define copygroup Example_PD Example_PS Example_MG_CR STANDARD
type=archive destination=archivepool retver=1 retinit=creation
v
Define copygroup Example_PD Example_PS Example_MG_0DAY STAN
DARD type=archive destination=archivepool retver=0
retinit=creation
v
Define copygroup Example_PD Example_PS Example_MG_10DAY STANDARD
type=archive destination=archivepool retver=10 retinit=creation
5. Assign one of the management classes created above as the default
management class for the policy domain.
6. Assign defmgmtclass Example_PD Example_PS Example_MG_CR.
7. Complete the definition of the policy domain by validating and activating the
policy set.
8. Validate policyset Example_PD Example_PS.
9. Activate policyset Example_PD Example_PS.
10. Create a node. This acts as a connection between IBM StoredIQ Platform and
the server and is used when defining IBM StoredIQ Platform volumes. For
example:
register node Example Domain password domain=Example_PD

Note: List NodesYou can also list nodes using the query node command.
Related concepts:
Configuring retention servers on page 40

NetApp SnapLock
This topic provides conceptual information regarding NetApp SnapLock retention
server configuration requirements.

Note: Manage files using IBM StoredIQ PlatformNetApp SnapLock retention


servers do not lock out user deletion and modification of files that do not have an
active retention period. Always use IBM StoredIQ Platform to manage files on
retention.

When preparing to add SnapLock retention volumes, you must have licensed
SnapLock, created at least one SnapLock volume and shared it using either

44 Administration Guide
Windows Share or NFS (or both), and initialized the SnapLock compliance clock.
Consult the NetApp administrator documentation for specific instructions.

When configuring the server, set the retention period settings. IBM StoredIQ
Platform recommends setting the minimum setting to zero while not entering a
maximum setting. Access the SnapLock server and enter the following commands:
vol options <vol-name> SnapLock_minimum_period 0d
vol options <vol-name> SnapLock_default_period min
Related concepts:
Configuring retention servers on page 40

Hitachi HCAP configuration requirements


This topic provides conceptual information regarding Hitachi HCAP retention
server configuration requirements.

The IBM StoredIQ Platform application accesses the Hitachi HCAP server using
HTTP. Consequently, the HTTP gateway must be enabled on the server. Depending
on the current allow/deny lists for the HTTP gateway, you may need to add the
IBM StoredIQ Platform data server's IP addresses to the Allow IP addresses list.
Related concepts:
Configuring retention servers on page 40

Configuring FileNet
By providing the configuration values for a FileNet domain, you are supplying the
values needed to bootstrap into a domain.

To configure the FileNet domain:


1. Go to Administration > Data sources > Specify Servers > FileNet domain
configurations.
2. Click Add new FileNet domain configuration, and the FileNet domain
configuration editor page appears.
3. In the FileNet domain configuration editor page, configure these fields:
v In the Configuration name text box, enter the configuration name for this
retention server.
v In the Server name text box, enter the server name.
v In the Connection list, select the desired connection type.
v In the Port text box, enter the port number.
v In the Path text box, enter the path for this retention server.
v In the Stanza text box, enter the stanza information for this retention server.
4. Click OK to save your changes.
Related concepts:
Configuring retention servers on page 40

Configuring Chatter messages


Within Chatter, the default administrator profile does not have the Manage Chatter
Messages permission, but the proper permissions are required in order to harvest
private messages.

There are certain administrative permissions that a user must have when that user
account is used in the Connect as text box in Chatter. When setting up a Chatter

Creating volumes and data sources 45


user account that should be used to harvest and run actions against Chatter, we
recommend that you use an account with the built-in System Administrator profile.
In general however, these administrative permissions must be assigned to the
account you use:
v API enabled
v Manager Chatter Messages (required if you wish to harvest Chatter Private
Messages)
v Manage Users
v Moderate Chatter
v View All Data

Note: For Chatter administrators using the Auth token option, read how to set up
a sandbox account.
Related concepts:
Creating volumes

Creating volumes
This section describes the types of volumes in IBM StoredIQ Platform as well as
how to create them as data sources within the platform.
Related concepts:
Creating volumes and data sources on page 33
Related reference:
Appendix B, Supported server platforms and protocols, on page 129

Creating primary volumes


This topic provides procedural information regarding how to create primary
volumes as data sources.

To add a primary volume:


1. Go to Administration > Data sources > Specify volumes > Volumes.
2. On the Primary volume list page, click Add primary volumes.
3. Enter the information described in the table below, based on your server type.
4. Click OK to save the volume.
5. Select one of the following:
v Add another volume on the same server.
v Add another volume on a different server.
v Finished adding volumes.
The following table describes the fields that are available in the Add volume
dialog box when configuring primary volumes.

Note: Case-sensitivity rules for each server type apply. Red asterisks within the
user interface denote required fields. For best-practice information regarding
editing volume definitions and system restarts, see Editing Volume Definitions.

46 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type

v For Documentum, you must v Windows Share


Server type In the Server type list, select the
specify the doc broker. See (CIFS)
server type.
Documentum. v NFS v2 and v3
v For Discovery Accelerator, you v NetApp
must first create a Discovery
v Exchange
Accelerator site and an Enterprise
Vault site. See Discovery v Domino
Accelerator. v Celerra
v For Domino, you must first v SharePoint
upload at least one user.id. See v Documentum
Adding Domino as a Primary
Volume. v Discovery
Accelerator
v For Desktop, the desktop agent
must be installed on that desktop v Enterprise Vault
and then pointed to the data v FileNet
server. Desktop will not appear v NewsGator
in the Server type list, but will
v Chatter
instead appear as an available
server or volume.

v Standard
Platform In the Platform list, select the
platform type. v Celerra
v NetApp

Version In the Version list, select the Exchange


appropriate version.

Site In the Site list, select the Enterprise Vault


appropriate site.

Vault store In the Vault store list, select the Enterprise Vault
desired vault store for the volume.

Creating volumes and data sources 47


Applicable Volume
Field Name Required Action Special Notes Type

v For Exchange primary volumes, v Windows Share


Server In the Server text box, enter the
this is the fully qualified domain (CIFS)
fully qualified name of the server
name where the OWA resides. v NFS v2 and v3
where the volume is available for
Multiple Client Access servers on
mounting. v NetApp
Exchange 2007 are supported.
Note that the server must be v Exchange
load-balanced at the IP or DNS v SharePoint
level. v Domino
v For NetApp, select the v Discovery
predefined server from the list. Accelerator
v For Domino, select the v NewsGator
appropriate user name, which
were entered using the
Configuration subtab in the
Lotus Notes user administration
area.
v For Celerra, select the predefined
server from the list.
v When adding SharePoint
volumes that contains spaces in
the URL, see Special Note:
Adding SharePoint Volumes.
v For Desktop, the desktop agent
must be installed on that desktop
and then pointed to the data
server. Desktop will not appear
in the Server type list, but will
instead appear as an available
server or volume.

FileNet config Use the FileNet config list to select For more information, see FileNet. FileNet
the FileNet server you would like
to use for this configuration.

Mailbox server When configuring multiple client For Exchange primary volumes, this Exchange
access servers, enter the name of is the fully qualified domain name
one or more mailbox servers, where the mailbox(es) to be
separated by a comma. harvested reside.

v Exchange
Active Directory In the Active Directory server text This must be a fully qualified
server box, enter the name of the Active Active Directory server. v Discovery
Directory server. Accelerator
v SharePoint

v Exchange
Protocol To use SSL, select the Protocol API client uses HTTP over SSL to
check box. communicate with Discovery v Discovery
Accelerator Server configuration. Accelerator
v SharePoint
v NewsGator

Doc base In the Doc base text box, enter the A Documentum repository contains Documentum
name of the Documentum cabinets, and cabinets contain
repository. folders and/or documents.

48 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type

v Windows Share
Connect as In the Connect as text box, enter For Domino, select the user name
(CIFS)
the logon ID used to connect and for the primary user ID. The user id
mount the defined volume. must be configured on the System v Exchange
Configuration screen under the v SharePoint
Lotus Notes user administration v Documentum
link.
v Domino
v FileNet
v Discovery
Accelerator
v NewsGator
v Chatter

v Windows Share
Password In the Password text box, enter the For Domino, enter the password for
(CIFS)
password used to connect and the primary user ID.
mount the defined volume. v Exchange
v SharePoint
v Documentum
v Discovery
Accelerator
v Domino
v Celerra Windows
Share
v FileNet
v NewsGator
v Chatter

Auth token In the Auth token text box, enter The auth token must match the user Chatter
the token used to authenticate the name used in the Connect as field.
Chatter volume. Auth tokens can be generated
online on Salesforce. See
Configuring Chatter Volumes.

Object store In the Object store list, select the The object store must exist prior to FileNet
desired object store. the creation of a FileNet primary
volume.

v For Exchange, Domino, v Windows Share


Volume In the Volume text box, enter the
Documentum, Enterprise Vault (CIFS)
name or names of the volume to be
(primary volume), and v NFS v2 and v3
mounted.
NewsGator, enter a friendly name
v NetApp
for the volume.
v Exchange
v When adding SharePoint
volumes that contains spaces in v Enterprise Vault
the URL, see Special Note: v Domino
Adding SharePoint Volumes. v Celerra
v FileNet
v NewsGator
v Chatter

Creating volumes and data sources 49


Applicable Volume
Field Name Required Action Special Notes Type

Discovery In the Discovery Accelerator case This text box is populated from Discovery
Accelerator case text box, enter the name of the Discovery Accelerator once Accelerator
Discovery Accelerator. connection is established.

Archive In the Archive list, select the Enterprise Vault


archive to which this Enterprise
Vault volume pertains.

v Documentum
Harvest This is the Documentum harvest This option obtains the list of all
option: known Domino users and their v Domino
v To enable harvesting, select NSFs. It then harvests those
Harvest all document versions. mailboxes unless it was pointed to
a single mailbox using the initial
These are the Domino harvest directory.
options:
v To harvest mailboxes, select the
Harvest mailboxes option.
v To harvest mail journals, select
the Harvest mail journals option.
v To harvest all applications, select
the Harvest all applications
option.

Folders In Folder, select either of the Exchange


Mailboxes or Public folders
options.

v This feature allows you to select v Windows Share


Initial directory In the Initial directory text box,
a volume further down the (CIFS)
enter the name of the initial
directory tree rather than v NFS v2 and v3
directory from which the harvest
selecting an entire volume.
should begin. v NetApp
v For Exchange, this field should
v Exchange
be left blank if you are harvesting
all mailboxes. If you are v Celerra
harvesting a single mailbox, enter v SharePoint
the email address for that v Documentum
mailbox.
v Chatter
v When adding SharePoint
volumes that contains spaces in
the URL, see Special Note:
Adding SharePoint Volumes.

Virtual root In Virtual root, change the default For Exchange, this option should be Exchange
name to match the Exchange server changed to match the server
designation. designation.

Personal archives To enable the collection of personal This option pertains only to Exchange 2010
archives, select the Harvest Exchange 2010 with SP1 applied.
personal archive check box.

50 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type

v Windows Share
Index options Select either or both of the Index These options are selected by
(CIFS)
options check boxes. default.
v NFS v2 and v3
v Include system metadata for
data objects within containers. v NetApp
v Include content tagging and v Exchange
full-text index. v SharePoint
v Documentum
v Celerra
v Enterprise Vault
v Discovery
Accelerator
v Domino
v FileNet
v NewsGator
v Chatter
v Desktop

v SharePoint
Subsites Select Recurse into subsites.

Versions Select Include all versions. IBM StoredIQ Platform supports SharePoint
indexing versions from SharePoint.
Versions For more information, see Special
Note: Adding SharePoint Volumes.

Remove journal When selected, the journal envelope Exchange


envelope is removed.

v Windows Share
Validation To validate volume accessibility, When selected (the default state),
(CIFS)
select Validation. IBM StoredIQ Platform tests to see
if the volume can be accessed. v NFS v2 and v3
v NetApp
v Exchange
v SharePoint
v Documentum
v Celerra
v Discovery
Accelerator
v Domino
v Chatter

Creating volumes and data sources 51


Applicable Volume
Field Name Required Action Special Notes Type

v These directories are defined as v Windows Share


Include In Include directories, specify a
sets of first node directories, (CIFS)
directories regular expression for included
relative to the specified (or v NFS v2 and v3
directories for each harvest (if it
or implied) starting directory, that
was specified). v NetApp
are considered part of the logical
Include research volume. v Exchange
folders v For Discovery Accelerator, this v SharePoint
regular expression helps restrict v Documentum
the volume to one or more v Celerra
Research Folders in the case.
v Discovery
Accelerator
v Domino
v FileNet
v Chatter
v Desktop

v Windows Share
Start directory In Start directory, designate a start
(CIFS)
directory for the harvest. The start
directory involves volume v NFS v2 and v3
partitioning in order to break up a v NetApp
large volume. The start directory v SharePoint
must be underneath the initial
v Domino
directory, if an initial directory is
defined. For example, in the case of v Celerra
directories EH, E would be the v FileNet
Start directory and H would be the v Chatter
End directory.

v Windows Share
End directory In End directory, determine the end
(CIFS)
directory for the harvest. The end
directory is also part of volume v NFS v2 and v3
partitioning and is the last directory v NetApp
harvested. v SharePoint
v Domino
v Celerra
v FileNet
v Chatter

v Windows Share
Access Times In Access times, select one of these
(CIFS)
options:
v NFS v2 and v3
v Reset access times but do not
synchronize them. (This is the v NetApp
default setting.) v Domino
v Do not reset or synchronize v Celerra
access times.
v Reset and synchronize access
times on incremental harvests.

52 Administration Guide
Applicable Volume
Field Name Required Action Special Notes Type

v Windows Share
Volume In Constraints, select one of these The Scope harvests on these
(CIFS)
constraints options: volumes by extension option
pertains only to these volume types: v NFS v2 and v3
v Only use __ connection process
or v NetApp
(es)Specify a limit for the v Windows Share
Constraints number of harvest connections to v NFS v2 and v3 v Exchange
this volume. If the server is also v Domino
v NetApp
being accessed for attribute and
v Celerra v Celerra
full-text searches, you may want
to regulate the load on the server v Discovery Accelerator v SharePoint
by limiting the harvester v Documentum
processes. The maximum number v Enterprise Vault
of harvest processes is
automatically shown. This v Discovery
maximum number is set on the Accelerator
system configuration tab. v FileNet
v Control the number of parallel v Chatter
data object readsDesignate the v Desktop
number of parallel data object
reads.
v Scope harvests on these volumes
by extensionInclude or exclude
data objects based on extension.

Configuring Exchange 2007 Client Access Server support


The system supports the harvest of multiple Client Access Servers (CAS) when
configuring Exchange 2007 primary volumes. This feature does not support
redirection to other CAS/Exchange clusters or autodiscovery protocol.

To include Client Access Servers in Primary Exchange 2007 Volumes:


1. Go to Administration > Data sources > Volumes > Primary > Add primary
volumes.
2. In the Server type list, select Exchange.
3. In the Version list, select 2007.
4. In the Server text box, type the name of the Exchange server. This server must
be load-balanced at the IP or DNS level.
5. In the Mailbox server: text box, enter the name(s) of one or more mailbox
servers, separated by a comma and a space.
6. Complete the remaining fields for the primary volume, and then click OK.
Related tasks:
Creating primary volumes on page 46

Adding Domino as a primary volume


This topic provides procedural information regarding how to add Domino as a
primary volume.

To add Domino as a primary volume:


1. Add a Lotus Notes user by uploading its user ID file in Lotus Notes User
Administration on the Administration > Configuration tab.
v If you would like to harvest a users mailbox, add the user ID file for that
user.

Creating volumes and data sources 53


If you would like to harvest multiple mailboxes within one volume
v
definition, add the administrators ID file.
v If the mailboxes have encrypted emails or NSFs, then you will need each
users user ID file in order to decrypt a given users data.
2. Point the volume to the Domino server. If a single mailbox must be harvested,
set the initial directory to be the path to the mailbox on the Domino server,
such as mail\USERNAME.
3. To harvest mailboxes, select the Harvest mailboxes option, which obtains the
list of all known Domino users and their NSFs. It then harvests those mailboxes
unless it was pointed to a single mailbox using the initial directory.
4. To harvest all mail journals, select the Harvest mail journals option.
5. To harvest all mail applications, select the Harvest all applications option,
which looks at all NSFs, including mail journals, on the Domino server.
Related tasks:
Creating primary volumes on page 46

Special note: adding SharePoint volumes


This topic provides conceptual information that must be used when adding
SharePoint volumes.

IBM StoredIQ Platform supports using the entire sites portion of a Sharepoint URL
for the volume /sites/main_site/sub_site in the Volume field when adding a
new SharePoint volume. However, if the SharePoint volume URL contains spaces,
then you must also utilize the Server, Volume, and Initial directory fields in the
Add volume dialog box. In addition to the required fields Server type, Server,
Connect as, and Password. For example, the SharePoint volume with the URL
http://shpt2010.reglab5.local/sitestest/autoteamsite1/Attribute Harvest
WikiPages Library/ would require using the fields in the following table because
of the spaces in the URL.

Primary Volume
Field Example

Server shpt2010.reglab5.local

Volume /sitestest/autoteamsite1

Initial directory Attribute Harvest Wiki Pages Library

Performance considerations for using versioning

When you add a primary volume, you define the volume by setting certain
properties. If a SharePoint volume is added, you have the option of indexing
different versions of data objects on that volume.

Since most versions of any given object will share full-text content and attributes,
there is a duplication of effort in processing them as well as the effort required to
maintain an updated context for the version history of an object in the index.
Additionally, if you enable versioning on a SharePoint volume, the API itself
causes extra overhead in fetching data and metadata for older versions.
v For each object, an additional API call must be made to get a list of all its
versions.

54 Administration Guide
v To fetch attributes for the older versions of an object, an API call must be made
for each attribute that needs to be indexed.
Related tasks:
Creating primary volumes on page 46

Configuring primary volumes using Enterprise Vault


Prior to creating primary volumes using Enterprise Vault, you must configure
Enterprise Vault sites so that certain configuration items can appear in the primary
volume configuration lists. When defining the Enterprise Vault site, DCOM
configuration is a prerequisite.

To configure an Enterprise Vault primary volume:


1. Go to Administration > Data sources > Volumes > Primary > Add primary
volumes.
2. In the Server Type list, select Enterprise Vault.
3. Click OK to save the volume, and then select one of the following:
v Add another volume on the same server
v Add another volume on a different server
v Finished adding volumes
Related tasks:
Enterprise Vault on page 37
Creating primary volumes on page 46

About Editing Volume Definitions


This topic provides conceptual information regarding the editing of volume
definitions.

Whenever volume definitions are edited or modified, you should restart the
system.
Related tasks:
Creating primary volumes on page 46
Restarting and rebooting IBM StoredIQ Platform on page 15

Creating retention volumes


This section provides both conceptual and procedural information regarding
retention volumes.

Retention volumes store data objects that have been placed under retention,
meaning that the object will be retained.

This is the process for using retention volumes to store such data:
1. Configure your retention servers. See Configuring Retention Servers.
2. Create management or retention classes, or, if you are using Centera retention
servers, create Centera pools as described in Creating Centera Pools.
3. If you are using Enterprise Vault, ensure you have defined Enterprise Vault
Sites (see Discovery Accelerator). You may also want to define Retention
Categories on the Enterprise Vault server.
4. Create retention volumes, as detailed in Adding a Retention Volume.
Related concepts:
Creating volumes on page 46

Creating volumes and data sources 55


Creating volumes and data sources on page 33

Adding a retention volume


This topic provides procedural information regarding how to add a retention
volume.

To add a retention volume:


1. Go to Administration > Data sources > Volumes, and then click Retention.
2. Depending on the type of retention server you are adding, complete the fields
as described in this table.
3. Click OK to save the volume.

Note: Case-sensitivity rules apply. Red asterisks within the user interface
denote required fields.

Applicable
Field Name Required Action Special Notes Volume Type

v Windows
Server type In the Server type list, select the
Share (CIFS)
server type.
v NFS v3
v Centera
v Hitachi
v IBM
Information
Archive
v Dell DX
Storage
v Enterprise
Vault
v NetApp
SnapLock
v NetApp
v FileNet

v Standard
Platform In the Platform list, select the
platform type.

v Centera
Pool In the Pool list, select the StoredIQ
pool profile name to provide access
to a specific Centera pool

Site In the Site list, select the appropriate


site.

Vault store In the Vault store list, select the


desired vault store for the volume.

Dell DX Storage In the Dell DX Storage Cluster list,


Cluster select the appropriate Dell DX
Storage Cluster.

56 Administration Guide
Applicable
Field Name Required Action Special Notes Volume Type

v Windows
Server In the Server text box, assign the
Share (CIFS)
server a name.
v NFS v3
v Hitachi
v IBM
Information
Archive
v NetApp
SnapLock

v IBM
Node name In the Node name text box, enter
Information
the name of the node.
Archive

v IBM
Node port In the Node port text box, enter the
Information
nodes port number.
Archive

v IBM
Node password In the Node password text box,
Information
enter the password for the node.
Archive

v FileNet
FileNet config In the FileNet config text box, enter
the name of the FileNet connection.

v Windows
Connect as In the Connect as text box, enter the
Share (CIFS)
login ID.
v IBM
Information
Archive
v FileNet

v Windows
Password In the Password text box, enter the
Share (CIFS)
password for the login ID.
v IBM
Information
Archive
v NetApp
SnapLock
v FileNet

v FileNet
Object store In the Object store list, select the The object store
desired object store. must exist prior
to the creation
of a FileNet
retention
volume.

Creating volumes and data sources 57


Applicable
Field Name Required Action Special Notes Volume Type

v Windows
Volume In the Volume text box, enter the
Share (CIFS)
name or names of the volume to be
mounted. v NFS v3
v Centera
v Hitachi
v IBM
Information
Archive
v Dell DX
Storage
v NetApp
SnapLock

v Windows
Index options Select either or both of the Index These options
Share (CIFS)
options check boxes. are selected by
v NFS v3
v Include system metadata for data default.
objects within containers. v Centera
v Include content tagging and v Hitachi
full-text index. v IBM
Information
Archive
v NetApp
SnapLock

v Enterprise
Matter archive In the Matter archive name text box,
Vault
name enter the name of the matter archive.

v Enterprise
Description In the Description text box, enter a
Vault
description.

Default In the Default retention category


retention list, select the default retention
category category.

Retention Select the Allow retention category


category to be overridden on policy.
override

Object creation Select the option to create an object FileNet


as a StoredIQ document.

58 Administration Guide
Applicable
Field Name Required Action Special Notes Volume Type
Note: For v Windows
Constraints In Constraints, select either or both
FileNet retention Share (CIFS)
of these options: volumes, you
v NFS v3
v Only use __ connection can select only
process(es)Specify a limit for the Only use __ v Centera
the number of harvest connections connection v Hitachi
to this volume. If the server is process(es) v IBM
also being accessed for attribute option. Information
and full-text searches, you may Archive
want to regulate the load on the
v NetApp
server by limiting the harvester
SnapLock
processes. The maximum number
of harvest processes is v FileNet
automatically shown. This
maximum number is set on the
system configuration tab.
v Control the number of parallel
data object readsDesignate the
number of parallel data object
reads.

Related concepts:
Creating retention volumes on page 55

Configuring Enterprise Vault retention volumes


This topic provides procedural information regarding how to configure Enterprise
Vault retention volumes.

To configure an Enterprise Vault retention volume:


1. Go to Administration > Data sources > Volumes > Retention.
2. Click Add retention volumes.
3. In the Server type list, select Enterprise Vault.
4. In thenSite list, select the Enterprise Vault site you created. See Enterprise
Vault.
5. Enter the information described in Creating Retention Volumes, based on your
server type.
6. Click OK to save the volume.
7. Select one of the following:
v Add another volume on the same server
v Add another volume on a different server
v Finished adding volumes

Creating discovery export volumes


This topic provides procedural information regarding how to create a discovery
export volume.

Discovery export volumes contain the data produced from a policy, which is kept
so that it can be exported as a load file and uploaded into a legal review tool.
Administrators can also configure discovery export volumes for managing harvest
results from cycles of a discovery export policy.

Creating volumes and data sources 59


To create a discovery export volume:
1. Go to Administration > Data sources > Specify volumes, and then click
Volumes.
2. Click Discovery export, and then click Add discovery export volumes.
3. Enter the information described in the table below, and then click OK to save
the volume.

Note: Case-sensitivity rules for each server type apply. Red asterisks within the
user interface denote required fields.

Applicable
Special Volume
Field Name Required Action Notes Type

v Windows
Type Using the Type list, select the type of server.
Share
(CIFS)
v NFS v2,
v3

v Windows
Server In the Server text box, enter the name of the
Share
server where the volume is available for
(CIFS)
mounting.
v NFS v2,
v3

v Windows
Connect as In the Connect as text box, enter the logon ID
Share
used to connect and mount the defined
(CIFS)
volume.

v Windows
Password In the Password text box, enter the password
Share
used to connect and mount the defined
(CIFS)
volume.

v Windows
Volume In the Volume text box, enter the name of the
Share
volume to be mounted.
(CIFS)
v NFS v2,
v3

v Windows
Constraints To utilize Constraints, select this option:
Share
v Only use __ connection process (CIFS)
(es)Specify a limit for the number of
v NFS v2,
harvest connections to this volume. If the
v3
server is also being accessed for attribute
and full-text searches, you may want to
regulate the load on the server by limiting
the harvester processes. The maximum
number of harvest processes is
automatically shown. This maximum
number is set on the system Configuration
tab.

Related concepts:
Creating retention volumes on page 55

60 Administration Guide
Creating system volumes
This topic provides procedural information regarding how to create a system
volume.

System volumes support volume export and import. When you export a volume,
data is stored on the system volume. When you import a volume, data is imported
from the system volume.

To add a system volume:


1. Go to Administration > Data sources > Specify volumes > Volumes.
2. Select the System tab, and then click Add system volumes.
3. Enter the information described in the table below, and then click OK to save
the volume.

Note: Case-sensitivity rules apply. Red asterisks within the user interface
denote required fields.

Applicable
Field Name Required Action Special Notes Volume Type

v Windows
Server type Using the Type list, select the type
Share (CIFS)
of server.
v NFS v2, v3

v Windows
Server In the Server text box, enter the
Share (CIFS)
name of the server where the
volume is available for mounting. v NFS v2, v3

v Windows
Connect as In the Connect as text box, enter the
Share (CIFS)
logon ID used to connect and
mount the defined volume.

v Windows
Password In the Password text box, enter the
Share (CIFS)
password used to connect and
mount the defined volume.

v Windows
Volume In the Volume text box, enter the
Share (CIFS)
name of the volume to be mounted.
v NFS v2, v3

v Windows
Constraints To utilize Constraints, select this
Share (CIFS)
option:
v NFS v2, v3
v Only use __ connection process
(es)Specify a limit for the
number of harvest connections to
this volume. If the server is also
being accessed for attribute and
full-text searches, you may want
to regulate the load on the server
by limiting the harvester
processes. The maximum number
of harvest processes is
automatically shown. This
maximum number is set on the
system Configuration tab.

Creating volumes and data sources 61


Related concepts:
Creating volumes on page 46
Configuring server platforms on page 33

Exporting and importing volume data


The systems export and import volume capabilities allow metadata and full-text
indexed data to be collected or exported from separate locations, such as data
servers located in various offices of the enterprise. Once the data is available, it can
be imported to a single location, such as a headquarter's office data server, where
selected files might be retained.

Only primary and retention volume data can be exported or imported using the
export/import feature. Discovery export and system volumes cannot be imported
or exported. The target location of an export or the source location of an import is
always the IBM StoredIQ Platform system volume.

Export and import volume processes run as jobs in the background. These jobs are
placed into their prospective queues, and they are executed sequentially. When one
job completes, the next one automatically starts. These jobs can be cancelled at any
time while they are running. Cancelling one import or export job also cancels all
the jobs that come after the one cancelled. Because the export jobs and import jobs
reside in separate queues, cancelling one type of job does not cancel jobs in the
other queue. The jobs are not restartable.

Volume data export to a system volume


This topic provides procedural information regarding exporting volume data to a
system volume.

The export process creates two files: a binary file and a metadata file, which
contains the exported data. These files' names contain the following information:
v Data server name and IP address
v Volume and server names
v Time stamp

The exported data consists of data from the selected volume and any related
information that describes that data with the exception of volume-specific audits.

The exported data has to be made available to the import data server before it can
be imported. This may require you to physically move the exported data to the
system volume of the import data server.

Licenses on the import appliances are enabled automatically if a feature of the


imported volume requires it (such as Centera or Exchange licenses).

To export a volumes data to the system volume:


1. Go to DSAdmin > Administration > Data sources > Volumes.
2. Select a volume of data to export from the list of volumes by clicking the
export link in the far right-hand column.
3. Complete the Export volumes details dialog, described in this table.
4. Click OK. A dialog appears to let you know that the data is being exported.
5. To monitor the export progress, click the Dashboard link. To cancel the export
process, under the Jobs in progress section of the dashboard, click Stop this
job. Alternately, click OK to return to the Volumes page.

62 Administration Guide
Note: The job cannot be restarted.

Name Description

Server Name of the server where the data resides.

Volume Name of the volume where the data resides.

Export path (on system Where to save the data on the System volume. The default
volume) path is /exports. You can edit the export path. The specified
location will be automatically created if necessary.

Description optional Enter a description of the exported data.

Export full-text index Select this option to export the volumes full-text index.
(Available only if the volume has full-text index.)

Volume data import to a system volume


This topic provides procedural information regarding importing volume data to a
system volume.

An imported volume looks, acts, and is, just like a volume originally defined and
harvested on the data server. Any action or relationship that is valid for a
non-imported volume is valid for an imported volume, with a few exceptions:
v Logs and audit trails that capture the activity on the volume before the import
are not available. However, the import itself is audited. See Import Audits.
v The imported volume can be reharvested as long as the appliance has the proper
network access and rights to the original source server and volume.
v The imported volume can be reharvested as long as the data server has the
proper network access and rights to the original source server and volume.
v The data viewer works only if the appliance has the proper network access and
rights to the source server and volume. You must have access and permission on
export servers and volumes if the file you want to view has been migrated to a
secondary server at the time of the export.

Note: When a volume with a licensed feature is imported into a data server that
does not utilize licensing, the license is imported along with the volume. To see the
licensed features, users will need to log out and then log back in to the data server.

To import data from the system volume:


1. Make sure that exported data file is present in the system volume of the import
appliance.
2. Go to Administration > Data sources > Volumes, and then click either the
Primary or Retention tab.
3. Click the Import volume link at the top of either the primary or the retention
volume lists. The Import volumes page appears, listing all of the volumes
available for import. By default, the data server searches for available volumes
to import in the /imports directory of the system volume. If you have placed
the exported data to another path, click Change path and enter the appropriate
path.
4. Click OK. The Import volumes page now displays the following information
about the imported volumes. See the table below.

Creating volumes and data sources 63


5. From the list of volumes, select a volume to import by clicking the Import link
in the last column on the right.
6. Select the Import full-text index check box to import the selected volume's
full-text index. (The check box is active only if full-text index is available.)
7. Select the Overwrite existing volume check box to replace the existing data of
the volume with the imported data.
8. Click OK. A dialog appears to let you know that the volume is being
imported. The OK button is not enabled if the volume already exists and
unless the Overwrite existing volume option is selected.
9. To view import progress, click the Dashboard link in the dialog. To cancel the
import process, under the Jobs in progress section of the dashboard, click Stop
this job, or click OK to return to the Manage volumes page.

Name Description

Server and volume Server name and volume name where the data physically
resides

Description A description added when the volume was exported

Volume type Exchange, SharePoint, and so on

Category Primary or Retention

Exported from Server name and IP address of the server from which the
data was exported

Export date The day and time the data was exported

Total data objects Total number of data objects exported for the exported
volume

Contains full-text index Whether or not the full-text index option was chosen when
the data was exported

Related concepts:
Creating volumes on page 46

Deleting volumes
Administrators can delete volumes from the list of available data sources, provided
that the data server is connected to the gateway.

Note the following regarding deleted volumes:


v Deleted volumes are removed from target sets.
v Deleted volumes are removed from all volume lists, both from IBM StoredIQ
Platform and IBM StoredIQ Administrator.
v Within created jobs, steps that reference deleted volumes are implicitly removed,
meaning that a job could contain no steps. The job itself is not deleted.
v Applicable object counts and sizes within IBM StoredIQ Administrator will
adjust automatically.

64 Administration Guide
v Object counts and sizes within user infosets will remain the same. Remember,
those user infosets were created at a specific point in time when this data source
was still available.
v Users exploring a specific data source and any generated reports will no longer
reference the deleted volume.
v No exceptions will be raised on previously executed actions. Instead, the data is
no longer available. For example, if an infoset is copied that contained data
objects from a volume that has been deleted, no exception is raised.
v If you mark a desktop volume for deletion, it will automatically be removed
from the Primary volume list; however, the status of that workstation is set to
uninstall in the background. When the desktop client next checks in, it will see
that change in status and uninstall itself.

Note: If retention volumes such as Centera, File Net, Hitachi, and so on contain
data, they cannot be deleted as IBM StoredIQ Platform is the source of record.
Instead, you will see the Under Management link.

To delete a volume:
1. Go to Administration > Data sources > Specify volumes > Volumes.
2. Click the tab of the volume type you would like to delete: Primary, Retention,
System, or Discovery export.
3. Click Delete, and in the confirmation dialog, click OK. The volume is deleted,
removing it from the list of available volumes.

Policy limitations for volume types


This conceptual topic denotes which volume types have policy limitations.

StoredIQ imposes some policy limitations on volume types, which are identified in
this table.
Table 1. Policy limitations for volume types

Policy Type Source/Target Limitations Other Restrictions

Copy from Centera, Windows Share SnapLock, NFS


SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Exchange, Enterprise Vault (Discovery
Accelerator), Dell DX Storage, FileNet ,
NewsGator

Creating volumes and data sources 65


Table 1. Policy limitations for volume types (continued)

Policy Type Source/Target Limitations Other Restrictions

Copy to Primary Volume


v CIFS
v NFS
v SharePoint
v Documentum

Retention Volume
v Centera
v NetApp SnapLock (CIFS/NFS)
v CIFS Celerra FLR (CIFS/NFS)
v Dell DX Object Store
v CIFS
v NFS
v IBM FileNet
v Documentum /w RPS
v Hitachi HCAP
v IBM Information Archive
v Symantec Enterprise Vault
Copy to Centera, Windows Share SnapLock, NFS
(Retention) SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Enterprise Vault, Dell DX Storage, FileNet

Discovery Export Centera, Windows Share SnapLock, NFS


from SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Exchange, Enterprise Vault (Discovery
Accelerator), Dell DX Storage, NewsGator

Discovery Export Windows Share, NFS Considered


to Discovery Export
volumes (category)
not harvested

Move from Centera, Windows Share SnapLock, NFS


SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Dell DX Storage

Move to Centera, Windows Share SnapLock, NFS


SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, SharePoint,
Dell DX Storage

Delete Centera, Windows Share SnapLock, NFS


SnapLock, Hitachi, IBM Information Archive,
Windows Share, NFS, Documentum, Dell DX
Storage

66 Administration Guide
Table 1. Policy limitations for volume types (continued)

Policy Type Source/Target Limitations Other Restrictions

Modify Security Choose single


volume (no volume
sets)

Related concepts:
Creating volumes on page 46

Creating volumes and data sources 67


68 Administration Guide
Harvesting data
This section provides information regarding harvesting data.

Understanding Harvests
This conceptual topic will help you to understand the harvest types available.

Harvesting (indexing) is the process or task by which IBM StoredIQ Platform


examines and classifies data in your network. Running an out-of-the-box Harvest
every volume job indexes all data objects on all volumes.
v A full harvest can be run on every volume or on individual volumes.
v An incremental harvest only harvests the changes on the requested volumes

These options are selected when you create a job for the harvest. A harvest must be
run before you can start searching for data objects or textual content. An
Administrator initiates a harvest by including a harvest step in a job.

Most harvesting parameters are selected from the Configuration subtab (see
Configuring Application Settings); you can specify the number of processes to use
during a harvest, whether a harvest should continue where it left off if it has been
interrupted, as well as many other parameters. There are also several standard
harvesting-related jobs that are provided in the IBM StoredIQ Platform system.

Harvesting with and without post-processing

You can separate harvesting activities into two steps: the initial harvest and harvest
post-processing. The separation of tasks gives Administrators the flexibility to
schedule the harvest or the post-process loading to run at times that do not impact
system performance for system users, who may, for example, be running queries.
These are examples of post-harvest activities:
v Loading all metadata for a volume.
v Computing all tags that are registered to a particular volume.
v Generating all reports for that volume.
v If configured, updating tags, and creating explorers in the harvest job.

Incremental harvests

Harvesting volumes takes time and taxes your organization's resources. You can
maintain the accuracy of the metadata repository quickly and easily with
incremental harvests. With both of these features, you can ensure that the
vocabulary for all volumes is consistent and up to date.

Once you have harvested a volume, you can speed up subsequent harvests by only
harvesting for data objects that have been changed or are new. An incremental
harvest indexes new, modified, and removed data objects on your volumes/file
servers.

Because the harvests are incremental, it takes less time to update the metadata
repository with the additional advantage of putting a lighter load on your systems
than the original harvests.

Copyright IBM Corp. 2001, 2013 69


Note: Harvesting NewsGator VolumesSince NewsGator objects are really just
events in a stream, an incremental harvest of a NewsGator volume only fetches
new events that have been added since the last harvest. To cover gaps due to
exceptions or to pick up deleted events, a full harvest may be required.
Related tasks:
Configuring harvester settings on page 26

Harvesting properties and libraries


This topic provides conceptual information regarding SharePoint volumes and their
permissions for mounting.

Note: Administrative permissions are required in order to harvest personal


information, libraries, and objects that have not been designated as being visible to
Everyone for user profiles.

The SharePoint volume needs to be mounted using administrative permissions. If


the harvest is performed without administrative permissions, then any of the user
profiles properties that SharePoint users marked as visible to the category other
than Everyone will not be visible in results. In order to harvest users personal
documents and information, volumes mounted without administrative permissions
must use credentials that have full control on all SharePoint site collections hosted
by the user profile service application.

To override this restriction, see http://technet.microsoft.com/en-us/library/


ee721057.aspx.

Performing a lightweight harvest


This section describes lightweight harvest as well as providing possible system
configuration suggestions.

IBM StoredIQ Platform allows you to perform many types of harvests, depending
on your data needs. While in-depth harvests are common, there are also instances
where you need an overview of the data and a systemwide picture of files types
and sizes. For example, at the beginning of a deployment, you may want to obtain
a high-level view of a substantial amount of data so that you can make better,
more informed decisions about how you want to handle harvesting or other
policies going forward. This section provides possible system configurations that
will allow the system to process the volumes data in the quickest manner possible.

Lightweight harvest parameters


The sections here denote only configuration changes that could be made in order
to perform a lightweight harvest.

Determining volume configuration settings


This topic denotes possible volume configuration changes that could be made in
order to perform a lightweight harvest.

Prior to modifying volume details, read Creating Primary Volumes.

Volume Details

When configuring data sources for a lightweight harvest, you do not need to
include content tagging and full-text indices. By clearing this option, the system

70 Administration Guide
will index the files metadata, not the entire content of those files. The system can
then execute and complete harvests very quickly, and you will obtain a large
amount of information regarding file types, the number of files, the age of the files,
file ownership, and so on.

To configure a volume for a lightweight harvest:


1. Go to Administration > Data sources > Specify volumes > Volumes.
2. On the Primary volume list page, click Add primary volumes or Edit to edit
an existing volume.
3. Verify that all of the Index options check boxes are cleared (some are selected
by default).
4. Click OK and then restart services. Show Advanced Details
In some cases, you may want to reduce the weight of a full-text harvest. In
these instances, you can simply adjust the processing involved with the various
harvest configuration controls.
Within volume configuration, the advanced settings are used to control what is
harvested within the volume. By only harvesting the directory structures in
which you are interested, you can exercise some control over the harvests
weight.
v Include DirectoryIf there is a subtree of the volume that you wish to
harvest rather than the whole volume, then you can enter the directory here.
This will eliminate the harvest of objects that you have already determined
are not relevant to your project.
v Start Directory/End DirectoryThis setting allows you to select a beginning
and end range of directories that will be harvested. Enter the start and end
directories as required.
v Constraints/Scope harvest by extensionThese options allow you to limit
the files harvested through connection processes, parallel data objects, or
scoping harvests by extension. For example, the Scope harvest by extension
setting allows you to limit the files that you harvest by using a set of
extensions. If you want to harvest only Microsoft Office files, you can
constrain the harvest to .DOC, .XLS, and .PPT files.

Note: This must be done from the Configuration tab.


Related tasks:
Creating primary volumes on page 46

Determining harvester configuration settings


This topic denotes possible harvester configuration changes that could be made in
order to perform a lightweight harvest.

Note: Prior to modifying the harvester settings, read Configuring Harvester


Settings.
1. Determine Skip Content Processing settings.

Note: This setting is relevant only for full-text harvests.


You may have a large number of files that are types for which you do not need
the contents, an example being .EXE files. In these instances, you can add these
file types to the list of files for which the content is not processed. There are
two points to consider when skipping content processing:
v You do not spend time harvesting unnecessary objects, which can be
beneficial from a time-saving perspective.

Harvesting data 71
v At a later date, you have the option of viewing the content of the skipped
files. This will create additional work, reharvesting these skipped files.
2. Determine which Locations to ignore.
There may be instances where large quantities of data are contained in
subdirectories, and that data is not relevant to your harvest strategy. For
example, you could have a directory with a tree of source code or software
archive that is not used as a companywide resource. In these cases, you can
eliminate these directories from the harvests by adding the directory to the
Locations to ignore. These locations are not specific to a volume, but can
instead be used for common directories across volumes.
3. Determine Limits.
v Maximum data object sizeThis setting is only relevant for full-text
harvests. In cases where there are a lot of very large files, you may want to
eliminate processing those files by setting the Maximum data object size to a
smaller number (the default value is 1,000,000,000). You will still collect the
metadata on the very large files, so you can search for them and determine
which files were missed due to the setting of this parameter.
4. Determine Binary Processing.
Binary processing is additional processing that can be performed if the
standard processing cannot index the contents of a file. For lightweight
harvests, the Run binary processing when text processing fails check box
should be cleared as this setting is only relevant for full-text harvests.

Determining full-text settings


This topic denotes possible full-text index configuration changes that could be
made in order to perform a lightweight harvest.

Note: Prior to modifying the full-text settings, see Configuring Full-text Index
Settings.

The full-text settings are only valid if you have full-text processing enabled for a
given volume. Consider these options for controlling the impact of a full-text
harvest on the systems performance.
1. Determine Limits. Limit the length of words to be harvested by selecting the
Limit the length of words index to __ characters option. The default value is
50, but you can reduce this number in order to reduce the quantity of indexed
words.
2. Determine Numbers. If there are large quantities of spreadsheet files, you can
control what numbers are indexed by the system.

Determining hash settings


This topic denotes possible hash setting configuration changes that could be made
in order to perform a lightweight harvest.

Note: Prior to modifying the hash settings, see Configuring Hash Settings.

Determine the hash settings. A files hash is a unique, calculated number based on
the content of the file. By selecting Partial data object content, you reduce the
required processing to create the hash. Be forewarned, however, that two different
data objects could create the same hash. This is a small but potential risk. This is
only relevant for full-text harvests.

72 Administration Guide
Configuring jobs
This section provides procedural information about configuring jobs within IBM
StoredIQ Platform.

Types of IBM StoredIQ Platform jobs


This topic lists and describes the types of IBM StoredIQ Platform jobs that are
available.

Jobs start tasks such as harvests. They can be run at the time of creation, or
scheduled to run at a designated future time and at regular intervals.

Jobs consist of either a single step or a series of steps. The actions available at each
step depend on the type of job being created. There are several of out-of-the-box
jobs included in IBM StoredIQ Platform; these and their locations in the interface
are described in this table.

StoredIQ Job Type Description


Centera deleted files This is an unscheduled, one-step job that synchronizes the
synchronizer deleted Centera files. This job is located in the Library/Jobs
folder.
Windows Share (CIFS)/NFS This is an unscheduled, one-step job that harvests the
retention volume deleted Windows Share/NFS retention volumes, looking for files that
files synchronizer require removal because the physical file has been deleted
from the retention file system. This job is located in the
Library/Jobs folder.
Database Compactor This is a scheduled job that helps to limit bloat
(unnecessary storage usage) in the database. While this job
runs, it must have exclusive, uninterrupted access to the
database. Administrators can override this job by logging in
and then proceed to use the system. This job is located in the
Library/Jobs folder.
Enterprise Vault retention This is an unscheduled, one-step job that harvests the
volume deleted files Enterprise Vault retention volumes, looking for files that
synchronizer require removal because the physical file has been deleted
from the retention file system. This job is located in the
Library/Jobs folder.
FileNet retention volume This is an unscheduled, one-step job that harvests the FileNet
deleted files synchronizer retention volumes, looking for files that require removal
because the physical file has been deleted from the retention
file system. This job is located in the Library/Jobs folder.
Harvest every volume This is an unscheduled, one-step job that harvests all
primary and retention volumes. This job is located in the
Workspace/Templates/Jobs folder.
Hitachi deleted files This is an unscheduled, one-step job that harvests the
synchronizer Hitachi volumes, looking for files that require removal
because the physical file has been deleted from the file
system. This job is located in the Library/Jobs folder.

Copyright IBM Corp. 2001, 2013 73


StoredIQ Job Type Description
IBM Information Archive This is an unscheduled, one-step job that harvests the IBM
deleted files synchronizer Information Archive volumes, looking for files that require
removal because the physical file has been deleted from the
file system. This job is located in the Library/Jobs folder.
System maintenance and This is a multistep job located in the Library/Jobs folder. The
cleanup system is configured to run a system maintenance and clean
up job once a day and includes the following:
v Email users about reports
v Email Administrators about reports
v Delete old reports
v Delete old harvests
v Load indexes
v Optimize full-text indexes.
Update Age Explorers This is a one-step job located in the Library/Jobs folder that
recalculates these items:
v Owner Explorer data for Access Date
v Owner Explorer data for Modified Date
v Created Date (API only) values

Related concepts:
Configuring jobs on page 73

Working with jobs


Defining jobs consists of two parts: naming the job and adding steps to the job.
This section details how to create and work with various IBM StoredIQ Platform
jobs.

The following tasks define how to create custom jobs.


Related concepts:
Configuring jobs on page 73

Creating a job
This topic provides procedural information regarding how to create a job.

To create a job:
1. From the Folders tab > Workspace folder, select New > Job.
2. Enter a unique job name.
3. In the Save in: list, select the appropriate folder, and the job is created.
4. Click OK. If you would like to view the job and add steps, click Yes.
5. On the View job page, click Add step and select a step type from the list.
6. For Run harvest jobs, on the Specify harvest and load options page, configure
the following options:
v Harvest these volumesSelect a volume from the list.
v Harvest typeSpecify the type of harvest:
Run a full harvest, meaning that all data objects on this volume will be
indexed.
Run an incremental harvest (default), meaning that only files or data
objects that have changed since the last harvest will indexed.

74 Administration Guide
v Harvest and load schedulingYou can separate harvest and load processes
to limit resource use. Select:
Load indexes when harvest completes.
Load indexes with next nightly system services job (This delays the
index loading to run with the next system-services job after the harvest
has completed. The system-services job is scheduled to run at midnight by
default.)
v Run harvest onlySelect this option if you plan to load harvested data into
indexes at a later time.
v Load indexes onlySelect this option to load previously harvested data into
indexes.
7. Harvest samplingSelect this option if you want to limit the harvest to a
smaller sample. This option skips every second, third, tenth, or other number
data object as entered in the text box.
8. Harvest limitsLimit the harvest by time or total number of data objects.
Enter the number of minutes or number of data objects.
9. Click OK.

When creating a job, you also have these options:


Related concepts:
Harvesting data on page 69

Creating a job to discover retention volumes


Jobs can be created with various types of steps. This topic provides procedural
information regarding how to create jobs to discover retention volumes.

Note: Prior to creating a job to discover retention volumes, you must have added
a retention volume.

To create a job to discover retention volumes:


1. From the Folders tab > Workspace folder, select New > Job.
2. Enter a unique job name.
3. In the Save in: list, select the appropriate folder, and the job is created.
4. Click OK. If you would like to view the job and add steps, click Yes.
5. On the View job page, click Add step.
6. Select Discover Retention volumes.
7. In the Discover Retention volume list, select the retention volume to be used
for this job.
8. Enter the duration that the harvest should run in Run harvest for __, and
then select Minutes or Hours.
9. Enter the number of data objects to be harvested in Only harvest __data
objects.
10. Click OK.
Related concepts:
Creating retention volumes on page 55

Editing a job
This topic provides procedural information regarding editing an existing job.

To edit a jobs steps:

Configuring jobs 75
1. From the Folders tab > Workspace folder, click the job you would like to edit.
The Job details page opens.
2. Click Edit job details, and the Edit job details dialog box appears, allowing
you to specify the time, date, and frequency for the job to run.
v In the Time: field, enter the time the job must start, or click Now to
populate the time field with the current time. You may want to add some
time if you have not specified all of the job steps.
v In the Date: field, enter the date on which to run the job, or click Today to
populate the date field with the current date.
3. Using the options in the Frequency field, specify how often the job must run.
If you select None, the job runs once, at the time and date provided.
4. Click OK.
5. To edit the job steps:
a. Add a step to by job by clicking Add step.
b. Edit an existing step by clicking Edit.
c. Remove an existing step by clicking Remove.
d. Change the order of existing steps by clicking Move up or Move down
icons.
6. Click OK.

Starting a job
This topic provides procedural information regarding starting a job.

To start a job:
1. To start a job, do either of the following:
v From the Folders tab > Workspace folder, click on the name of the job you
would like to start, and in the Job details page, click Start job.
v From the Folders tab > Workspace folder, right-click the job and select Start
job.
2. In the Schedule area, a started job will be displayed as This job is running
now.
3. Click OK to return to the Folders tab. The started job will display Running in
the Status column.

Saving a job
This topic provides procedural information regarding saving a job.

To save a job as:


1. From the Folders tab > Workspace folder, click on the name of the job you
would like to save. The Job details page opens.
2. Click Save as, and the Save job as dialog box appears.
3. In the Job name text box, verify that the jobs name is correct. Each job must
have a unique name if it is saved in the same folder as another job.
4. In the Save in list, select the folder in which you would like to save the job.
5. Click OK to close the Save job as dialog box and to return to the Folders tab.

76 Administration Guide
Running a predefined job
This topic provides procedural information regarding running a predefined job.

To run a predefined job:


1. Go to Folders > Library.
2. Click Jobs folder to open the job list.
3. Click the predefined job that you would like to edit.
4. To set the schedule, click Edit job details, completed the desired changes, and
then click OK.
5. Alternately, click Start job (selected jobs only) in the bottom right-hand area of
the pane to start the job immediately.

Deleting a job
This topic provides procedural information regarding deleting a job.

To delete a job:
1. From the Folders tab > Workspace folder, select Filter by... Jobs.
2. Click the job name to open the job details.
3. Click Delete in the lower-left hand corner of the screen. Click OK.

Monitoring processing
You can track the systems processing on your harvest/policy and discovery export
tasks using the View cache details feature. The appliance gathers data in
increments and caches the data as it gathers it. If a collection is interrupted, the
appliance can resume collection at the point that it was interrupted, instead of
starting over from the beginning of the task.

To monitor processing:

From Administration > Dashboard, in the Appliance status pane, click View
cache details. The View cache details page appears. To see harvest/policy progress,
click the Volume cache tab, or to see discovery export job progress, click the
Discovery export cache tab.

Note: Note that information for a job is only available while the job is running.
Once a task is completed, the job disappears from the list.
Table 2. Harvest/Volume Cache Details

Column Description Values

Name The name of the volume


being harvested

Start date The time the job started

v Copy
Type Type of job being executed
v Harvestfull
v Harvestincremental

Configuring jobs 77
Table 2. Harvest/Volume Cache Details (continued)

Column Description Values

v CachingThe volume
State Status of the process
cache is currently being
created/updated by a
harvest or policy
v CachedCreation or
update of volume cache is
complete (harvest only)
v LoadingVolume cache
contents are being
successfully loaded into
the volume cluster

Full-text Indicates whether a full-text Yes or No


harvest is being performed

View audit link details Link to the harvest/policy


audit page

Table 3. IBM StoredIQ Platform Discovery Export Cache Details

Column Description Values

Name Name of the volume being


processed

Start date Starting date/time for the


process

Type The type of file being Discovery export


prepared for export

State Status of the discovery AbortedDiscovery export


export job policy has been cancelled or
deleted by the user.

CachingThe volume cache


is currently being
created/updated by a
harvest or policy.

CachedCreation or update
of volume cache is complete
(harvest only).

LoadingVolume cache
contents are being
successfully loaded into the
volume cluster

Full-text Indicates whether a full-text Yes or No


harvest is being performed

Related concepts:

78 Administration Guide
Configuring jobs on page 73

Deleting a volume cache


This topic provides procedural information regarding the deletion of a volume
cache.

To delete a volume cache:


1. From the volume cache or discovery export cache list, select the check box of
the cache you would like to delete.
2. Click Delete. A confirmation dialog appears. Click OK.

Determining if a harvest is stuck


The speed of a harvest is dependent on volume size and processing speed;
however, harvests do occasionally become stuck and are unable to complete
successfully. Use the procedures outlined here to troubleshoot the harvest process
and determine if and when to contact technical support.

To troubleshoot a harvest:
1. Click Administration > Dashboard > Jobs in Progress to verify that your job
continues to run.
2. In Jobs in Progress, note the Total data objects encountered number.
3. Wait 15 minutes, letting the harvest continue to run.
4. Note the new value the Total data objects encountered, and then compare it to
that value denoted previously.
5. Go to Question 1: Is the Total data objects encountered counter increasing?
Question 1: Is the Total data objects encountered counter increasing?
v YesIf the number of encountered data objects continues to increase, then
the harvest is running correctly.
v NoIf the number of encountered objects remains the same, then go to
Question 2: Is the load average up?
Question 2: Is the load average up?
a. To view load averages, on Appliance status > About appliance > View
details > System services, look at the load averages in the Basic system
information area.
v YesIf the load averages number is up, the harvest may be stuck. Call
technical support to report that the harvest is stuck on files.
v NoThe job is not really running, meaning that the job must be restarted.
Go to Question 3: Did the job complete on the second pass?
Question 3: Did the job complete on the second pass?
v YesIf the job completed successfully after it was restarted, then the harvest
is not stuck.
v NoThe job did not complete successfully. Call technical support to report a
job that does not complete.

Configuring jobs 79
80 Administration Guide
Utilizing desktop collection
When configuring desktop settings, you are enabling or disabling encryption
within IBM StoredIQ Platform. The IBM Desktop Data Collector (desktop client or
client) enables desktops as a volume type or data source, allowing them to be used
just as other types of added data sources. The IBM Desktop Data Collector is
provided as a standard MSI file and is installed according to the typical method
(such as Microsoft Systems Management Service (SMS)) used within your
organization. The IBM Desktop Data Collector can collect PSTs and .ZIP files as
well as other data objects and is capable of removing itself once its work is
completed.

Once the desktop client has been installed on a desktop and then connects and
registers with the data server, that desktop is available as a data source within the
list of primary volumes. Additionally, while the snippet support and the snippet
step-up action are supported by IBM Desktop Data Collector, it should be noted
that a desktop cannot be the target or destination of an action.

Desktop collection processes


The following sections describe how to configure and perform desktop collection.
Related concepts:
Utilizing desktop collection

IBM Desktop Data Collector client installation


The IBM Desktop Data Collector agent works with the following operating
systems:
v Windows XP 32- and 64-bit, Service Pack 2 or later
v Windows 7 32- and 64-bit
v Windows Vista 32- and 64-bit
v Windows server 2003, 2008

Installation requires administrative privileges on the desktop.

Prior to using the IBM Desktop Data Collector, the Administrator may want to
notify end users that desktop collection is going to be performed and make them
aware of the following:
v That the desktop must be connected over the network during data collection. If
the connection is interrupted, IBM Desktop Data Collector will resume its work
from the point at which it stopped.
v That users might notice a slight change in performance speed, but that they can
continue working normally. Desktop collection will not interfere with work
processes.
v That certain actions can be taken from the tray icon: Right-click for About,
Restart, Status, and Email Logs (which packages logs into single file and
launches the email client so that the user can mail them to the IBM StoredIQ
Platform Administrator).

Note that all communications are outbound from the client. The appliance never
pushes data or requests to the desktop. The IBM Desktop Data Collector pings the

Copyright IBM Corp. 2001, 2013 81


appliance about every 60 seconds, and the Last Known Contact time statistic is
updated approximately every 30 minutes. Additionally, the IBM Desktop Data
Collector checks for task assignments every five minutes.

One can download the installer application from the application in the
Configuration tab. Also, the Administrator can temporarily disable the client
service on all desktops registered to the data server from the Configuration tab.

IBM Desktop Data Collector installation methods


During installation, the hostname and IP address of the IBM StoredIQ Platform
must be supplied. If the installation will be performed manually by end users, you
must provide this information to them using email, a text file, or another method.
The IBM Desktop Data Collector can be installed using the following methods.

Mass Distribution Method (SMS)


v The appliance id as part of the distribution configuration. This method supports
passing installation arguments as MSI properties.

Required
v SERVERACTIONNODEADDRESSIP address or hostname for the Action
node. When the install is not silent, the user is prompted for IP address or
hostname. The default is the value of this argument. This field must be entered
accurately or manual correction will be required in the desktop config file.

Optional
v SERVERACTIONNODEPORTPort number for the Agent on the Action node.
Defaults to 21000, and should only be changed when the agent will connect on a
different port that is subsequently mapped to 21000.
v NOTRAYICONSpecifies whether the agent displays the IBM Desktop Data
Collector tray icon while running. Changing the setting to 1 forces the agent to
run silently and not display a tray icon.
v MSI
v Emailing linksSend a link within an email such as file:\\g:\group\install\
Client-install.vbs. The link may be to any executable file format such as .BAT,
.VBS, or .MSI. The .BAT/.VBS formats can be used to pass client arguments to
an .MSI. The user executing the link must have administrative privileges.
v NT Logon Script, in which a .BAT file or .VBS script invokes msiexec. Examples
are given here:
/iinstall
/x {7E9E08F1-571B-4888-AC08-CEA8A076F5F9}uninstall the agent. The
product code must be present.
/quietinstall/uninstall runs silently. When specifying this option,
SERVERACTIONNODEADDRESS must be supplied as an argument.
Set WshShell = CreateObject("WScript.Shell")

WshShell.Run "%windir%\System32\msiexec.exe /i G:\group\install\desktopclient.msi


NOTRAYICON=0 SERVERACTIONNODEADDRESS=clust017.test.local /q"Set WshShell
= CreateObject("WScript.Shell")

Set WshShell = Nothing

Batch file
msiexec /i G:\group\install\desktopclient.msi NOTRAYICON=1
SERVERACTIONNODEADDRESS=clust017.test.local /q

82 Administration Guide
Related concepts:
Installing the IBM Desktop Data Collector in stealth mode
Desktop collection processes on page 81

Installing the IBM Desktop Data Collector in stealth mode


The IBM Desktop Data Collector can also be installed in stealth mode, meaning
that the IBM Desktop Data Collector icon does not appear in the system tray and
the end user will not know that data is being collected. To enable stealth node for
the IBM Desktop Data Collector, launch the command-line interface and edit the
msiexec parameters as follows:
msiexec /i \\server\volume\DesktopClient.msi NOTRAYICON=1
SERVERACTIONNODEADDRESS=address SERVERACTIONPORT=21000 /q
Related concepts:
Desktop collection processes on page 81

Configuring IBM Desktop Data Collector collection


Desktop collection configuration includes the IBM Desktop Data Collector installer
and Encrypted File System.

For procedural information on downloading the IBM Desktop Data Collector


installer, see Downloading the IBM Desktop Data Collector installer from the
application on page 31.

To configure Encrypted File System desktop collection:


1. Complete the procedure outlined in Using the Encrypted File-System Recovery
Agent.
2. Restart the service and begin collection, noting the following points:
v If the computer is not part of a domain and it is running any version of
Windows earlier than 7.0, then the user name should simply be the user
name.
v If the computer is not part of a domain and it is running Windows 7.0 or
later, then the user name should be the name of the PC and the domain.
Related concepts:
Utilizing desktop collection on page 81

Using the delete policy with the IBM Desktop Data Collector: special
notes
When you use the IBM Desktop Data Collector to delete files from a desktop, they
are removed permanently. They are not transferred to the appliance or backed up
to any other location. Consequently, you must carefully review the infoset of
affected data objects prior to executing a delete action. Your organization may use
custom applications or other files that you may not want to delete. In reviewing
the returned list, do not allow the following to be deleted.
v Anything under this directory: c:\Windows
v Anything under this directory: Documents and Settings, with these extensions:
c:\Documents and Settings\<username>\UserData\ and extension *.xml
c:\Documents and Settings\<username>\Cookies\ and extension *.txt
c:\Documents and Settings\<username>\Start Menu\Programs\ and extension *.lnk
v Executable files
Utilizing desktop collection 83
*.dll
*.exe
*.ocx
v Drivers
*.sys
*.inf
*.pnf
v Installers
*.msi
*.mst
v Important data files
*.dat
*.ini
*.old
*.cat
v These file names
desktop.ini
ntuser.dat
index.dat
ntuser.pol
ntuser.dat.log

Some important system files can be automatically excluded from harvest


operations. Enabling the Harvester setting Determine whether data objects have
NSRL digital signature excludes some well-known system files.
Related concepts:
Utilizing desktop collection on page 81

84 Administration Guide
Using Folders
This section provides both conceptual and procedural information regarding
folders and their usage.

Understanding folder types


This topic contains conceptual information regarding both the Library and
Workspace folders.

The Folders tab displays two types of folders: Library and Workspace.

Library folder

The Library folder contains the Jobs folder.

Note: This folder cannot be renamed, moved, or deleted.

Workspace folder

The Workspace folder is a custom folder that reflects your use of the system. By
default, it contains a folder entitled Templates. If you are using the system for IT
purposes, you may want to create folders for each locale or function.

Note: These folders can be renamed, moved, or deleted, and you also have the
options of setting folder security.

Creating a folder
To create a folder:
1. From the Folders tab, select New > New Folder. The Create new folder dialog
appears.
2. In the Name: field, give a name that represents the folders purpose (legal
matter, local, business unit, or the like).
3. In the Description: field, type a description for the folder.
4. In the Create in: field, use the list to select a place for the folder. Note that all
custom folders must be placed in the Workspace or a Workspace subdirectory;
you cannot add folders to the Library folder.
5. Click OK. If you wish to open the folder, click OK in the dialog that appears.
Related concepts:
Understanding folder types

Deleting a folder
When deleting folders, note that only empty folders can be deleted.

To delete a folder:
1. From within the Folders tab, Workspace folder, do either of the following:
a. Select the check box next to the folder you want to delete, and in the
Actions list, select Delete.

Copyright IBM Corp. 2001, 2013 85


b. Right-click on a folder name and select Delete.
2. In the confirmation box that appears, click OK.
Related concepts:
Understanding folder types on page 85

Moving a folder
To move a folder:
1. From within the Folders tab, Workspace folder, do either of the following:
a. Select the check box next to the folder you want to move, and in the
Actions list, select Move.
b. Right-click on the folder name and select Move.
2. Click OK, and in the Move items dialog, select the new location from the list.
3. Click OK.
Related concepts:
Understanding folder types on page 85

Renaming a folder
To rename a folder:
1. Right-click on the folder name and select Rename.
2. In the Rename folder dialog, change the Name and/or the Description.
3. Click OK.
Related concepts:
Understanding folder types on page 85

Copying items to different folders


To copy an item from one folder to another:
1. From within Folders > Workspace, right-click the item you want to copy.
2. Select Copy.
3. In the Copy dialog, assign a new name (if appropriate) in the Name field.
Note that you cannot reuse the same name for an item within a single folder.
4. In the Description field, type a description.
5. In the Save in list, choose the location for the copied item.
6. Click Save.
Related concepts:
Understanding folder types on page 85

Saving items into different folders


Jobs can be renamed and saved into other folders.

To save items into a different folder:


1. From the items editor pane, click Save as.
2. In the Save [item] as... pane, type a name in the [Item] name: field.
3. In the Description: field, type a description.
4. In the Save in: field, use the list to select a location for the item.

86 Administration Guide
5. Click Save.
Related concepts:
Understanding folder types on page 85

Filtering items within the folder view


To filter items within the folder view:
1. From within Folders, click Filter by...
2. Select the component you want to display.
Related concepts:
Understanding folder types on page 85

Using Folders 87
88 Administration Guide
Using audits and logs
This section describes the audit and log categories in the IBM StoredIQ Platform
system, including descriptions of the various audit types as well as how to view
and download details.

Understanding harvest audits


Harvest audits provide a summary of the harvest, including status, date, duration,
average harvest speed, and average data object size. They can be viewed in two
ways: by volume name or by the date and time of the last harvest.

Data objects can be skipped during a harvest for a variety of reasons such as the
object being unavailable or a selected user option that excludes the data object
from the harvest. The Harvest details page lists all skipped data objects based on
file-system metadata level and content level.

All skipped harvest audit data and other files that have not been processed can be
downloaded for analysis.

This section lists the different fields seen when harvesting audits.
v Harvest audit by volume
v Harvest audit by time
v Harvest audit overview, including summary options, results options, and
detailed results
v Skipped data objects details

Harvest Audit by Volume Fields and Descriptions


Server Volume Harvest type Last harvested
v This is the server name. v This is the volume name. v This indicates the type of v This is the date and time
harvest performed: Full of the last harvest.
Harvest, ACL only, or
Incremental.
Total system data objects Data objects fully Data objects previously Processing exceptions
v This is the total number processed processed v This is the number of
of system data objects v This is the number of v This is the number of exceptions thrown
encountered. data objects that were data objects that were during processing.
fully processed. previously processed.
Binary processed Harvest duration Status Average harvest speed
v This is the number of v This is the length of time v This is the harvests v This is the average
processed binaries. of the harvests duration. status: Complete or harvest speed, given in
Incomplete. terms of data objects
processed per second.
Average data object size
v This is the average size
of encountered data
objects.

Copyright IBM Corp. 2001, 2013 89


Harvest Audit by Time Fields and Descriptions
Harvest start Harvest type Total system data objects Data objects fully
v This is the time and date v This indicates the type of v This is the total number processed
at which the harvest was harvest performed: Full of system data objects v This is the total number
started. Harvest, ACL only, or that were found. of system data objects
Incremental. that were fully processed.
Data objects previously Processing exceptions Binary processed Harvest duration
processed v This is the total number v This is the total number v This is the length of time
v This is the total number of encountered of processed binaries. of the harvests duration.
of system data objects processing exceptions.
that were previously
processed.
Status Average harvest speed Average data object size
v This is the harvests v This is the average v This is the average size
status: Complete or harvest speed, given in of encountered data
Incomplete. terms of data objects objects.
processed per second.

Harvest Overview Summary Options: Fields and Descriptions


Harvest type Harvest status Harvest date
v This indicates the type of harvest v This indicates the harvests status. v This lists the date and time of the
performed: Full Harvest, ACL only, Options are Complete or harvest.
or Incremental. Incomplete.
Harvest duration Average harvest speed Average data object size
v This is the length of time of the v This is the average harvest speed, v This is the average size of
harvests duration. given in terms of data objects encountered data objects.
processed per second.

Harvest Overview Results Options: Fields and Descriptions


Total system data objects Total contained data objects Total data objects
v This is the total number of system v This lists the total number of v This lists the total number of
data objects that were found. contained data objects. encountered data objects.

Harvest Overview Detailed Results: Fields and Descriptions


Skipped - previously Fully processed Skipped - cannot access data
processed v This is the number of fully object
v This is the number of processed data objects. v This is the number of data
skipped objects that were objects that were skipped
previously processed. as they could not be
accessed.
Skipped - user Skipped directories Content skipped - user
configuration v This is the number of data configuration
v This is the number of data objects in skipped v This is the number of data
objects that were skipped directories. objects where the content
because of their user was skipped due to user
configuration. configuration.

90 Administration Guide
Harvest Overview Detailed Results: Fields and Descriptions
Content type known, partial Content type known, but Content type known, but
processing complete error processing content cannot extract content
v This is the number of data v This is the number of data v This is the number of data
objects for which the objects for which the objects for which the
content type is known and content type is known, but content type is known, but
partial processing is an error was thrown while the content could not be
complete. processing content. extracted.
Content type unknown, not Binary text extracted, full Binary text extracted, partial
processed processing complete processing complete
v This is the number of data v This is the number of data v This is the number of data
objects for which the objects for which the objects for which the
content type is unknown binary text has been binary text has been
and has not been extracted and full extracted and partial
processed. processing has been processing has been
completed. completed.
Error processing binary Total
content v This lists the total number
v This is the number of data of data objects.
objects for which an error
was thrown while
processing binary content.

Related concepts:
Using audits and logs on page 89

Viewing harvest audits


This topic provides procedural information regarding how to view harvest audits.

To view harvest audits:


1. Go to Audit > Harvests > View all harvests. The Harvest audit by volume
page opens, which lists recent harvests and includes details about them.
2. In the Volume column, click the volume name link to see the harvest audit by
time page for that particular volume. The harvest audit by time page lists all
recent harvests for the chosen volume and includes details about each harvest.
3. In the Harvest start column, click the harvest start time link to see the harvest
overview page for the volume. You can also access this page by clicking the
Last harvested time link on the Harvest audit by volume page. The Harvest
overview page provides:
v SummaryHarvest type, status, date and time, duration, average harvest
speed, and average data object size.
v ResultsTotal system data objects, total contained data objects, total data
objects
v Detailed resultsSkipped - previously processed; fully processed; skipped -
cannot access data object; skipped - user configuration; skipped directories;
content skipped - user configuration; content type known, partial processing
complete; content type known, but error processing content; content type
known, but cannot extract content; content type unknown, not processed;
binary text extracted, full processing complete; binary text extracted, partial
processing complete; error processing binary content; error-gathering ACLs;
and total.
4. To view details on data objects, click the link next to the data objects under
Detailed results.

Using audits and logs 91


v With the exceptions of skipped - previously processed, fully processed, and
the total, all other results with more than zero results listed have clickable
links that allow you to view and download results.
v The skipped data object lists includes object name, path, and reason skipped.
Data objects can be skipped at the file system metadata level or at the
content level. Data objects skipped at the content level are based on
attributes associated with the data object or its contents. Skipped Data
Objects Results Details provides details about skipped data objects.
If data objects were not harvested (such as skipped, not fully processed, or had
errors in processing), you may want to download the data object's harvest audit
list details for further analysis.
Related concepts:
Understanding harvest audits on page 89

Downloading harvest list details


This topic provides procedural information regarding how to download harvest list
details.

To download harvest list details:


1. From the Harvest details page, click the active link next to the data objects
under Detailed results. A page named for the detailed result chosen (such as
Skipped - user configuration or binary text extracted, full processing complete)
appears.
2. Click the Download list in CSV format link on the upper left side of the page.
A dialog informs you that the results are being prepared for download.
3. Click OK. A new dialog appears, prompting you to save the open or save the
.CSV file. Information in the downloaded CSV file includes:
v Object name
v System path
v Container path
v Message explaining why data object was skipped
v Server name
v Volume name
Related concepts:
Understanding harvest audits on page 89

Understanding import audits


Volume-import audits provide information about the volume import, including the
number of data objects imported, the IBM StoredIQ Platform system exported
from, the time and date of the volume import, whether or not the imported
volume overwrote an existing volume, and status. The volume name links to the
Import details page.

This section lists the different fields seen when viewing import audits.

Imports by Volumes Details: Fields and Descriptions


Volume Exported from Import date
v This is the name of the imported v This is the source server of the v This is the date and time on which
volume. imported volume. the import occurred.

92 Administration Guide
Imports by Volumes Details: Fields and Descriptions
Total data objects imported Overwrite existing Status
v This is the total number of v If the import overwrote an existing v This is the status of the import:
imported data objects. volume, the status is Yes. If the Complete or Incomplete.
import did not overwrite an
existing volume, the status is No.

Related concepts:
Using audits and logs on page 89

Viewing volume import audit details


This topic provides procedural information regarding how to view volume import
audit details

To view volume import audit details:


1. Go to Audit > Imports, and click View all imports. The Imports by volume
page opens, which lists volume imports and import information.
2. Click a volume name link in the Volume column to view the audit details for
that particular import.

Understanding event logs


Every action taken by the system and its users is captured by the event logs, which
document actions that succeed and fail.

Every action taken by the system and its users is captured by the event logs. These
actions include creating draft and published queries and tags, running policies,
publishing queries, deleting objects, configuring settings, and any other action
taken through the IBM StoredIQ Platform interface. A detailed of list of log entries
is provided in the event log messages.

You can view event logs for the current day or review saved logs from previous
days, and up to 30 days worth of logs can be viewed through the interface. If you
select and clear a day of logs, those logs are removed from the system.
Related reference:
Appendix C, Event log messages, on page 133

Working with event logs


This topic provides procedural information regarding how to work with event
logs, including viewing logs, subscribing to an event, clearing the current event
log, or downloading an event log.

Viewing event logs


To view the event log:
1. Perform either of the following:
v Click Administration > Dashboard, and then locate the Event log section on
the dashboard. The current days log displays there by default.
v Click the Audit tab and locate the Event logs section.
2. To view a previous day's log on the dashboard, use the View all event logs list
to select the day for which you would like to view an event log.

Using audits and logs 93


3. Select a different day from the view event log from the list. This menu
displays the event log dates for the past 30 days. Each log is listed by date in
YYYY-MM-DD format.

Subscribing to an event
To subscribe to an event:
1. Go to Audit > Event logs.
2. Click View all event logs, and the Event log for today page opens.
3. To the right of the event log to which you would like to subscribe, click
Subscribe. The Edit notification page appears.
4. In Destination, select the method by which you would like to be notified of
this event log. If you select Email address, be certain to use commas to
separate multiple email addresses.
5. Click OK.

Note: You can also subscribe to an event on the Dashboard. In the Event log
area, click Subscribe to the right of the event.

Clearing the current event log


To clear the current event log:

On the Administration > Dashboard, click Clear for the current view.

Downloading an event log


To download an event log:
1. When viewing an event log, select the Download link for saving the data to a
text file.
2. Select to save the file from the prompt. Enter a name and select a location to
save the file.

Understanding policy audits


This topic provides conceptual information regarding how to view and understand
policy audits.

Policy audits provide a detailed history of the policy, including type of action, date
last executed, start and end dates with times, average speed, total data objects, and
data object counts. They can be viewed by name, volume, time, and by discovery
export.

This section lists the different fields seen when harvesting audits.

Policy Audit by Name: Fields and Descriptions


Policy name Policy Status Number of times Most recent date
v This is the policy v This is the policys executed executed
name. status. v This is the number v This is the date on
of times that the which the policy
policy was was last executed.
executed.

94 Administration Guide
Policy Audit by Volume: Fields and Descriptions
Volume Most recent date a policy Number of policies executed
v This is the name of the was executed v This is the number of
volume on which the v This is the most recent policies that were
policy was executed. date on which the policy executed.
was last executed.

Policy Audit by Time: Fields and Descriptions


Policy name Policy Status Start End
v This is the policy v This is the policys v This is the time at v This is the time at
name. status: Complete or which the policys which the policys
Incomplete. execution was execution was
started. completed.
Success count Failure count Warning count Other count
v This is the number v This is the number v This is the number v This is the number
of processed of processed of processed of processed
messages that have messages that have messages that have messages that have
been classified as a been classified as a been classified as a been classified as
success. failure. warning. other.
Total data objects Action type Avg. actions/second
v This is the total v This is the type of v This is the average
number of data policy that took number of action
objects. place. per second.

Policy Audit by Discovery Export: Fields and Descriptions


Discovery export name Number runs Most recent export status
v This is the name of the v This is the number of v This is the status of the
Discovery export. times the policy has run. most recent Discovery
export.
Most recent load file status Most recent date executed
v This is the status of the v This is the date of the
most recent load file. most recent policy
execution.

Discovery Export Runs by Discovery Export: Fields and Descriptions


Discovery export run Number of Success count Failure count
v This is the name of executions v This is the number v This is the number
the discovery v This is the number of processed of processed
export run. of times the run messages that have messages that have
was executed. been classified as a been classified as a
success. failure.
Warning count Other count Total data objects Export status
v This is the number v This is the number v This is the total v This is the status
of processed of processed number of data of the export:
messages that have messages that have objects. Complete or
been classified as a been classified as Incomplete.
warning. other.

Using audits and logs 95


Discovery Export Runs by Discovery Export: Fields and Descriptions
Load file status
v This is the status
of the load file:
Complete or
Incomplete.

Note: A warning in a policy audit trail is a success with the following conditions:
v If you copy an Exchange item such as re:, the re is copied, not the :. This will
generate a warning.
v The copied file is renamed.
v The file system to which you are copying does not accept characters in the file
name.

Viewing policy audit details


This topic provides procedural information regarding viewing policy audit details.

To view policy audit details:


1. Go to Audit > Policies, and then click Name. The Policy audit by name page
provides policy name and status, the number of times it has been executed, and
the time and date of the most recent execution.
2. Click a policy name to open the Policy executions by time page.
3. Click a policy name to open the Policy execution results page.

Note: To view the list of data objects, click on the [#] data objects link. To
create a report, click Create XML or Create PDF.
v Click Volume to open the policy audit by volume page.
v Click on a volume link to go to the Policy audit by time page.
v Click Time to see Audit by time page for the policy.
v On the Policy audit by time page, click the policy name to open the Policy
execution results page.

Note: To view the list of data objects, click the [#] data objects link. To create
a report, click Create XML or Create PDF.
v Click Discovery export.
v On the Policy audit by discovery export page, click the discovery export
name to open the Discovery export runs by production page. The page
details further information according to the incremental runs of the policy.
v Click a policy name to open the Policy executions by time page.
v Click on a policy name to open the Policy execution results page.

Note: To view the list of data objects, click the [#] data objects link. To create
a report, click Create XML or Create PDF.
As you review audit results through these pages, you can continue clicking
through to review various levels of information, from the volume and policy
execution level down to the data objects. To view more policy execution details,
click on the policy name in the execution summary page, which can be
accessed by any of the above policy views. As you continue browsing, IBM
StoredIQ Platform provides more detailed information such as:
v Source and destination settings

96 Administration Guide
v Policy optionsDetails of the policy actionThis section reflects the options
selected when creating the policy. Most attributes that appear depend upon
the type of policy run and the options available in the policy editor.
v Query (eitherIBM StoredIQ Platform or user-defined)
v View metadata linkThe view metadata page describes security details for
source and destination locations of the policy action.
Related concepts:
Understanding policy audits on page 94

Viewing a policy audit by name


To view a policy audit by name:
1. Go to Audit > Policies.
2. Click Name. The Policy audit by name page provides policy name and status,
the number of times it has been executed, and the time and date of the most
recent execution.
3. Click a policy name to open the Policy execution by time page.
Related concepts:
Understanding policy audits on page 94

Viewing a policy audit by volume


To view a policy audit by volume:
1. Go to Audit > Policies.
2. Click Volume. The Policy audit by page provides policy name, its most recent
time and date of execution, and the number of policies that were executed.
3. Click a policy name to open the Policy executions by time page.
4. Click the policy name to open the Policy executions by results page.
Related concepts:
Understanding policy audits on page 94

Viewing a policy audit by time


To view a policy audit by time:
1. Go to Audit > Policies.
2. Click Time. The Policy audit by time provides policy name, its status, the
executions start and end time, its success count, failure count, warning count,
other count, the total number of data objects, the action type, and the average
number of actions per second.
3. On the Policy audit by time page, click the policy name to open the Policy
execution results page.
Related concepts:
Understanding policy audits on page 94

Viewing a policy audit by discovery export


To view a policy audit by discovery export:
1. Click Discovery export.
2. On the Policy audit by discovery export page, click the discovery export name
to open the Discovery export runs by production page. The page details further
information according to the incremental runs of the policy.

Using audits and logs 97


3. Click a policy name to open the Policy executions by time page.
4. Click on a policy name to open the Policy execution results page.
Related concepts:
Understanding policy audits on page 94

Understanding the search audit feature


With the search audit feature, you can search audit trails by entering either by
Policy Details, Execution Details, or Data Object Details.

Policy details

Policy audits can be searched using any of these details.

Policy Audit Details: Fields and Descriptions


Audit search by policy details Specify search criteria Audit search criteria
v In this area, select search criteria, v In this area, specify the Policy v In the Find audits that match list,
define their values, and then add name, the Policy state, and the select either Any of the following
them to the list to search across all Action type. or All of the following.
audits.

Execution details

Policy audits can be searched using any of these execution details.

Policy Audit Execution Details: Fields and Descriptions


Audit search by execution details Specify search criteria Audit search criteria
v In this area, select search criteria, v In this area, specify the Action v In the Find audits that match list,
define their values, and then add type, Action status, Action start select either Any of the following
them to the list to search across all date, Action end date, Total count, or All of the following.
audits. Success count, Failure count,
Warning count, Source volume,
Destination volume, or Query
name.

Data object details

Policy audits can be searched using any of these data-object details.

Policy Audit Data Object Details: Fields and Descriptions


Audit search by data object details Specify search criteria Audit search criteria
v In this area, select search criteria, v In this area, specify the Source v In the Find audits that match list,
define their values, and then add volume, Destination volume, select either Any of the following
them to the list to search across all Source object name, Destination or All of the following.
audits. object name, Source system path,
Destination system path, or
Action result.

Related concepts:
Understanding policy audits on page 94

98 Administration Guide
Saving results from an audit
You can save the results of policy executions into PDF and XML files. The
information can be saved as PDF and XML files. The exporting of information
appears as a running job on the dashboard until completed.

To save results from an audit trail:


1. Go to Audit > Policies.
2. In the Browse by options, click Time.
3. Click the policy name.
4. In the Results pane, click Data objects to see items that were responsive to the
policy. To download the material in .CSV, click CSV.
5. On the Policy execution results page, select Create PDF to generate a PDF or
Create XML to generate an XML file of the results.
6. Access the report through the inbox on the navigation page.
Related concepts:
Understanding policy audits on page 94

Policy audit messages


A policy audit shows the number of data objects that have been processed during
the policy execution. Processed data objects are divided into these categories:
Success, Warnings, Failures, and Other (discovery export policies only).

Policy audit success messages

Data objects can receive a success message for these reasons:

Policy Audit Success Messages

Success Data object is a duplicate of Data object skipped but Data object is a duplicate
[object name] will be loaded in load file. produced in a previous run
This applies to intermediate (discovery export only)
and files archives produced
during a discovery export
policy.

Policy audit warning messages

Data objects can receive a warning during a policy action if they fail to do any of
the following:

Policy Audit Warnings

Set directory attributes Reset time stamps Set attributes Set time stamps

Set security descriptor Set access modes (Windows Set owner information Set group information
(Windows Share) Share) (NFS)

Set security permissions Create a link after a Find template to create a Extract text for the object
migrate (Windows Share, shortcut (Windows Share) (Discovery export policy)
NFS)

Using audits and logs 99


Note: A data object whose name was modified to respect file system rules on the
destination volume will also appear in the warning category.

Policy audit failure messages

Data objects can receive one of the following failures during a policy action.

Data-Object Failures

Failed to create target Source does not exist Failed to find a new name
directory structure for the incoming object

Target is a directory File copy failed Could not create target

Error copying data to target Could not copy due to Could not delete source after
network errors move

Target disk is full Source equals target on a Insufficient permissions in


copy or move general to perform an action

All modify actions failed File timed out waiting in the File under retention; cannot
pipeline be deleted (retention server)

Data object is a constituent of


a container that already
encountered failure
(Discovery export policy)

Other policy audit messages

Data objects are categorized in the other category during a discovery export policy
when:
v A data object is a member that makes its container responsive.
v A data object is a non-responsive member of a container.
Related concepts:
Understanding policy audits on page 94

100 Administration Guide


Appendix A. Supported file types
This section provides a comprehensive list of the file types that can be harvested
and processed by IBM StoredIQ Platform, organized by name and by category. You
can also view SharePoint attributes.

Supported file types by name


This topic lists all supported file types by name.

Format Extension Category Version

Adobe Acrobat PDF graphic 2.1

3.07.0

Japanese

Adobe FrameMaker FMV graphic vector/raster through 5.0


Graphics

Adobe FrameMaker MIF word processing 3.06.0


Interchange Format

Adobe Illustrator graphic through 7.0

9.0

Adobe Photoshop PSD graphic 4.0

Ami Draw SDW graphic all

ANSI TXT text and markup 7- and 8-bit

ASCII TXT text and markup 7- and 8-bit

AutoCAD DWG CAD 2.52.6

9.014.0

2002

2004

2005

AutoShade Rendering RND graphic 2.0

Binary Group 3 Fax graphic all

Bitmap BMP, RLE, ICO, CUR, DIB, graphic all


WARP

Copyright IBM Corp. 2001, 2013 101


Format Extension Category Version

CALS Raster GP4 graphic Type I, II

Comma Separated Values CSV spreadsheet

Computer Graphics CGM graphic ANSI


Metafile
CALS

NIST 3.0

Corel Clipart CMX graphic 56

Corel Draw CDR graphic 3.x8.x

Corel Draw (CDR with Tiff graphic 2.x9.x


header)

Corel Presentations SHW presentation through 12.0

X3

Corel WordPerfect WPD word processing through 12.0


Windows
X3

DataEase database 4.X

dBase Database database through 5.0

dBXL database 1.3

DEC WPS PLUS DX word processing through 4.0

DEC WPS PLUS WPL word processing through 4.1

DisplayWrite (2 and 3) IP word processing all

DisplayWrite (4 and 5) word processing through 2.0

DOS command executable COM system

Dynamic link library files DLL system

EBCDIC text and markup all

ENABLE word processing 3.0

4.0

4.5

102 Administration Guide


Format Extension Category Version

ENABLE database 3.0

4.0

4.5

ENABLE Spreadsheet SSF spreadsheet 3.0

4.0

4.5

Encapsulated PostScript EPS graphic TIFF header


(raster)

Executable files EXE system

First Choice database through 3.0

First Choice word processing through 3.0

First Choice spreadsheet through 3.0

FoxBase database 2.1

Framework database 3.0

Framework word processing 3.0

Framework spreadsheet 3.0

GEM Bit Image IMG graphic all

Graphics Interchange GIF graphic all


Format

Graphics Environment GEM VDI graphic bitmap and vector


Manager

Gzip GZ archive all

Haansoft Hangul HWP word processing 1997

2002

Harvard Graphics (DOS) graphic 2.x

3.x

Harvard Graphics graphic all


(Windows)

Appendix A. Supported file types 103


Format Extension Category Version

Hewlett-Packard Graphics HPGL graphic 2


Language

HTML HTM text and markup through 3.0

IBM FFT text and markup all

IBM Graphics Data Format GDF graphic 1.0

IBM Picture Interchange PIF graphic 1.0


Format

IBM Revisable Form Text text and markup all

IBM Writing Assistant word processing 1.01

Initial Graphics Exchange IGES graphic 5.1


Spec

Java class files CLASS system

JPEG (not in TIFF format) JFIF graphic all

JPEG JPEG graphic all

JustSystems Ichitaro JTD word processing 5.0

6.0

8.013.0

2004

JustSystems Write word processing through 3.0

Kodak Flash Pix FPX graphic all

Kodak Photo CD PCD graphic 1.0

Legacy word processing through 1.1

Legato Email Extender EMX email

Lotus 1-2-3 WK4 spreadsheet through 5.0

Lotus 1-2-3 (OS/2) spreadsheet through 2.0

Lotus 1-2-3 Charts 123 spreadsheet through 5.0

Lotus 1-2-3 for SmartSuite spreadsheet 1997Millennium 9.6

Lotus AMI Pro SAM word processing through 3.1

104 Administration Guide


Format Extension Category Version

Lotus Freelance Graphics PRZ presentation through Millennium

Lotus Freelance Graphics PRE presentation through 2.0


(OS/2)

Lotus Manuscript word processing 2.0

Lotus Notes NSF email

Lotus Pic PIC graphic all

Lotus Snapshot graphic all

Lotus Symphony spreadsheet 1.0

1.1

2.0

Lotus Word Pro LWP word processing 19969.6

LZA Self Extracting archive all


Compress

LZH Compress archive all

Macintosh PICT1/2 PICT1/PICT1 graphic bitmap only

MacPaint PNTG graphic n/a

MacWrite II word processing 1.1

Macromedia Flash SWF presentation text only

MASS-11 word processing through 8.0

Micrografx Designer DRW graphic through 3.1

Micrografx Designer DSF graphic Win95, 6.0

Micrografx Draw DRW graphic through 4.0

MPEG-1 Audio layer 3 MP3 multimedia ID3 metadata only.

These files can be


harvested, but there is no
data in them that can be
used in tags.

MS Access MDB database through 2.0

MS Binder archive 7.01997

Appendix A. Supported file types 105


Format Extension Category Version

MS Excel XLS spreadsheet 2.22007

MS Excel Charts spreadsheet 2.x7.0

MS Excel (Macintosh) XLS spreadsheet 3.04.0

1998

2001

2004

MS Excel XML XLSX spreadsheet

MS MultiPlan spreadsheet 4.0

MS Outlook Express EML email 19972003

MS Outlook Form Template OFT email 19972003

MS Outlook Message MSG email all

MS Outlook Offline Folder OST email 19972003

MS Outlook Personal PST email 19972007


Folder

MS PowerPoint (Macintosh) PPT presentation 4.02004

MS PowerPoint (Windows) PPT presentation 3.02007

MS PowerPoint XML PPTX presentation

MS Project MPP database 19982003

MS Windows XML DOCX word processing

MS Word (Macintosh) DOC word processing 3.04.0

1998

2001

MS Word (PC) DOC word processing through 6.0

MS Word (Windows) DOC word processing through 2007

MS WordPad word processing all

MS Works S30/S40 spreadsheet through 2.0

MS Works WPS word processing through 4.0

106 Administration Guide


Format Extension Category Version

MS Works (Macintosh) word processing through 2.0

MS Works Database database through 2.0


(Macintosh)

MS Works Database (PC) database through 2.0

MS Works Database database through 4.0


(Windows)

MS Write word processing through 3.0

Mosaic Twin spreadsheet 2.5

MultiMate 4.0 word processing through 4.0

Navy DIF word processing all

Nota Bene word processing 3.0

Novell Perfect Works word processing 2.0

Novell Perfect Works spreadsheet 2.0

Novell Perfect Works graphic 2.0


(Draw)

Novell WordPerfect word processing through 6.1

Novell WordPerfect word processing 1.023.0


(Macintosh)

Office Writer word processing 4.06.0

OpenOffice Calc SXC/ODS spreadsheet 1.1

2.0

OpenOffice Draw graphic 1.1

2.0

OpenOffice Impress SXI/SXP/ODP presentation 1.1

2.0

OpenOffice Writer SXW/ODT word processing 1.1

2.0

OS/2 PMMetafile Graphics MET graphic 3.0

Appendix A. Supported file types 107


Format Extension Category Version

Paint Shop Pro 6 PSP graphic 5.06.0

Paradox Database (PC) database through 4.0

Paradox (Windows) database through 1.0

PC-File Letter word processing through 5.0

PC-File+Letter word processing through 3.0

PC PaintBrush PCX, DCX graphic all

PFS: Professional Plan spreadsheet 1.0

PFS: Write word processing A, B, C

Portable Bitmap Utilities PBM graphic all

Portable Greymap PGM graphic n/a

Portable Network Graphics PNG graphic 1.0

Portable Pixmap Utilities PPM graphic n/a

PostScript File PS graphic level II

Professional Write word processing through 2.1

Professional Write Plus word processing 1.0

Progressive JPEG graphic n/a

Q &A (database) database through 2.0

Q & A (DOS) word processing 2.0

Q & A (Windows) word processing 2.0

Q & A Write word processing 3.0

Quattro Pro (DOS) spreadsheet through 5.0

Quattro Pro (Windows) spreadsheet through 12.0

X3

R:BASE 5000 database through 3.1

R:BASE (Personal) database 1.0

R:BASE System V database 1.0

108 Administration Guide


Format Extension Category Version

RAR RAR archive

Reflex Database database 2.0

Rich Text Format RTF text and markup all

SAMNA Word IV word processing

Smart Ware II database 1.02

Smart Ware II word processing 1.02

Smart Ware II spreadsheet 1.02

Sprint word processing 1.0

StarOffice Calc SXC/ODS spreadsheet 5.2

6.x

7.x

8.0

StarOffice Draw graphic 5.2

6.x

7.x

8.0

StarOffice Impress SXI/SXP/ODP presentation 5.2

6.x

7.x

8.0

StarOffice Writer SXW/ODT word processing 5.2

6.x

7.x

8.0

Sun Raster Image RS graphic n/a

Supercalc Spreadsheet spreadsheet 4.0

Text Mail (MIME) various email

Total Word word processing 1.2

Appendix A. Supported file types 109


Format Extension Category Version

Truevision Image TIFF graphic through 6

Truevision Targa TGA graphic 2

Unicode Text TXT text and markup all

Unix TAR (tape archive) TAR archive n/a

Unix Compressed Z archive n/a

UUEncoding UUE archive n/a

vCard word processing 2.1

Visio (preview) graphic 4

Visio 2003 graphic 5

2000

2002

Volkswriter word processing through 1.0

VP Planner 3D spreadsheet 1.0

WANG PC word processing through 2.6

WBMP graphic n/a

Windows Enhanced EMF graphic n/a


Metafile

Windows Metafile WMF graphic n/a

Winzip ZIP archive

WML text and markup 5.2

WordMARC word word processing through composer


processor

WordPerfect Graphics WPG, WPG2 graphic through 2.0, 7, 10

WordStar word processing through 7.0

WordStar 2000 word processing through 3.0

X Bitmap XBM graphic x10

X Dump XWD graphic x10

110 Administration Guide


Format Extension Category Version

X Pixmap XPM graphic x10

XML (generic) XML text and markup

XyWrite XY4 word processing through III Plus

Yahoo! IM Archive archive

ZIP ZIP archive PKWARE2.04g

Related reference:
Appendix A, Supported file types, on page 101

Supported file types by category


This topic lists supported file types by category.

Category Format Extension Version

Archive Gzip GZ all

LZA Self Extracting all


Compress

LZH Compress all

MS Binder 7.01997

RAR RAR

Unix TAR (tape archive) TAR n/a

Unix Compressed Z n/a

UUEncoding UUE n/a

Winzip ZIP

Yahoo! IM Archive n/a

ZIP ZIP PKWARE2.04g

CAD AutoCAD DWG 2.52.6

9.014.0

2002

2004

2005

Appendix A. Supported file types 111


Category Format Extension Version

Database DataEase 4.x

dBase Database through 5.0

dBXL 1.3

ENABLE 3.0

4.0

4.5

First Choice through 3.0

FoxBase 2.1

Framework 3.0

MS Access MDB through 2.0

MS Project MPP 19982003

MS Works Database through 2.0


(Macintosh)

MS Works Database (PC) through 2.0

MS Works Database through 4.0


(Windows)

Paradox Database (PC) through 4.0

Paradox Database through 1.0


(Windows)

Q & A (database) through 2.0

R:BASE 5000 through 3.1

R:BASE (personal) 1.0

R:BASE System V 1.0

Reflex Database 2.0

Smart Ware II 1.02

Email Legato Email Extender EMX

Lotus Notes NSF

MS Outlook Express EML 19972003

112 Administration Guide


Category Format Extension Version

MS Outlook Form Template OFT 19972003

MS Outlook Message MSG all

MS Outlook Offline Folder OST 19972003

MS Outlook Personal PST 19972007


Folder

Text Mail (MIME) various

Graphic Adobe Acrobat PDF 2.1

3.07.0

Japanese

Adobe FrameMaker FMV vector/raster5.0


Graphics

Adobe Illustrator through 7.0

9.0

Adobe Photoshop PSD 4.0

Ami Draw SDW all

AutoShade Rendering RND 2.0

Binary Group 3 Fax all

Bitmap BMP, RLE, ICO, CUR, DIB, all


WARP

CALS Raster GP4 Type I, II

Computer Graphics CGM ANSI


Metafile
CALS

NIST 3.0

Corel Clipart CMX 56

Corel Draw CDR 3.x8.x

Corel Draw (CDR with Tiff 2.x9.x


header)

Encapsulated Post Script EPS TIFF header


(raster)

Appendix A. Supported file types 113


Category Format Extension Version

GEM Bit Image IMG all

Graphics Interchange GIF all


Format

Graphics Environment GEM VDI bitmap and vector


Manager

Harvard Graphics (DOS) 2.x

3.x

Harvard Graphics all


(Windows)

Hewlett-Packard Graphics HPGL 2


Language

IBM Graphics Data Format GDF 1.0

IBM Picture Interchange PIF 1.0


Format

Initial Graphics Exchange IGES 5.1


Spec

JPEG (not in TIFF format) JFIF all

JPEG JPEG all

Kodak Flash PIX FPX all

Kodak Photo CD PCD 1.0

Lotus Pic PIC all

Lotus Snapshot all

Macintosh PICT1/2 PICT1/PICT2 bitmap only

MacPaint PNTG n/a

Micrografx Designer DRW through 3.1

Micrografx Designer DSF Win95

6.0

Micrografx Draw DRW through 4.0

Novell Perfect Works 2.0


(Draw)

114 Administration Guide


Category Format Extension Version

OpenOffice Draw 1.1

2.0

OS/2 PM Metafile Graphics MET 3.0

Paint Shop Pro 6 PSP 5.06.0

PC PaintBrush PCX, DCX all

Portable Bitmap Utilities PBM all

Portable Greymap PGM n/a

Portable Network Graphics PNG 1.0

Portable Pixmap Utilities PPM n/a

PostScript PS level II

Progressive JPEG n/a

StarOffice Draw 5.2

6.x

7.x

8.0

Sun Raster Image RS n/a

Truevision Image TIFF through 6

Truevision Targa TGA 2

Visio (preview) 4

Visio 2003 5

2000

2002

WBMP n/a

Windows Enhanced EMF n/a


Metafile

Windows Metafile WMF n/a

Appendix A. Supported file types 115


Category Format Extension Version

WordPerfect Graphics WPG, WPG2 through 2.0

10

X Bitmap XBM x10

X Dump XWD x10

x Pixmap XPM x10

Archive Gzip GZ all

LZA Self Extracting all


Compress

LZH Compress all

MS Binder 7.01997

RAR RAR

Unix TAR (tape archive) TAR n/a

Unix Compressed Z n/a

UUEncoding UUE n/a

Winzip ZIP

Yahoo! IM Archive n/a

ZIP ZIP PKWARE2.04g

CAD AutoCAD DWG 2.52.6

9.014.0

2002

2004

2005

Database DataEase 4.x

dBase Database through 5.0

dBXL 1.3

116 Administration Guide


Category Format Extension Version

ENABLE 3.0

4.0

4.5

First Choice through 3.0

FoxBase 2.1

Framework 3.0

MS Access MDB through 2.0

MS Project MPP 19982003

MS Works Database through 2.0


(Macintosh)

MS Works Database (PC) through 2.0

MS Works Database through 4.0


(Windows)

Paradox Database (PC) through 4.0

Paradox Database through 1.0


(Windows)

Q & A (database) through 2.0

R:BASE 5000 through 3.1

R:BASE (personal) 1.0

R:BASE System V 1.0

Reflex Database 2.0

Smart Ware II 1.02

Email Legato Email Extender EMX

Lotus Notes NSF

MS Outlook Express EML 19972003

MS Outlook Form Template OFT 19972003

MS Outlook Message MSG all

MS Outlook Offline Folder OST 19972003

Appendix A. Supported file types 117


Category Format Extension Version

MS Outlook Personal PST 19972007


Folder

Text Mail (MIME) various

Graphic Adobe Acrobat PDF 2.1

3.07.0

Japanese

Adobe FrameMaker FMV vector/raster5.0


Graphics

Adobe Illustrator through 7.0

9.0

Adobe Photoshop PSD 4.0

Ami Draw SDW all

AutoShade Rendering RND 2.0

Binary Group 3 Fax all

Bitmap BMP, RLE, ICO, CUR, DIB, all


WARP

CALS Raster GP4 Type I, II

Computer Graphics CGM ANSI


Metafile
CALS

NIST 3.0

Corel Clipart CMX 56

Corel Draw CDR 3.x8.x

Corel Draw (CDR with Tiff 2.x9.x


header)

Encapsulated Post Script EPS TIFF header


(raster)

GEM Bit Image IMG all

Graphics Interchange GIF all


Format

118 Administration Guide


Category Format Extension Version

Graphics Environment GEM VDI bitmap and vector


Manager

Harvard Graphics (DOS) 2.x

3.x

Harvard Graphics all


(Windows)

Hewlett-Packard Graphics HPGL 2


Language

IBM Graphics Data Format GDF 1.0

IBM Picture Interchange PIF 1.0


Format

Initial Graphics Exchange IGES 5.1


Spec

JPEG (not in TIFF format) JFIF all

JPEG JPEG all

Kodak Flash PIX FPX all

Kodac Photo CD PCD 1.0

Lotus Pic PIC all

Lotus Snapshot all

Macintosh PICT1/2 PICT1/PICT2 bitmap only

MacPaint PNTG n/a

Micrografx Designer DRW through 3.1

Micrografx Designer DSF Win95

6.0

Micrografx Draw DRW through 4.0

Novell Perfect Works 2.0


(Draw)

OpenOffice Draw 1.1

2.0

Appendix A. Supported file types 119


Category Format Extension Version

OS/2 PM Metafile Graphics MET 3.0

Paint Shop Pro 6 PSP 5.06.0

PC PaintBrush PCX, DCX all

Portable Bitmap Utilities PBM all

Portable Greymap PGM n/a

Portable Network Graphics PNG 1.0

Portable Pixmap Utilities PPM n/a

PostScript PS level II

Progressive JPEG n/a

StarOffice Draw 5.2

6.x

7.x

8.0

Sun Raster Image RS n/a

Truevision Image TIFF through 6

Truevision Targa TGA 2

Visio (preview) 4

Visio 2003 5

2000

2002

WBMP n/a

Windows Enhanced EMF n/a


Metafile

Windows Metafile WMF n/a

WordPerfect Graphics WPG, WPG2 through 2.0

10

120 Administration Guide


Category Format Extension Version

X Bitmap XBM x10

X Dump XWD x10

x Pixmap XPM x10

Multimedia MPEG-1 Audio layer 3 MP3 ID3 metadata only.

These files can be


harvested, but there is no
data in them that can be
used in tags.

Presentation Corel Presentations SHW through 12.0

X3

Lotus Freelance Graphics PRZ through Millennium

Lotus Freelance Graphics PRE through 2.0


(OS/2)

Macromedia Flash SWF text only

MS PowerPoint (Macintosh) PPT 4.02004

MS PowerPoint (Windows) PPT 3.02007

MS PowerPoint XML PPTX

OpenOffice Impress SXI/SXP/ODP 1.1

2.0

StarOffice Impress SXI/SXP/ODP 5.2

6.x

7.x

8.0

Spreadsheet Comma Separated Values CSV

ENABLE Spreadsheet SSF 3.0

4.0

4.5

FIrst Choice through 3.0

Framework 3.0

Appendix A. Supported file types 121


Category Format Extension Version

Lotus 1-2-3 WK4 through 5.0

Lotus 1-2-3 (OS/2) through 2.0

Lotus 1-2-3 Charts 123 through 5.0

Lotus 1-2-3 for SmartSuite 19979.6

Lotus Symphony 1.0

1.1

2.0

MS Excel XLS 2.22007

MS Excel Charts 2.x7.0

MS Excel (Macintosh) XLS 3.04.0

1998

2001

2004

MS Excel XML XLSX

MS MultiPlan 4.0

MS Works S30/S40 through 2.0

Mosaic Twin 2.5

Novell Perfect Works 2.0

OpenOffice Calc SXC/ODS 1.1

2.0

PFS: Professional Plan 1.0

Quattro Pro (DOS) through 5.0

Quattro Pro (Windows) through 12.0

X3

Smart Ware II 1.02

122 Administration Guide


Category Format Extension Version

StarOffice Calc SXC/ODS 5.2

6.x

7.x

8.0

Supercalc Spreadsheet 4.0

VP Planner 3D 1.0

System

Executable files .EXE

Dynamic link library files .DLL

Java class files .class

DOS command executables .COM

Text and markup ANSI TXT 7- and 8-bit

ASCII TXT 7- and 8-bit

EBCDIC text all

HTML HTM through 3.0

IBM FFT all

IBM Revisable Form Text all

Rich Text Format RTF all

Unicode Text TXT all

WML 5.2

XML (generic) XML

Word processing Adobe FrameMaker MIF 3.06.0


Interchange Format

Corel WordPerfect WPD through 12.0


Windows
X3

DEC WPS PLUS DX through 4.0

DEC WPS PLUS WPL through 4.1

Appendix A. Supported file types 123


Category Format Extension Version

Display Write (2 and 3) IP all

Display Write (4 and 5) through 2.0

ENABLE 3.0

4.0

4.5

First Choice through 3.0

Framework 3.0

Haansoft Hangul HWP 1997

2002

IBM Writing Assistant 1.01

JustSystems Ichitaro JTD 5.0

6.0

8.013.0

2004

JustSystems Write through 3.0

Legacy through 1.1

Lotus AMI Pro SAM through 3.1

Lotus Manuscript 2.0

Lotus Word Pro LWP 19969.6

MacWrite II 1.1

MASS-11 through 8.0

MS Windows XML DOCX

MS Word (Macintosh) DOC 3.04.0

1998

2001

MS Word (PC) DOC through 6.0

MS Word (Windows) DOC through 2007

124 Administration Guide


Category Format Extension Version

MS WordPad all versions

MS Works WPS through 4.0

MS Works (Macintosh) through 2.0

MS Write through 3.0

MultiMate 4.0 through 4.0

Navy DIF all versions

Nota Bene 3.0

Novell Perfect Works 2.0

Novell WordPerfect through 6.1

Novel WordPerfect 1.023.0


(Macintosh)

Office Writer 4.06.0

OpenOffice Writer SXW/ODT 1.1

2.0

PC-File Letter through 5.0

PC- File + Letter through 3.0

PFS: Write A

Professional Write through 2.1

Professional Write Plus 1.0

Q & A (DOS) 2.0

Q & A (Windows) 2.0

Q & A Write 3.0

SAMNA Word IV

Smart Ware II 1.02

Sprint 1.0

Appendix A. Supported file types 125


Category Format Extension Version

StarOffice Writer SXW/ODT 5.2, 6.x, 7.x, 8.0

Total Word 1.2

Related reference:
Appendix A, Supported file types, on page 101

SharePoint attributes
This section describes the various SharePoint data object types and their properties
currently supported by IBM StoredIQ Platform.

Supported SharePoint object types

These types of SharePoint objects are supported:

Supported SharePoint Supported SharePoint Supported SharePoint


Object Types Object Types Object Types

Blog posts and comments Discussion board Calendar

Tasks Project tasks Contacts

Wiki pages Issue tracker Announcements

Survey Links Document libraries

Picture libraries Records center

Notes regarding SharePoint object types

Calendar

Recurring calendar events are indexed as a single object in IBM StoredIQ Platform.
Each recurring calendar events will have multiple Event Date and End Date
attribute values, one pair per recurrence. For instance, if there is an event defined
for American Independence Day and is set to recur yearly, it will be indexed with
Event Dates 2010-07-04, 2011-07-04, 2012-07-04, and so on.

Survey

Only individual responses to a survey are indexed as system-level objects. Each


response is a given user's feedback to all questions in the survey, and each
question in the survey that was answered for a given response is indexed as an
attribute of the response in the IBM StoredIQ Platform index. The name of the
attribute is the string forming the question while the value is the reply entered.

Surveys have no full-text indexable body, and they are always indexed with
size=0.

126 Administration Guide


Hash computation

The hash of a full-text indexed object is generally computed using the full-text
indexable body of the object. However, in the case of SharePoint list item objects
(excluding documents and pictures), the full-text indexable body might be empty
or too simplistic, meaning that you could easily obtain duplicate items across
otherwise two completely different objects. For this reason, other attributes are
included in the hash computation algorithm.

These attributes are included while computing the hash for the SharePoint data
objects, excluding documents and pictures.
Table 4. Hash-computation attributes
Attributes Types
Generic attributes v Title (SharePoint)
v Content Type (SharePoint)
v Description (SharePoint)
Blog post attributes v Post category (SharePoint)
Wiki page attributes v Wiki page comment
Calendar event attributes v Event category (SharePoint)
v Event date (SharePoint)
v Event end date (SharePoint)
v Event location (SharePoint)
Task or project task attributes v Task start date (SharePoint)
v Task due date (SharePoint)
v Task assigned to (SharePoint)
Contact attributes v Contact full name (SharePoint)
v Contact email (SharePoint)
v Contact job title (SharePoint)
v Contact work address (SharePoint)
v Contact work phone (SharePoint)
v Contact home phone (SharePoint)
v Contact mobile phone (SharePoint)
Link attributes v Link URL (SharePoint)
Survey attributes All survey questions and answers in the response are
included in the hash.

Related reference:
Appendix A, Supported file types, on page 101

Appendix A. Supported file types 127


128 Administration Guide
Appendix B. Supported server platforms and protocols
This section lists the supported server platforms by volume type and the protocols
for supported systems.

Supported server platforms by volume type


Table 5. Supported server platforms by volume type
Primary Volumes v Windows Share v NFS v2 and v3 v Exchange (2003, v Domino for Email
v Windows Share v NFS (Celerra)* SP2-2007, 2010)* v Enterprise Vault,
(Celerra)* v NFS (NetApp)* v SharePoint v8.0.3, 9.0x, and
v Windows Share (2003/2007/2010)* 10.0
(NetApp)* v Symantec v IBM FileNet
Discovery v NewsGator
Accelerator, v8.0.3,
v Jive 5.0.2 and later
9.0x, and 10.0
v Chatter, including
private messages
Secondary Volumes v Windows Share v NFS v2 and v3 v Centera*
Retention Volumes v Windows Share v Enterprise Vault, v IBM FileNet v NFS (NetApp
v NFS v3 v8.0.3, 9.0x, and v Windows Share SnapLock)*
10.0 (NetApp v NFS (Celerra FLR)*
v Centera*
v Dell DX Object SnapLock)*
v Hitachi HCAP
Storage Platform v Windows Share
v IBM Information
(Celerra FLR)*
Archive*
Discovery Export v NFS v2 and v3 v Windows Share
Volumes
System Volumes v NFS v2 and v3

* Available through license only.

Protocols for supported systems


Table 6. Supported system and protocol version

Data Source Protocol System Version StoredIQ Version Notes


File Servers

Windows servers Windows Share V1.0 Windows NT 4.4.2

Windows 2000

Windows 2003

Windows 2008

UNIX servers NFS V2/V3 All 4.4.2 Supported, not tested

Copyright IBM Corp. 2001, 2013 129


Table 6. Supported system and protocol version (continued)

Data Source Protocol System Version StoredIQ Version Notes

EMC Celerra Windows Share V1.0 DART 5.5 or later 4.4.2 When back-up
operator is set,
Celerra does not reset
access times.

NFS V2/V3 DART 5.5 or later 4.4.2 Supported, not tested

NetApp Filer Windows Share V1.0 ONTAP 7.0.0 or later 4.4.2 Tested against 7.x
servers. NetApp 6.x
support until proven
otherwise.

NFS V2/V3 ONTAP 7.0.0 or later 4.4.2 Tested against 7.x


servers. NetApp 6.x
support until proven
otherwise.

Other Windows Share V1.0 All 4.4.2 Supported, not tested

Other NFS V2/V3 All 4.4.2 Supported, not tested


Email Servers

Exchange WebDav Exchange 2003 4.4.2

Exchange Web Exchange 2007 4.6.0


Services (Soap
XML-RPC)

Exchange 2010 5.0.4

Notes Domino NRPC over TCP/IP Lotus Domino/Notes 5.0.4


Email 6.x, 7.x, and 8.x

Email Archives

Symantec Enterprise Discovery Accelerator Discovery Accelerator 4.70


Vault/Discovery Web Services (Soap v. 9.0, 9.0.1, and 10.0
Accelerator XML-RPC)

Email Extender 4.7.0 StoredIQ understands


the Email Extender
format, but does not
communicate with
Extender.

Desktops

Windows Windows Share V1.0 Windows XP, Vista 4.4.2

Macintosh Windows Share V1.0 4.4.2

130 Administration Guide


Table 6. Supported system and protocol version (continued)

Data Source Protocol System Version StoredIQ Version Notes

NFS V2/V3 4.4.2

Unix/Linux NFS V2/V3 4.4.2


Document Management

Documentum TCP/IP 5.3, 6.0, 6.5 4.5.5

SharePoint SharePoint Web SharePoint 2003 4.4.2


Services (Soap
XML-RPC)

SharePoint SharePoint Web SharePoint 2007 4.6.0


Services (Soap
XML-RPC)

SharePoint 2010 5.0.4


Retention Servers

Centera TCP/IP CentraStar 3.1 4.4.2

Dell DX Object HTTP (SCSP subset) 4.0 5.0.4


Storage Platform

EMC Celerra w/FLR Windows Share V1.0 DART 5.5 4.6.0

EMC Celerra w/FLR NFS V2/V3 DART 5.5 4.6.0

Enterprise Vault DCOM/RPC 8.0 SP3 5.0.3

NetApp SnapLock Windows Share V1.0 ONTAP 7.2.1 and 4.6.0


later

NetApp SnapLock NFS V2/V3 ONTAP 7.2.1 and 4.6.0


later

IBM Information TCP/IP TSM Client 5.3.4.0, 4.4.2 Tivoli Storage


Archive TSM Server 5.3.4.0 Manager

Hitachi HCAP HTTP/HTTPS 1.1 HCA 1.8 4.4.2 Hitachi Content


Archiver
Enterprise Social Solutions
NewsGator HTTP REST 2.1.1229 or later 6.3.0.0

Appendix B. Supported server platforms and protocols 131


132 Administration Guide
Appendix C. Event log messages
This section contains a list of the messages that may appear in the Event Log of
the IBM StoredIQ Platform Console. Messages are sorted by type (ERROR, INFO,
WARN) and event number.

ERROR event log messages


This topic contains a complete listing of all ERROR event-log message, reasons for
occurrence, sample messages, and any required customer action.

Event
Type Number Reason Sample Message Customer Action

ERROR 1001 Harvester was unable to Harvester could not Log into UTIL and restart
open a socket for listening allocate listen port after the application server.
to child processes. <number> attempts. Restart the data server.
Cannot kickstart Contact Customer Support.
interrogators. (1001)

ERROR 9083 Unexpected error while Exporting volume Contact Customer Support.
exporting a volume. 'dataserver:/mnt/demo-A'
(1357) has failed (9083)

ERROR 9086 Unexpected error while Importing volume Contact Customer Support.
importing a volume 'dataserver:/mnt/demo-A'
(1357) failed (9086)

ERROR 15001 No volumes are able to be No volumes harvested. Make sure IBM StoredIQ
harvested in a given job. (15001) Platform still has
For instance, all of the appropriate permissions to
mounts fail due to a a volume. Verify there is
network issue. network connectivity
between the data server
and your volume. Contact
Customer Support.

ERROR 15002 Could not mount the Error mounting volume Make sure the data server
volume. Check <share> <startdir> on still has appropriate
permissions and network server <server-name>. permissions to a volume.
settings. Reported <reason>. (15002) Verify there is network
connectivity between the
data server and your
volume. Contact Customer
Support.

ERROR 15021 Error saving harvest Failed to save Contact Customer Support.
record HarvestRecord for This message occurs due
qa1:auto-A (15021) to a database error.

ERROR 17001 Unhandled fatal exception Centera Harvester fatal Contact Customer Support.
in Centera Discovery. failure: <exception
description> (17001)

Copyright IBM Corp. 2001, 2013 133


Event
Type Number Reason Sample Message Customer Action

ERROR 17012 Error while trying to create Unable to create Centera Contact Customer Support.
a volume during Centera Volume This message occurs due
Discovery Company_jpool_2009_ to a database error.
FEB_1 in pool jpool.
Error:<database error
description> (17012)

ERROR 17501 Generic retention Generic retention Contact Customer Support.


discovery failed in a discovery fatal failure:
catastrophic manner. <17501>

ERROR 17503 Generic retention Error creating/loading Contact Customer Support.


discovery creates volume volumeset for
sets associated with <server>:<share>
primary volumes. When
that fails, IBM StoredIQ
Platform send this
message. This failure likely
occurred due to database
errors.

ERROR 17505 Unable to query object Unable to determine object Contact Customer Support.
count for a discovered count for <server>:<share>
volume. This is likely due
to a database error.

ERROR 17506 Generic retention Error creating volume Contact Customer Support.
discovery could not create <server>:<share:>
discovered volume.

ERROR 18001 SMB connection fails. Windows Share Protocol Make sure IBM StoredIQ
Exception when Platform still has
connecting to the server appropriate permissions to
<server-name> : <reason>. a volume. Verify there is
(18001) network connectivity
between the data server
and your volume. Contact
Customer Support.

ERROR 18002 SMB volume mount failed. Windows Share Protocol Verify the name of the
Check the share name. Exception when server and volume to
connecting to the share make sure they are correct.
<share-name> on If this message persists,
<server-name> : <reason>. then Contact Customer
(18002) Support.

ERROR 18003 There is no volume Windows Share Protocol Contact Customer Support.
manager. Exception while
initializing the data object
manager: <reason>.
(18003)

134 Administration Guide


Event
Type Number Reason Sample Message Customer Action

ERROR 18006 Grazer volume crawl Grazer._run : Unknown Verify the user that
threw an exception. error during walk. (18006) mounted the specified
volume has permissions
equivalent to your current
backup solution. If this
message continues, contact
Customer Support.

ERROR 18021 An unexpected error from Unable to fetch trailing Check to ensure the
the server prevented the activity stream from NewsGator server has
harvest to reach the end of NewsGator volume. Will sufficient resources (disk
the activity stream on the retry in next harvest. space, memory, and so on).
NewsGator data source (18021) It is very likely that this
being harvested. The next error is transient. If the
incremental harvest will error persists across
attempt to pick up from multiple harvests, contact
where the current harvest Customer Support.
was interrupted.

ERROR 18018 Start directory has escape Cannot graze the volume, Consider turning off
characters, and the data root directory Nez has escape character checking.
server is configured to escape characters (18018)
skip them.

ERROR 19001 An exception occurred Interrogator.__init__ Contact Customer Support.


during interrogator exception: <reason>.
initialization. (19001)

ERROR 19002 An unknown exception Interrogator.__init__ Contact Customer Support.


occurred during exception: unknown.
interrogator initialization. (19002)

ERROR 19003 An exception occurred Interrogator.process Contact Customer Support.


during interrogator exception (<volumeid>,
processing. <epoch>): <reason>.
(19003)

ERROR 19004 An unknown exception Interrogator.process Contact Customer Support.


occurred during exception (<volumeid>,
interrogator processing. <epoch>). (19004)

ERROR 19005 An exception occurred Viewer.__init__: Exception Contact Customer Support.


during viewer - <reason>. (19005)
initialization.

ERROR 19006 An unknown exception Viewer.__init__: Unknown Contact Customer Support.


occurred during viewer exception. (19006)
initialization.

Appendix C. Event log messages 135


Event
Type Number Reason Sample Message Customer Action

ERROR 33003 Could not mount the Unable to mount the Verify user name and
volume. Check volume: <error reason>. password used for
permissions and network (33003) mounting the volume are
settings. accurate. Check the user
data object for appropriate
permissions to the volume.
Make sure the volume is
accessible via one of our
built-in protocols (NFS,
Windows Share, or
Exchange). Verify that the
network is properly
configured for the data
server to reach the
volume. Verify the data
server has appropriate
DNS settings to resolve the
server name.

ERROR 33004 Volume could not be Unmounting volume failed Reboot the data server. If
unmounted. from mount point : the problem persists, then
<mount point>. (33004) Contact Customer Support.

ERROR 33005 Data server was unable to Unable to create Reboot the data server. If
create a local mounting mount_point using the problem persists, then
point for the volume. primitive.threadSafe contact Customer Support.
Makedirs(). (33005)

ERROR 33010 Failed to make SMB Mounting Windows Share Verify user name and
connection to Windows volume failed with the password used for
Share server. error : <system error mounting the volume are
message>. (33010) accurate. Check the user
data object for appropriate
permissions to the volume.
Make sure the volume is
accessible via one of our
built-in protocols
(Windows Share). Verify
that the network is
properly configured for
the data server to reach
the volume. Verify the
data server has
appropriate DNS settings
to resolve the server name.

ERROR 33011 Internal error. Problem Unable to open Reboot the data server. If
accessing local /proc/mounts. Cannot test the problem persists, then
/proc/mounts if volume was already contact Customer Support.
mounted. (33011)

ERROR 33012 Database problems when An exception occurred Contact Customer Support.
deleting a Volume while working with
HARVESTS_TABLE in
Volume._delete(). (33012)

136 Administration Guide


Event
Type Number Reason Sample Message Customer Action

ERROR 33013 No volume set was found Unable to load volume set Contact Customer Support.
for the given volumes set by its name. (33013)
name.

ERROR 33014 System could not An error occurred while Contact Customer Support.
determine when this performing the
volume was last harvested. last_harvest operation.
(33014)

ERROR 33018 An error occurred Mounting Exchange Server Verify user name and
mounting the Exchange failed : <reason>. (33018) password used for
share. mounting the share are
accurate. Check for
appropriate permissions to
the share. Make sure the
share is accessible. Verify
that the network is
properly configured for
the data server to reach
the share. Verify the data
server has appropriate
DNS settings to resolve the
server name.

ERROR 33019 Failed to connect and Mounting HCAP volume Ensure the connectivity,
authenticate to the Hitachi failed : Cannot connect to credentials and
Archivas Content Archive HCAP share. (33019) permissions to the Hitachi
server. volume and retry.

ERROR 33020 Failed to connect and Mounting IBM Information Ensure the connectivity,
authenticate to the IBM Archive volume failed credentials and
Information Archive with the error: Server permissions to the IBM
retention volume. unreachable. (33020) Information Archive
volume and retry.

ERROR 33022 Failed to connect to the Mounting Discovery Verify the information
Discovery Accelerator Accelerator volume failed used to add the volume
using the information for with the error: insufficient and ensure all details have
the Volume. permissions to review been entered correctly
CaseOne (33022) before retrying. If the error
points to network issues
with connectivity, address
them and retry.

ERROR 33027 The attempt to connect Mounting IBM FileNet Ensure the connectivity,
and authenticate to the retention volume failed : credentials, and
IBM FileNet server failed. <reason>. (33027) permissions to the FileNet
volume and retry.

ERROR 34002 Could not complete the Copy Action aborted as Verify there is space
copy action because the the target disk has run out available on your policy
target disk was full. of space (34002) destination and try again.

Appendix C. Event log messages 137


Event
Type Number Reason Sample Message Customer Action

ERROR 34009 Could not complete the Move Action aborted as Verify there is space
move action due to full the target disk has run out available on your policy
target disk. of space. (34009) destination, then run
another harvest before
executing your policy.
When the harvest
completes, try running the
policy again.

ERROR 34015 The policy audit could not Error Deleting Policy Contact Customer Support.
be deleted for some Audit: <error message>
reason. (34016)

ERROR 34020 The copy to Centera action Copy to Centera failed as Check permissions on the
could not be executed we do not have read/write access profile provided for
because of insufficient permissions on the access the Centera pool on which
permissions. profile used. (34020) the volume has been
defined, and check if the
proper permissions have
been provided.

ERROR 34021 The move to Centera Move to Centera failed as Check permissions on the
action could not be we do not have read/write access profile provided for
executed because of permissions on the access the Centera pool on which
insufficient permissions. profile used. (34020) the volume has been
defined, and check if the
proper permissions have
been provided.

ERROR 34026 HSM action cannot recall a HSM Stub action aborted Verify there is space
file because the primary because the primary disk available on the volume
volume does not have has run out of space that the file is being
enough free space. (34026) recalled to and run the
HSM action again.

ERROR 34030 Discovery export policy is Production Run action Create sufficient space on
aborted since it detected aborted because the target target disk and run
the target disk is full. disk has run out of space. discovery export policy
(34030) again.

ERROR 34034 The target volume for the Copy objects failed, unable Ensure the connectivity,
policy could not be to mount volume: login credentials and
mounted. The policy will QA1.COMPANY. permissions to the target
be aborted. COM:SHARE. (34034) volume for the policy and
retry.

ERROR 41004 The job terminated <job-name> ended Try to run the job again. If
abnormally. unexpectedly. (41004) it fails again, contact
Customer Support.

138 Administration Guide


Event
Type Number Reason Sample Message Customer Action

ERROR 41007 Job has failed. [Job name] has failed Look at previous messages
(41007). to see why it failed and
refer to that message ID to
pinpoint the error. Contact
Customer Support.

ERROR 42001 The copy action could not Copy data objects did not Contact Customer Support.
run because of parameter run. Errors
errors. occurred:<error-
description>. (42001)

ERROR 42002 The copy action was Copy data objects failed, Check permissions on the
unable to create a target unable to create target target. Make sure the
directory. dir:<target- permissions that are
directory-name>. (42002) configured to mount the
target volume have write
access to the volume.

ERROR 42004 An unexpected error Copy data objects Contact Customer Support.
occurred. terminated abnormally.
(42004)

ERROR 42006 The move action could not Move data objects did not Contact Customer Support.
run because of parameter run. Errors
errors. occurred:<error-
description>. (42006)

ERROR 42007 The move action was Move data objects failed, Check permissions on the
unable to create a target unable to create target target. Make sure the
directory. dir:<target-directory- permissions that are
name>. (42007) configured to mount the
target volume have write
access to the volume.

ERROR 42009 An unexpected error Move data objects Contact Customer Support.
occurred. terminated abnormally.
(42009)

ERROR 42017 An unexpected error Delete data objects Contact Customer Support.
occurred. terminated abnormally.
(42017)

ERROR 42025 The policy action could Policy cannot execute. Contact Customer Support.
not run because of Attribute verification
parameter errors. failed. (42025)

ERROR 42027 An unexpected error Policy terminated Contact Customer Support.


occurred. abnormally. (42027)

Appendix C. Event log messages 139


Event
Type Number Reason Sample Message Customer Action

ERROR 42050 The data synchronizer Content Data Synchronizer Contact Customer Support.
could not run because of synchronization of
an unexpected error. <server-name>:
<volume-name> failed
fatally.

ERROR 42059 Illegal set of parameters Production Run on objects Contact Customer Support.
passed to discovery export did not run. Errors
policy. occurred: The following
parameters are missing:
action_limit. (42059)

ERROR 42060 Discovery export policy Production Run on objects Verify the Discovery
failed to create target (Copying native objects) export volume has write
directory for the export. failed, unable to create permission and re-execute
target dir: production/10. policy.
(42060)

ERROR 42062 Discovery export policy Production Run on objects Contact Customer Support.
was terminated (Copying native objects)
abnormally. terminated abnormally.
(42062)

ERROR 42088 The full-text optimization Full-text optimization Contact Customer Support.
process failed; however, failed on volume
the index is most likely <volume-name> (42088)
still usable for queries.

ERROR 45802 A full-text index is already Time allocated to gain Contact Customer Support.
being modified. exclusive access to
in-memory index for
volume= 1357 has expired
(45802)

ERROR 45803 The index for the specified Index '/deepfs/fulltext/ No user intervention is
volume does not exist. volume_index/ required.
This message may occur volume_1357' not found
under normal conditions. (45803)

ERROR 45804 Programming error. A Transaction of client: Contact Customer Support.


transaction was never node.client.com_FINDEX_
initiated or was closed QUEUE_
early. 1357_1172515222_3_ 2 is
not the writer (45804)

ERROR 45805 The query has not been Query ID: 123 does not No user intervention is
started or has expired The exist (45805) required.
forme.r is a programming
error, the latter is normal.

ERROR 45806 The query expression is Failed to parse 'dog pre\3 Revise your full-text query.
invalid or not supported. bar' (45806)

140 Administration Guide


Event
Type Number Reason Sample Message Customer Action

ERROR 45807 Programming error. A Client: Contact Customer Support.


transaction has already node.client.com_FINDEX_QUEUE
been started for the client. _1357_1172515222_3_2 is
already active (45807)

ERROR 45808 A transaction has never No transaction for client: No user intervention is
been started or has node.client.com_FINDEX_ required. The system
expired. QUEUE_1357_1172515222_3_2 handles this condition
(45808) internally.

ERROR 45810 Programming error. Invalid volumeId. Contact Customer Support.


Expected: 1357
Received:2468 (45810)

ERROR 45812 A File I/O error occurred Failed to write disk Try your query again.
while accessing index data. (45812). Contact Customer Support
for additional assistance if
necessary.

ERROR 45814 The query expression is Query: 'a* b* c* d* e*' is Refine your full-text query.
too long. too complex (45814)

ERROR 45815 The file that is being Java heap exhausted while Check the skipped file list
indexed is too large or the indexing node with ID: in the audit log for files
query expression is too '10f4179cd5ff22f2a6b that failed to load due to
complex. The engine has 79a1bc3aef247fd94ccff' their sizes. Revise your
temporarily run out of (45815) query expression and retry.
memory.

ERROR 46023 Tar command failed while Failed to back up fulltext Check disk space and
persisting full-text data to data for server:share. permissions.
Windows Share or NFS Reason: <reason>. (46023)
share.

ERROR 46024 Unhandled fatal exception Exception <exception> Contact Customer Support.
while persisting full-text while backing up fulltext
data into a .tgz file. data for server:share
(46024)

ERROR 46025 Was not able to delete Failed to unlink Check permissions.
partial .tgz file after a incomplete backup image.
failed full-text backup. Reason: <reason>. (46025)

ERROR 47002 Synchronization failed on Synchronization failed for Contact Customer Support.
a query. query '<query-name>' on
volume
'<server-and-volume>
(47002)

ERROR 47101 An error occurred during Cannot process full-text Restart services and
the query of a full-text expression (Failed to read contact Customer Support.
expression. from disk (45812) (47101)

Appendix C. Event log messages 141


Event
Type Number Reason Sample Message Customer Action

ERROR 47203 No more database Database connections Contact Customer Support.


connections are available. exhausted (512/511)
(47203)

ERROR 47207 User is running out of disk Disk usage exceeds Contact Customer Support.
space. threshold. (%d) In rare cases, this message
can indicate a program
error leaking disk space. In
most cases, however, disk
space is almost full, and
additional storage is
required.

ERROR 47212 Interrogator crashed while Harvester 1 Does not exist. If the problem persists
processing a file. The Action taken : restart. (that is, the system crashes
current file will be missing (47212) on the same file or type of
from the volume cluster. files), contact Customer
Support.

ERROR 47214 SNMP notification sender Unable to resolve host Check spelling and DNS
is unable to resolve the name setup.
trap host name. nomachine.nowhere.com
(47214)

ERROR 50011 The DDL/DML files Database version control Contact Customer Support.
required for the database SQL file not found. (50011)
versioning were not found
in the expected location on
the data server.

ERROR 50018 Indicates that the Database restore is Contact Customer Support.
pre-upgrade database unsuccessful. Contact
restoration failed, which Customer Support. (50018)
was attempted as a result
of a database upgrade
failure.

ERROR 50020 Indicates the current Versions do not match! Contact Customer Support.
database requirements do Expected current database
not meet those specified version: <dbversion>.
for the upgrade and (50020)
cannot proceed with the
upgrade.

ERROR 50021 Indicates that the full Database backup failed. Contact Customer Support.
database backup failed (50021)
when attempting a
data-object level database
backup.

ERROR 61003 Discovery export policy Production policy failed to


failed to mount volume. mount volume. Aborting.
Aborting. (61003)

142 Administration Guide


Event
Type Number Reason Sample Message Customer Action

ERROR 61005 The discovery export load Production load file Contact Customer Support.
file generation fails generation failed. Load
unexpectedly. The load files may be produced, but
files may be produced post-processing may be
correctly, but incomplete. (61005)
post-processing actions
like updating audit trails
and generating report files
may not have completed.

ERROR 61006 The discovery export load Production load file Free up space on the target
file generation was generation interrupted. disk, void the discovery
interrupted because the Target disk full. (61006) export run and run the
target disk is full. policy once more.
ERROR 68001 The gateway and data Gateway connection failed Update your data server to
server must be on the due to unsupported data the same build number as
same version in order to server version. the gateway and restart
connect. services. If your encounter
issues, contact Customer
Support.
ERROR 68003 The data server has failed The data-server connection Contact Customer Support.
to connect to the gateway to the gateway cannot be
over an extended period of established.
time.

ERROR 80002 The system failed to open Failed to connect to the The "maximum database
a connection to the database (80002). connections" configuration
database. parameter of the Database
engine may need to be
increased. Contact
Customer Support.

Related reference:
Appendix C, Event log messages, on page 133

INFO event log messages


This topic contains a complete listing of all INFO event-log messages, reasons for
occurrence, sample messages, and any required customer action.

Event
Type Number Reason Sample Message Customer Action

INFO 9001 No conditions have been Harvester: Query <query Add conditions to the
added to a query. name> cannot be inferred specified query.
because no condition for it
has been defined (9001).

Appendix C. Event log messages 143


Event
Type Number Reason Sample Message Customer Action

INFO 9002 One or more conditions in Harvester: Query < query Verify that regular
a query were incorrect. name> cannot be inferred expressions are properly
because of regular formed.
expression or other
condition error (9002).

INFO 9003 Volume Harvest has Volume statistics No user intervention is


completed and explorers computation started (9003). required.
are being calculated.

INFO 9004 Explorer calculations have Volume statistics No user intervention is


completed. computation completed required.
(9004).

INFO 9005 Query membership Query inference will be No user intervention is


calculations started. done in <number> steps required.
(9005).

INFO 9006 Query membership Query inference step No user intervention is


calculations progress <number> done (9006). required.
information.

INFO 9007 Query membership Query inference completed No user intervention is


calculations completed. (9007). required.

INFO 9012 Indicates the end of Dump of Volume cache(s) No user intervention is
dumping the content of completed (9012). required.
the volume cache.

INFO 9013 Indicates the beginning of Postprocessing for volume No user intervention is
the load process. 'Company Data required.
Server:/mnt/demo-A'
started (9013).

INFO 9067 Indicates load progress. System metadata and No user intervention is
tagged values were required.
successfully loaded for
volume 'server:volume'
(9067).

INFO 9069 Indicates load progress. Volume 'data server: No user intervention is
/mnt/demo-A': System required.
metadata, tagged values
and full-text index were
successfully loaded (9069).

INFO 9084 The volume export Exporting volume 'data No user intervention is
finished. server:/mnt/demo-A' required.
(1357) completed (9084)

INFO 9087 The volume import Importing volume No user intervention is


finished. 'dataserver:/mnt/demo-A' required.
(1357) completed (9087)

144 Administration Guide


Event
Type Number Reason Sample Message Customer Action

INFO 9091 The load process has been Load aborted due to user No user intervention is
aborted by the user. request (9091). required.

INFO 15008 The volume load step has Post processing skipped No user intervention is
been skipped, per user for volume required.
request. <server>:<volume>.
(15008)

INFO 15009 The volume load step has Harvest skipped for No user intervention is
been executed but the volume required.
harvest step has been <server>:<volume>.
skipped, per user request. (15009)

INFO 15012 The policy running on the Volume <volume> on No user intervention is
volume has completed and server <server> is free required.
the volume load can now now. Proceeding with load.
proceed. (15012)

INFO 15013 The configured time limit Harvest time limit reached No user intervention is
on a harvest was reached. for server:share. Ending required.
harvest now. (15013)

INFO 15014 The configured object Object count limit reached No user intervention is
count limit on a harvest for server:share. Ending required.
was reached. harvest now. (15014)

INFO 15017 Check box selected for Deferring post processing No user intervention is
nightly load job. for volume server:vol required.
(15017)

INFO 15018 Harvest size and/or time Harvest limit reached on No user intervention is
limit is reached. server:volume. Synthetic required.
deletes will not be
computed. (15018)

INFO 15019 User stops harvest process. Harvest stopped by user No user intervention is
while processing volume required.
dpfsvr:vol1. Rest of
volumes will be skipped.
(15019)

INFO 15020 Harvest vocabulary has Vocabulary for Full harvest should be run
changed. Full harvest dpfsvr:jhaide-A has instead of an incremental
should run instead of changed. A full harvest is harvest.
incremental. recommended (15020).

INFO 15022 The user is trying to Permission-only harvest: No action needed as the
execute an ACL-only permission checks not volume is skipped.
harvest on a volume that supported for
is not a Windows Share or <server>:<share>
SharePoint volume.

Appendix C. Event log messages 145


Event
Type Number Reason Sample Message Customer Action

INFO 15023 The user is trying to Permission-only harvest: No action needed as the
execute an ACL-only volume <server>:<share> volume is skipped.
harvest on a volume that has no associated user list.
has not been assigned a
user list.

INFO 17002 Sent when Centera Centera External Iterator : No user intervention is
Discovery sends the query Starting to populate using required.
to the Centera server pool QAPOOL. (17002)

INFO 17003 Sent when Centera <servername>: Created No user intervention is


Discovery autocreates a new volume required.
new volume QAPOOL:_QAPOOL_2009_JAN_1
(17003)

INFO 17004 A Centera Discovery Object limit reached for No user intervention is
auto-created volume has _QAPOOL_2009_JAN_1, required.
reached the preconfigured starting new volume
limit, starting a new one. (17004)

INFO 17007 Pending data return from Centera Harvester: No Check if a Centera node is
Centera. items returned from down. A Centera cluster
Centera for over 5 may be overloaded.
minutes. Still waiting.
(17007)

INFO 17009 Configured time limit Centera Discovery : time No user intervention is
reached for Centera limit for discovery required.
discovery step. reached. Ending this run.
(17009)

INFO 17010 Configured object count Centera Discovery: No user intervention is


limit reached for Centera configured item count required.
discovery step. limit for discovery
reached. Ending this run.
(17010)

INFO 17507 Limit (time or object Retention discovery limit Contact Customer Support.
count) reached for generic reached for
retention discovery. <server>:<share>

INFO 17508 Generic retention No new items discovered. No user intervention is


discovery found no new Post-processing skipped required unless the user is
items for this master for volume certain that new items
volume. <server>:<share> should have been
discovered.

INFO 17509 Generic retention Created new discovered No user intervention is


discovery created a new volume <server>:<share> required.
volume. in volume set
<autodiscovered volume
set name>.

146 Administration Guide


Event
Type Number Reason Sample Message Customer Action

INFO 18004 Job was stopped. Walker._processFile: No user intervention is


Grazer Stopped. (18004) required.

INFO 18005 Grazer queue was closed. Walker._processFile: No user intervention is


Grazer Closed. (18005) required.

INFO 18016 Displays the list of top Choosing top level No user intervention is
level directories selected directories: <directories> required.
by matching the start (18016)
directory regular
expression. Displays at the
beginning of a harvest.

INFO 34001 Marks current progress of <volume>: <count> data No user intervention is
a copy action. objects processed by copy required.
action. (34001)

INFO 34004 Marks current progress of <volume>: <count> data No user intervention is
a delete action. objects processed by delete required.
action. (34004)

INFO 34008 Marks current progress of <volume>: <count> data No user intervention is
a move action. objects processed by move required.
action. (34008)

INFO 34014 A policy audit was Deleting Policy Audit # No user intervention is
deleted. <audit id> <policy name> required.
<start time> (34014)

INFO 34015 A policy audit was Deleted Policy Audit # No user intervention is
deleted. <audit id> <policy name> required.
<start time> (34015)

INFO 34031 Progress update of the Winserver:topshare : 30000 No user intervention is


discovery export policy, data objects processed by required.
every 10000 objects production action. (34031)
processed.

INFO 41001 A job was started either <jobname> started. (41001) No user intervention is
manually or was required.
scheduled.

INFO 41002 The user stopped a job <jobname> stopped at user No user intervention is
that was running. request (41002) required.

INFO 41003 A job completed normally <jobname> completed. No user intervention is


with or without success. (41003) required.

INFO 41006 Rebooting or restarting Service shutdown. Rerun jobs after reboot or
services on the controller Stopping outstanding jobs. restart if you want the jobs
or compute node causes (41006) to complete.
all jobs to stop.

Appendix C. Event log messages 147


Event
Type Number Reason Sample Message Customer Action

INFO 41008 Database compactor Database compactor was Set the database
(vacuum) job can not run not run because other jobs compactor's job schedule
while there is database are active (41008). so that it doesnt conflict
activity. with long-running jobs.

INFO 42005 The action completed or Copy complete: <number> No user intervention is
was aborted. Shows results data objects required.
of copy action. copied,<number>
collisions found. (42005)

INFO 42010 The action completed or Move complete: <number> No user intervention is
was aborted. Shows results data objects required.
of move action. moved,<number>
collisions found. (42010)

INFO 42018 The action completed or Copy data objects No user intervention is
was aborted. Shows results complete: <number> data required.
of deleted action. objects copied,<number>
collisions found. (42018)

INFO 42020 Third-party application Centera Deleted Files No user intervention is


deleted clips on discovered Synchronizer complete: required.
Centera volumes. 10000 objects passed. 2
objects were missing from
the Centera cluster. (42020)

INFO 42024 The synchronizer Content Data Synchronizer No user intervention is


completed normally. complete. (42024) required.

INFO 42028 The action completed or Policy completed (42028). No user intervention is
was aborted. Shows results required.
of policy action.

INFO 42032 The action completed or <report name> completed No user intervention is
was aborted. Shows results (42032). required.
of report action.

INFO 42033 The synchronizer started Content Data Synchronizer No user intervention is
automatically or manually started. (42033) required.
with the GUI button.

INFO 42048 Reports that the Content Data Synchronizer No user intervention is
synchronizer is skipping a skipping required.
volume if synchronization <server-name>:<volume-
is determined not to be name> as it does not need
required. synchronization. (42048)

INFO 42049 Reports that the Content Data Synchronizer No user intervention is
synchronizer has started starting synchronization required.
synchronization of a for volume
volume <server-name>:<volume-
name>

148 Administration Guide


Event
Type Number Reason Sample Message Customer Action

INFO 42053 The policy, that was Proceeding with execution No user intervention is
waiting for participant of <policy-name>. required.
volumes to be loaded
before continuing, is now
starting.

INFO 42063 Report on completion of Production Run on objects No user intervention is


discovery export policy (Copying native objects) required.
execution phase. completed: 2003 data
objects copied, 25
duplicates found. (42063)

INFO 42065 A discovery export policy Proceeding with execution No user intervention is
that was held up for want of 'Production case One'. required.
of resources, is now done (42065)
waiting, and will begin
execution.

INFO 42066 A new discovery export New run number 10 Note the new run number
run has started. started for production in order to tie the current
Production Case 23221. run with the
(42066) corresponding audit trail.

INFO 42067 Discovery export policy is Production Run producing No user intervention is
preparing the audit trail in Audit Trail XML. (42067) required.
XML format. This may
take a few minutes.

INFO 42074 A query or tag has been Successfully sent query No user intervention is
replicated to a member 'Custodian: Joe' to member required.
data server successfully. data server San Jose Office
(42074)

INFO 46001 The backup process has Backup Process Started. No user intervention is
begun. Any selected (46001) required.
backups in the system
configuration screen will
be run if necessary.

INFO 46002 The backup process did Backup Process Failed: Check your backup
not complete all its tasks <error-description>. volume.
successfully. One or more (46002)
backup types did not
occur.

INFO 46003 The backup process Backup Process Finished. No user intervention is
successfully completed (46003) required.
attempting all the
necessary tasks. Any parts
of the overall process add
their own log entries.

Appendix C. Event log messages 149


Event
Type Number Reason Sample Message Customer Action

INFO 46004 The Application Data Application Data backup Check your backup
backup, as part of the failed. (46004) volume. Look at the setup
overall backup process, for the Application Data
needed to run but did not backup. If backups
succeed. continue to fail, Contact
Customer Support.

INFO 46005 The Application Data Application Data backup No user intervention is
backup, as part of the finished. (46005) required.
overall backup process,
needed to run and
succeeded.

INFO 46006 The Application Data Application Data backup No user intervention is
backup, as part of the not configured, skipped. required.
overall backup process, (46006)
was not configured.

INFO 46007 The Harvested Volume Harvested Volume Data Check your backup
Data backup, as part of the backup failed. (46007) volume. Look at the setup
overall backup process, for the Harvested Volume
needed to run but did not Data backup. If backups
succeed. continue to fail, contact
Customer Support.

INFO 46008 The Harvested Volume Harvested Volume Data No user intervention is
Data backup, as part of the backup finished. (46008) required.
overall backup process,
needed to run and
succeeded.

INFO 46009 The Harvested Volume Harvested Volume Data No user intervention is
Data backup, as part of the backup not configured, required.
overall backup process, skipped. (46009)
was not configured

INFO 46010 The System Configuration System Configuration Check your backup
backup, as part of the backup failed. (46010) volume. Look at the setup
overall backup process, for the System
needed to run but did not Configuration backup. If
succeed. backups continue to fail,
contact Customer Support.

INFO 46011 The System Configuration System Configuration No user intervention is


backup, as part of the backup finished. (46011) required.
overall backup process,
needed to run and
succeeded.

INFO 46012 The System Configuration System Configuration No user intervention is


backup, as part of the backup not configured, required.
overall backup process, skipped. (46012)
was not configured.

150 Administration Guide


Event
Type Number Reason Sample Message Customer Action

INFO 46013 The Audit Trail backup, as Policy Audit Trail backup Check your backup
part of the overall backup failed. (46013) volume. Look at the setup
process, needed to run but for the Audit Trail backup.
did not succeed. If backups continue to
fail, Contact Customer
Support.

INFO 46014 The Audit Trail backup, as Policy Audit Trail backup No user intervention is
part of the overall backup finished. (46014) required.
process, needed to run and
succeeded.

INFO 46015 The Audit Trail backup, as Policy Audit Trail backup No user intervention is
part of the overall backup not configured, skipped. required.
process was not (46015)
configured.

INFO 46019 Volume cluster backup Indexed Data backup Contact Customer Support.
failed. failed: <specific error>
(46019)

INFO 46020 Volume cluster backup Indexed Data backup No user intervention is
finished. finished. (46020) required.

INFO 46021 Volume is not configured Indexed Data backup not No user intervention is
for indexed data backups. configured, skipped. required.
(46021)

INFO 46022 Full-text data was Successfully backed up No user intervention is


successfully backed up. full-text data for required.
server:share (46022)

INFO 47213 Interrogator was Harvester 1 is now No user intervention is


successfully restarted. running. (47213) required.

INFO 60001 The user updates an object Query cities was updated No user intervention is
on the system. This by the administrator required.
includes any object type account (60001).
on the data server,
including the updating of
volumes.

INFO 60002 The user creates an object. Query cities was created No user intervention is
This includes any object by the administrator required.
type on the data server, account (60002).
including the creation of
volumes.

INFO 60003 The user deletes an object. Query cities was deleted No user intervention is
This includes any object by The administrator required.
type on the data server, account (60003).
including the deletion of
volumes.

Appendix C. Event log messages 151


Event
Type Number Reason Sample Message Customer Action

INFO 60004 The user publishes a Query cities draft was No user intervention is
full-text query set or a published by the required.
query. administrator account
(60004).

INFO 60005 The user tags an object. Query tagging for cities No user intervention is
This includes a published class was started by the required.
query, a draft query, or administrator account
tag. (60005).

INFO 60006 A user restarted services Application services restart No user intervention is
on the data server. for all data servers was required.
requested by the
administrator account
(60006).

INFO 61001 Concordance discovery Preparing for upload of No user intervention is


export is now preparing load file(s). (61001) required.
the load file(s).

INFO 61002 Concordance discovery Load file(s) ready for No user intervention is
export is ready to upload upload. (61002) required.
the load file(s).

INFO 65000 The log file has finished Log file download No user intervention is
downloading complete (65000) required.

Related reference:
Appendix C, Event log messages, on page 133

WARN event log messages


This topic contains a complete listing of all WARN event-log messages, reasons for
occurrence, sample messages, and any required customer action.

Event
Type Number Reason Sample Message Customer Action

WARN 1002 An Interrogator process Processing could not be Classify the document
died because of an completed on object, manually and Contact
unknown error. The data interrogator died : <data Customer Support.
object it was processing object name>. (1002)
will be skipped. A new
process will be created to
replace it.

WARN 1003 Interrogator child process Interrogator terminated Try to re-add the volume
did not properly get before accessing data being harvested. If that
started. There may have objects. (1003) fails, Contact Customer
been problems accessing Support.
the Volume to be
harvested.

152 Administration Guide


Event
Type Number Reason Sample Message Customer Action

WARN 1004 Interrogator child process Processing was not Contact Customer Support.
was terminated because it completed on object,
was no longer responding. interrogator killed : <data
The data object it was object name>. (1004)
processing will be skipped.
A new process will be
created to replace it.

WARN 6001 A user email could not be Failed to send an email to Verify that your SMTP
sent. The mail server user <email address>; server is configured
settings are incorrect. check mail server correctly. Make sure the IP
configuration settings address configured for the
(6001). data server is permitted to
relay on the configured
SMTP server.

WARN 8001 The database needs to be The Database is Run the Database
vacuumed. approaching an maintenance task to
operational limit. Please vacuum the database.
run the Database
maintenance task using the
Console interface (8001)

WARN 9068 Tagged values were System metadata and Contact Customer Support.
loaded, but full-text index tagged values were loaded
loading failed. successfully for volume
'server:volume', but
loading the full-text index
failed (9068)

WARN 9070 Tagged values and full-text Loading system metadata, Contact Customer Support.
index loading failed. tagged values and the
full-text index failed for
volume 'server:volume'
(9070)

WARN 15003 The volume mount Volume <volume name> Contact Customer Support.
appeared to succeed, but on server <server name> is
the test for mount failed. not mounted. Skipping.
(15003)

WARN 15004 A component cleanup [<component>] Cleanup Contact Customer Support.


failure on stop or failure on stop. (15004)
completion.

WARN 15005 There was a component [<component>] Run Contact Customer Support.
run failure. failure. (15005)

WARN 15006 Cleanup failed for [<component>] Cleanup Contact Customer Support.
component after a run failure on abort. (15006)
failure.

Appendix C. Event log messages 153


Event
Type Number Reason Sample Message Customer Action

WARN 15007 A component timed out Component Try your action again. If
and needs to be stopped. [<component>] this error continues,
unresponsive; contact Customer Support.
autostopping triggered.
(15007)

WARN 15010 The same volume cannot Volume <volume-name> No user intervention is
be harvested in parallel. on server <server-name> is required. You may wish to
The harvest will be already being harvested. verify that the volume
skipped and the next one, Skipping. (15010) harvest has completed.
if any are in queue,
started.

WARN 15011 A volume cannot be Volume <volume-name> No user intervention is


harvested if it is being on server <server-name> is required.
used by another job. The being used by another job.
harvest will continue when Waiting before proceeding
the job has completed. with load. (15011)

WARN 15015 Configured harvest time Time limit for harvest Reconfigure harvest time
limit reached. reached. Skipping Volume limit.
v1 on server s1. 1 (15015)

WARN 15016 Configured harvest object Object count limit for Reconfigure harvest data
count limit reached. harvest reached. Skipping object limit.
Volume v1 on server s1
(15016)

WARN 17008 Query ran to discover Centera External Iterator : Contact Customer Support.
Centera items terminated Centera Query terminated
unexpectedly. unexpectedly (<error
description>). (17008)

WARN 17011 Running discovery on the Pool Jpool appears to have Make sure that two jobs
same pool in parallel is another discovery running. are not running at the
not allowed. Skipping. (17011). same time that discover
the same pool.

WARN 17502 Generic retention Volume <server>:<share> No user intervention is


discovery is already appears to have another required as the next step,
running for this master discovery running. if any, within the job will
volume. Skipping. be executed.

WARN 17504 Sent when a retention Volume <server>:<share> Contact Customer Support.
discovery is run on any is not supported for
volume other than a discovery. Skipping.
Windows Share retention
volume.

154 Administration Guide


Event
Type Number Reason Sample Message Customer Action

WARN 18007 Directory listing or Walker._walktree: OSError Make sure the appliance
processing of data object - <path><reason> (18007) still has appropriate
failed in Grazer. permissions to a volume.
Verify there is network
connectivity between the
appliance and your
volume. Contact Customer
Support.

WARN 18008 Unknown error occurred Walker._walktree: Contact Customer Support.


while processing data Unknown exception -
object or listing directory. <path>. (18008)

WARN 18009 Grazer timed out Walker._processFile: Contact Customer Support.


processing an object. Grazer Timed Out. (18009)

WARN 18010 The skipdirs file is either Unable to open skipdirs Contact Customer Support.
not present or not readable file: <filename>. Cannot
by root. skip directories as
configured. (18010)

WARN 18011 An error occurred reading Grazer._run: couldn't read Contact Customer Support.
the known extensions list extensions - <reason>.
from the database. (18011)

WARN 18012 An unknown error Grazer._run: couldn't read Contact Customer Support.
occurred reading the extensions. (18012)
known extensions list from
the database.

WARN 18015 NFS initialization warning NIS Mapping not User name and group
that NIS is not available. available. (18015) names may be inaccurate.
Please check that your NIS
server is available and
properly configured in the
data server.

WARN 18017 A folder in an Enterprise Skipping Folder (ID=3) in If all folders were expected
Vault case being harvested volume to be harvested in the
was skipped because of evdiscaccel.company.com: volume, verify that the
insufficient permissions. CaseOne due to username used to add the
insufficient permissions. Volume has Folder Review
Check Review permissions permission on all folders.
on the folder. (18017)

WARN 18019 The checkpoint saved from Unable to load checkpoint If the message repeats in
the last harvest of the for NewsGator volume. A subsequent harvests,
NewsGator data source full harvest will be contact Customer Support.
failed to load. Instead of performed instead. (18019)
performing an incremental
harvest, a full harvest will
be executed.

Appendix C. Event log messages 155


Event
Type Number Reason Sample Message Customer Action

WARN 18020 The checkpoint noted for Unable to save checkpoint If the message repeats in
the current harvest of the for NewsGator harvest of subsequent harvests,
NewsGator data source volume. (18020) contact Customer Support.
could not be saved. The
next incremental harvest of
the data source will not be
able to pick up from this
checkpoint.

WARN 33016 System could not unmount Windows Share Protocol Server administrators may
this volume. Session teardown failed. see connections left
(33016) hanging for a predefined
period of time. These
connections will drop off
after they time out. No
user intervention required.

WARN 33017 System encountered an An error occurred while Contact Customer Support.
error while trying to figure retrieving the query
out what query use this instances pointing to a
volume. volume. (33017)

WARN 33021 The teardown operation of IBM Information Archive None


the connection to a IBM tear-down failed. (33021)
Information Archive
volume failed. Some
connections may be left
open on the IBM
Information Archive server
until they are timed out.

WARN 33023 A connection to the HTTP Connection None


Discovery Accelerator tear-down to Discovery
could not be torn down Accelerator failed. (33023)
completely, and a few
resources on the server
might be tied up for a
short while.

WARN 33028 The tear-down operation IBM FileNet tear-down None


of the connection to a operation failed. (33028)
FileNet volume failed.
Some connections may be
left open on the FileNet
server until they have
timed out.

WARN 34003 Skipped a copy data object Copy action error :- Target Please verify there is space
because disk full error. disk full, skipping copy : available on your policy
<source volume> to destination and try again.
<target volume>. (34003)

156 Administration Guide


Event
Type Number Reason Sample Message Customer Action

WARN 34010 Skipped a move data Move action error :- Target Please verify there is space
object because disk full disk full, skipping copy : available on your policy
error. <source volume> to destination. After verifying
<target volume>. (34010) space is available, please
run another harvest before
executing your policy.
Upon harvest completion,
please try running the
policy again.

WARN 34029 Discovery export policy Discovery export Run Create sufficient space on
detects the target disk is action error: Target disk target disk and run
full and skips production full, skipping discovery discovery export policy
of an object. export: again.
share-1/saved/years.pdf
to production/10/
documents/1/0x0866e
5d6c898d9ffdbea720b0
90a6f46d3058605.txt.
(34029)

WARN 34032 The policy being executed No volumes in scope for Check policy query and
has no volumes in scope policy. Skipping policy scoping configuration, and
based on the configured execution. (34032) re-execute policy.
query and scoping. The
policy will not be
executed.

WARN 34033 Celerra data mover error. Data mover returned Consult the Celerra
There are a large number NO_MATCHING_CONNECTION Administrator Manual.
of possible causes; the : The specified
actual one is listed in the OFFLINE_PATH was not
error message. The sample able to map to an existing
provides the most connection to a secondary
common one. server. (34033)

WARN 34035 If the global hash setting Copy objects : Target hash If target hashes need to be
for the system is set to not will not be computed computed for the policy
compute data object hash, because Hashing is audit trail, turn on the
no hash can be computed disabled for system. global hash setting before
for the target objects (34035) executing the policy.
during a policy action.

WARN 34036 The policy has no source The policy has no source Confirm that the query
volume(s) in scope, volume(s) in scope. Wait used by the policy has one
meaning that the policy for the query to update or more volumes in scope.
cannot be executed. before executing the policy.
(34036)

WARN 42003 The job containing this Copy data objects stopped No user intervention is
action is stopped by the at user request. (42003) required.
user.

Appendix C. Event log messages 157


Event
Type Number Reason Sample Message Customer Action

WARN 42008 The job containing this Move data objects stopped No user intervention is
action is stopped by the at user request. (42008) required.
user.

WARN 42016 The job containing this Delete data objects No user intervention is
action is stopped by the stopped at user request. required.
user. (42016)

WARN 42026 The job containing this Policy stopped at user No user intervention is
action is stopped by the request. (42026) required.
user.

WARN 42035 When the job containing Set security for data No user intervention is
this action is stopped by objects stopped at user required.
the user. request. (42035)

WARN 42051 Two instances of the same Policy <policy-name> is No user intervention is
policy cannot run at the already running. Skipping. required.
same time. (42051)

WARN 42052 Policies will wait to One or more volume(s) No user intervention is
execute until after volumes needed by policy required.
have loaded, if those <policy-name> are being
volumes are participants to loaded. Waiting for the
the policy by virtue of bulk load(s) to finish.
being in the query. (42052)

WARN 42061 Discovery export policy Discovery export run on No user intervention is
was stopped by user. objects (Copying native required.
objects) stopped at user
request. (42061)

WARN 42064 Discovery export policy A Discovery export run The discovery export
execution has been related to policy policy execution is held up
delayed because a 'Discovery export case for required resources.
conflicting discovery One' is in progress. Execution should begin as
export run is in progress. Waiting for it to finish. soon as resource becomes
(42064) available.

WARN 42068 Policy failed to set Copy objects warning, Note that policy may not
appropriate permissions unable to set permissions be able to set appropriate
on the target directory. on target directory: permissions on the objects
Objects created from the share-1/saved. (42068) it creates. If this is not
policy may not have acceptable, verify that
appropriate permissions target volume has proper
set. write permissions and
re-execute.

158 Administration Guide


Event
Type Number Reason Sample Message Customer Action

WARN 42069 If the Copy data objects Discovery export If the modified objects
modified since last DAT_Export is configured need to be acted upon,
harvest option is selected to act on members of either use a discovery
for a discovery export containers, and cannot act export acting only on the
policy, it is only valid if on objects modified after original file/email archive,
the discovery export itself the last harvest. Discovery or perform an incremental
is defined to act on the export run X will skip harvest on the source
original file/email archive, modified objects. (42069) volume(s).
as opposed to their
members. If this is not
the case, the warning lets
the user know that
modified objects will still
be skipped.

WARN 46026 Volume is being harvested Volume volume:share is in Rerun backup when
or policies are running use. Unable to back up volume is not in use.
against it. If there are other full-text index. Will retry
full-text indexes to be later. (46026)
backed up, the system
works on those and retries
this volume later.

WARN 47201 Database connections are Database connections at No user intervention is


down to a healthy level. normal level again required.
(512/100) (47201)

WARN 47202 The system is starting to Database connections Contact Customer Support.
run low on database usage seems excessive
connections. This is (512/415) (47202)
abnormal, and an
indication of process
restarts and connections
not being cleared.

WARN 47215 Someone internally or SSHD: Failed password for Contact your local IT
externally is trying (and root from 172.17.18.185 manager. This could be
failing) to SSH into the port 57982. (47125) either a mistyped
data server. password by a legitimate
user or in the worst case
scenario, a genuine
break-in attempt.

WARN 61003 One of the load files Failed to mount Some of the load files will
cannot be uploaded transaction cache dump be missing after the
because the compute node '/deepfs/postgres/ discovery export
could not be accessed to production_cache'. (61003) completes. These load files
obtain. will be reproduced on a
new run. If problem
persists across runs,
Contact Customer Support.

Appendix C. Event log messages 159


Event
Type Number Reason Sample Message Customer Action

WARN 61004 Warns the user that one of Transaction Cache Dump Run the discovery export
the transaction cache failed with error - policy that saw the error
dump processes Validation failed during again. If the error persists,
encountered an error. In creation of load file. and you cannot find any
the case of a discovery (61004) cluster/data server
export run, this means that configuration issues,
the discovery export will contact Customer Support.
fail to produce one of the
load files.

Note: If multiple dumps


fail, there will be one
warning per failed dump.

Related reference:
Appendix C, Event log messages, on page 133

160 Administration Guide


Notices
This information was developed for products and services offered in the U.S.A.

IBM may not offer the products, services, or features discussed in this document in
other countries. Consult your local IBM representative for information on the
products and services currently available in your area. Any reference to an IBM
product, program, or service is not intended to state or imply that only that IBM
product, program, or service may be used. Any functionally equivalent product,
program, or service that does not infringe any IBM intellectual property right may
be used instead. However, it is the user's responsibility to evaluate and verify the
operation of any non-IBM product, program, or service.

IBM may have patents or pending patent applications covering subject matter
described in this document. The furnishing of this document does not grant you
any license to these patents. You can send license inquiries, in writing, to:

IBM Director of Licensing


IBM Corporation
North Castle Drive
Armonk, NY 10504-1785
U.S.A.

For license inquiries regarding double-byte character set (DBCS) information,


contact the IBM Intellectual Property Department in your country or send
inquiries, in writing, to:

Intellectual Property Licensing


Legal and Intellectual Property Law
IBM Japan Ltd.
19-21, Nihonbashi-Hakozakicho, Chuo-ku
Tokyo 103-8510, Japan

The following paragraph does not apply to the United Kingdom or any other
country where such provisions are inconsistent with local law:
INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS
PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER
EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS
FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or
implied warranties in certain transactions, therefore, this statement may not apply
to you.

This information could include technical inaccuracies or typographical errors.


Changes are periodically made to the information herein; these changes will be
incorporated in new editions of the publication. IBM may make improvements
and/or changes in the product(s) and/or the program(s) described in this
publication at any time without notice.

Any references in this information to non-IBM Web sites are provided for
convenience only and do not in any manner serve as an endorsement of those Web
sites. The materials at those Web sites are not part of the materials for this IBM
product and use of those Web sites is at your own risk.

Copyright IBM Corp. 2001, 2013 161


IBM may use or distribute any of the information you supply in any way it
believes appropriate without incurring any obligation to you.

Licensees of this program who wish to have information about it for the purpose
of enabling: (i) the exchange of information between independently created
programs and other programs (including this one) and (ii) the mutual use of the
information which has been exchanged, should contact:

IBM Corporation J46A/G4


555 Bailey Avenue
San Jose, CA 95141-1003 U.S.A.

Such information may be available, subject to appropriate terms and conditions,


including in some cases, payment of a fee.

The licensed program described in this document and all licensed material
available for it are provided by IBM under terms of the IBM Customer Agreement,
IBM International Program License Agreement or any equivalent agreement
between us.

Any performance data contained herein was determined in a controlled


environment. Therefore, the results obtained in other operating environments may
vary significantly. Some measurements may have been made on development-level
systems and there is no guarantee that these measurements will be the same on
generally available systems. Furthermore, some measurements may have been
estimated through extrapolation. Actual results may vary. Users of this document
should verify the applicable data for their specific environment.

Information concerning non-IBM products was obtained from the suppliers of


those products, their published announcements or other publicly available sources.
IBM has not tested those products and cannot confirm the accuracy of
performance, compatibility or any other claims related to non-IBM products.
Questions on the capabilities of non-IBM products should be addressed to the
suppliers of those products.

This information contains examples of data and reports used in daily business
operations. To illustrate them as completely as possible, the examples include the
names of individuals, companies, brands, and products. All of these names are
fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.

COPYRIGHT LICENSE:

This information contains sample application programs in source language, which


illustrate programming techniques on various operating platforms. You may copy,
modify, and distribute these sample programs in any form without payment to
IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating
platform for which the sample programs are written. These examples have not
been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or
imply reliability, serviceability, or function of these programs. The sample
programs are provided "AS IS", without warranty of any kind. IBM shall not be
liable for any damages arising out of your use of the sample programs.

Each copy or any portion of these sample programs or any derivative work, must
include a copyright notice as follows:

162 Administration Guide


(your company name) (year). Portions of this code are derived from IBM Corp.
Sample Programs. Copyright IBM Corp. 2011.

If you are viewing this information softcopy, the photographs and color
illustrations may not appear.

Trademark

IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of
International Business Machines Corp., registered in many jurisdictions worldwide.
Other product and service names might be trademarks of IBM or other companies.
A current list of IBM trademarks is available on the Web at "Copyright and
trademark information" at http://www.ibm.com/legal/copytrade.shtml.

Adobe, the Adobe logo, PostScript, and the PostScript logo are either registered
trademarks or trademarks of Adobe Systems Incorporated in the United States,
and/or other countries.

Linux is a registered trademark of Linus Torvalds in the United States, other


countries, or both.

Microsoft and Windows are trademarks of Microsoft Corporation in the United


States, other countries, or both.

Java and all Java-based trademarks and logos are trademarks or registered
trademarks of Oracle and/or its affiliates.

UNIX is a registered trademark of The Open Group in the United States and other
countries.

Other company, product, and service names may be trademarks or service marks
of others.

Privacy policy considerations

IBM Software products, including software as a service solutions, (Software


Offerings) may use cookies or other technologies to collect product usage
information, to help improve the end user experience, to tailor interactions with
the end user or for other purposes. In many cases no personally identifiable
information is collected by the Software Offerings. Some of our Software Offerings
can help enable you to collect personally identifiable information. If this Software
Offering uses cookies to collect personally identifiable information, specific
information about this offerings use of cookies is set forth below.

This Software Offering does not use cookies or other technologies to collect
personally identifiable information.

If the configurations deployed for this Software Offering provide you as customer
the ability to collect personally identifiable information from end users via cookies
and other technologies, you should seek your own legal advice about any laws
applicable to such data collection, including any requirements for notice and
consent.

For more information about the use of various technologies, including cookies, for
these purposes, See IBMs Privacy Policy at http://www.ibm.com/privacy and
IBMs Online Privacy Statement at http://www.ibm.com/privacy/details the

Notices 163
section entitled Cookies, Web Beacons and Other Technologies and the IBM
Software Products and Software-as-a-Service Privacy Statement at
http://www.ibm.com/software/info/product-privacy.

164 Administration Guide


Index
A emails
hashing 30
Hitachi HCAP 45

action 2 Encrypted File System


admin account
changing 23
configuring 83
encrypted file-system user
I
audit settings IBM Desktop Data Collector
adding 32
configuring 30 installation 81
Enterprise Vault
audit trail installation in stealth mode 83
configuring 37
saving results from 99 installation methods 82
configuring primary volume 55
audit trails using delete policy 83
configuring site 39
searching 98 IBM Desktop Data Collector installer 31
retention volumes 42
audits downloading 83
event
harvest 89 IBM Information Archive
subscribing to 94
policy 94 configuring retention server 43
event log
IBM StoredIQ Data Workbench
clearing current 94
supporting 4
downloading 94
C viewing 93
IBM StoredIQ Platform
checking status 13
Centera event logs 93
rebooting 15
advanced retention 40 exceptions 2
IBM StoredIQ Policy Manager
Centera pool 41 Exchange
about 5
creating 41 servers 34
import audits 92
clients incremental harvests 69
upgrading for registered items
workstations 32 F saving into different folders 86
failure messages
policy audit 100
D FileNet J
DA Gateway settings bootstrapping into a domain 45
job
configuring 18 folder
creating 74
data object types 30 copying items to different 86
deleting 77
Data Workbench creating 85
discovery retention volumes 75
about 3 deleting 85
editing 75
potential uses of 3 filtering items within view 87
predefined 77
supporting 1 Library 85
saving as 76
DCOM traffic moving 86
starting 76
allowing through Windows renaming 86
jobs
firewall 43 Workspace 85
available types 73
Dell DX Storage Cluster full-text index settings 28
creating 41 configuring 28
desktop service
disabling 31 L
desktop services H lightweight harvest
full-text settings 72
enabling 31 harvest
desktop settings, configuring 31 harvester settings 71
full 69
desktop volume, deleting 65 hash settings 72
incremental 69
Discovery Accelerator 38 parameters 70
lightweight 70
configuring customer settings 39 performing 70
post-processing 69
required permissions 38 volume configuration 71
troubleshooting 79
discovery export volume logging in
harvest audits 89
creating 60 to Web application 23
viewing 91
Documentum logging out
harvest list
servers 37 to Web application 23
downloading 92
harvester settings 26
harvesting
E about 69 M
eDiscovery harvests mail settings
about 4 incremental 69 configuring 20
potential uses of 4 hash computation messages
prerequisites for using 5 SharePoint 127 event log 133
hash settings 30

Copyright IBM Corp. 2001, 2013 165


N SNMP settings
configuring 20
NetApp SnapLock 44 success messages
NewsGator policy audit 99
required privileges 40 supported file types
NFS by category 111
servers 34 by name 101
notifications supported system protocols 129
configuring 21 system time and date
NSF files setting 22
importing from Lotus Notes 25 system volume
adding 61

O
OCR image processing 26 U
OCR processing 27 user
Optical Character Recognition 27 creating new 23
editing 24
user account
P deleting 24
policy audit locking 24
viewing by discovery export 97 unlocking 24
viewing by name 97
viewing by time 97
viewing by volume 97 V
viewing details 96 volume
policy audit failure messages 100 adding primary 46
policy audit success messages 99 discovery export 60
policy audit warning messages 99 system 61
policy audits 94 volume cache
predefined job deleting 79
running 77 volume data
processing exporting 62
monitoring 77 exporting to a system volume 62
importing 62
importing to a system volume 63
R volume definitions
recovery agent 32 editing 55
remote DCOM volume import audit
enabling 42 viewing 93
retention volume volume indexing 33
adding 56 volume-import audits 92
Enterprise Vault 59 volume, deleting 65
retention volumes 55 volumes
Enterprise Vault 42 policy limitations by type 65
retention 55

S
search depth W
volume indexing 33 warning messages
server platforms policy audit 99
supported by volume type 129 Windows authentication
services enabling integration on Exchange
restarting 15 servers 35
SharePoint Windows Share
alternate-access mappings 36 server 34
Privileges 36
privileges for social data 36
Secure Connection 35
servers 35
SharePoint objects
supported types 126
SharePoint volume
performance considerations with
versioning 54

166 Administration Guide




Product Number: 5725M84, 5725M85, 5725M86

SC27-5692-00

Potrebbero piacerti anche