Contact
To learn more about Intella™, please contact us using the contact information below, or contact an Intella Channel Partner.

image

Office Phone
+1 888-291-7201

Postal Address
10643 N Frank Lloyd Wright Blvd, Suite 101
Scottsdale, AZ 85259
U.S.A.

We will be pleased to provide additional information concerning Intella and schedule a demonstration at your convenience.

To become an Intella reseller, please contact us!

For user and technical support please visit our website: https://www.vound-software.com

Vound Colorado (“Vound”).
© 2023 Vound. All rights reserved.

The information in this User Manual is subject to change without notice. Every effort has been made to ensure that the information in this manual is accurate. Vound is not responsible for printing or clerical errors.

VOUND PROVIDES THIS DOCUMENT “AS IS” WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED AND SHALL NOT BE LIABLE FOR TECHNICAL OR EDITORIAL ERRORS OR OMISSIONS CONTAINED HEREIN; NOR FOR INCIDENTAL OR CONSEQUENTIAL DAMAGES RESULTING FROM THE FURNISHING, PERFORMANCE, OR USE OF THIS MATERIAL.

Other company and product names mentioned herein are trademarks of their respective companies. It is the responsibility of the user to comply with all applicable copyright laws.

Mention of third-party products is for informational purposes only and constitutes neither an endorsement nor a recommendation. Vound assumes no responsibility regarding the performance or use of these products. Under the copyright laws, this manual may not be copied, in whole or in part, without the written consent of Vound.

Your rights to the software are governed by the accompanying software license agreement. The Vound logo is a trademark of Vound. Use of the Vound logo for commercial purposes without the prior written consent of Vound may constitute trademark infringement and unfair competition in violation of federal and state laws.

All rights reserved by Vound. Intella is a trademark of Vound.

1. Preface

Intella is designed to be an investigation and eDiscovery tool for electronically stored information (ESI). It is ideally suited for use by enterprise, law enforcement, and regulatory agencies in civil, criminal, or policy-related investigations. Intella is an excellent tool to prepare ESI for discovery.

Intella’s powerful indexing search engine and its unique visual presentation will let you quickly and easily search and review emails, files, and other information types, to find critical evidence and visualize relevant relationships.

With Intella, you can…​

  • Gain deeper insight through visualizations and statistics.

  • Search a wide range of file types.

  • Drill deeply using Intella’s unique facets.

  • Group and trace conversations.

  • Reveal the social graph of a person or group of persons of interest.

  • Plot items on a geographical map.

  • Preview, cull, and deduplicate email and data.

  • Export results in a variety of formats for reporting, follow-up investigation, eDiscovery, or later use.

image

1.1. Training

This manual outlines the features incorporated in the Intella products. Its focus is to explain the rudimentary functions of each Intella feature. It should not be seen as explaining how to manage data or cases.

While Intella is an easy and intuitive software package to use in the fields of forensic search, data analysis and eDiscovery, the user is required to have a firm grasp of how Intella treats and manages certain information/data types as applicable to these fields.

As with any software, the user must understand the issues and actions required that may arise prior to and while using Intella, particularly in the following areas:

  • Different data types.

  • Emails and attachments (parent-child relationships).

  • Search parameters.

  • Date formats.

  • Inclusions and exclusions.

  • Chain of custody.

  • Legal and privacy issues.

  • How to cross-verify results to ensure the accuracy of those results before they are relied upon.

  • How to identify inconsistent results.

  • The necessity to pre-process or convert certain data types prior to processing.

The user should understand that his manual does not seek, nor can it be an exhaustive list of the usage of Intella. This manual is structured to explain the use of certain, but not all, features at a basic level. This manual does not consider specific user requirements when explaining those features. Furthermore, this manual does not outline the steps required to be undertaken prior to processing data to ensure the accuracy of all results. The user should always ensure that they are personally aware of any special circumstances or steps required with the data, prior to processing and searching that data.

This is critical to being able to get the most out of Intella and undertake your investigation free from mistakes. This manual cannot and does not offer this information. We do however offer training that will help the user to have a better understanding of these issues.

We highly recommend that the user takes advantage of this training on the correct use of Intella. This can be critical for any matter where the user will rely upon the results produced in Intella as part of an action or investigation. Failure to undertake adequate training may cause unreliable results.

Please contact us for additional information at training@vound-software.com or visit us at https://www.vound-software.com/training.

1.2. Document conventions

The following section introduces you to conventions used throughout the Intella documentation.

Menu Functions
For functions that can be reached through menus, the different menu levels are illustrated as follows:

Menu > Menu entry

Important Entries
Some text will be shown as follows:

Important information on Intella.

These entries discuss a key concept or technical information that should, or must, be followed or considered. Please pay special attention to these entries.

Warning Entries
Some text will be shown as follows:

Warning information when using Intella.

Notes
Some sections provide additional information that will assist your use of Intella. These are displayed as shown below:

Information on function or parameter.

Keyboard Shortcuts
Some Intella functions can be activated or accessed through keyboard shortcuts. They are shown as follows:

CTRL+E

Tips
Several shortcuts, alternative methods, or general working tips are included throughout the documentation. These may help your workflow or provide additional information on other uses of functions. Tips are shown as below:

Information on Intella.

Folder and file names
Folder and file names are shown as below:

C:\Program Files\Vound\Intella\

2. An introduction to Intella

Intella is an instrument for data and email investigation and eDiscovery. It helps you search and explore information stored on computers, network disks, email archives and cloud sources. Intella is being used by law enforcement, legal and regulatory bodies to do all the above.

Intella indexes all places where you expect valuable information and provides powerful means for retrieving that information. Facets allow you to find items based on more than just keywords. Visualizations allow you to see how files and emails are related to your query.

Setting up Intella on your computer takes little time. Install the software, define the sources to search and explore and let Intella index the sources.

Searching with Intella is also easy. Start as if you are using a familiar search engine by entering a search term, or choose any value from the information facets. Let Intella help you to refine your question with the suggested refinements shown in its facets.

2.1. Key benefits

  • Easy to use interface means cutting down on training expenses and time, allowing a broad group of investigators to join in an investigation.

  • Visualizations of search results provide you with deeper insight. See how files, emails and cellphone items relate to parts of your query.

  • Facets, like Type, Date, and Language, help you to drill down in the evidence and focus on the information you need.

  • Search email attachments and archives such as zip files.

  • Searching is simple and requires very little training.

  • Export the search results for later use and for creation of reports.

2.2. Intella editions

Intella comes in five different product editions. The table below shows the most important features of these editions.

10 GB 100 GB 250 GB Professional Viewer

Preparation

Evidence size limit

10 GB

100 GB

250 GB

none

none

Create new cases

Index evidence files

Compound cases

Crawler scripts

Investigation

Search, filter & review

Preview items

Flag & tag items

Export items

Cooperation

Connect to Shared Cases

Export Cases

Import Cases

The evidence size limit relates to the cumulative file size of the evidence files as reported by Windows Explorer.

2.3. Supported file formats

Content and metadata of the following file formats can be extracted:

  • Mail formats:

    • Microsoft Outlook PST/OST. Versions: 97, 98, 2000, 2002, 2003, 2007, 2010, 2013, 2016, 2019, 365.

    • Microsoft Outlook Express DBX, MBX. Versions: 4, 5 and 6.

    • Microsoft Outlook for Mac OLM and OLK15* files.

    • Microsoft Exchange EDB files. Versions: 2003, 2007, 2010, 2013, 2016.

    • HCL/IBM Notes NSF (formerly known as Lotus Notes or IBM Lotus Notes). Notes 8.5.x or higher needs to be installed on the computer performing the indexing to process the NSF files. All NSF files are supported that can be processed by the installed HCL/IBM Notes version. Notes 9.0.1FP8 or higher needs to be installed to decrypt messages in a non-encrypted NSF.

    • Mbox (e.g. Thunderbird, Foxmail, Apple Mail)

    • Windows 10 Mail (POP accounts).

    • Saved emails (.eml, .msg)

    • Apple Mail (.emlx). Versions: 2 (Yosemite), 3 (El Capitan), 4 (Sierra), 5 (High Sierra) and 6 (Mojave). Testing concentrated mostly on versions 2, 5 and 6.

    • TNEF-encoded files (“winmail.dat” files).

    • Bloomberg XML dumps

  • Cellphone extraction formats:

    • Cellebrite UFED XML export or UFDR file. Tested up to version 7.44.

    • UFDR file made with MOBILedit 7.

    • Micro Systemation XRY XML and Extended XML exports. Tested up to version 7.4.1.
      (Extended XML is strongly recommended)

    • Oxygen Forensic Suite XML export. Tested up to version 13.

    • iTunes backups. iOS versions 8, 9 and 10 backed up with iTunes 12. Other versions may work but have not been tested.

  • Disk image formats:

    • EnCase images (E01, Ex01, L01, Lx01* and S01 files)

    • FTK images (AD1 files), version 3 and 4

    • DMG. Supported compression formats: ADC, LZFSE, ZLIB, BZIP2. Supported compressed image formats: UDCO, UDZO, UDBZ, UDCo. Supported uncompressed image formats: RdWr, Rdxx, UDRO.

    • DD images

    • AFF4 images:

      • AFF4 physical images containing images of the ContiguousImage or Diskimage AFF4 lexicon type.

      • AFF4 logical images:

        • Pre-release 1.0 version.

        • Version 1.1. Supporting all AFF4 lexicon types described in "AFF4-L: A Scalable Open Logical Evidence Container".

        • Version 1.2. Supporting password-based encryption.

    • MacQuisition images (RAW, .00001 files)

    • ISO images (ISO 9660 and UDF formats)

    • VMware images (VMDK files). Supported types are RAW (flat), COWD version 1 (sparse) and VMDK version 1, 2 and 3 (sparse). Not supported are images that use a physical storage device.

    • VHD disk images. Supported type is VHD version 1.

    • BitLocker-encrypted volumes.

    • Volume shadow copies.

    • X-Ways image files (.CTR files)

(*) Lx01 images created with recent EnCase versions may not be supported.
  • Document formats:

    • MS Office: Word, Excel, PowerPoint, Visio, Publisher, OneNote,
      both old (e.g., .doc) and new (.docx) formats, up to MS Office 2019 and Microsoft 365.
      MS OneNote 2007 is not supported.

    • OpenOffice: both OpenDocument and legacy OpenOffice/StarOffice formats

    • Hangul word processor (.hwp files)

    • Corel Office: WordPerfect, Quattro, Presentations

    • MS Works

    • Plain text

    • HTML

    • RTF

    • PDF (incl. entered form data)

    • XPS

  • Archives:

    • Zip. Supported compression methods: deflate, deflate64, bzip2, lzma and ppmd.

    • 7-Zip. Supported compression methods: lzma, lzma2, bzip2 and ppmd.

    • Gzip

    • Bzip2

    • ZipX

    • Tar

    • Rar

    • RPM Package Manager (RPM)

    • Cpio

    • ARJ

    • Cabinet (CAB)

    • DEB

    • XZ

  • Web-browser artifacts:

    • Google Chrome: history, keyword search, typed URLs, cookies, form history, bookmarks, logins, downloads

    • Mozilla Firefox: history, keyword search, typed URLs, cookies, form history, bookmarks, downloads

    • Microsoft Internet Explorer (6-11): history, keyword search, typed URLs, cookies (partial support)

    • Microsoft Edge: history, keyword search, typed URLs

    • Apple Safari: history, bookmarks

  • Search Warrant Results:

    • Hotmail (uses a HTML-based collection of files)

    • Gmail and Yahoo (uses an Mbox variant)

  • Instant Messaging

    • Skype SQLite databases, versions 7.x (stable), 8.x, 11.x, 12.x and 14.x.

    • Slack exports. Both channel exports and user exports are supported.

    • HCL/IBM Notes Sametime chats

    • Pidgin account stores

    • Note that cellphone extraction reports typically also contain instant messaging fragments that may be picked up during indexing.

    • RSMF 1.0.0

    • RSMF 2.0.0 (partial)

  • Databases

    • SQLite databases, version 3.
      Note that Skype SQLite databases get processed differently.

    • Mac OS property lists (.plist and .bplist files), in ASCII, XML or binary form.

  • Cryptocurrency (detection only):

    • Bitcoin wallets and blockchains

    • Dogecoin wallets and blockchains

    • Litecoin wallets and blockchains

    • Multibit Classic wallets and blockchains

    • Multibit HD wallets and blockchains

  • Miscellaneous formats:

    • iCal

    • vCard

    • XML

    • URL files (Internet shortcuts)

    • HCL/IBM Notes deletion stubs

    • Microsoft Teams conversations as part of PST export (limited indexing of attachments)

The following registry and system artifacts are extracted:

  • System:

    • Installed operating systems. Windows 7, 8, 10 and 11 have been tested.

    • User accounts.

    • User sessions: logon and logoff dates.

    • Windows event log entries. Supported Windows versions: 7, 8.1, 10 and 11.

    • Windows 10 Timeline entries.

  • Programs:

    • Installed programs.

    • Startup programs.

    • Launched programs extracted from User Assist, BAM (Background Activity Moderator), RecentApps registry keys, and Prefetch files.

  • Devices:

    • USB devices.

    • USB device activity extracted from Windows Event Log (connect and disconnect events).

    • Network interfaces.

    • Network profiles including Wi-Fi network names.

  • Files and folders:

    • Recently accessed folders (Shell Bags).

    • Recently accessed files (LNK, Jump Lists and RecentApps registry key).

    • Files and folders deleted to the Recycle Bin.

The following types of encrypted files and items can be decrypted, if the required access keys (passwords, certificates, ID files) are provided in the Key Store:

  • PST/OST

  • NSF (*)

  • PDF

  • DOC

  • XLS

  • PPT

  • OpenXML (.docx, .xlsx, .pptx)

  • PDF

  • ZIP

  • RAR

  • 7-Zip

  • S-MIME-encrypted emails

  • PGP-encrypted emails

  • BitLocker volumes (**)

  • APFS file systems

(*) Encrypted fields of NSF items are only decrypted if the NSF as a whole is encrypted too.

(**) Only BitLocker images using a password, recovery key or recovery file are supported. Other methods, such as smart cards or TPM, are not supported.

Intella will detect Outlook IRM-protected emails (restricted-permission message, rpmsg) in PST, EML and MSG files and mark them as encrypted. Decryption of such emails is not supported yet.

Supported image formats:

Format Type Identification Metadata Extraction Preview & Export to PDF OCR

Adobe Photoshop (PSD)

Apple Icon (ICNS)

Apple PICT

BMP

DjVu

Gif

HEIF/HEIC

Icon (ICO)

Interleaved Bitmap (IFF)

JBIG2

JPEG

JPEG-2000 (JP2)

PCX/DCX (DCX not tested)

PNG

Radiance HDR

SVG

TIFF

WebP

WMF / EMF (partial)

Some HEIC/HEIF types are not supported for preview, export to PDF and OCR.

When indexing plain text file formats, essentially all character encodings supported by the Java 8 platform. This relates to regular text files and to email bodies encoded in plain text format. See https://docs.oracle.com/en/java/javase/11/intl/supported-encodings.html for a complete listing.

When the encoding is not specified, the application will try to heuristically determine the encoding. The following encodings are then supported:

  • UTF-7

  • UTF-8

  • UTF-16BE

  • UTF-16LE

  • UTF-32BE

  • UTF-32LE

  • Shift_JIS Japanese

  • ISO-2022-JP Japanese

  • ISO-2022-CN Simplified Chinese

  • ISO-2022-KR Korean

  • GB18030 Chinese

  • Big5 Traditional Chinese

  • EUC-JP Japanese

  • EUC-KR Korean

  • ISO-8859-1 Danish, Dutch, English, French, German, Italian, Norwegian, Portuguese, Swedish

  • ISO-8859-2 Czech, Hungarian, Polish, Romanian

  • ISO-8859-5 Russian

  • ISO-8859-6 Arabic

  • ISO-8859-7 Greek

  • ISO-8859-8 Hebrew

  • ISO-8859-9 Turkish

  • windows-1250 Czech, Hungarian, Polish, Romanian

  • windows-1251 Russian

  • windows-1252 Danish, Dutch, English, French, German, Italian, Norwegian, Portuguese, Swedish

  • windows-1253 Greek

  • windows-1254 Turkish

  • windows-1255 Hebrew

  • windows-1256 Arabic

  • KOI8-R Russian

  • IBM420 Arabic

  • IBM424 Hebrew

Several file formats are processed by applying heuristic string extraction algorithms, rather than proper parsing and interpretation of the binary contents of the file. This is due to a lack of proper libraries for interpreting these file formats. Experiments with these heuristic algorithms have shown that their output is still useful for indexing and full-text search. It typically will produce a lot of extra gibberish data, visible in the Previewer, and there is no guarantee that the extracted text is complete and correct. The affected formats are:

  • Corel Office: WordPerfect, Quattro, Presentations

  • Harvard Graphics Presentation

  • Microsoft Project

  • Microsoft Publisher

  • Microsoft Works

  • StarOffice

2.4. Supported sources

File or Folder
Files on local and network file systems can be indexed. Please check the list of supported file formats. The use of external and network drives is not supported, both for stability and performance reasons.

Load files
Load files stored in Concordance, Relativity, and CSV format can be indexed.

Hotmail Search Warrant Result
The mail packages delivered by Microsoft when responding to a search warrant can be indexed.

Disk images
Several disk image file formats can be indexed, including the EnCase, FTK (AD1) and DD formats. Their contents will be indexed as if they were mounted and indexed as a regular Folder source. Optionally, files and folders can be recovered from the Master File Table (MFT). Carving of unallocated space and slack space is not supported.

MS Exchange EDB Archive
Use this option to index an MS Exchange EDB files and restrict indexing to a specific set of mailboxes. Indexing an EDB file in its entirety can be done by using the File or Folder source type.

Vound W4 Case
Entire cases created by Vound W4 can be imported.

IMAP account
Email accounts on an IMAP email server can be indexed, including all emails and attachments.

Dropbox
Both personal Dropbox and Dropbox for Business accounts can be accessed. Folders and files stored in that account will be retreived and indexed.

Google
Google accounts can be accessed and indexed, including all emails, attachments and other items in the selected services in that account.

SharePoint
Both local and cloud SharePoint instance can be accessed, indexing one or more of the sites in that instance.

Microsoft 365
The complete contents of a Microsoft 365 account can be accessed and indexed, incl. the Outlook, OneDrive, and SharePoint services of that account.

iCloud
Apple iCloud accounts can be accessed and indexed, including all information synced to that account from an associated device or entered on icloud.com.

2.5. Supported languages

As Vound applications are entirely based on Unicode, they can index and provide keyword search for texts from any language. There is no specific support for the handling of diacritics. E.g., characters like é and ç will be indexed and displayed, but these characters will not match with 'e' and 'c' in full-text queries.

The Language facet supports detection of the following languages:

af

Afrikaans

he

Hebrew

nl

Dutch

th

Thai

ar

Arabic

hi

Hindi

no

Norwegian

tl

Tagalog

bg

Bulgarian

hr

Croatian

pa

Punjabi

tr

Turkish

bn

Bengali

hu

Hungarian

pl

Polish

uk

Ukrainian

cs

Czech

id

Indonesian

pt

Portuguese

ur

Urdu

da

Danish

it

Italian

ro

Romanian

vi

Vietnamese

de

German

ja

Japanese

ru

Russian

zh-cn

Simplified Chinese

el

Greek

kn

Kannada

sk

Slovak

zh-tw

Traditional Chinese

en

English

ko

Korean

sl

Slovene

es

Spanish

lt

Lithuanian

so

Somali

et

Estonian

lv

Latvian

sq

Albanian

fa

Persian

mk

Macedonian

sv

Swedish

fi

Finnish

ml

Malayalam

sw

Swahili

fr

French

mr

Marathi

ta

Tamil

gu

Gujarati

ne

Nepali

te

Telugu

2.6. Supported platforms

We support and test our products on Windows 8/8.1, 10 and 11. A 64-bit operating system is required. The “Home” or “Starter” editions are not recommended as they limit the maximum amount of memory and CPUs. Please use the “Pro”, “Enterprise” or “Ultimate” versions instead.

Intella is tested on the abovementioned operating systems. That said, we have customers who are running Intella on the Windows Server platform, versions 2008, 2012, 2016, 2019 and 2022. Note that there may be security settings that need to be configured on the server to allow Intella to run on it. This needs to be addressed by your IT team; we cannot provide advice on these settings.

For detailed instructions about installing and running Intella, please read section 4: Installation and configuration.

2.7. Strong cryptography

Intella bundles and uses the Java Runtime Environment (JRE). This JRE contains the JCE Unlimited Strength Jurisdiction Policy files for decrypting certain types of encryption. Furthermore, it has been configured to allow the use of unlimited strength cryptography, by enabling this option in the java.security file:

crypto.policy=unlimited

We recommend that you check your local regulations to ensure that the use of encryption is permitted.

2.8. Feedback

We take great care in providing our customers with a pleasant experience, and therefore greatly value your feedback. You can contact us through the form on http://support.vound-software.com/ or by mailing to one of the email addresses on the Contact page.

3. Getting support

3.1. Different ways to get support

Vound offers four support options designed to assist users that experience problems while working with Intella™:

  1. Standard technical support

  2. User support contract

  3. Vound User Support portal

  4. Certified Intella™ training courses

3.1.1. Standard technical support

Standard technical support is offered free of charge to all Vound customers that have a current support and maintenance contract.

Standard technical support can be requested at the Vound support page, http://support.vound-software.com.

Support is provided on business days, Monday through Friday. We attempt to give you a first answer within 2 business days.

All communication will be remote – e-mail, GoToMeeting, and other means – and not in person unless otherwise arranged.

Standard technical support will only be provided if your computer and operating system meet the minimum recommended specifications listed in the latest version of the Intella™ manual.

Who is eligible for technical support?
Our goal at Vound is to provide our customers high quality and timely technical support. To do this we limit technical support to the registered owners of Intella. Companies that allow a third party to use their Intella licenses must have that third-party channel all technical support through the original registered owner of the software.

To ensure that we support our customers, Vound regrets it cannot support users who are not the original registered owner of Intella.

What technical support is included?

  • Installation and set-up support limited to one computer in your environment.

  • Configuration technical support and user support on use for standard Intella™ options.

  • Support for errors in the software (bugs).

Please note that Vound will make reasonable efforts to correct identified software errors. However, this may not be achievable until a later date or version release. If this is the case, the user should make efforts and take responsibility to achieve the required outcomes via other methods. Where the errors relate to or are caused by corrupt data (within source files), Vound reserves the right to charge for the work needed to rectify the issue.

No support can be provided…

  • When your computer does not meet the minimum or essential system requirements.

  • When you made any kind of modifications to the installed software.

  • When you are not using the software for its intended purpose.

  • When 3rd party applications, like virus scanners, firewalls, and other forensic applications, interfere with Intella™.

  • Explaining the method needed to use each feature to achieve a set outcome.

At no time should Vound technical support be seen as legal or forensic advice. Our support is given with no knowledge of the specific case or matter Intella is being used on. Technical support is focused on the correct installation and usage of Intella features. We do not warrant that we are aware of all facts around the case that may be under investigation. As such, our replies should not be seen as advice or the only way to achieve the required outcome.

3.1.2. User support contract

A paid user support contract is offered to those customers that want additional user support. The user support contract provides assistance that falls outside the standard support package (see 3.1.1 Standard technical support).

What can be included in the user support contract?

  • Help with the case or setup configuration of Intella™.

  • Assistance in using the basic and advanced features of Intella™ such as searching, tagging, and exporting.

  • Help with the installation of Intella™, or help with the configuration and set-up of your computer that runs Intella™.

  • Detailed explanation of Intella™ case management and help with Intella™ case setup.

  • Help with the export of search results found with Intella™ for use with other applications.

  • Support for using Intella™ in combination with software from other vendors.

  • Support for issues that a newer Intella™ release has addressed.

How to buy to a user support contract?
User support contracts are based on your specific needs. If you want to know more, please contact your nearest Vound representative or your local Intella™ reseller.

3.1.3. Certified Intella training courses

Vound offers several paid training courses for its product. These courses are designed to expand your effectiveness and output when using Intella™. It is recommended that all users take a minimum basic training course to ensure they are correctly using the product.

Users who have taken a recent training course for their Intella™ product will be offered a discount on a paid user support contract.

For more information on types of training and available dates please visit http://www.vound-software.com/training.

3.2. Working with Vound support

It is highly recommended that customers and users take advantage of the Vound support page when seeking assistance. The support portal takes care of collecting all necessary information such as the Intella version, Windows version, source types used, etc. and will suggest relevant articles from the Intella knowledge base.

3.3. Upgrade contract

Vound customers that purchased an Intella license are entitled to install free upgrades of the software for a period of one-year. In other words: an Intella™ license comes with a one-year upgrade contract.

After this period purchasing an upgrade subscription will continue the upgrade contract. Please contact your nearest Vound representative for more information.

Please know that you will only have access to standard technical support if you have an upgrade contract.

4. Installation and configuration

4.1. Installation

4.1.1. Step 1: Check the hardware requirements

CPU, memory, and disk space requirements depend on how Intella is intended to be used:

Indexing

  • As a rule of thumb, the case folder requires between 150% and 200% of the size of the combined evidence data, depending on data complexity and amount of compression used on the evidence data. When caching of original evidence items is turned off, this reduces the amount of disk space.

  • For better indexing performance, we suggest storing the case data folder on a physically different disk than the one with the evidence data. Ideally, aim for the following setup of physical disks:

    • 1 disk for the operating system.

    • 1 disk for the evidence folder.

    • 1 disk for the case folder.

    • 1 disk for the optimization folder.

  • Disk access times for the case indexes are critical for performance. We therefore strongly suggest not using USB or network drives for the case data folder.

  • See the section on Storage Recommendations for more storage-related tips.

  • When indexing MS Exchange EDB files, the memory sizes in the table below should be doubled and the memory settings will need to be adjusted (see the Memory Settings section).

Main memory and CPU requirements for indexing:

Evidence size Minimum memory Recommended memory Number of CPU cores

Up to 10 GB

4 GB

8 GB

2

10 to 100 GB

8 GB

16 GB

4

100 to 500 GB

16 GB

32 GB or more

4 or more

Using Viewer to connect to cases shared with Intella Connect or Intella Investigator

While technically Intella will work over slow network connections, a local and fast (gigabit) network is preferable, especially when working with large cases or with large reviewer teams.

Main memory and CPU requirements for connecting to a shared case:

Evidence size Minimum memory Recommended memory Number of CPU cores

Up to 10 GB

2 GB

4 GB

2

10 to 100 GB

4 GB

8 GB

2

100 to 500 GB

8 GB

16 GB or more

4

4.1.2. Step 2: Check the software requirements

Intella is supported on Windows 8/8.1, 10 and 11. Starting with the Intella 2.4 release, it is only available for the 64-bit editions.

Although our products can be installed on a number of Windows Server products such as Server 2012, 2016, 2019 and 2022, our products do not require a server operating system, and they run perfectly well on the listed desktop operating systems. For server installations, we only support our applications. We do not provide support for the server itself. Server security settings may need to be configured, and ports may need to be opened, for our products to operate on a server platform. These settings need to be addressed by your IT team to ensure that security of the system is maintained.

We do not support our products when installed on an operating system deemed end of life by its manufacturer. For example, these would include platforms such as Windows Vista and Windows Server 2008.

The following external applications may also be necessary to use some of Intella’s functionalities:

  • HCL/IBM Notes

  • Microsoft .NET

  • kCura Relativity SDK

HCL/IBM Notes
To index NSF files, HCL/IBM Notes 8.5 or higher is required. Only the application files are necessary, HCL/IBM Notes does not have to be fully set up and configured. In principle, all HCL/IBM Notes 8.5.x versions or later can be used, but the following versions will produce a warning:

  • 8.5.3 FP 3

  • 8.5.3 FP 4

  • 8.5.3 FP 5

  • 9.0

These versions contain a bug described here that cause emails with multiple “Received” headers to be altered: all Received headers will get the value of the first header. At the time of writing HCL/IBM Notes 9.0.1 was available, in which this bug has been fixed.

To index files made with HCL/IBM Notes 9.x, we recommend installing HCL/IBM Notes 9.x.

Notes 9.0.1FP8 or higher needs to be installed to decrypt messages in a non-encrypted NSF. Other versions will work, but encrypted messages will not be decrypted. In order to use an older version you need to select the Enable using unsupported version of HCL/IBM Notes checkbox.

Intella needs to know the location of HCL/IBM Notes to index NSF files. Please go to File > Preferences > HCL/IBM Notes to check if the location is validated.

Microsoft .NET and kCura Relativity SDK
To be able to export directly to a Relativity server, i.e. without having to handle Relativity load files, Microsoft .NET and the kCura Relativity SDK need to be installed.

The following versions of Relativity are supported:

  • Relativity 8.2

  • Relativity 9.7 - 10.3

Microsoft .NET can be obtained from the Microsoft website:

  • Microsoft .NET 4.5 for Relativity 8.2

  • Microsoft .NET 4.6.2 for Relativity 9.7 - 10.3

The Relativity SDK can be obtained from the Relativity website. This functionality was tested with version 8.2 of the SDK. After running the SDK installer, copy all 20 DLLs from this folder:

C:\Program Files\kCura Corporation\Relativity SDK\ImportAPI\Client\x64

to this folder for Relativity 8.2:

C:\Program Files\Vound\Intella X.Y.Z\bin\relativity\8.2\

In order to install Relativity SDK 10.3, do the following:

  • Go to the Intella program folder, e.g. C:\Program Files\Vound\Intella X.Y.Z

  • Go to the sub folder bin\relativity

  • Run the script download_sdk_10.bat

  • Script will download and extract the required files from the public repository.

  • Restart Intella.

4.1.3. Step 3: Learn about licenses and dongles

Notes on the trial license that is bundled with the software that you have downloaded:

14-Day evaluation period
The trial version runs under a HASP Software License, which gives you the ability to use Intella for 14 days. The
14-day evaluation period cannot be extended. The only way to continue using Intella is to purchase a dongle.

Trial restrictions
Besides the 14 days of usage, the trial only allows 10 GB of evidence files per case. Also, exporting is limited to maximally 1000 items per export.

Continue working with a USB dongle
If you would like to continue using Intella after this 14-day period, you will need to buy a license. After buying the license you will receive a USB dongle that will allow you to continue using the version you already installed. A dongle provides a perpetual license without export restrictions. Evidence size restrictions may still apply, based on the licensed product.

System clock
Changing the clock on your system will cause the trial to automatically expire. When this occurs, the only way to continue using Intella will be to purchase a license.

Virtual Machines, VMware
The evaluation version not work in VMware without a dongle.

RDP (Remote Desktop Protocol) connection
When using RDP, the dongle or trial license must be in/on the computer running the Intella software, not in the computer running the RDP viewer.

Other dongle-protected software must be closed
All other HASP protected software, like EnCase (Guidance), Smart Mount (ASR Data), HBGary and i2 products, must be closed when installing Intella.

4.1.4. Step 4: Install the software

  1. Download Intella through the Downloads page on the Vound website: https://www.vound-software.com/software-downloads. Note that a direct link to the download is also included on the USB flash drive that came with your purchase. Please use this drive to archive the Intella installer once you have downloaded it as you may need it in the future.

  2. Double-click on the downloaded .exe file to launch the installer. You must be logged on as an administrator to install Intella.

  3. Accept the software license when prompted.

  4. Enter the location to store the application files and shortcuts or accept the default settings. All files will be extracted to the location of your choosing and an Intella shortcut is (optionally) placed on your desktop and in your Start menu.

Running the installer with the /S switch runs it in silent mode. This way, it runs as a command-line application that requires no user interaction. Note that the uppercase 'S' must be used.

The application folder contains an executable called “Intella.exe” that can be used to launch the application. The desktop and menu shortcuts also start this executable. The program will start with the Case Manager window.

Intella will not install in an installation folder of an earlier version. Install a new version of Intella in a folder with a new name, for example: C:\Program Files\Vound\Intella 2.2.2\
It is possible to install multiple Intella versions side by side.

4.2. Storage considerations

Besides the memory and CPU requirements above, there are other hardware considerations that impact performance.

Use of USB drives
Our testing shows that USB drives are generally slower than internal hard drives or eSATA drives.

Please note that Windows allows you to use USB drives in two performance modes: the default “Quick Removal” mode and the “Better Performance” mode. Using the latter helps a lot to achieve better performance, but you will have to make sure to properly remove the drive in Windows before unplugging the drive. Not doing so means you risk damaging your case files beyond repair.

Evidence on external drives
Many users like to keep their evidence data on an external drive, for a variety of reasons. A common question is whether they can still use the case when this drive is disconnected after indexing. This is certainly possible. Access to the original evidence files is only necessary when you want to export the original evidence files themselves and have disabled the “Cache original evidence files” option when you added the source. For the rest, the case folder is completely self-contained as all extracted items are stored in the case folder and can be exported without access to the original evidence files.

For example, when you index a folder with PST files, any email and other embedded items extracted from those PST files are stored in the case folder and can always be exported. The PST files themselves are not copied into the case folder, unless the “Cache original evidence files” setting is selected.

Selection and configuration of hard drives
Because Intella is an intensive user of a system’s hard drive, we recommend careful selection and configuration of the hard drives to optimize performance. Generally, newer hard drives will outperform older drives in that they benefit from design improvements and new technology. Consider the following when using Intella:

  • Separate disks for evidence and case indexes. During indexing, Intella accesses the database continually performing read and write functions. To use the hardware resources more efficiently, it is recommended that the evidence data and the case data be allocated to separate hard drives. For example, put the case data on the “C” Drive and the evidence data on the “E” Drive. See the hardware requirements section for appropriate drive sizes given the case at hand.

  • Optimization folder. Since Intella 1.8 the case creator can specify a third folder for optimization purposes in the case details. Currently this folder is used for storing temporary indexing data that else would be stored in the case folder. When the optimization folder resides on a different drive than the case folder or evidence folder(s), this can further improve indexing performance.

  • Disk space. The amount of disk space needed to store your case depends heavily on the nature of your evidence data. As a rule of thumb, you should reserve twice the size of your evidence data for storing the case folder. The optimization folder has the same storage requirements as the case folder.

  • Proper connection. To realize maximum benefit from Intella’s multi-disk optimization architecture, ensure that the hard drives are appropriately connected to the computer’s motherboard to benefit from the higher available bandwidth. For example, connect the drives to the SATA-300 or SATA-600 connector rather than the smaller bandwidth carrying SATA-150.

  • Configure the system’s BIOS correctly. Typically, the computer’s BIOS defaults to the lowest common denominator to facilitate compatibility for connected hardware components. As a result, performance and speed can suffer. To address this possibility, check the BIOS to:

    • Ensure the hard drive supports Native Command Queuing – it should!

    • Confirm that the SATA control mode is set to either AHCI or RAID.
      Note: if the setting is at IDE (typically the default), Intella’s performance will suffer with slower indexing and searching as a result.

  • Use of external and/or network drives. Internal drives are always the preferred option for Intella. Intella’s indexing and search performance can deteriorate significantly when used with external or network drives.

  • If required, external drives such as a USB can be used to hold the evidence data. However, it is recommended that the fastest available connection option be used. USB 3.0 or eSATA should offer acceptable performance. Avoid USB 2.0 drives as they are significantly slower for any evidence or case file greater than 2-5 GB.

  • Network drives may be acceptable for holding evidence files if on a fast network. When using network drives, it is imperative that no other users access the files at the same time. You should also ensure that no network antivirus or filtering software blocks the indexing processes.

  • When processing a large case (> 100 GB of evidence files), it is advisable to format the NTFS disk with a cluster size that is larger than the default (usually 4 KB). This reduces the chance of defragmentation issues during indexing. Furthermore, it is recommended to turn off disk compression.

4.3. Installation troubleshooting

4.3.1. Error code 7 (H0007)

“HASP key not found (H0007)”

This error code might be caused by other HASP dongle protected programs. Please close all HASP related programs (i.e. EnCase, Smart Mount) and reinstall Intella.

4.3.2. Error code 31 (H0031)

“Could not find a valid Intella license, please insert a dongle”

This error message is shown when your trial license has expired, or when you unplug your dongle while Intella is running and it cannot fall back to a non-expired trial license. You can only continue using Intella by inserting a dongle.

4.3.3. Error code 33 (H0033)

“Unable to access HASP SRM Run-Time Environment (H0033)”

This error code may be triggered if you run antivirus software. It is probably due to the antivirus software incorrectly blocking access to the HASP install. Please update your antivirus software to the latest virus definition file.

If this problem persists, reboot your computer, open a Command Prompt, and run (as administrator)

<intella-dir>\bin\haspdinst.exe -i -kp

and restart Intella.

4.3.4. Error code 37 (H0037)

Other HASP dongle protected software may cause this error. Please close all HASP related programs (i.e. EnCase, Smart Mount) and reinstall Intella.

If this problem persists, open a Command Prompt, and run (as administrator)

<intella-dir>\bin\haspdinst.exe -i -kp

and restart Intella.

If problem persists after running this command, please open a Command Prompt as administrator and run

net start hasplms

4.3.5. Error code 41 (H0041)

“Your Intella (trial) license has expired (H0041)”

This error will be triggered if Intella is run and your trial license has expired. Once the trial has expired, you can only continue using Intella by inserting a dongle.

4.3.6. Error code 51 (H0051)

“Virtual machine detected, cannot run without a dongle (H0051)”

To protect our intellectual property, the evaluation version of Intella WILL NOT run in a virtual machine (VM) environment. A “stand-alone” machine is required. This is only true for the evaluation version; Intella will run in a VM environment using a dongle.

Solution 1: Reconnect the USB dongle to your computer
Solution 2: Install the Intella evaluation version outside a virtual machine

4.3.7. Memory, crawler count and timeout settings

The Intella process and its child processes (one for each case that you open + additional processes during indexing and exporting) are limited by the amount of RAM that the process can maximally use, despite how much memory is installed in the machine. On some data sets this limitation can cause issues when indexing or reviewing the data. These issues can be recognized by errors in the log files containing the text “OutOfMemoryError” or “java heap space”.

When such errors occur, a workaround may be to increase the automatically managed memory settings, especially when the machine meets the recommended hardware settings (at least 8 GB of RAM).

To increase these limits, select the case in the Case Manager and click Edit… button. Change the “Memory allocation” setting from Auto to Manual and increase the value. Note that you can never specify more than half of the available system RAM. This is to make sure that Intella’s child processes and the OS still have sufficient memory available to them.

When the memory issue relates to the processing of evidence files (you may need to contact tech support for that diagnosis) or to exporting, you can change the “Service memory allocation” setting from Auto to Manual and increase the value as suggested to you by tech support. Make sure that you do not assign more than your machine and OS support, taking into account that the other processes and Windows itself will also need sufficient memory. For the processing of MS Exchange EDB files, a minimum of 3 GB will be necessary for the service memory allocation setting.

By default, Intella will use up to 4 parallel crawlers when processing evidence files. In some cases, the limit can be increased by changing the Crawler Count setting from Auto to Manual. The number of crawlers should never exceed the number of CPU cores on your PC. Setting a too high number might result in non-optimal performance.

By default, Intella will cancel processing of an item if takes an extremely long time. It happens when a crawler does not produce any items in more than one hour. This number can be changed via the Crawler Timeout setting. It is also possible to disable the timeout mechanism completely by setting the value to 0.

4.3.8. Where are Intella’s data files located?

There is an Intella data folder in your home folder. Typically, it is in

C:\Users\<USERNAME>\AppData\Roaming\Vound\Intella

4.3.9. Where can I find Intella’s log files?

Intella has two types of log files:

  • Case-specific log files. These will contain any messages (errors, warnings, status messages) relating to your activities in the case, such as indexing, searching, and exporting. They are in

…\Intella\cases\<CASE FOLDER>\logs
  • Log files of operations performed in the Case Manager, such as exporting or importing a case. These are in

…\Intella\logs

The log files can be opened in any text editor like TextPad or Notepad++. Be aware that Windows’ default text editor Notepad may have issues opening large files.

Click Help > Open Log Folder to open the log folder of the current case.

4.3.10. Impact of running Intella under a different user account

This section describes situation where Intella was being run under certain user account in the past, but a different user now needs to work on cases and is trying to run Intella under different Windows user account than in the past:

  • Case integrity is not affected. However, due to the case list being stored in C:\Users\<USERNAME>\AppData\Roaming\Intella\cases.xml, the cases list will be empty. The user can either copy the cases.xml file from the original Windows user account or import each case by hand using Case Manager > Add…​ > Add an existing case.

  • The application preferences, which are not case specific, are stored in C:\Users\<USERNAME>\AppData\Roaming\Intella and will therefore be reset to their default values. This includes aspects such as the display language, backup folder, Geolocation tile server and databases such as the GeoLite2 database and hash filters.

4.4. Migrating Intella to new machine

This section describes the steps required to migrate Intella from one machine to another either for backup or hardware upgrade purposes.

Please leave Intella and all cases intact on the first machine until you are ensured that all cases on the second machine are working properly and that migration was successful.

In order to move cases to the second machine please follow these steps:

  1. Install Intella on the second machine.

  2. Close cases you want to move on the first machine - this step is required as changes made to case files during the copying of the case could result in damaged copy.

  3. Copy all cases and original evidence files to the second machine.

  4. (optional) If you want to move configuration settings, these folders also have to be moved (Important: please backup original files first):

    • Intella Home Folder: C:\Users\[User]\AppData\Roaming\Intella

  5. Open Intella Case Manager on the second machine and add each of the newly copied cases by pressing 'Add…​' → 'Add an existing case'. (Skip this step if configuration files were moved in above step. Note that file C:\Users\[User]\AppData\Roaming\Intella\cases.xml contains cases list, which gets copied in above step.)

  6. Open each of the cases with Intella Professional and check out that Evidence paths are set-up properly. You can do this by going to 'Sources' → 'Edit evidence paths'.

  7. The dongle with license can be moved physically to a new machine. Software-based licenses can be migrated using haspupdate.exe. You will find haspupdate.exe in the bin folder in the installation folder. Once haspupdate.exe has been started, go to tab Transfer License and follow the instructions on this tab. See section [dongles] for more information.

  8. Start Intella

.

4.5. Upgrading to the latest version of Intella

This section describes how to upgrade to the latest version of Intella, keep all of the settings and what to look out for.

Why upgrade to the latest version:

It is always best to install and use the latest version of Intella . As is with any software development, it is near impossible to test every scenario in which the software will be used, and what type of data is indexed with the tool. Although there is vigorous testing regime for all of our products, some customers find issues which they report back to support. These issues are generally resolved and added to the next release. Therefore, using the latest version will give you all of the updates from all previous versions.

Another good reason to upgrade is because the latest version has a number of new features that are not in previous versions. These features can make processing faster, can make analysis of the data easier, and adds more functionality to the tool.

Upgrading Intella:

There is no problem with installing the latest version of Intella on the same server. Note that it will need to be installed next to the current version. E.g. as long as the new version is installed in a different folder, the existing version should not interfere with the newer version. In addition, there is no need to uninstall the previous version.

When installing a new version of Intella, we make sure that any configurations from the previous version are also migrated over. We often keep old configuration as backup as well, so your previous configurations are not lost.

Installing the latest version of Intella is quite straightforward, but you should be aware of these aspects:

  1. Make sure that you are always using the same Windows Account when installing different versions of Intella

  2. The configuration and settings for your current version are stored in user-specific location, and those locations will not be available to other user accounts. E.g., we have seen cases when users were installing version 2.0 with the "John" user account, then later installed version 2.1 with the "Administrator" user account. They were surprised to see that they ended up with a clean instance of Intella, with all default configurations and settings.

  3. It is always best to run the latest version of all of our tools.

Before you start the upgrade:

You should consider the following before you start the Intella upgrade process:

  1. With every release of Intella we provide Release Notes. The very last section of the release notes is the 'Upgrade Notes' section. In that section we list information regarding backwards compatibility with earlier case versions. This section also points out any features which may be limited due to the version upgrade etc.

  2. We always suggest backing up your Intella systems before undertaking any upgrades. This minimises the risk of downtime, as you have an avenue to go back should you have any issues with the upgrade process.

  3. You should make a backup of these folders (which contain entire configurations) prior to proceeding with the upgrade. C:/Users/USER/AppData/Roaming/Intella

New major versions require a dongle update. See section [dongles] for more information.

After the upgrade is complete:

Once the upgrade process is complete, start Intella and check that it is reporting the correct version.

Migrating fonts:

When upgrading Intella, then the fonts copied to "Font" folder will need to be copied to the upgraded version.

5. Frequently asked questions

How is a file type determined?
Intella looks for certain binary markers (so-called magic numbers) that identify certain file types regardless of the file extension (e.g., .pst, .doc, etc.). When this detection process fails to produce a detected file type, Intella uses a list of known file types by file extensions. Intella may not be able to determine the file type of files with non-standard (unknown) file extensions.

Should I re-index a case when I want to add a new source?
No, to add a new source to case you do not need to re-index the whole case.

When you add the source, make sure that the option “Yes, I want to index this source now (recommended)” is selected on the last page of the “Add new source” wizard. Intella will index only the new source when you click “Finish”.

When you define the new source without the “Yes, I want to index this source now” option selected, you can use the “Index new data” button in the Sources view. This will scan selected sources for new evidence items, including sources that have not been indexed at all.

Can I re-index a single source in my case?
No, you can only re-index the entire case.

When the information in one of the sources has changed, and re-indexing the entire case is undesirable (e.g. because of the time needed), you can work around this by adding a new source and masking the old one.

For example, when you have a source named “Evidence 1”, which is one of several evidence folders in the case, and only the files in “Evidence 1” have changed, you can do the following:

  1. Rename the source folder “Evidence 1”, e.g. to “Evidence 1 (updated)”.

  2. Add it as a new source to the case and keep the “Yes, I want to index this source now” selected when you click “Finish”.

  3. Exclude the old source “Evidence 1” using the “Location” facet: select the node, click on the arrows in the “Search” button and click “Exclude”.

Even though the old data is still in the case, all search operations will filter out the results from the old “Evidence 1” source.

When the items in the old source have any annotations (tags, comments, etc.), these will not be copied to the items obtained from the new source. You will need to transfer them manually, e.g. using MD5 and message hash lists. When there is a substantial amount of such annotations, you may want to reconsider re-indexing the entire case, as this is a fully automatic operation.

Will I lose my tags and comments after re-indexing my case?
No, all your existing tags and comments will remain in the case after re-indexing.

Will I lose my OCR results after re-indexing my case?
No, all your existing OCR results will remain in the case after re-indexing. Be aware though that the OCR results will be linked to the re-indexed items by MD5. If the MD5s of the original evidence files have been changed, e.g. because damaged files have been replaced by their repaired versions, the OCR results may not be fully restored.

Why are some characters ignored in search queries?
This is caused by what is called the analyzer: before an item can be indexed, the analyzer breaks down the text to determine the individual words used in it. This analyzer discards white space, punctuation characters, etc. The same analyzer is also used to break down your query into individual terms.

As non-letters and non-digits are ignored, for example, the queries “searchterm,” “searchterm/” and “searchterm ” (with an extra space at the end) all end up being equivalent.

Why does the number of messages in the ($All) folder in my case not match the number of messages in the “All Documents” folder in HCL/IBM Notes?

Intella collects all items from all folders and lists them in the Location facet. The only exception is “($All)” folder. This is a special folder that usually contains all items from all folders – the other folders are essentially a selection of items from the ($All) folder. Intella won’t attribute a copy found in the ($All) folder when it is already present in another folder, to prevent duplication.

Can Intella perform live indexing?
Some cases may require you to index files while the computer is being used, or across a network. For such cases, we have made Intella to work with the best-of-breed application F-Response, by Matt Shannon. This combination provides you with a live forensic solution for under $300.

You can obtain F-Response at www.f-response.com.

Does Intella index attachments?
Intella will search both the email and the attachment for the keyword(s) and metadata.

Can Intella deduplicate results?
Yes, Intella can deduplicate search results. During indexing, the checksum (hash) of every item is stored. Intella can be set to show or hide duplicates while you use it. Intella uses the MD5 hash to calculate checksums of binary items. For emails and SMS messages a more specialized algorithm is used that can deduplicate across sources and source types.

Why do Chinese/Japanese/Korean queries give imprecise search results?
Documents written in Chinese, Japanese and Korean (often referred to as the “CJK languages”) differ from western languages in that the use of whitespace characters in CJK texts is optional. This makes it harder to create indexing software, as it typically uses whitespace, punctuation, and other character classes to determine the words in a text that need to be stored in the index.

“Proper” segmentation of CJK texts into words is still an open research issue and every method has its drawbacks.

A solution could for example be to index all characters from the CJK character sets as independent words. This would be trivial to implement, but has as a drawback that words that do consist of multiple characters will be much harder to find due to the large number of false positives that this method generates.

The solution used in Intella is to index the texts using bi-grams: every combination of two adjacent CJK characters in the text is seen and indexed as a word. In practice, this method gives reasonable performance:

  • It is simple to create and does not rely on e.g. expensive word dictionaries and perfect document language identification.

  • It is quick to process and produces a small text index.

  • The resulting index will find all occurrences of the entered terms, but with some number of false positives; this method favors recall over precision.

A way to find out how a certain piece of text is processed by Intella’s indexing engine is to create a short document with this text, index it, open the item in the Previewer and look at the Words tab. This tab shows a table with all terms extracted from the document and stored in the full-text index. Sort the table by the Field column and look for the words in all rows that have “text” as value in the Field column.

How can I print and export PDF reports with characters of my language?
By default, Intella supports printing and PDF generation for the basic Latin character set only. To enable printing and PDF export for a language that uses another character set, you need to install an additional Unicode font that supports that language.

  • Download the font file and install it in your system

  • Copy the font file to the font subfolder of your Intella installation: C:\Program Files\Vound\Intella 2.6\font

  • Restart Intella

The font must be a Unicode TrueType or OpenType font with “.ttf” or “.otf” file extension. The font folder must contain a single font file only. Using more than one font at the same time is not supported at the moment.

Recommendations for font selection:

  • For Chinese, Japanese, or Korean languages it is recommended to install a language-specific font. A large list of fonts for different languages and writing systems is available at http://www.wazu.jp/. If you already have the native font installed on your Windows system, you can copy it from “C:\Windows\fonts” to the Intella “font” folder.

  • For languages other than Chinese, Japanese or Korean, it is possible to install a single universal font supporting a broad range of character sets. You can try the GNU FreeFont font collection at http://www.gnu.org/software/freefont/.

Why is Korean text not rendering properly in properties tab of previewer?

In order to display Korean text properly, the GulimChe font needs to be installed in Windows. Follow these steps to install Korean supplemental font:

  • In Windows start menu, type "Apps & features", and then click Manage optional features.

  • If you don’t see Korean Supplemental Fonts in the list of installed features, click the plus sign (+) to add a feature.

  • Select Korean Supplemental Fonts in the list, and then click Install.

Note that this method may also apply to other languages as well.

Why are characters shown on my display tiny or overlapping?
This is usually related to Windows scaling and resolution settings. We see the issue mostly on newer systems that have 4K Ultra HD monitors capable of 3840×2160 resolution, but it can also occur on lower resolution monitors such as 1920x1080 (Full HD) or below.

There are two settings in Windows that affects the display of Intella:

  1. On 4K displays, we notice that the default value of the scaling control in Windows (which controls the size of text, apps, and other items) can be set to a value higher than 100%. We have seen this set as high as 250% on one system. This is to ensure that fonts are displayed at a normal visual size.

    Intella works best when the scaling is set to 100%, as it is not yet DPI-aware. In Windows 10, right-click on the Desktop and select “Display settings”. Under the “Display” tab you will see a slider that can be adjusted to 100%.

    It is not that Intella will not work when you increase the scaling level. For example, you may see a small difference if you set the scaling control to 125%, which may still be workable for many people. As you go higher, the characters get condensed more, making it hard to read.

  1. When the scaling setting is set to 100%, you may find that the font is so small that it becomes difficult to read. This can be changed by adjusting the resolution that is being displayed. The resolution for a 4K monitor could be set to a very high resolution such as 3840×2160. It is up to the user which resolution setting suits them, but we find that 1920x1080 is quite good as a rule of thumb. If the resolution is set higher, particularly on a smaller laptop screen, the font may be too small to use the computer comfortably.

    To change the resolution, click on the “Advanced display settings” option at the bottom of the screen with the scaling slider. Here you can change the resolution for the screen using the dropdown.

6. Dongles

Vound licenses are typically delivered in the form of a dongle. Dongles have several benefits over software-based license keys. For example, users can easily move software licenses from one machine to another by simply plugging the dongle into the other machine, there is no loss of license when the operating system is reinstalled or reverted from an image, changes to the hardware (new motherboard etc.) do not lock the license, hard drive failures do not result in the loss of licenses, etc.

To protect our intellectual property, dongles may not be activated when shipped by Vound or one of its resellers. In that case, it is necessary to activate your dongle to use the application, using one of the update mechanisms below.

By default, users are supplied with a single user dongle for every ordered copy of Intella. Optionally, a network dongle can be delivered instead. This type of dongle allows for consolidating the licenses of multiple users on a single dongle, which then is typically installed on a physically secured, always-on machine. See the section on network dongles below for how to configure your systems to use a network dongle.

6.1. Dongle Manager

Each Vound application ships with a Dongle Manager application. The Dongle Manager will list all connected Vound dongles and the products they currently contain. When the PC running the Dongle Manager is connected to the Internet, it can also contact the Vound license server to check for any updates for a dongle. These updates are then downloaded and applied automatically.

The Dongle Manager is in the root application folder:

C:\Program Files\Vound\Intella 2.5.1

A shortcut to the Dongle Manager can also be found in the Start menu. After starting the Dongle Manager, the following screen will appear:

image

This screenshot shows a typical setup where only one Vound dongle is connected. When multiple dongles are present, they will each be listed separately in this list.

Click on Blink to see to which physical dongle an entry in the list corresponds. This will cause the LED in the represented dongle to blink rapidly. This can be useful when you have multiple Vound dongles plugged in or are using HASP dongles from a different Vendor.

Show Products will list the licensed products on that dongle. All products typically have a perpetual license; hence no license restrictions are displayed by the Dongle Manager.

Show Products also shows a list of expiration dates. These reflect the end date after which you will not be able to receive technical product support and license updates. These end dates do not affect the ability to use the existing licenses on your dongle.

C2V will create a file that reflects the identity and content of your dongle. Please use this button when instucted by Vound Support to do so.

To activate or update your dongle, click on Check for Updates. This will contact the Vound license server and download and apply any updates. When the process has finished, the Dongle Manager will show which products, if any, have been added to the dongle. The update procedure will only add new licenses to the dongle; it will leave your existing licenses untouched.

An active Maintenance Agreement with at least 60 days remaining until the expiration date is necessary to qualify for maintenance updates.

When you are on a network using a proxy, the Dongle Manager will automatically try to detect and use it. If this fails, the proxy settings can still be set using the “Configure proxy settings…”. Consult your IT admin for further instructions.

Note that this update procedure can be done on any Internet-connected computer; it does not have to be done on the computer on which the licensed application will be run.

6.2. Dongle updates with haspupdate.exe

If the dongle cannot be activated or updated in this fashion, e.g. because external network connections are not allowed and the dongle can not be temporarily moved to a machine that can make such connections, please follow the steps below.

Step 1: Collect your dongle and license information and upload it to the Dongle Updates website.

image

  1. Plug your dongle into an available USB port.

  2. Start haspupdate.exe. You will find haspupdate.exe in the bin folder in the installation folder.

  3. Select the Collect Status Information tab.

  4. Select the Update of existing protection key option and click Collect Information.

  5. In the next dialog that may appear , you will be asked to select a Sentinel Protection Key. Be sure to select the one listed as HL in the Key Type column. image

  6. Next, the Save Key Status As dialog will appear, asking you to enter a file name and location. Please save the file with your company name. If you are activating more than one dongle, please number the files. The file(s) you create will have a c2v file extension.
    Example:
    ACME_Forensics_1.c2v
    ACME_Forensics_2.c2v

  7. Record the dongle ID numbers for each dongle. This will help when applying the update files.

  8. Open https://www.vound-software.com/dongle-update in a web browser.

  9. For each C2V file, use the Choose File button on this page to select the C2V file, and click the Submit button. When updates are available for a dongle, a list of one or more V2C files will be shown. Download all V2C files and keep track of which dongle/C2V file they corresponded to.

Step 2: Apply the license update file(s) you receive from the Dongle Updates website.

image

  1. Make sure your dongle is connected to the computer where you will apply the license update file(s).

  2. Start haspupdate.exe as before.

  3. Click the Apply License File tab.

  4. Click the Browse button labeled "…" next to the Update File field. This opens a file selector dialog.

  5. Select the .v2c file in the file selector and click Open.

  6. Click the Apply update button.

Your Vound dongle is now activated!

When the Dongle Updates website listed multiple V2C files for an uploaded C2V file, be sure to apply them in that order.

In case of questions or problems, please contact Vound Support at http://support.vound-software.com/.

6.3. Network dongles

A prerequisite for using network dongles is that the so-called "Sentinel LDK driver" is installed on both the client and the server.

This driver is in fact known under several different names due to historic reasons. When it is installed, it is typically visible as "Sentinel LDK License Manager" in the Windows Services application and under that same name or as hasplms.exe in the Windows Task Manager and Windows Resource Monitor.

On a standalone PC, the driver provides a bridge between the licensed application (e.g., Intella or Intella Connect) and the dongle holding the license. Furthermore, it handles software-based licenses such as the bundled trial license. In case of a network dongle, the drivers on the machines stretch that bridge across the network, making the products on the network dongle available to other PCs in the network.

Getting this driver installed is best achieved by simply running the installer on both machines, as it includes the installation of the Sentinel driver.

Once the Sentinel driver is up and running on both machines, the drivers will communicate with each other automatically, or after a bit of network configuration (see below). When the application starts on the client and requests a license from its local driver, the driver will communicate with the server’s driver and exchange information about the network-enabled licenses on the server’s dongle, making the licenses also available to the client. The server’s driver will register that one more user is using the application, or refuse the operation (and block the client machine from starting the application) when the allotted maximum number of concurrent users has been reached.

Network dongles often work out-of-the-box, but may in some cases require a small amount of network configuration. This depends mostly on the locality of the client running the application and the server holding the network dongle.

6.3.1. Client and server in single subnet

When the client and server are within the same subnet, no network setup is usually necessary. The drivers on both machines will usually find each other automatically and the client will be able to use the licenses on the network dongle.

For example, in the following setup:

Server IP address: 172.168.12.223
Client IP address: 172.168.12.26
Subnet Mask (Class-C): 255.255.255.0

the drivers will be able to communicate directly, if port 1947 is not blocked.

When the application is not able to use the network dongle’s licenses, please follow the steps below for setting up usage with different subnets. This may resolve the issue.

6.3.2. Client and server in different subnets

Given the following setup:

Server IP address: 172.168.12.223
Client IP address: 172.168.16.46
Subnet Mask (Class-C): 255.255.255.0

the drivers will require some configuration for the client and the server to be able to find each other.

Step 1: Make sure that port 1947 (used by the drivers) is not blocked by any firewall. The drivers use this port to communicate with each other and with the application. Both TCP and UDP communication need to be enabled.

Step 2: Ensure that the server and client machines can "ping" each other.

Step 3: Plug the network dongle into the server. Make sure that the key is detected when viewing the Admin Control Center on http://localhost:1947 on the server, like this:

image

Depending on the exact network dongle type, the value in the Key Type column can vary. E.g., HASP HL Net 10, HASP HL Net 50 or Sentinel HL Net 50.

Step 4: On the server, do the following:

  • On http://localhost:1947, click on Configuration.

  • Select the Access from Remote Clients tab.

  • In the Allow Access from Remote Clients section, make sure that the All licenses are accessible without need of identity option is selected.

  • Click Submit if a change was made.

image

Step 5a: When client and server are on the same subnet, then on the client:

  • On http://localhost:1947, click on Configuration.

  • Select the Access to Remote License Managers tab.

  • Make sure that the Allow Access to Remote Licenses checkbox is selected.

  • Make sure that the Broadcast Search for Remote Licenses checkbox is selected.

  • Click Submit if a change was made.

image

Step 5b: When client and server are on different subnets, then on the client:

  • On http://localhost:1947, click on Configuration.

  • Select the Access to Remote License Managers tab.

  • Make sure that the Allow Access to Remote Licenses checkbox is selected.

  • Enter the IP address of the server holding the network dongle in the Remote License Search Parameters box.

  • Click Submit if a change was made.

image

Step 6: On the client:

  • On http://localhost:1947, click on Sentinel Keys.

  • Verify that the network dongle plugged into the server is now listed here. This verifies that the client and server can communicate properly, and that the licenses on the dongle can be used. If any changes were made in the previous steps, this may in some cases take a few minutes to propagate.

You should now be able to start the application on the client, using a license from the network dongle. You can verify this by checking the Case Manager window; it should list the network dongle’s ID:

image

7. Products and workflow

7.1. Feature overview

The following table lists the five different Intella desktop applications and their features:

10 GB 100 GB 250 GB Professional Viewer

Preparation

Evidence size limit

10 GB

100 GB

250 GB

none

none

Create new cases

Index evidence files

Compound cases

Crawler scripts

Investigation

Search, filter & review

Preview items

Flag & tag items

Export items

Cooperation

Export Cases

Import Cases

Connect to Shared Cases

7.2. Standalone use

The following Intella products can be used for standalone use:

  • Intella 10

  • Intella 100

  • Intella 250 GB

  • Intella Professional

They allow a user to create cases, index evidence files, search, filter, flag, tag or otherwise annotate and export items.

The number in the 10/100/250 products indicates the number of gigabytes of evidence files that each case can hold in a case. Intella Professional has no such limit.

The cases created by these products can also be reviewed by the following products:

  • Intella Viewer: a desktop product with reduced functionality.

  • Intella Connect and Intella Investigator: separate, server-based products accessible via a web browser.

The workflow for standalone use is as follows:

  1. The investigator creates a case in the case manager of Intella and indexes evidence files.

  2. The investigator flags and tags items and gives comments to items of interest.

  3. The investigator exports the results for further processing of the case.

In principle, it is possible cooperate on a case by giving other investigators a copy of the case folder. While technically this will work (a case is in no way tied to a specific end user license or machine), the main challenge will be to coordinate that joint investigation. A case copy essentially starts a life on its own, meaning that tags and other annotations exist only within that copy. Ideally the tags are visible to all other investigators, perhaps filtered based on a permissions model.

7.3. Compound cases

A compound case is a virtual case that instantly combines two or more cases ("sub-cases") into a single unified case. The compound cases are available for Intella Professional product.

Compound cases offer several benefits:

  • Users can search, review and export items across all sub-cases, without having to repeat these actions manually for each sub-case.

  • When processing very large data sets, indexing can be spread across multiple machines.

    • Once each machine has indexed its data subset, the resulting cases can easily be combined through a compound case.

    • The data in the sub-cases is not duplicated (copied) in this process. This makes compound cases very fast to create and requiring less disk space, compared to when the sub-cases are fully merged with the "Export to Intella Case" functionality.

  • One may need to add additional evidence data to a case. Compound cases make this possible without having to take the active case down. The new evidence data is processed as a separate case and then joined with the current "live" case with minimal downtime.

7.3.1. Caveats and limitations

Only local cases can be included in a compound case; remote cases shared by Intella Connect or Intella Investigator are not available for inclusion in a compound case.

Only cases created with Intella 2.5 or newer can be included in compound cases. Older cases will first need to be converted to the current case format.

Evidence data sources cannot be added directly to the compound cases, but they can be added to the sub-cases separately. New data will be available when the compound case is opened next time.

It is not possible to use a compound case and any of its sub-cases at the same time. Opening a compound case will lock all its sub-cases. Similarly, opening a sub-case separately will lock all compound cases with this sub-case.

User annotations (e.g., tags, flags, comments, and custodians) that are added in a compound case are specific to that compound case and do not affect the sub-cases.

The following annotations that are added in the sub cases will be visible (read-only) in the compound case:

  • Tags

  • Custodians

  • Comments

  • OCR

The following annotations that are added in the sub cases will NOT be visible in the compound case:

  • Batches and Coding

  • Content Analysis

  • Custom ID

  • Duplicate Custodians and Locations

  • Email Threading

  • Excluded Fragments

  • Export sets

  • Generated thumbnails

  • Identities

  • Keyword Lists (including MD5 and Item ID lists)

  • Near-Duplicate Detection

  • Redactions

  • Saved searched

  • Tasks

One can perform these actions, but their results stay local to the compound case. The compound case also maintains its own Event Log, independent from its sub-cases.

Thumbnails generated in a sub-case will be visible in the compound case. But thumbnails generated in the compound case will not be visible in the sub-cases.

The following functionalities are not available in compound cases:

  • Any form of source management, including viewing, adding and removing sources. That also includes fatal source errors (the top-level Errors tab).

  • Indexing and re-indexing.

  • Load file import (both regular load files and overlays)

  • Exporting items to another Intella case.

  • Changing primary date rules.

  • Changing message hash options.

  • Changing show parents options.

  • Any form of key store management, including viewing and changing passwords, certificates, etc.

  • Importing item texts via command line options (-importText option).

  • W4 import tab.

  • Custom column management.

Text imported into a sub case via '-importText' command line option will be visible in the compound case.

Notes on item IDs in compound cases:

  • An ID of an item in a compound case can, and generally will, be different from the ID of the corresponding item in the sub-case.

  • Item IDs in the compound case will always remain the same (stable) no matter whether a new sub-case is added or removed, or any source is added, removed or re-indexed in a sub-case.

7.3.2. Important note on case deletion

When a sub-case is deleted, evidence items of this case will no longer be available in a compound case that refers to it. If the compound case had annotations or other review data associated with these items, this data will be unavailable.

Deletion of a compound case will not cause the deletion of any sub-cases. If desired, these need to be deleted individually and separately.

8. Managing cases

A case is a collection of evidence sources that can be searched by Intella as a single collection. You use cases to organize your investigations.

When you start Intella, the Intella Case Manager will first show up. Here you can define new cases, open, and edit existing cases, remove old ones, and share and export cases.

image

The icons represent local cases, compound cases, remote cases and old cases that are no longer supported (grayed-out folder icon).

A case can also be marked as “review only”. This happens when the Intella version used to create the case is still close to the current Intella version. Such cases can be opened but have restricted indexing options, such as adding new sources, re-indexing/refreshing, OCR-ing, etc.

Above the case list is a field for entering the Investigator name. This name will be used as the default user name when creating new cases and connecting to remote cases. Also, when opening a local case made by someone else, all user actions like previewing, tagging, and exporting will be associated with this user name. The initial value used here is your Windows user name.

Use “Sort by” options to change the order of cases in the list:

  • Last opened - by last opening time (the default)

  • Case name - alphabetically by name

  • Case folder - alphabetically by folder path

  • Created - by creation time

  • Creator - alphabetically by Investigator name of a user who created the case

  • Evidence size - by the total size of evidence files

  • Case version - by Intella version used to create the case

Below the cases list you can see the ID of your dongle. This can be relevant in conversations with Vound’s support department. When you are using a trial license, this line will reflect that.

When your dongle is inserted but you still see a line indicating that you are using a trial license message, this could indicate technical problems with accessing the dongle, but also that your dongle needs to be updated to run with this Intella version.

8.1. Adding cases

image

To create a new case, select “Add” in the Case Manager window. The Add Case options will appear. It shows the four ways of adding a new case to Intella:

  1. Create a new, local case from scratch. Use this to index a new set of evidence files on your machine.

  2. Create a compound case. With this option, you can combine multiple existing local cases to new virtual case.

  3. Open a shared case. With this option, you can connect to a case that is shared by Intella Connect or Intella Investigator. This option is only available when running Intella with a Viewer license.

  4. Add an existing case. Use this when you have a case folder already on your system but it is not yet in the list of cases shown by the Case Manager.

  5. Import a case. Use this when you have received a copy of a case from another investigator as an ICF file. Importing the ICF file will extract its contents into a local case folder and add the case to the Case Manager’s list.

8.1.1. Creating a new case

image

Choose “Create a new case” to create a new local case from scratch. When the Create New Case dialog is displayed, give the case a name, enter an optional description, enter the name of the investigator creating the case and select a location where you want to store the data that belongs to this case.

The default location for data storage, visible when you click the Suggest button, is
C:\Users\<username>\AppData\Roaming\Intella\cases
When you use a different parent folder, subsequent cases will default to a subfolder in that parent folder.

The selected case folder will be checked for being a hard disk formatted with the NTFS file system. A warning is displayed when this is not the case, e.g. when a FAT file system is used, which has file size limitations unusable for Intella, or when a USB flash drive is detected, which is not recommended for various reasons.

When processing a large case (> 100 GB of evidence files), it is advisable to format the NTFS disk with a cluster size that is larger than the default (usually 4 KB). This reduces the chance of defragmentation issues during indexing. Furthermore, it is recommended to turn off disk compression.

Clicking on the “Advanced” button adds more options that normally are only necessary if you want to apply a case template or when dealing with very large cases (hundreds or GBs or more):

  • In the “Case template” field you can specify an Intella Case Template (ICT) file that will be applied to the new case. See the “Case templates” section for details.

  • The optimization folder can be used to speed up indexing by distributing certain database files during indexing across the case folder drive and the optimization folder drive. See the “Storage considerations” section for more details.

  • The memory allocation and crawler count settings can normally stay at Auto. See the “Memory, crawler count and timeout settings” section for instructions on how to use this.

8.1.2. Creating a compound case

image

Choose “Create a compound case”. In the “Create new case” dialog, specify creator name, case name, optional description and case folder, as described for a local case above.

In “Include cases:” list, choose one or more local cases to be included in the compound case.

Only local cases can be included in a compound case. Remote shared cases and other compound cases are not available.
Only cases created with Intella 2.5 or newer can be included in compound cases. Older cases will first need to be converted to the current case format.

Click on the “Advanced” button to access memory allocation settings for new case. See the “Memory, crawler count and timeout settings” section for instructions on how to use this.

8.1.3. Opening a shared case

In the Add Case dialog, select “Open a shared case” to open a case on another machine that has been shared by a Intella Connect or Intella Investigator user.

A wizard for creating remote case will open. Depending on the setup, it will ask for a Case URL, investigator name, passphrase and more. This information should be provided to you by the case administrator (typically the Intella Connect or Intella Investigator administrative user).

The first “Case URL” step asks for URL under which the case is shared.

desktop add remote case case url

If the case is shared by Intella Connect or Intella Investigator without any Single-Sign-On (SSO) providers configured, then the “Login credentials” step will ask for investigator name and passphrase.

desktop add remote case login credentials without sso

Check the “Remember passphrase” checkbox if you want to store the password locally, so that you don’t have to re-enter it each time you select the case in the Case Manager and click Open.

On the other hand, if SSO providers were configured in the server product, then they will be shown on the top separated from investigator name and passphrase fields.

desktop add remote case login credentials

Your case administrator will let you know whether to use SSO provider button or investigator name and passphrase fields.

Clicking on SSO provider button will show login page of the SSO provider similar to how it is shown in the (Edge, Firefox or Chrome) browser.

desktop add remote case sso login

You will be asked to provide login credentials which are configured on the SSO provider side.

Alternatively, when logging in via local or LDAP account by providing investigator name and passphrase, then in case of local account you may be asked for "Verification code".

desktop add remote case two factor auth

The verification code is required when the account is configured with Two Factor Authentication. Use your device with two factor authentication enabled to obtain the verification code.

To summarize, there are three ways to login to a remote case by using:

  • local or LDAP account - only investigator name and passphrase are required

  • local account with 2FA - investigator name, passphrase and verification code are required

  • SSO account - login credentials which are configured on the SSO provider side

Once correct login credentials were provided and verified, then the last step of this wizard will be shown.

desktop add remote case case settings

On the top, the case name and description will be loaded from the server. These are only for information and to verify that the correct case is being logged into.

Click Suggest or enter a custom data folder for local storage. This folder will hold local log files, user preferences, etc.

The memory allocation settings can typically be left to their default value. See the Memory Settings section for more information on this.

Now the case will be added to the Case Manager list and will open instantly when you keep the “Open case immediately” checkbox selected.

8.1.4. Opening an existing case not in the list

In the Add Case dialog, select “Add an existing case”. This will open a dialog prompting you to choose a case file (case.xml). This file is in the top-level case folder. Choose the case file and click on Open. The case will now be added to the Case Manager’s list and can be opened by selecting the case and clicking the Open button.

8.1.5. Importing a case

In the Add Case dialog, select “Import a case”. This will open a dialog prompting you to choose an Intella case file (.icf file). Choose the case file and click on Import. Once importing has completed, the case will be added to your Case Manager’s case list.

8.2. Opening a case

Opening a case is merely a matter of selecting the case in the Case Manager’s list and clicking the Open button. The Case Manager window will disappear and Intella’s main screen will be opened. This may take some time, depending on factors like disk speed, case size and concurrent tasks performed by the PC.

In some scenarios, a case may be grayed out:

  • There may be a lock icon next to the case name, with the text “Case in use”. This indicates that there is already a process running that is accessing this case. Access to the case is blocked for other processes to prevent damage to the case databases. This locked status will disappear as soon as the other process has ended.

    If there is no other Intella window visible, it may be that an earlier Intella session did not exit correctly and completely, still holding the lock. A reboot of the machine is the simplest solution to fix this situation.

    If the case folder is on a shared network drive, then it may also mean that an Intella process on another machine is accessing the case. In that scenario rebooting the local machine will not help; the process on the other machine must be ended first before other processes can open the case again.

  • The case data folder that the cases.xml file points to cannot be found or does not contain a case.xml file. In that scenario, the case will be listed as “Unnamed case”.

  • All other grayed out and disabled cases are cases made with older Intella versions.

    • Cases made with the 2.5.x version and older are marked with the text “Old case format, conversion required”. These cases need to be converted to the latest case format before they can be opened. See the Release Notes for more details and limitations.

    • Cases made with Intella 1.8.x and older are not supported by Intella 2.1 and newer.

    • Cases made with Intella 2.0.x and older are not supported by Intella 2.6 and newer.

Before attempting to convert a case, it is recommended to open the case with the Intella version that was used to create it, and check that all evidence paths are correct using the Sources view.

8.3. Editing a case

In the Case Manager, use “Edit…” to open the “Edit case” dialog to change the case name or description, or to alter the memory settings.

For compound cases, it is possible to change the set of included cases (add new or remove existing sub-cases).

You cannot change the Data folder and Creator. For remote cases, the case name and case description cannot be changed.

8.4. Removing a case

image

In the Case Manager, use “Remove…” to remove the selected case(s) from the Case Manager’s cases list. You will be asked to confirm the removal.

By default, only the reference to the case is removed, the case folder is left intact. By checking “Also remove the related case folders from disk”, the case folder will be permanently removed as well.

If the case was a part of a compound case, it would also be removed from the compound case.

Removal of the case folder cannot be undone. Also, all files that you may have placed manually in the case folder will also be removed.

8.5. Converting a case

In the Case Manager, use “Convert” to start a case conversion process. Cases that have been created by an older Intella version can be converted to the latest case format to enable the use of features that have been added since. Case conversion is available for Intella 2.1.x cases and newer.

Case conversion always creates a copy of the existing cases and updates that copy. The original case will not change. Make sure that you have sufficient disk space to store this copy. You can specify the folder for the case copy in the case conversion dialog.

Case conversion ensures that case databases can technically be opened in a newer version. Re-indexing ensures that case databases are rebuild from the ground up with that newer version, employing all the improvements that are present in that newer version. It is recommended to re-index a converted case to take advantage of the improvements of a newer version and to make sure that non-trivial queries work as expected.

8.5.1. Converting Compound Cases

When converting a compound case, Intella will ask to specify new paths for all sub cases and convert them automatically. At the end, all converted sub cases will be added to the newly converted compound case. The procedure is fully automatic.

8.6. Exporting a case

In the Case Manager, use “Export…” to export the selected case. Choose a location for the ICF file in the “Choose file to export the case” dialog. You can also select the number of threads (CPU cores) used during the export, and whether or not to split the output file into smaller segments. To start the export process click "Export".

Once the case file has been created, a dialog is shown that lists the location of the created ICF file. This file is to be handed out to the investigators that need to work on this case.

The dialog also lists the location(s) of the evidence file(s) used in the case. These only need to be distributed when the receiver of the ICF file needs to be able to re-index the case. For all other tasks, including exporting, the case is fully self-contained.

8.7. Case templates

A case template is a collection of configuration settings, preferences and case metadata that can be exported from an existing Intella case and re-used for the creation of other cases. Case templates allow for initializing new cases quickly with predefined sets of tags, keyword lists, tasks, column settings, etc.

image

To create a new case template, select a case in the Case Manager and click the “Export as template…” button. This opens a dialog box where you can specify the path and name of the Intella Case Template (ICT) file and select the template components to include:

The available template components are:

  • Case preferences. These are mostly settings in the “File → Preferences” window. Also included are the number of settings that are specified in the "Advanced" section of the Case Manager (optimization folder, memory allocation and crawler configuration).

  • Saved Searches

  • Column presets

  • Indexing tasks

  • Tags

  • Custodians

  • Content Analysis categories

  • Keyword Lists

  • MD5 lists

  • Export templates

  • Redaction profiles

  • Excluded paragraphs

  • Custom columns

  • Load File import templates

  • Coding layouts

Press the “Export” button to save the ICT file and close the dialog box.

To apply a template when creating a new case, click the “Advanced…” button in the “Create new case” dialog to show the additional options. Then enter the path and name of the ICT file in the “Case template” field. You can use the “Browse” button to find the template file or select one of the recently used templates from the drop-down list.

9. Interoperability with other Intella products

9.1. Reviewing a case using Intella Viewer

It is possible to use Intella Viewer products to connect to a case shared with Intella Connect or Intella Investigator. This can be especially useful if certain features available in the former tools need to be used in a case which is actively reviewed. In such scenarios unsharing a case might not be an option. Thankfully, Intella Connect and Intella Investigator use remote APIs which are compatible with Intella Viewer, so such connection is possible.

In Intella Case Manager, choose "Add…​" → "Open a shared case" to get "Create new case" window.

Create new case

Case link field needs to be filled with case URL, example:

http://192.168.1.115:9999/shared/master-outlook

Investigator and Passphrase fields need user and his password, which was granted access to that case.

Use "Check connection" button to verify if Case link, Investigator and Passphrase field are correct and the shared case can be reached. If those fields will be correct and shared case can be reached, then message "Connection OK" will be shown next to "Check connection" button. Otherwise check if Case link, Investigator and Passphrase field are correct or if firewall is not blocking communication. If the case still cannot be reached, check if it can be reviewed in browser as described in [reviewing-with-intella] section.

After filling in Local Case folder and clicking Ok, the shared case will open and the reviewer can start reviewing it.

9.2. Uploading a case to Intella Connect or Intella Investigator using Intella Professional

Intella Professional client can upload its local cases straight to Intella Connect or Intella Investigator, given that the latter has been properly configured beforehand. This process relies on a proprietary case format called Intella Case File with an *.icf extension.

9.2.1. Uploading an *.icf file

The process is initiated from the Case Manager window of Intella Professional client, by pressing Upload…​ button. The following dialog will appear.

ICF file upload dialog

The Server URL field should specify a full URL point to your Intella Connect or Intella Investigator instance, including a protocol and port (ex. https://connect.mycompany.com:443). User name should match a valid Intella Connect or Intella Investigator user account. That account should have appropriate permissions to either manage the Intella Connect or Intella Investigator instance as an administrator or at minimum manage cases. The Passphrase should match the password used when signing in using Intella Connect or Intella Investigator login form.

In the ICF area, please select the appropriate option depending if you want to use existing *.icf file or create a new one:

  • Create new - allows to specify a location where a new ICF file will be created. This file can be as large as the case folder itself, so please make sure to specify a disk with sufficient free disk space.

  • Use existing - allows to select existing ICF file, which could be created using Intella Professional beforehand.

After pressing Upload the procedure will start. You can monitor its progress inside the black text area.

Once this operation finishes, you can import the ICF file to extract the contents of the file to a disk on Intella Connect or Intella Investigator server. This feature is available under the Cases / Import / Intella Case File (*.icf) subview.

9.2.2. Windows configuration

The process of uploading *.icf files to Intella Connect or Intella Investigator server is delegated to a secure, well known sftp protocol. Therefore in order to make this feature work properly, the Windows operating system installed on the machine where the product is hosted needs to be correctly configured.

First thing which needs to be done is to install Open SSH Server on machine hosting Intella Connect or Intella Investigator. Depending on which operating system you are running, this process may vary. Newer versions of Windows Server allow to install the server using Windows Settings and Optional Features. Older versions may require to install the binaries from a 3rd party repository. The detailed steps for this process are out of scope of this manual, but below is a list of handy resources which may help you through the process:

Once the Open SSH Server is up and running on the default port (22 or other), make sure to allow the network traffic on Windows Firewall. It doesn’t have to be open to the internet, but you must make sure that connections between your server and the machine running Intella Professional are not interrupted.

You can now move to additional configuration of the Open SSH server.

Open SSH: changing default location of uploaded ICF files

You need to change the default SFTP folder location so that it matches shared system configured in Intella Connect or Intella Investigator. This is usually done by editing the sshd_config file which is created during the installation of Open SSH server. The default location should be C:\ProgramData\ssh\sshd_config, but if it doesn’t work you may need to consult documentation of the Open SSH server for alternatives. Find a line where sftp sftp-server.exe is listed and change it to the following format:

sftp sftp-server.exe -d "D:\configured_shared_file_system\cases"

Make sure to substitute the path according to your file system structure and that the Windows account running Intella Connect or Intella Investigator process has sufficient permissions to access that directory.

At this point you can restart the Open SSH Server Windows service for changes to take effect.

Make sure that your path does not contain whitespaces, as the Open SSH server used to have a problem with handling them. In case of you need more information or a place to ask questions please check this post: https://github.com/PowerShell/Win32-OpenSSH/issues/730

Open SSH: configuring access to "authorized users file"

Uploading of ICF files is handled by private-public key pair encryption, well known and supported by Open SSH server. This happens transparent to the end user and both types of keys are additionally encrypted when stored on disk or when in transfer. This, however, requires Open SSH server to be additionally configured so that the product can update the list of public SSH keys recognized by the server.

Since the product is started by a Windows account belonging to Administrators group, the configuration file that is of our interest should is: C:\ProgramData\ssh\administrators_authorized_keys It may happen that this file doesn’t exist (by default), so if that is the case you should create it. It is of vital importance that only three types of users can access that file: Administrators, System and user account running Intella Connect or Intella Investigator. The following PowerShell script can be used to set those permissions properly (you must run it as an administrator):

$acl = Get-Acl C:\ProgramData\ssh\administrators_authorized_keys
$acl.SetAccessRuleProtection($true, $false)
$administratorsRule = New-Object system.security.accesscontrol.filesystemaccessrule("Administrators","FullControl","Allow")
$systemRule = New-Object system.security.accesscontrol.filesystemaccessrule("SYSTEM","FullControl","Allow")
$intellaAccountRule = New-Object system.security.accesscontrol.filesystemaccessrule("NAME OF ACCOUNT RUNNING INTELLA PRODUCT","FullControl","Allow")
$acl.SetAccessRule($administratorsRule)
$acl.SetAccessRule($systemRule)
$acl.SetAccessRule($intellaAccountRule)
$acl | Set-Acl

Please make sure to change NAME OF ACCOUNT RUNNING INTELLA PRODUCT to a valid Windows Account name which starts Intella Connect or Intella Investigator process.

Proper set of permissions is presented in the following picture:

Open SSH correct file permissions

At this point Open SSH server installed on your server becomes as every other SSH server. If you wish to test that everything is fine at this point, you may create a new SSH key pair on a different machine, then append your public key to the mentioned administrators_authorized_keys file and try to connect. This process is well described here: https://docs.microsoft.com/en-us/windows-server/administration/openssh/openssh_keymanagement

Depending on the Open SSH Server that you are using, the configuration guide presented above may not be 100% complete. You may also want to change other parameters of the SSH server configuration, like: disabling password authentication, changing Strict Mode, etc. Note that this extra configuration and troubleshooting Open SSH Server issues is not supported by our Support terms and conditions.

10. Overview of the Intella interface

Intella’s main window consists of the following tabs, or "views": Welcome, Sources, Insight, Search, Keywords and Identities. Optionally, one or more Review views, W4 Import view, and Errors view can be present. Together they give access to all information stored in an Intella case.

Another prominent window is the Previewer, showing detailed information about an item. The Previewer can be opened by clicking or double-clicking on certain elements within the Insight or Search tabs.

This chapter gives a brief overview of these user interface parts. More detailed information is provided in later chapters.

10.1. Welcome view

The Welcome view is the first and default tab in a new case. It shows what features and other improvements have been added in this version of Intella, and offers links to various types of documentation, support, and common actions.

image

The Welcome view offers a checkbox that lets the user suppress this tab the next time this case is opened. This setting is stored per case, i.e. new cases will again show the Welcome view.

10.2. Sources view

The Sources view is a sources management panel, where data sources, such as evidence files and internet resources can be added to the current case.

image

The panel also provides source management functions, such as updating source parameters, reindexing and removing. Please refer to the "Sources" section for details.

10.3. Insight view

image

The Insight view shows notable aspects of the indexed evidence files and possible next steps to take. The overview given here can help an investigator get a grasp of the case’s contents, such as the encountered item types and their volumes, date ranges, web browser activity, potentially privacy-sensitive information, etc. This will help formulating follow-up questions for further research.

Most elements in this view can be clicked or double-clicked, which starts a search in the Search tab or opens a corresponding item in the Previewer.

10.4. Search view

image

The Search view allows for arbitrary searches in the case data using keywords or one of the navigation facets such as date, location, item type, etc.

  1. The Keyword Search panel is the place to enter search terms or phrases.

  2. The Facet Search panel shows a list of facets for searching and filtering results. Each facet represents a different dimension in which the items can be discerned. Select a facet from the list to see the navigation options offered by that facet, shown beneath the list.

  3. The Searches panel shows the user’s keyword and facet queries, together with their result count.

  4. The Results panel shows the search result sets of these queries in various ways, by grouping the items in a certain way.

  5. The Details panel shows a table, list, thumbnail view, or timeline view of the results in a selected element in the Results view. It is populated by selecting elements in the Cluster Map, Geolocation view, Histogram, Social Graph, or Searches list. Click or double-click (depending on the chosen view) on an item to view in in full detail in the Previewer.

10.5. Keywords view

The Keywords view allows the user to gather statistics on the keywords in a keyword list. After selecting or adding a new keyword list, the user can choose on what document fields the keywords need to be evaluated, e.g. the document text, email headers, authors, etc. Furthermore, the user can choose what statistics need to be calculated: the number of items matching the query, the number of hits (occurrences) of each query, the item count per custodian, etc.

image

Once all options are configured as desired, the user can click the Calculate button. This will populate the table row by row.

10.6. Identities view

image

The identities view lets one build an “address book” of the persons of interest in a case. An identity bundles the communication aliases used by a person, such as email addresses, phone numbers and chat accounts, into a single unit. The identity is given a Full Name and can be annotated with other properties. This information is used to enhance the querying and display of items in other parts of the user interface.

10.7. Previewer

The Previewer is typically opened by clicking on elements in the Insight or Search tab, but it can also be opened by clicking on hyperlinks in the Previewer’s own Tree tab, or by using the “Preview item…” option in the View menu and entering the item’s ID.

image

  1. Use the tabs at the top to inspect an item’s contents, headers, properties, attachments, thumbnails, tree structure, extracted terms, comments and performed user actions. The tabs shown for a specific item depends on the item type and its data. Bold tab names indicate the presence of a keyword search hit in the text inside that tab.

  2. The Contents tab always starts with a summary of important information of the item, followed by a document or message body, image content, etc.

  3. When the Search view shows the results of one or more keyword queries, the status bar at the bottom will show the keywords found in the current item and offer buttons to navigate from hit to hit.

  4. The toolbar on the left lets one navigate to and search for related items, annotate the current item in various ways and produce the current item in several formats.

10.8. W4 Import view

The W4 Import view shows various details of all imported W4 cases that are present in the current Intella case. This tab will not be present when the case does not contain any W4 sources. Use the “W4 case source” drop-down menu to select a source to display.

The tab consists of the following sections:

  • A “case details” section, showing basic case information such as size and creation date.

  • A “Delta” chart, showing the additional items found and tagged by Intella, compared to the items in the original W4 case.

  • A “Top-level Categories” chart, showing the additional items found by Intella for each top-level category.

  • An “All Categories” tree table, breaking down the additional items by type. Clicking on a number in the table shows the corresponding items in the Search tab.

  • A “Tags” section, showing the additional items by tag.

  • A “Keyword Lists” section, doing the same for each keyword list.

For more details, see the “Source types” section – Vound W4 case.

10.9. Errors view

The Errors view is visible when critical errors occurred during source indexing. It lists item-specific processing errors, grouped by source, as well as general post-processing errors in the special “Post-processing” category.

desktop errors tab

The bottom part of the Errors view contains an explanation text for the selected error.

To preview an item related to an error, double-click the item’s row in the table. This option is only available for item-specific errors.

To export the Errors view content to Excel or CSV format, click on the Export button and specify the file and format in the file chooser that opens.

Click the "Dismiss" button to clear and hide the Errors view until any new indexing errors occur.

11. Sources

Sources are one of the key concepts of Intella. They represent the locations where items such as emails, documents and images can be found. The user explicitly defines the sources, providing full control over what information is searched.

11.1. Source types

desktop add new source

Intella distinguishes between various types of sources. The Add New Source wizard organizes them in two columns : sources dealing with local evidence files on the left and sources dealing with cloud or server-based data on the right ;

The supported types are:

  • File or Folder: A single file or folder with source files on a local hard drive or on a shared/network drive. Such source files could be:

    • Regular loose files like MS Word, Excel, and PDF files.

    • Email containers such as MS Outlook PST/OST, HCL/IBM Notes NSF files, Mbox files.

    • Cellphone XML and UFDR reports such as made by Cellebrite UFED, MicroSystemation’s XRY and Oxygen Software’s Forensic Suite.

    • Even large containers like EDB files and disk images can be indexed this way, together with many other files in one go. The downside of doing this is that any EDB- or disk image-specific configuration options are not available this way.

  • Load file: A Concordance, Relativity, or CSV load file.

  • Hotmail Search Warrant Result (experimental): a collection of files in HTML and other formats, provided by Microsoft pursuant to a search warrant.

  • Disk Image: One or more disk images in E01, Ex01, L01, Lx01, S01, AD1, AFF4, VHD, VHDX, AVHDX, VMDK, X-Ways or DD format.

  • MS Exchange EDB Archive: a single MS Exchange EDB file.

  • Vound W4 Case: A case created in the Vound W4 application.

  • IMAP account: An email account on an IMAP email server.

  • Dropbox: all files stored in a personal Dropbox or DropBox for Business account.

  • Google: A Google account: Gmail, Drive, Contacts, Calendars and Tasks.

  • SharePoint: The complete contents of a SharePoint instance.

  • Microsoft 365: The complete contents of a Microsoft 365 account, incl. the Outlook, OneDrive and SharePoint services of that account.

  • iCloud: The complete contents of an iCloud account, incl. iCloud Drive, Mail, Calendar, Contacts, and other services.

  • AWS S3: The complete contents of an Amazon AWS S3 bucket.

Notes on mail formats
Intella supports PST and OST files created by the following versions of Microsoft Outlook: 97, 98, 2000, 2002, 2003, 2007, 2010, 2013, 2016, 365. Make sure that Intella has exclusive access to the PST or OST file; it cannot be open in Outlook or other application at the same time.

Intella will try to recover the deleted items from the file. Recovered items will be placed in a special folder named <RECOVERED>. Furthermore, Intella may encounter items outside the regular root folder. Any such items are placed in a special folder called <ORPHAN ITEMS>. Recovered emails may contain traces of other emails. This should be considered when reviewing such items.

Orphan items may contain unreliable data. For example, some orphan items can contain pieces of the message body, and message metadata from different emails. This may be due to the way the email client caches message data in the email container.

You should consider whether this information should be included in exports. Some clients may not want this information exported due to its unreliable nature.

To index NSF files, HCL/IBM Notes 8.5 or higher needs to be installed. For NSF files made with HCL/IBM Notes 9 it is recommended to install HCL/IBM Notes 9. Intella supports all NSF files that can be processed by the installed HCL/IBM Notes version. Make sure that Intella has exclusive access to the NSF file; it cannot be open in a Notes client or other application at the same time. Only NSF files containing emails are supported by Intella, all other types are not supported. Make sure to use a default Notes installation and user configuration. A "corporate" Notes installation is often problematic for indexing, e.g. because of installed plugins interfering with access to the NSF file, the installation being tied to the corporate identify management system, etc.

The HCL/IBM Notes tool nupdall.exe can be used to convert older NSF files to NSF files that can be processed by HCL/IBM Notes 8.5 and higher.

Notes 9.0.1FP8 or higher needs to be installed to decrypt messages in a non-encrypted NSF.

Intella supports Windows 10 Mail mailboxes, provided that the account uses the POP protocol. Accounts that use the IMAP protocol are not supported, as only POP accounts store mails locally. Furthermore, Windows 10 mails do not keep track of BCC-ed email addresses and of the email headers.

Intella supports DBX files created by the following versions of Microsoft Outlook Express: 4.0, 5.0, 6.0.

Intella has been tested on Thunderbird Mbox files.

Intella supports MS Exchange EDB files of Exchange versions 2003, 2007, 2010, 2013 and 2016.

Some items may turn out to only contain email headers and are lacking an email body. Examples of such items are messages typically sent back by mail servers to indicate undeliverable mails, e.g. due to an unknown recipient or a mailbox quota that has been reached. Such items are typed as "Email Headers" rather than "Email Message".

Notes on cellphone formats
When indexing Cellebrite, MicroSystemation or Oxygen cellphone reports, each report should be in its own subfolder. Any additional files that were produced together with the XML report, such as audio, video, and image files, should have the same relative location to the XML file as the exporting application produced them. These two requirements are crucial for correctly linking the binary files with the XML report. Finally, no other evidence files should be placed in these folders, as they will be ignored.

The folder should reside in the local file system or in a disk image, i.e. not in a ZIP file or other type of archive, as quick random access is needed to be able to process the files linked from this report.

A folder with the XML report and its related files can in principle be indexed straight away. However, most XML reports will often only contain the external numbers related to the calls and messages, i.e. the number of the phone itself is not in the report. This has valid technical reasons (e.g. it cannot be guaranteed that the current SIM card was used for these calls and messages), but it makes analysis of the communication a lot harder. Also, Intella functionalities like message deduplication require this information. When the investigator knows the number, e.g. obtained from the network provider, it may be specified through a separate text file:

  1. Create a text file named after the XML report. For example, if the report is called report.xml, the text file should be named report.numbers.txt.

  2. Put it in the same folder as the XML report.

  3. Store the phone’s own number in this file.
    When the XML report holds information about multiple phones, enter the number of each phone on a separate line, like this:
    number1
    number2
    …​

  1. The first line will be used for the first phone found in the report, the second line for the second phone, and so on.

When indexing XRY’s XML reports, we recommend using the Extended XML report introduced in XRY 6.4. This new format solves many issues with the encodings of dates and other fields. Furthermore, the older XML format did not support exporting binary items. To get binary items with the Extended XML report, you need to select the "Export media files and manifest" option.

The XML formats used by these cellphone extraction vendors are often evolving over time and are not fully documented. While we strive to extract all information from these reports as completely and correctly as we can, we can only offer this functionality on a best-effort basis. We recommend that you verify any results that you may rely on in your report with the original cellphone extraction software. Please let us know if you find any issues with processing reports made with a certain version of a cellphone extraction suite, and we will add that version to our development roadmap.

Notes on instant messages
When instant message items (SMS/MMS/iMessage/Skype/Jabber/etc.) have a timestamp and the sender and receiver(s) are all known, Intella will bundle all messages of that group of participants into "conversation items". A conversation item bundles the messages between a group of people on a day-by-day basis. All messages of a single day are now placed below each other in the Previewer’s Contents tab, rather than being presented as one message per item. Hyperlinks are provided to navigate to the previous and next day in the conversation.

Compared to emails, instant message texts are typically very short and do not contain the previous thread. Therefore, bundling messages in this way greatly improves reviewing of instant messages.

Other instant messages, which do not have enough metadata to be bundled into conversations, will be reported as conversations consisting of a single message.

Notes on IBM Sametime dumps
IBM Sametime dumps must be located in the local file system or in a disk image, i.e. not in a ZIP file, as quick random access is needed to be able to process the files linked from this report.

Common file locations
MS Outlook PST and OST files are typically located in the following folder:

  • Windows Vista, Windows 7, Windows 8/8.1, and Windows 10:
    C:\Users\<username>\AppData\Local\Microsoft\Outlook

  • Windows 2000 and XP:
    C:\Documents and Settings\<username>\Local Settings\Application Data\Microsoft\Outlook

MS Outlook Express DBX files are typically located in the following folder:

  • Windows 2000 and XP:
    C:\Documents & Settings\<username>\Local Settings\Application Data\Identities\
    \{<arbitrary string>}\Microsoft\Outlook Express

HCL/IBM Notes NSF files are typically found in the following folder:

  • Version 7.x:
    C:\Program Files\Lotus\Notes\Data

  • Version 8.x:
    C:\Program Files\IBM\Lotus\Notes\Data

  • Version 9.x and 10.x:
    C:\Program Files\IBM\Notes

  • Version 11.x:
    C:\Program Files\HCL\Notes

Notes on cloud sources
Each of the supported cloud services (Dropbox, Google, SharePoint, and Microsoft 365) provides a so-called REST API for data retrieval. Access to a cloud service via this API often requires an authorization token, rather than or in addition to a username and password. Each cloud service provides a web portal where users can register the client application (in this case: Intella) and obtain the authorization token.

Depending on what the REST API supports, Intella uses read-only data operations wherever possible, as to minimize changes to server-side data. Nevertheless, access may be visible to the cloud service and to the account holder, e.g. due to the presence of an authorization token in the server settings, access logging, altered metadata, etc.

Notes on document length
The indexing of a document text for keyword search can consume a considerable amount of RAM. With multiple documents being processed in parallel, this carries the risk of one of Intella’s processes running out of memory. To combat this, Intella imposes a maximum length to the document text. This way, typically problematic textual files such as large server logs and database dumps in CSV format can be processed without terminating the indexing abruptly.

By default, the maximum length is set to 50M (52,428,800) characters. Any text beyond that point is skipped. Consequently, the document will not be returned when using query terms that only occur after this point. Affected documents can be located using the "Exception Items" category, "Truncated text" branch in the Features facet.

The limit can be adjusted on a case-specific basis via the case.prefs file. For example, alter or add the following line to set the limit to 100M characters:

ItemTextMaxCharCount=100M

A future Intella version will make this configurable via the user interface. The limit can also be adjusted globally via the Intella.l4j.ini file:

-Dintella.itemTextMaxCharCount=100M

11.2. Sources view

The top-level "Sources" view of Intella main window provides a user interface for adding and managing data sources.

image

On the left side, the view contains a list of all sources defined in the case. With checkboxes, one or more sources can be chosen in the list for operations, specified by the toolbar buttons ("Re-index", "Index new data", "Exceptions report" and "Remove"). Click on "Select All" checkbox to mark/unmark all sources in the list.

The right side contains source information panel, where various details of the currently selected source can be reviewed and edited (see "Editing sources" section below).

11.3. Adding sources

Adding sources to Intella is done with the Add New Source wizard. You can start this wizard by clicking "Add New…​" button in the Sources view.

11.3.1. Files and Folders

Follow these steps to add a File or Folder source to Intella:

  • Source type

    Select "File or Folder" source type and click "Next". A folder tree will be displayed next.

  • Specify file or folder

    Select the folder or file from the tree that you want to index, or enter the folder or file name in the text field above the tree. When selecting a folder, all files in the selected folder will be indexed. When the "Include subfolders" checkbox is selected, files in all subfolders (and sub-subfolders, etc.) will also be indexed. When the "Include hidden folders and files" checkbox is selected, hidden files and folders will be indexed as well.

Folder trees containing many items may take some time to be displayed. Please be patient.

Click "Next" to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

Though disk images can be added and indexed via File or Folder source, there is a limitation on the maximum number of parts (files). For E01 format the maximum number of parts is 4,831 (e01-e99, eAA-eZZ, fAA-kZZ), for other EnCase formats (L01, Lx01, Ex01) the maximum number is 775 (L01-L99, LAA-LZZ). If the disk image contains more parts then it is strongly recommended to use Disk Image source type instead.

11.3.2. Load files

The built-in export and import templates "Intella Standard Relativity Export (All Columns)" and "Intella Standard Relativity Import" can be used to export items and re-import them in another case, effectively creating a subset of the original case. Please note that not all metadata fields are supported.

Follow these steps to add a load file to an Intella case:

  • Source type

    Select "Load file" source type and click "Next".

  • Import load file

    • Select the import operation: New Data or Overlay. When New Data is selected, Intella will import new items to the case. An Overlay operation is used to import tags, comments and tag columns into existing items.

    • Add the file name and location of the load file that you wish to investigate; use the tree component to browse for the file. If the load file comes with an Opticon image file, then you should specify it in the "Opticon image file" field.

    • Specify the source name.

    • Specify the custodian. If the custodian information is stored in one of the columns, then leave the text field empty and use the column chooser on the "Map fields" page instead.

    • Specify the time zone. By entering the time zone, all dates associated with items from this load file will be displayed in that time zone, rather than the time zone of the investigator’s system.

    • You can use a previously saved import template.

    • Click "Next" to continue.

  • Configure delimiters

    On the "Configure delimiters" page you can set the file encoding and delimiter settings for:

    • Column delimiter – the character that separates the columns in the load file.

    • Text qualifier – the character that marks the beginning and end of each field.

    • New line – the character that marks the end of a line inside a text field.

    • Multi-value delimiter – the character that separates distinct values in a column.

    • Escape character – the character that is used for escaping a separator or quote.

    • Strict quotes – sets if characters outside the quotes are ignored.

    • Use absolute path – select this option when the load file uses absolute paths rather than relative paths.

      You can click the Detect button when you are not sure about the encoding used in the load file.

      You can specify date, time and number formats in the right part of the screen. The Size unit option allows to change the way how the Size field is imported.

      Intella will validate the load file using these settings and display the validation result in the status line. When the file can be validated successfully, the number of columns found in the load file will be displayed. When validation fails, a reason will be given in this line.

      The "Load file preview" table can be used to make sure that you have specified the correct parameters for the load file. Additionally, the "Image preview" panel will show the first image associated with the selected table record. It can be used to ensure that the Opticon file is correctly loaded. The "Text preview" shows the raw text of the load file and can be used to check the delimiters.

      Click "Next".

  • Map fields

    • Overlay options: this is only used when Import operation is set to Overlay. See the "Importing an overlay file" section for details

    • External files:

      • Select the "Load native files" checkbox if you want to import original format files associated with the load file into the case. Specify the column containing the paths to the native files. When the native files are imported, you will be able to use functions such as Preview tab and Open in External Application.

      • If you select the "Extract type information from native files" check box, then Intella will analyze the native files and import the type information into the Mime Type and Type columns. This option may be useful in case the load file does not have any type information such as File Extension.

      • Select "Load extracted text" when you want to import the extracted or OCRed text of the document. Select the "Extracted text column is a link to an external file" checkbox when the column contains a link to the text file rather than the text itself. Select "Analyze paragraphs" to let Intella determine the paragraph boundaries and to let it build a database registering which paragraph occurs in which item and where (see section [last-steps-in-a-source-definition] for more details). When the extracted text is imported, it will be shown in the Contents tab of the Previewer.

    • Field mapping – You can see the Field chooser in the bottom part of the panel. The table on the left shows all fields in the load file ("Load file field") and the Intella columns they are mapped to. In the table on the right you can see the list of all Intella columns available for mapping. To map a column:

      • Select one of the load file fields on the left.

      • Select one of the columns on the right.

      • Click the left arrow button. That will move the selected column from the right to the left table.

        Click the right arrow button to remove the selected mapping.

        When the load file contains a field that cannot be mapped to any existing columns, then you can create a tag or custom column and map the field to it. Click the "Add" button to add a new column to the right table. Click the "Remove" button to remove the selected column. Note that a tag or custom column can only be removed if there is no data in the case associated with it. Tag columns should only be used for importing tag-like data where the number of unique values is not high. In all other cases custom columns should be used instead.

        Click the "Clear all" button to remove all the selected columns from the right table. Click the "Save template" button to save the current settings as an import template which can be reused later. Select the "Extract text and metadata from native files" checkbox when you want to extract the text and metadata from the native file. The button with a gears icon can be used to adjust the processing options. See [last-steps-in-a-source-definition] section for more details about the processing options. Note that Intella will replace any original metadata from the load file with the new metadata extracted from the native file. The option is turned off by default.

        It is highly recommended to resolve all errors by clicking the "Check for errors" button before importing the load file. That will let Intella validate the load file using the entered settings. Among other things, it will check each row and ensure that:

    • The Document ID is unique and not empty.

    • The Parent ID refers to an existing record.

    • Native and extracted text paths are correct.

    • Date and time fields can be parsed using the selected date and time formats.

    • The MD5 field contains a valid MD5 hash.

    • Number fields such as File Size and Page Count contain a valid number.

    • Boolean fields such as Encrypted and Decrypted contain either "true" or "false".

    • The Source IP field contains a valid IP address.

      Select the "Skip error records" checkbox to instruct Intella to skip items with errors during import.

Date and time values (separate columns) will be merged into one column.

Important notes on load file importing

There are several aspects to be aware of when importing a load file into an Intella case:

  • All paths in the load file to external resources should be relative to the load file, unless the "Use absolute paths" checkbox is selected.

  • The original load file record identifiers will be imported into the "Document ID", "BegAttach / Parent ID" and "EndAttach" columns and can be used in a subsequent load file export.

  • Imported images can be viewed in the "Image" tab in the Previewer.

You can save the specified load file import options as a template for later usage on the last page using the button Save Template. All import templates are stored as XML files in the "<Intella Home Folder>import-templates" folder.

11.3.3. Hotmail Search Warrant Results

This source type is still in an experimental stage. We welcome any feedback; please visit our support portal at http://support.vound-software.com/.

Follow these steps to add a Hotmail Search Warrant Result to Intella:

  • Prepare evidence files

    The evidence files you have received may consist of a folder containing a "Click Here.html" file and some legal files related to the search warrant, with a subfolder for each account involved. It may also be that you have only one of those account subfolders, recognizable by a "Folders.html" and "Messages" file in this folder. In case you have received a ZIP file or some other type of archive file, please unpack this archive file first.

  • Source type

    Select "Hotmail Search Warrant Result" source type and click "Next".

  • Specify file

    Select the folder holding the Hotmail Search Warrant Result files that you wish to investigate in the folder tree. Make sure to select the top-level folder of the provided file collection. Click "Next" to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

11.3.4. Disk Images

Follow these steps to add a Disk Image source to Intella:

  • Source type

    Select "Disk image" source type and click "Next".

  • Select disk image file

    Click on the Browse button to navigate the folder tree in order to select the disk image file. You can also paste the location to the input field. Once an image is specified Intella will automatically detect other parts of this image, if there are any. You will see basic information about discovered paths and total disk image size.

After you click Next button in the wizard footer, Intella will validate selected disk image. This process may take long time depending on the complexity of the data.

If the disk image contains encrypted volumes, such as BitLocker or APFS, a notification will be shown instructing to update Keystore of this case with matching passwords or recovery keys to access the image.

  • Select folders to process

    Specify folders that need to be processed. Initially Intella will only show the first three levels of the folders. If you wish to load all folders in the disk image, press the "Load all" button. Note that scanning all folders in the disk image might take a while.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

A single disk image source should only contain the files relating to a single conceptual image. Files relating to a different image should be entered as a separate source.

Filtering disk image content

A disk image often contains a lot irrelevant files, such as executables, DLLs. These files add to the processing time and disk space that the case will consume. It is possible to define a set of rules to filter out unnecessary files and folders, to save processing time and disk space.

Filtering disk image content is not possible for DMG images.

Note that search results can also be filtered after indexing, using the Hide Irrelevant filter option in the Details tab.

Supported disk image formats

The Disk image source type supports EnCase E01, Ex01, L01, Lx01 and S01 files. Password-protected files are supported and indexed without manual interaction, except for FTK-encrypted files.

DD images are supported, but when a Folder source is used, they need to use the .dd file extension to be detected and processed as DD images. Because of potential issues with DD image detection, we recommend using the Disk Image source directly. This is also required when you want to index a multi-volume DD image

Supported file systems and partition types

The following file systems have been tested: FAT16, FAT32, ExFAT, NTFS, Ext2, Ext3, Ext4, HFS, HFS+, APFS and ISO 9660. Other file systems such as YAFFS2, ISO 13346 (UDF), UFS 1 and UFS 2 may work but have not been tested yet.

MBR and GUID partition tables (GPT) partitions are supported. Apple Partition Maps (APM) have been tested but results were mixed. When an image cannot be indexed, we recommend mounting it manually and indexing the mounted drive using a “File or Folder” source.

APFS and BitLocker encrypted volumes are supported. When an encrypted volume is detected, a dialog will be shown where it’s possible to enter a password or recovery key. BitLocker volumes with suspended protection (also known as "clear key") will be indexed automatically without a password prompt. If a BitLocker volume is protected with multiple keys, you can enter any key.

Multi-volume files

When using a Folder source to index multiple image files, Intella will rely on the following file name convention to determine which files together make up a single image:

image1.e01        (first volume of image 1)
image1.e02        (second volume of image 1)
image1.e03        (third volume of image 1)
…
image2.e01        (first volume of image 2)
image2.e02        (second volume of image 2)
image2.e03        (third volume of image 2)
…
image2.e99        (99th volume of image 2)
image2.eaa        (100th volume of image 2)
image2.eab        (101st volume of image 2)
…

Volume shadow copies

Enabling volume shadow copies processing might considerably slow down the indexing process.
Volume shadow copies can only be processed when using the Disk Image source. If disk image is added via File or Folder source, volume shadow copies will not be processed.

Volume shadow copies (VCS) is a mechanism in Windows OS that preserves previous versions of files in a special hidden area on the disk. A new VSC snapshot is often created automatically by Windows when installing major system updates or drivers.

When Intella detects that the disk image contains VCS, the Specify Volume Shadow Copies page will be shown. On this page you can select specific snapshots that need to be processed.

By default, Intella will only extract the files that were changed between snapshots. That allows to save a lot of processing time and disk space by not indexing the same file several times:

  • Select the "Prefer oldest files" option to extract all files from the oldest snapshot and only the changed files from the newer snapshots.

  • Select the "Prefer newest files" option to extract all files from the current file system and only the changed files from the older snapshots.

Intella uses the last modified date of the file to determine whether it has changed. It is also possible to take the last access date into account.

The "Has Shadow Copies" option in the Features facet can be used to see all items that have other versions in shadow copy volumes.

To see all items extracted from all volume shadow copies, use the "Recovered → Recovered from volume shadow copy" option in the Features facet.

11.3.5. MS Exchange EDB Archives

Processing an EDB archive may require to adjust memory settings. Please see the "Memory settings" section for detailed instructions.

The currently supported MS Exchange versions are 2003, 2007, 2010, 2013 and 2016.

Follow these steps to add a MS Exchange EDB Archive source to Intella:

  • Source type

    Select "MS Exchange EDB Archive" source type and click "Next".

  • Specify EDB file

    Specify the location of the EDB file you wish to investigate either by typing it’s location or by selecting it in the folder tree. Click "Next" to continue.

  • Select mailboxes

    Check all mailboxes that you wish to process. Click "Next" to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

When an EDB source has been added and not all mailboxes were selected, it is still possible to index additional mailboxes in that EDB file at a later stage. To do that, the following steps should be performed: 1. Click on the "Edit" button for the respective source on Sources page. 2. Indicate which mailboxes should be processed. Note that you cannot unselect or remove already processed mailboxes. Click OK. 4. Use the "Index new data" button option to index the new mailboxes.

11.3.6. Vound W4 Case

This source type lets one import an entire case created with the Vound W4 application into the current Intella case.

Follow these steps to add and process a W4 case source:

  1. Source Type
    Start the Add New Source wizard from the Sources view. Select "Vound W4 Case" and click Next.

  2. Select W4 Case
    Click the Browse button to specify the location of the "case.json" file in the root folder of the W4 case that you wish to add. After the file is selected, Intella will validate the W4 case and check its availability. Once it has validated successfully, click Next to continue.
    Note that the case cannot be imported while it is still open in W4. If case validation fails with a "W4 case is currently in use" message, close W4, click Back to return to the Source Type selection page, then click Next and Browse…​ to select the "case.json" file again.

  3. W4 Case Options
    On this page, you can specify which parts of the W4 case data should be imported. Furthermore, you can configure post-processing tasks to enhance the original W4 data in Intella.

    • Tags
      Import the tags assigned to the items in W4 case.

    • Item notes as comments
      Import the notes assigned to the items in W4 case and represent them as Intella comments.

    • Keyword Lists
      Import all keyword lists from the W4 case.

    • Auto-tag items using imported keyword lists
      When keyword lists are imported from the W4 case, this option performs auto-tagging of all items with these keyword lists (see the "Keyword Lists" section for details).

The last steps in the definition of a source are almost the same for all types. They are described in the section "Last steps in a source definition".

W4 cases that contain a local disk source cannot be imported by Intella. Such functionality may be added in a future version.

11.3.7. IMAP accounts

The IMAP standard is implemented in many ways. Furthermore, some mail servers may throttle the network connection during mass downloads. We tested Intella on several IMAP servers with good response. However, we cannot guarantee that Intella can create IMAP account sources for every IMAP server.

We recommend using a mail client to download the entire mailbox and indexing the resulting PST or Mbox file instead, rather than using Intella to download the mailbox. This way a copy of the mailbox is created outside of the case. This results in a cleaner and better auditable workflow, allowing e.g. cross-validation of the investigation results with other forensic tools or indexing with future Intella versions.

Follow these steps to add an IMAP Account source to Intella:

  • Source type

    Select "IMAP account" source type and click "Next".

  • Specify account

    Enter the settings for the target email account, e.g., "mail.my-isp.com" with the username and password. Select the "use secure connection (SSL)" checkbox if you want or need a secure connection to the mail server. This is recommended, because without a secure connection your password will be sent as plain text. Click "Next" to continue.

  • Select folders

    In the next step, Intella will contact the specified email server to retrieve the folder tree of the target mail account. You can then select the folders that you want to make searchable by placing a check in the box next to the desired folders. When you want to index subfolders, you will need to select them; otherwise they will be ignored. The wizard has two convenient buttons for selecting and deselecting all folders. Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

11.3.8. Dropbox accounts

A Dropbox source reconstructs the entire folder tree in a Dropbox account and downloads current and past revisions of the files in the account.

The official Dropbox REST API used by Intella limits this to a maximum of 10 revisions per file. All revisions except for the last one have their file names decorated with the revision identifier. Furthermore, additional Dropbox-specific metadata is retrieved for both files and folders. These are displayed in the Previewer’s Raw Data tab and are subject to full-text indexing.

Intella uses the OAuth2 (Open Authorization) protocol to access the account. Prior to defining the source, the investigator needs to obtain an OAuth2 token for the account.

This process is described in detail in the following Knowledge Base Article: Collecting data from a DropBox source.

Next, follow these steps to add a Dropbox source to Intella:

  • Source type

    Select "Dropbox" source type and click "Next".

  • Connect to Dropbox

    Follow steps from the knowledge base article to fill required fields.

    Click Next to continue.

  • Select files or folders

    Besides indexing of the entire account, it is also possible to index specific files or folders only. The next wizard sheet shows the folder tree of the account. Nested folders are loaded on demand when the parent folder is expanded. Click the checkboxes of the desired files or folders. Selecting a folder automatically marks all nested elements as selected.

    Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

11.3.9. Google accounts

A Google source allows to download items from the following Google services:

  • Gmail

  • Drive

  • Contacts

  • Calendars (including tasks)

More Google services will be added in the next versions.

Different services require different APIs to be enabled:

  • Gmail - Gmail API

  • Drive - Google Drive API

  • Contacts - People API

  • Calendars - Calendar and Tasks API

Optionally, the set of retrieved items can be restricted to a certain date range.

Benefits of using the Google/Gmail source over the generic IMAP source are: faster performance, more accurate data representation (e.g. folders vs. Gmail’s Labels, threads), and a read-only data connection ensuring that no data is altered on the server.

Intella uses the OAuth2 (Open Authorization) protocol to access the account. Prior to defining the source, the investigator needs to obtain an OAuth2 token for the account. The token will be downloaded as a JSON file, which Intella can use to access the account. This process is described in detail in the following Knowledge Base Article: Collecting data from a Google source.

Next, follow these steps to add a Google source to Intella:

  • Source type

    Select "Google" source type and click "Next".

  • Select Google services

    Select "Google" services and click "Next".

  • Connect to Google

    Click the Select button and select the JSON file saved above in the file chooser that opens. Alternatively you can also drag and drop this file directly to the file upload box. Click Connect to Google.

    A connection will be established and the token will be validated. A browser window will automatically open, through which Google will request permission to continue. If the token validation is successful, basic information about the account such as the account owner’s email address and the total number of emails will be shown beneath the OAuth2 File field.

    Note the Help button below the upload box. Clicking it will display the steps required to create the OAuth2 file.

    Click Next to continue.

  • Configure download

    Select whether all items are to be downloaded or whether a date filter is to be applied. If so, enter the desired date range.

    The end date is included, so that items on that day are also retrieved. Both the start and end dates are optional, making it possible to enter a half-open date range, e.g. "all emails since May 1st, 2015".

    Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

11.3.10. SharePoint

SharePoint is organized as a recursive hierarchical tree of sites. A SharePoint source will reflect all subsites, document libraries, discussion boards containing posts and Microfeed posts. Furthermore, a SharePoint source extract all user accounts, including both system accounts and regular user accounts. What information is required to retrieve information from a SharePoint instance depends on whether it is an on-premise instance or an instance hosted in the Azure Cloud (as a separate Microsoft 365 service). For both types, a server URL, username and password are required. For SharePoint instances in the Azure Cloud, a client ID token is additionally required.

This process is described in detail in the following Knowledge Base Article: Collecting data from a Microsoft 365 or a SharePoint Source.

Next, follow these steps to add a SharePoint source to Intella:

  • Source type

    Select "SharePoint" source type and click "Next".

  • Connect to SharePoint

    Enter the server URL, username and password of the SharePoint account. For Azure Cloud-hosted SharePoint instances, click the Hosted in Azure Cloud checkbox and enter the Client ID in the field that appears. Click Connect to SharePoint.

A connection will be established and the token will be validated. If the token validation is successful, basic information about the account will be shown in the wizard.

The following authentication methods are supported: OAuth2 (for cloud instances), Kerberos, NTLM and basic authentication.

Click Next to continue.

  • Select files or folders

    The next wizard sheet shows the site and folder tree of the account. Nested folders are loaded on demand when the parent folder is expanded. Click the checkboxes of the desired sites.

In this version, only entire sites can be retrieved. A future version may add retrieval of parts of a site.

Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition] .

11.3.11. Microsoft 365

The Microsoft 365 source types allows for retrieving both user account and user groups. For each user account, used to access Microsoft 365, the source can retrieve data from Outlook, OneDrive and SharePoint. For each user group, the source retrieves titled conversations containing emails.

Before a source can be added, the Microsoft 365 account must be properly configured. This process is described in detail in the following Knowledge Base Article: Collecting data from a Microsoft 365 or a SharePoint Source.

Once the credentials are established, follow these steps to add a Microsoft 365 source to Intella:

  • Source type

    Select "Microsoft 365" source type and click "Next".

  • Connect to Microsoft 365

    Enter the username, password and client ID obtained above. Click Connect to Microsoft 365.

A connection will be established and the credentials will be validated. If credentials validation is successful, basic information about the account such as the tenant name and location will be shown beneath the configuration fields.

Note the Help button at the top of the screen. Clicking it will display the steps required to create the client ID.

Click Next to continue.

  • Select items

    The next screen shows the available accounts. Select the accounts that you wish to retrieve.

Selective indexing of part of the account data is not possible at this moment.

Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition] .

11.3.12. iCloud

The iCloud source type is used for indexing the contents of an iCloud account, such as emails, photos and notes.

Prior to defining an iCloud source, the investigator must obtain the Apple ID and password used by the account owner. When the account has been configured to use two-factor authentication (2FA), iCloud additionally sends a verification token. The verification token is sent only if a valid phone number is set for the Apple ID. Hence, the investigator needs to have access to one of the physical device (an iPhone or an iPad) associated with the account, including the passcode to unlock the device.

Intella supports the retrieval of the following data from an iCloud account:

  • Contacts

  • Emails

  • iCloud Drive

  • Reminders

  • Calendar

  • Event notifications

  • Photos

  • Account settings

  • “Find my phone” data

  • Notes

Follow these steps to add an iCloud source to Intella:

  • Source Type

    Select "iCloud" source type and click "Next".

  • Connect to iCloud

    Enter the Apple ID and password of the account. Click Connect to iCloud.

When this account requires two-factor authentication, Intella will extend the form with an option to choose the verification delivery method: SMS or Idmsa.

Both methods are equally capable of providing access to the account’s data. When the account is linked to an iPhone and/or iPad, the Idmsa method is recommended. When the account is linked to a non-Apple device (e.g. a cellphone or tablet from a different vendor), SMS is the only way to obtain the verification code. Even when using an Apple device, SMS can be selected as the preferred method for delivering the verification code. In that case, the registered device may receive multiple notifications from Apple’s identify management service (IDMSA). Such notifications should then be ignored and the code from the SMS message should be used.

Choose the desired delivery method and click Get Verification Code. A six-digit verification code will be either sent as an SMS or show up as a native iOS notification on the Apple device. The controls for choosing the delivery method will be replaced by a Verification Code field. Enter the received verification code in this field. Click Connect to iCloud.

When the credentials and the verification code are all valid, Intella will list some account info such as the Full Name of the account owner. Click Next to continue.

  • Select items

    In the next step, the available iCloud services for this account are listed. The user can choose whether to retrieve one or more specific services, or whether to retrieve all account data.

The last steps in the definition of a source are almost the same for all types. They are described in the section [last-steps-in-a-source-definition] .

When Intella establishes a connection to iCloud using the account credentials, it will obtain a trust token. This token allows Intella to connect to iCloud at a later point in time without requiring the user to re-enter the credentials and perform any two-factor authentication steps. The trust token has a limited validity period. iCloud sources can be indexed and re-indexed during the validity period of the token. Once the token has expired, the source must be re-created; there is no way to refresh the token of an existing source.

Documents in Keynote, Pages and Numbers format are converted by the iCloud web service to MS Word, MS PowerPoint, and MS Excel format respectively when they are retrieved by Intella. Processing of the documents in their native format may be added in a future release.

The Notes branch currently lists Note items in a flat list; folders are not reported. This may be addressed in a future release.

11.3.13. AWS S3

An AWS S3 source reconstructs the entire folder tree in the selected S3 buckets and downloads current and past revisions of the files.

All revisions except for the last one have their file names decorated with the revision identifier. Furthermore, additional S3-specific metadata is retrieved for both files and folders. These are displayed in the Previewer’s Raw Data tab and are subject to full-text indexing.

Prior to defining the source, the investigator needs to obtain an access key for the account. This process is described in detail in the following Knowledge Base Article: Creating AWS access keys.

Next, follow these steps to add an S3 source to Intella:

  • Source type

    Select "AWS S3" source type and click "Next".

  • Connect to AWS

    Follow steps from the knowledge base article to fill required fields.

    Click Next to continue.

  • Select buckets

    Besides indexing of the entire account, it is also possible to index specific buckets only. The next wizard sheet shows the bucket list of the account. Click the checkboxes of the desired buckets.

    Click Next to continue.

The last steps in the definition of a source type are almost the same for all types. They are described in the section [last-steps-in-a-source-definition].

11.3.14. Last steps in a source definition

The sheets described in the following sections are the same for all source types.

Source Name and Time Zone

In the Source Name and Time Zone sheet you are asked to enter a name for the source. The name will be shown in the list of sources in the Sources panel and functions purely as a label for your reference.

Furthermore, a suspected system base time zone can be entered. This setting indicates the time zone of the system from which the evidence file(s) were obtained. By specifying this time zone, all dates associated with items from this source will be displayed in that time zone, rather than the time zone of the investigator’s system. This often makes it easier to correctly interpret those dates, e.g. determine whether a given timestamp falls inside regular business hours. By default, the local time zone is used for new sources. Time zones supporting Daylight Savings Time (DST) are marked with an asterisk (*).

File type settings

In the File type settings sheet you can specify which item types need to be included in the case. Leave the checkboxes selected for those types or categories that you want to include in the case, and deselect the checkboxes for those you want to be excluded.

Excluded items will be added as stub items that only contain the item’s type and file name (if available).

Items embedded in excluded items will still be processed and included. An example: if there is an email with a PDF attachment and the email item type is not included, the attachment will be processed as usual. The parent email will be added as a stub item.
For technical reasons it is not possible to select the following types: Exchange EDB, Internet Explorer History File (ESE DB) and Windows Search Database. Instead, you can select their parent type: Microsoft ESE database.
MD5 Hash Filters

MD5 hash filters can be used to exclude items that have a specific known MD5 hash from a case. The so-called "De-NISTing" of evidence data is the most well known application of such hash lists: it excludes many files that belong to the operating system or common software applications from your case. But you can also add other types of MD5 hash lists, or create your own.

When selecting one or more of the hash filters for the source, Intella will ignore any items that have an MD5 hash that is in at least one of the filters. After the source has been indexed, such items will not be visible in your case. A future Intella release will add the ability to add "stubs" for such items.

Intella can create an MD5 hash filter from a CSV file, where the MD5 hash is encoded as a hexadecimal value. To do so, click "Create" or drag the CSV file from Windows Explorer to this sheet to open the "Create MD5 hash filter" dialog. After specifying the path to the CSV file Intella will analyze the CSV file and show you the values for the first few lines. If there’s a single column that contains MD5 hash values then that column will be automatically selected. After specifying an appropriate name for the hash filter you can start the filter creation by clicking "Create hash filter".

Intella can process plain CSV files, but also CSV files that are compressed using ZIP or GZIP. Processing the files in compressed form is often preferable as the uncompressed files can be very large (multiple gigabytes).

The Reference Data Set (RDS) that is made available by the National Institute of Standards and Technology (NIST) comes in the form of an ISO file. You will need to extract the NSRLFile.txt.zip file that is stored in this ISO. This NSRLFile.txt.zip file is a ZIP-compressed CSV file that can be processed by Intella. You can find the most recent versions of the RDS at https://www.nist.gov/itl/ssd/software-quality-group/national-software-reference-library-nsrl/nsrl-download/current-rds. For the "Modern RDS" set the "minimal" version is the smallest download that still contains the complete set of hashes.

Any MD5 hash filters that you create will also be available for use in other Intella cases. They are stored in the folder C:\Users\<USERNAME>\AppData\Roaming\Vound\Intella\hash-filters (click Open folder to open this folder in Windows Explorer). The files in this folder can be copied to/from other computers to make them available there as well. Clicking Rescan folder will update the list of available filters.
Deleting MD5 hash filter files will affect the ability to re-index other cases that use the same hash filter.
Items

Intella makes the indexing of certain complex file types optional. You can disable this to improve indexing performance at the cost of fewer results.

  • Select Index mail archives if you want to extract all emails and attachments from mail archives like PST and NSF files. Subsequent processing of documents, archives and other items found in the attachments are still subject to the other options.

  • Select Index chat messages if you want to index chat messages inside Skype SQLite databases, Pidgin accounts and Bloomberg XML dumps. This also controls what happens with Skype, WhatsApp messages etc. in cellphone reports.

  • Select Index archives if you want Intella to index files inside archives such as ZIP and RAR files.

  • Select Index images embedded in emails and documents if you want to extract images embedded in emails, MS Office, OpenOffice, PDF, and XPS documents. This will make these images separately searchable and viewable.

  • Select Index databases to enable the indexing of all tables in SQLite databases.

  • Select Index Windows registry to make all keys and values in a Windows registry file searchable by full-text keyword search. When turned off, a limited amount of registry indexing necessary for populating the Insight tab will still take place. The overhead for this is negligible.

  • Select Index Windows event log to let Intella process the contents of Windows event log files (evtx).

  • Select Index browser history to let Intella process the contents of web browser histories.

  • Select Recovered deleted emails, files and Notes deletion stubs to enable the processing of deleted emails from MS Outlook (PST, OST) and MS Exchange (EDB) files, deleted files and folders from disk images, and deletion stubs in HCL/IBM Notes files (NSF).

  • Select Extract text fragments from unsupported and unrecognized file types to enable heuristic string processing on all items whose type is not recognized by Intella (they are binary blobs) or whose type is not supported apart from type detection (e.g., executable files).

Options

This sheet provides additional options affecting the time needed for indexing.

  • Select Cache original evidence files to copy all evidence files into the case folder. Use this option if you want to create a self-contained case where the evidence files can be opened or exported even when they are not found in their original locations, for instance when the case is moved to another system).

When this option is turned on, additional processing time (especially for compression) and disk space is needed.

This setting has no effect on storing of the items extracted from these evidence files (e.g. the mails, attachments and other embedded items extracted from a PST file), as these are always stored in the case folder after extraction.

  • Select Analyze paragraphs to let Intella determine the paragraph boundaries and to let it build a database registering which paragraph occurs in which item and where. This enables various search and review options at the expense of additional processing time. The required storage space is negligible. For subsequent sources this setting is forced to be same as what has been used for the first source.

  • Select Determine the geographic location of an email sender’s IP address to let Intella estimate the geographic location of an email senders IP address, using a lookup table. To be able to use this feature, a GeoIP2 or GeoLite2 database needs to be present. See the Preferences section for how to properly set up a GeoIP2/GeoLite2 database.

  • Present chat messages as option controls how chat messages will be represented - i.e. what kind of items will be produced: Conversations and Message, Only Conversations, Only Messages

  • Split chat conversations option controls how messages inside chat conversations will be bundled. Possible values are: Per day, Per week, Per month and Per year.

  • Limit number of messages per conversation controls the maximum number of messages that will be contained in each conversation. If this limit is reached, the conversation is split independently of Split chat conversations option.

Last two options will be visible only when Index chat messages option is turned on at the Items step.
Intella creates artificial Conversations items for easier review of conversations based on the two options above. Another reason for such approach is to make it possible to use AND, OR search operators when searching for multiple terms across the messages bundled inside conversation. Note also, that for each message contained in the conversation also separate Chat Message item is created. That makes it possible to annotate or export a particular message in isolation.
  • Select Enable execution of a crawler script to assign a custom script that will be executed for each processed item. Crawler scripts can be used for filtering out irrelevant items, data enrichment and integration with external systems. Supported languages: Java, Groovy and Python. See documentation and samples on GitHub page: https://github.com/vound-software/intella-crawler-scripts

Tasks

This sheet lets the user define post-processing steps that need to take place once all evidence files have been crawled and all indices have been build. See the Tasks section for more details.

Completed Source Definition

Finally you will be presented with a dialog to inform you that you have successfully defined a new source. You may optionally start indexing the source. Indexing is required to be able to search and explore the items in this source. Once you click the Finish button, the indexing process will proceed per the options you have selected.

Because the active indexing process prevents you from interacting with the rest of the program until finished, you may wish to skip this part now (e.g., to define more new sources) and index the sources later by clicking the Re-index button in the Sources view.
At any time except before the step "Completed Source Definition" you can click the Cancel button to return to the Intella interface without having added a new source to the case.

11.4. Indexing

After defining a source, Intella will index it. During indexing it will inspect all items (emails, files etc.) that it can find in the source file(s), enabling Intella to return instantaneous results during your investigation for relevant evidence.

Having anti-virus software active during indexing can lead to certain items not being indexed. This will usually be restricted to the files that are blocked by the anti-virus software, but this cannot be guaranteed. Running anti-virus software may also affect indexing performance.

During indexing, you will see an overlay displaying various types of information:

  • Statistics on indexing speed.

  • Statistics on encountered file types.

  • The amount of data that is being indexed and how much has been indexed already.

  • The number of indexing steps to perform, which current step is being performed and (for some steps) a progress percentage.

You will not be able to interact with the rest of the program while this dialog is shown. Resizing and minimizing the main window remains possible though, as is stopping the index process.

You can stop the index process at any time by clicking the Stop button. Intella will finish processing the current item and then complete its case databases with the information that has been extracted thus far. Afterwards it will let you close the dialog.

It is not recommended to stop indexing and continue it at a later point in time using the Index New Data method. Evidence files may be missed when this is done. The graceful stopping procedure is only designed to make sure that what has been crawled up until the point of stopping can be searched afterwards. A full re-index is still needed if a complete index is desired.

We recommend that you back up the case after indexing it. A case back-up can be used to restore annotations (tags, comments, etc.) if the case happens to become corrupt at a later point in time, e.g. due to a power outage. See the Restoring Annotations section for more information.

Re-indexing a case
There may be circumstances when you want to re-index individual sources or the entire case, e.g. to use extraction features offered by a newer Intella version or fix a broken index.

To rebuild the case index from scratch, choose one or more sources in the Sources view and click on the Re-index button. Intella will remove all indices it has previously created and create new ones.

For this to work, all evidence files must be present at the location they had during the initial indexing.

Any tags, tag columns, flags, and comments will be retained during re-indexing. This includes annotations obtained via an overlay file.

When Content Analysis results, Email Threading results or Near-Duplicate groups are detected in the case, options to clear these information types will be offered in a dialog. Select the checkboxes for the types of analysis results that you want to discard during re-indexing.

Starting with the Intella 2.1 release, the item IDs will remain the same after re-indexing. Also, starting with this release, OCR results will be retained and relinked with the items after indexing completes.

Updating a case
Alternatively, there may be times when you want to update an index, e.g. in the following scenarios:

  • Files and/or folders have been added to folders that have already been indexed.

  • New sources have been defined but were not indexed immediately.

  • The set of mailboxes to index in an EDB source has been extended.

  • You interrupted indexing using the Stop button. See note below for caveat.

In these cases, the "Index new data" operation in the Sources view will scan selected sources for new evidence items. Items that have already been indexed are not changed, also when their original evidence items are no longer available.

In a scenario when container file (e.g. PST file, ZIP file, or disk image) is being processed, and you interrupted indexing using the Stop button it may interrupt processing of that file, leaving it partially indexed. The "Index New Data" operation will not pick that up, it will only index new files, or files that have not been indexed at all yet. The partially indexed files will stay in that state. Re-indexing is required to fully index containers in this situation.

11.5. Automatic item decryption

Intella can automatically decrypt several file formats, if the required credentials are supplied before indexing starts. Therefore, you may want to uncheck the checkbox in the Add Source wizard that starts indexing and use the Re-index option (see above) after these credentials have been entered.

Intella will store decrypted versions of emails and documents in the case. For more details about exporting decrypted data see section "Preferred content type options".

11.5.1. Supported formats

The following file formats can be decrypted by Intella when the credentials are specified before indexing:

  • HCL/IBM Notes NSF files, including encrypted messages in non-encrypted NSF.

  • S/MIME- and PGP-encrypted emails, regardless of the container type they reside in (e.g. EML, MSG, PST, OST, NSF, Mbox, DBX).

  • PDF documents.

  • Old format MS Word documents (.doc), MS Excel spreadsheets (.xls) and MS PowerPoint presentations (.ppt).

  • MS Office 2007 formats (OpenXML): .docx, .xlsx, .pptx, …​

  • ZIP, RAR and 7-Zip archives.

  • Partial support for ZipX.

  • BitLocker volumes.

  • APFS file systems.

Furthermore, password-protected PST files can be automatically decrypted without specifying any passwords.

11.5.2. Supplying access credentials

To let Intella automatically decrypt the encrypted items that it encounters, their keys (passwords, certificates, etc.) need to be added to the Key Store first. Click File > Key Store and follow the instructions below. Afterwards you can (re)index your data and let the items be decrypted automatically.

All credentials that you enter will be tried on all encrypted files to which they can apply. It is therefore not necessary to specify e.g. which password applies to which file or file type.

After indexing you can see which items were successfully decrypted by using the "Decrypted" category in the Features facet or by using the "Decrypted" column in the Details table. Note: due to technical reasons, decrypted NSF files will not be marked as such.

Password-protected files
Passwords are the simplest type of key. They are used for decrypting PDF and MS Office documents and archives.

You can either add passwords one by one, or load them in batch from a text file: specify a password per line and use UTF-8 encoding for the file.

HCL/IBM Notes NSF files
To decrypt HCL/IBM Notes NSF files, so-called ID files need to be added to the key store. Go to the "HCL/IBM Notes ID Files" tab and click "Add…​". Enter the location of an ID file and the password associated with the file. Click OK to add it to the store. Intella will validate the ID file to make sure you entered the password correct. Repeat this for all ID files.

Intella will also try to decrypt encrypted messages in non-encrypted NSF files using the provided ID files.

S/MIME-encrypted emails
To decrypt emails with S/MIME encryption, one or more X.509 certificates and private keys need to be added. Go to the "X.509 Certificates" tab and click Import, then select a PKCS12 archive file (*.p12 or *.pfx file) that contains the keys. Intella will analyze the key file and import all found certificates and keys.

Usually you can export the certificates and keys from a mail client in this format. Do not forget to include private keys as they are critical for decrypting the emails.

PGP-encrypted emails
To index PGP-encrypted emails you will need to import the PGP private keys. Go to the "PGP Keys" tab and click Import. Intella can import ASCII armored PGP private keys (*.asc files), but it is also possible to import key in binary format.

An ASCII armored PGP private key usually starts with the following text:

-----BEGIN PGP PRIVATE KEY BLOCK-----

Importing multiple .p12 files
At the moment it is not possible to enter multiple .p12 files in a single action, they need to be entered one by one. We have put this feature request on our roadmap for future development.

Please note that .p12 files can contain multiple certificates. Therefore, if your environment is able to export multiple certificates into a single .p12 file, or you can find a third party tool that merges them, you can effectively import multiple certificates at once.

Furthermore, note that you can copy the keystore files to another case. That way you can reuse the entered credentials if they apply to other cases/evidence sets as well.

Encrypted volumes in disk images
To decrypt BitLocker and APFS volumes in disk images, a correct password, recovery key, or recovery file needs to be added. Passwords can be added via the "Passwords" tab. "BitLocker Recovery Keys" and "BitLocker Recovery File" tabs should be used to add BitLocker-specific credentials.

Note that BitLocker images protected with other methods, such as a Smart Card or TPM, are not supported.

11.6. Custom columns

There may be a need to extract specific metadata fields and put them into separate columns. This is what the custom columns functionality can be used for. An example is extracting the "X-Mailer" header and putting it into a dedicated column "Email Client" column.

Custom columns are populated during indexing. Therefore, they need to be set up prior to indexing. If the custom column is added after indexing, re-indexing will be required to populate the custom columns.

To add Custom columns, select Custom Columns option from the File menu.

Click on the New button to add a new custom column. Specify the name, description, and type of the column. The following column types are supported:

  • Text — The column can hold arbitrary text of any length.

  • Integer — The column can hold a whole number in range between -2,147,483,648 and 2,147,483,647.

  • Long Integer — The column can hold a whole number in range between -9,223,372,036,854,775,808 and 9,223,372,036,854,775,807.

  • Floating Point — The column can hold a 64-bit floating point number.

  • Boolean — The column can hold either True or False.

  • Date — The column can hold a date-time value.

The Extract Data section below shows which metadata fields will be used to populate the column. Click the plus sign button to add a new rule:

  • The Type option allows to restrict the rule to a specific file type or type category. By default, the type is set to "Any" which means that the rule applies to any item.

  • The From option defines where the metadata field comes from: Raw Data or Headers.

  • The Field option defines the name of the metadata field that should be extracted.

  • The Date Format option allows to specify a date format if this is a date column. The Language option can be used to tell which language should be used when parsing certain date elements, like day of the week or month names.

  • The Case Sensitive option can be used to specify whether the Field name is case sensitive.

It is possible to add more than one rule to a custom column. In this case the option below the table defines the way in which the extracted values should be merged. By default, the first non-empty value will be used. For text columns the values can also be joined together.

Once added, the new custom columns will appear in the column chooser of the Details table, in the Custom Columns section. The date columns will also appear in the Date facet panel, Histogram and Timeline views and can be used when defining Primary Date rules. Custom columns will also be shown in the Properties tab of the Previewer.

Examples of using custom columns:

  • Extract the email client information into an "Email Client" column:

  • Name: Email Client.

  • Type: Text.

  • Extract data from:

    • Type: Email Message

    • From: Headers

    • Field: X-Mailer

  • Extract the camera model information from JPEG photos into a "Camera Model" column:

  • Name: Camera Model.

  • Type: Text.

  • Extract data from:

    • Type: JPEG Image

    • From: Raw Data

    • Field: Model

11.7. Post-processing

Once indexing completes, the case owner can opt to refine the indexing results in several ways. These steps are kept separate from indexing as they typically contribute considerably to the processing time and disk space usage and, depending on the case at hand, may not be needed.

11.7.1. Tasks

Intella allows for the definition of "tasks". These are essentially compound processing steps such as searching for all items that match a certain keyword or keyword list and tag or export the results. These tasks can be defined and selected during source creation, which will run these tasks right after indexing. The tasks editor can also be reached by selecting Tasks from the File menu, which allows for defining and running the tasks at any point in time after index creation.

Each task consists of conditions, post-conditions, and actions. A task must have at least one condition and one action.

A condition (Step 1 in the task dialog) defines a search query that select items from the case. Currently the following conditions can be defined:

  • A keyword search.

  • A keyword list search.

  • An MD5 list search.

  • An arbitrary Saved Search, which can combine all Intella’s search facets.

  • A tag, possibly assigned by one of the tasks executed earlier.

  • A date range search on all date fields.

  • A build-in query that selects all items in the case.

  • An OCR Candidates search. It facilitates the selecting of various categories of images and documents that are usually subjected to OCR.

image

A task may combine any number of conditions. The match option controls if the items should match all specified criteria or at least one of them, i.e. a Boolean AND or OR of the specified conditions.

An optional list of post-conditions (Step 2) specify how to transform the item set retrieved in the previous step. Possible post-condition steps are:

  • Deduplicate results

  • Identify parents of the retrieved items

  • Identify children of the retrieved items

  • Suppress irrelevant items

It is possible to define multiple post-conditions for a single task. The first post-condition is applied on the set of items resulting from the conditions in Step 1. Subsequent post-conditions are applied on the outcome of the preceding post-condition.

Finally, task actions (Step 3) define the operations that will be applied to the items resulting from the previous steps. The following actions can be defined:

  • Tag all found items with one or more tags. The tag(s) can optionally be inherited by items in the same family hierarchy and/or by duplicates of the found items.

  • Set custodian attributes.

  • Flag all found items.

  • Add a comment to all found items.

  • Export all found items using an export template.

  • Export the metadata of all found items to a CSV file. Click the Configure button to set the CSV file name and path and to select the metadata fields that are to be included.

  • Start an OCR process on the found items using the embedded ABBYY FineReader engine, an external OCR tool, or by connecting to an ABBYY Recognition Server.

  • Start a Content Analysis process on the found items for the selected entity types.

  • Start the email threading process on the found items.

  • Expand existing tags to similar items, identified by the Smart Search algorithm. A detailed explanation of the options in this task can be found in the "Vound W4 Case" section.

  • Generate custom IDs for the found items.

  • Generate duplicate custodians and locations.

  • Generate thumbnails for image items

Every task may define multiple actions that will be applied sequentially to the determined item set.

Tasks can be exported to a file so that they can be reused in other cases. These files are self-contained, i.e. when the task involves MD5 hash lists or keyword lists, these lists are embedded in the task file.

Tasks are executed in the order they have in the task list. This makes it possible to "pipeline" tasks, e.g. use one task to assign specific tags to a subset of the items and use a subsequent task that is based on those tags. The order can be changed by selecting a task and using the "Move Up" and "Move Down" buttons.

11.7.2. Custodians

The Custodian attribute can be assigned to items after indexing. This can be used to represent the custodian of the evidence items. To enable automated assigning of multiple custodians in a folder source (10.2.1), the root folder should organize the evidence in subfolders, one subfolder for every custodian.

If the evidence folder is structured in this way, the "Indexing Tasks" sheet in the Source Wizard will contain a "Configure custodians…​" button that opens the below dialog:

image

By default, the custodian names are set to equal the subfolder names. It is possible to alter the used custodian names by double-clicking the values in the table. This Custodian value will be assigned to all items obtained from the evidence files within the respective subfolder.

For other types of sources, the "Indexing Tasks" tab contains a text field for setting a single custodian name.

Besides the above method, the Custodian attributes can also be set or changed using the "Set Custodian" indexing task with an arbitrary condition, or edited manually in the Details…​ right-click menu.

11.7.3. OCR

OCRing, or applying Optical Character Recognition techniques, is a common way to make the text inside bitmap images responsive to keyword searches. Intella’s OCR support is documented in the next chapter.

11.7.4. Thumbnail generation

Cases that rely heavily on viewing collections of images in the Thumbnail view will benefit from pre-creating the thumbnail images in advance.

Especially when dealing with digital camera images that each are multiple megabytes in size, the time needed to generate the thumbnail image can make the Thumbnails view appear sluggish. When the thumbnails have been pre-generated, the time needed to populate the view will be a lot faster and it will be constant regarding the number of visible images, i.e. the file size of the original image is no longer a factor.

To pre-generate the thumbnail images, select the "Generate Thumbnails…​" option from the File menu. The thumbnail generation process will start immediately and show its progress in the main window.

The thumbnail generation process can be cancelled at any point. The thumbnail images that have been generated will be kept. When the user starts the process once more at a later point in time, it will reuse the existing thumbnails and only create the missing thumbnails.

11.7.5. Importing an overlay file

An overlay file is a file that contains additional information about the current items in a case. By importing the overlay file, the metadata of these items can be extended.

Intella currently only supports the importing of tags, tag columns, comments, and metadata columns (both regular and custom). Importing overlay images, texts, and natives may be added in a future release.

The following file formats are supported for overlay files:

  • Concordance/Relativity load file (.DAT)

  • Comma Separated Values file (.CSV)

To import an overlay file, select File > Import Load File…​ in the main menu. Next, set the Import operation to Overlay and specify the location of the file. You can optionally use a previously saved template.

On the "Configure Delimiters" page you can set the file encoding, delimiter settings and date formats. Please see the "Adding a load file source" section for a description of these options.

On the "Map Fields" page you need to specify the identifier field and type. This is how Intella will match items in the overlay file with the existing items in the case. There are four options for matching items:

  • By Document ID, also known as DocID. This is the most common way to import new tags and comments into previously imported load file.

  • The Item ID is the internal item identifier used by Intella. This is the simplest way to process your data using an external tool and then import the result back into Intella.

  • By MD5 Hash. This is the most flexible way of matching items. Using the MD5 hash it is possible to transfer tags from one case to another. Note that the imported tags will be applied to all copies.

  • The Item URI is an internal item identifier that is not changed after re-indexing the case, but it may be changed when re-indexed with a newer Intella version due to changes in the crawling software. This method can be used to transfer tags when other options are not suitable, e.g. when migrating tags from a case backup to a live case that has been re-indexed in the meantime.

The "Also overlay metadata shared with duplicates" option is used to control whether the imported metadata will be applied to all duplicates as well (see the limitations below for this setting).

Current limitations of the overlay functionality are:

  • Overlaying images, texts and natives is not supported.

  • Location and MD5 columns cannot be overlaid.

  • It is not always possible to overlay metadata for regular items that do not originate from a load file. For example, if an item from a non-load file source has duplicates, then the overlaid metadata should be applied to all duplicates as well using the "Also overlay metadata shared with duplicates" option. Otherwise, the overlaid metadata might not be applied. There is no such limitation when overlaying data to items from a load file source. In this case, each record in the overlay is unique and the "Also overlay metadata shared with duplicates" option should be unchecked.

  • Metadata imported into regular and custom columns will be lost after re-indexing the case.

Please see the "Adding a load file source" section for a description of the remaining options on this page.

11.7.6. Content analysis

The Content Analysis facet allows you to search items based on specific types of entities that have been found in the textual content of these items. Three of the categories in this facet are populated automatically during indexing and are available immediately afterwards. These are:

  • Credit card numbers

  • Social security numbers (SSNs)

  • Phone numbers

The other categories are more computationally expensive to calculate and therefore require an explicitly triggered post-processing step. These categories are:

  • Person names

  • Organizations (e.g. company names)

  • Locations (e.g. city and country names)

  • Monetary amounts

  • Time (words and phrases related to the hours, minutes, weekdays, dates, etc.)

  • Sentiment analysis (sub-categorized as Negative, Neutral and Positive based on the text in documents or messages)

  • Custom regular expressions (for searching e.g. bank account numbers, patent numbers and other types of codes that can be formally described as a regular expression)

To start the content analysis procedure, select one or more items in the Details view and select "Content Analysis…​" in the context menu. This will open a dialog like the one below:

image

Select the desired entity categories in the list by clicking the checkboxes. Click the "Select text fields" button to specify which item fields should be analysed, e.g. the header, title, and summary fields. By default, only the document text is analysed.

Check the "Replace existing facet values" option selected if you want to clear the results of the previous analysis or keep it unselected to add new results to the existing content of the selected categories.

Select the "Ignore excluded paragraphs" option if you don’t want the content of excluded paragraphs to be processed by the content analyzer.

Click the Run button to start the analysis. The results of the analysis will appear in the Content Analysis facet.

The content analysis procedure will open a separate window showing the progress of the procedure. The procedure can be stopped at any time by clicking on the Cancel button. In this case, the categories will contain information from the items processed up to the moment of cancellation. You can continue the processing later by repeating the steps above on the same set of items and choose not to clear the existing results (unselect the "Replace existing facet values" option).

The items that have been analysed can be found by using the "Content Analysed" category in the Features facet.

Content analysis also supports doing a test run, without storing the results in the facet database. To use this mode, click the Preview button. After the analysis has completed, the Content Analysis Preview window will appear where the extracted entities and associated items can be reviewed. This mode can be especially useful for testing and refinement of the custom regular expression-based categories (see below).

There are some important caveats and disclaimers concerning Content Analysis:

  • Content analysis is a heuristic procedure based on typical patterns and correlations that occur in natural language texts. Therefore, the quality of the output may vary within a broad probability range.

  • Content analysis works best on English texts. The quality of the output may be poor on texts in other languages.

  • Content analysis works best on texts containing properly formulated natural language sentences. Unstructured texts (e.g. spreadsheets) usually lead to poor quality of the output.

  • Content analysis is both CPU- and memory-intensive. For adequate performance, please make sure that your computer meets the system requirements and that no other processes are taxing your system at the same time. Use of the 64-bit version of Intella is highly recommended, especially for analyzing large quantities of items. In our experiments, the amount of time needed for processing an entire case was roughly similar to the amount of time it took to index the case.

  • Sentiment analysis generates a sentiment score for a text and subsequently categorizes items into Positive, Negative or Neutral categories. It is based on a dictionary analysis text method that focuses on extracted emotional words. Currently, Sentiment Analysis is an experimental function that only supports texts written in English. The best analysis results are usually obtained for e-mails, chat messages and short documents. This may be extended to other languages and a wider array of document types in a future release.

When the case has been created with Intella version 2.1 or earlier, on the first Content Analysis run you will be asked to upgrade the Content Analysis data store. This is needed to enable Content Analysis data columns in the Details table.

When opting for an upgrade, the new Content Analysis results will not be available in older Intella versions. If this is not what you want, you can click the "Continue…​" button to skip the upgrade. In this case, backward compatibility with older versions will be preserved, but Content Analysis results cannot be shown as columns in the Details table.

Along with the predefined categories for built-in entity types such as Person names, Organizations, etc., it is possible to define custom Content Analysis categories populated by scanning the text of selected items for specific text patterns. The text patterns are defined using IEEE POSIX regular expressions syntax. See http://pubs.opengroup.org/onlinepubs/9699919799/basedefs/V1_chap09.html for documentation on this syntax. This provides an effective and versatile mechanism for extracting user-defined entities, such as national passport numbers, bank account information etc.

To create a new custom category, click the New button in the Content Analysis window. In the dialog that opens, enter the name of the new category and a regular expression that defines the pattern to search for. You also can use the following search options:

  • Case sensitive — Lower- vs. uppercase should be considered by the search operation.

  • Dot (.) can match multiple lines — Enables multiline matches of expressions using the dot (.) symbol.

Click the Regex Assistant button to open the Regular Expression Assistant window. This is where you can test your expression on a custom text fragment, choose one from the examples library and get quick help on the regular expression syntax. The "Import keywords from file" button will add regular expression that will match all keywords from loaded file. The "Load from file…​" button will load text from file for testing of regular expression. The "Load item content…​" button will load text from case item for testing of regular expression.

Once the new custom category has been created, it can be selected in the Content Analysis dialog and populated by analysis of the selected item set, just like the other (predefined) categories.

To edit a custom category, select it in the list and click the Edit button. In the dialog that appears, edit the name and/or the regular expression of the category.

To delete a custom category, select it in the list and click the Delete button.

An alternative way to create custom categories is to click the New button in the Content Analysis facet panel. The custom categories can also be edited and deleted through the context menu of that facet.

11.7.7. Image analysis

To start the image analysis procedure, select one or more items in the Details view and select Process > Image Analysis in the context menu. This will open a dialog like the one below:

Image analysis dialog

Select the desired categories of entities in the list by clicking the checkboxes.

Check the "Replace existing facet values" option selected if you want to clear the results of the previous analysis or keep it unselected to add new results to the existing content of the selected categories.

Click the Run button to start the analysis. The results of the analysis will appear in the Image Analysis facet.

The content analysis procedure will open a separate window showing the progress of the procedure. The procedure can be stopped at any time by clicking on the Cancel button. In this case, the categories will contain information from the items processed up to the moment of cancellation. You can continue the processing later by repeating the steps above on the same set of items and choose not to clear the existing results (unselect the "Replace existing facet values" option).

The items that have been analysed can be found by using the “Images Analysed” category in the Features facet.

There are some important caveats and disclaimers concerning Image Analysis:

  • Image analysis is both CPU- and memory-intensive. For adequate performance, please make sure that your computer meets the system requirements and that no other processes are taxing your system at the same time. In our experiments the amount of time needed for processing an entire case was roughly similar to the amount of time it took to index the case.

  • Skin tone analysis is based on a colorimetry method that is proven to be good for detection of pictures exposing natural variations of human skin colors under various light conditions. However, the output may include photos of other objects colored similarly, for example sandstone walls, some ceramic sculptures, etc. On the other hand, the algorithm may miss some skin photos if the amount of the skin tone pixels is too small or if the skin has an unnatural tint, for example due to wrong white balance or exposure settings of the camera.

  • Image categorization and object detection are based on machine learning models that were trained on sample images. Depending on the quality and variety of the images in the source data, and any biases that the training data may have had, images may incorrectly be put into certain categories, or left out of the categories that they do belong in. Please ensure to review your results before relying on them.

11.7.8. Email Threading

The linear review of emails is often a time-consuming and expensive task to perform. One factor is that emails may quote the text of previous emails in the thread, resulting in a lot of redundant text. Take for example these three emails:

image

Marked in red is the redundant text. The text of the first two emails is quoted in full in the last email. When a reviewer reads the last email, he or she has read everything there is to read in this thread. The reality is often more complex, e.g. because people respond to the same root email, remove part of the quoted text, forward it to new recipients, or even alter the quoted text to cover up certain facts. Therefore, it is not always as simple as reading the last email in the thread.

Intella helps with this type of review through the process of email threading. First, it identities the emails that belong to the same thread. Within each thread, it links the replies and forwards to their parent emails, constructing a graph of how the conversation unfolded. All duplicates of a mail will be represented by the same node in this graph. Next, it compares the emails within the thread and determines the set of "inclusive" and "non-inclusive" emails. By default, a mail will be marked as inclusive. When Intella detects that one of the follow-ups of a mail (a reply or a forward) contains all its text and attachments, it will be marked as non-inclusive, as reading the latter email implies having read the first as well. Reading all the inclusive emails and their attachments in a thread implies having read everything there is to read in the thread. This can greatly reduce the time needed to review a large collection of emails.

Besides separating inclusive from non-inclusive emails, email threading enables several other functionalities:

  • Sort the emails in a thread in the Details view, to read the entire email thread sequentially.

  • Group the emails in the Details view by thread.

  • Visualize a specific email thread in the Email Thread tab of the Previewer. This shows how the previewed email relates to the other emails in the thread, e.g. what email did it reply to, what replies did it trigger, are there different branches in the thread, how was its content forwarded, etc.

  • Tag all emails in a thread at once.

  • Identify missing emails in a thread. These are emails that are referred to in the email headers or in the metadata embedded in an email body, but that cannot be found in the current evidence data. This may indicate missing evidence data that an investigator may still be able to acquire, e.g. from other custodians or from a backup. If additional evidence becomes available later, it can be added to the case. The email threading processing will then attempt to use the new emails to resolve the missing emails.

  • List the normalized subjects of the email threads in the Email Thread facet.

Each email item that was processed by the Email Threading analysis is assigned the following properties:

  • Threaded - Indicates whether the item has been subjected to email thread analysis.

  • Inclusive - Indicates whether the email is inclusive.

  • Non-Inclusive - Indicates whether the email is non-inclusive.

  • Missing Email Referent - Indicates that the threading process has detected that the email item is a reply to another email or a forwarded email, but the email that was replied to or forwarded is not available in the case.

  • Email Thread ID - The unique identifier of the thread that the email has been placed in.

  • Email Thread Name - The normalized subject of the thread that the email has been placed in.

  • Email Thread Node Count - The number of nodes in the thread that the email has been placed in.

Furthermore, the algorithm establishes for each follow-up email if it is a Reply, Reply All, or Forward. This status is derived from the sender and receiver information, rather than from e.g. the Subject line. A loose but conceptually practical definition is:

  • If the set of participants of the response email is the same as the email that it is responding to (the previous email in the thread), it is a Reply All, unless this is a conversation between only two people, in which case it is a Reply.

  • If the response email is going to one or more people, and none of them was involved in the original email, it is a Forward.

  • In all other cases, it is a Reply.

As email threading is a computationally expensive algorithm, it requires an explicitly triggered post-processing step. To start the Email Threading procedure, select one or more items in the Details view and select Process > Email Threading…​ in the right-click menu. This will open the dialog shown below:

image

Select "Discard existing email threading data" if you want to clear the Email Thread facet and all the data generated as part of previous runs of Email Threading procedure.

Select "Analyze headers embedded in email body" if you want the algorithm to take the headers embedded in the email body into account. Such headers are typically placed above the quoted text, referencing the original author and time of the quoted text and sometimes other metadata. This can be used to link emails together when the SMTP or mail container-specific metadata is missing or incomplete. This option may produce better results but is computationally expensive. When speed is not of the essence, we recommend turning this feature on.

Click the "Run" button to start the email threading process.

Once the process is done, the Email Thread facet will be populated and the email items that were part of the threading analysis will be augmented with the threading-related information.

Besides processing the selected items, Intella will automatically process all duplicate items and parent items as well.

The "Analyze paragraphs" indexing option is a prerequisite for determining the inclusiveness of emails. If this option was not used during indexing, all emails will be marked as Inclusive.
By default, the maximum depth of an email thread that Intella can process is limited to 500. Email threads which depth exceeds this value will be split into smaller ones. This is to avoid too long processing times. It can be changed via "EmailThreadMaxDepth" property in the case.prefs file. Note that setting the property to a too high value (especially more than 1,000) might result into processing errors.
Email threading analysis is a heuristic process, particularly when the analysis of headers embedded in email bodies is used. In some cases, this may lead to incorrect results, such as grouping emails together that do not belong to the same thread. We advise users to use this functionality with care and always validate results before relying on them.

11.7.9. Near-duplicates Analysis

A technique to reduce the reviewing time is Near-duplicates Analysis. It splits a selected set of items into groups based on the similarity of their text content. Every group is centered around a "master item" which is the most common near-duplicate for other items in the group (usually, an item with the largest text size). Other items are included in the group if they are determined to have an appropriately high similarity score to the master item. The similarity score is based on an amount of co-occurrent text fragments and is a number between 0.0 and 1.0. The master item and its exact duplicates are assigned a score of 1.0. The rest of the group items have scores between 1.0 and a threshold value specified by the user before the analysis.

To start the Near-duplicates Analysis process, select multiple items in the Details view and select "Process > Near-Duplicate Detection" in the right-click menu. In the dialog window, move the "Similarity threshold" slider to set the desired minimum similarity score for items to be included in near-duplicate groups. Select the "Ignore excluded paragraph" option if you don’t want the content of excluded paragraphs to be considered by the similarity calculation algorithm.

The dialog window allows the user to select a text analysis method. We recommend choosing the "Word-based" option for documents written in languages in which the representation of meanings is contained in words. The “Character-based” option is intended for languages in which the semantic representation is represented by morphemes (Chinese, Japanese, Vietnamese, Korean). Typically, these are languages in which the use of white space characters is optional. Choosing the appropriate algorithm for a data set will improve the quality of the analysis results.

desktop nd analysis

After the analysis is complete, a window with basic information about the results (such as number of groups or average similarity) will appear. This data can be exported to a .pdf or .docx report file by clicking on the "Generate report" button.

desktop ndd report

Upon completion, near-duplicate groups are available for search in the "Near-duplicates" facet (see the "Near-duplicates" section for details). Additionally, "Near-Duplicate Group", "Near-Duplicate Master Item" and "Near-Duplicate Score" columns can be made visible in the Details table to show the group names, master item IDs and similarity scores of items included in near-duplicate groups.

To query for Near-Duplicates of specific items that are subject to Near-duplicates Analysis, select the item in the Details view , right-click, and choose "Show Near-Duplicates". This option will be enabled only when the selected item has at least one Near-Duplicate.

The total set of analyzed items and items included in near-duplicate groups can be retrieved via the "Analyzed for Near-Duplicates" and "Has Near-Duplicates" nodes in the Features facet.

11.7.10. Custom IDs

The "Generate Custom IDs" task lets one assign each item a unique custom ID that takes the item families into account. Such IDs are often used in load file exports. It can also help with identifying an item’s position or role in its family.

Items are processed in hierarchical order, starting from the roots and exploring as far as possible along each branch before backtracking (i.e. a depth-first search). Sibling items, i.e. items that are on the same level in the hierarchy, are processed in the order defined by the Sort Order setting. If the selected items do not contain complete families, this task will add the remaining items automatically.

Click the Configure button on the task action panel to configure the numbering settings:

  • Prefix defines the prefix for each custom ID.

  • Start at defines the starting number. If the Auto option is selected, Intella will use the next available number for this prefix or 1 if the prefix has not been used before. The Manual option allows to set a custom starting number.

  • Number of digits defines the number of leading zeroes that will be used in the number.

  • Child numbering defines the way child documents are numbered relative to their parents:

    • Add suffix - The child document ID is derived from its direct parent’s ID by adding the Child Suffix Delimiter (see below) and a child number, starting at 1. For example, if the parent item is ABC123, then its children will be numbered as ABC123.001, ABC123.002 and so on.

    • Use sequential number after parent - The child document ID will use the next consecutive number after its parent. For example, if the parent item is ABC123, then its children will be numbered as ABC124, ABC125 and so on.

  • Child Suffix Delimiter defines the delimiter that is used to separate the parent and child IDs when the Add Suffix option is selected.

  • Sort Order defines the attribute by which items located on the same hierarchy level will be sorted.

  • Family defines how the Custom Family ID column is constructed:

    • Use Parent ID - The Custom Family ID is the Custom ID of the top-level parent in this family.

    • Use Family Range (Start-End) - The Custom Family ID is the Custom IDs of the first and last items in this family, separated by a hyphen.

  • When the Overwrite Existing option is selected, Intella will overwrite any existing Custom IDs and Custom Family IDs.

Generated Custom IDs can be used in load file exports and can be imported from a load file. Custom IDs do not change when the case is re-indexed, provided that the case is re-indexed using the same Intella version.

11.7.11. Duplicate custodians and locations

The "Generate duplicate custodians and locations" task generates values for the following duplicates-related columns:

  • All Custodians - The custodians of all duplicate items in the case (including this item).

  • All Locations - The locations of all duplicate items in the case (including this item).

  • Duplicate Custodians - The custodians of all duplicate items in the case (excluding this item).

  • Duplicate Locations - The locations of all duplicate items in the case (excluding this item).

11.8. Editing sources

To see the configuration of a source, go to the Sources view and click on a source name in the list. The right panel of the Sources view will be updated to show details of the selected source.

The source details are organized in tabs:

  • General: Basic source parameters, such as evidence paths and suspect base time zone.

  • Index options: Various indexing options, defined during source creation.

  • Hash filtering: Any MD5 Hash filters specified for this source.

  • File types: Types of items included in the case.

  • Indexing tasks: Indexing tasks assigned to this source.

  • Errors: Any source-level errors happened during the last indexing (also available in the top-level Errors view). This tab does not appear if the source is indexed without errors.

The source parameters on the tabs (except for "Errors") are editable. When you change the parameters, the Apply button will become enabled. Changes will only be applied when you click Apply. If you select a different source or leave the Sources view without first clicking Apply, a dialog will appear to prompt you to apply the changes, discard the changes, or cancel the operation.

The Discard button reverts all unapplied changes in the source parameters.

Changes to many configuration options do not take effect until the source is reindexed. To indicate that the source configuration is changed, the source will get a "Reindexing required" status in the list.

11.9. Exceptions report

An indexing exceptions report can be produced by choosing one or more sources in the Sources view and clicking the Exceptions Report button. This produces a XLSX or CSV file that lists all items that had issues during indexing. This can range from minor issues such as date parsing problems to file corruptions that affect the entire item and all nested items.

For every item, the following information is listed:

  • The item ID. This can be used to quickly locate the item using View > Preview Item…​ The Previewer will also show a warning icon when displaying such an exception item.

  • The MD5 hash. This can be used to locate duplicates of the item within the case or in other cases.

  • The source to which this item belongs.

  • The file name, file size and detected file type of the problematic item.

  • The name of the source in which the item was found.

  • The location of the problematic item. This includes both the path to the containing evidence file (e.g. a PST file) as well as the path within that file (e.g. the mail folder and parent email, when the exception occurred on an attachment).

  • Information on the parent email if there is any: its item ID, the sender, sent date and subject.

  • A warning scope, warning code and warning description. The scope and code are the most useful for end users and are documented below. The description provides a low-level error message that is also contained in the log file and can be used for error diagnosis by Vound’s technical support team.

The warning scope indicates the type of data that is affected by the exception. Possible values are:

  • Item — the entire item is affected.

  • Text — the extracted text is affected.

  • Metadata — the extracted metadata is affected.

  • Embedded — embedded items such as attachments and archive entries are affected. An example is a document that internally references an embedded image but the image is not present in the file, resulting in an error when processing the embedded items of the document. In that case the document gets an error with "Embedded items" as the Warning Scope.

The warning code indicates the nature of the issue. Possible values are:

  • Unprocessable data — The data cannot be processed because it is corrupt, malformed, or not understood by the processor. Retrying will most likely result in the same result.

  • I/O errors — The processing failed due to I/O errors. The processing might succeed in a repeated processing attempt. There can be a lot of reasons for such errors, e.g. a drive that fails to respond, or permissions blocking Intella from accessing it. The indexing logs will have the full error. The difference with the other errors is that the reason is typically external to Intella, which is why retrying indexing may sometimes resolve the issue.

  • Decryption failed — The data cannot be processed because it is encrypted and a matching decryption key is not available. The processing might succeed in a repeated processing attempt when the required decryption key is supplied.

  • Timeout — The processing took too long and was aborted. See more details on how to configure crawler timeout in "Memory, crawler count and timeout settings" chapter.

  • Out of memory — The processing failed due to a lack of memory.

  • Processing error — The processing failed due to a problem in the processor. The description should contain the stack trace.

  • Truncated text — The item text was not fully processed due to one of the following reasons:

    • The item text was larger than the imposed item text limit and any additional text was ignored. See the Sources section for a description of this limit and how to alter it.

    • Binary content was removed from the item text. Intella will try to detect and remove so-called binary content from all processed text to reduce memory usage when processing corrupt or recovered files. It includes any control and non-printable characters that are not normally present in regular texts. Items with binary content removed will have an error description: "Binary content detected".

    • The item text could not be extracted because the format is not fully supported yet.

  • Crawler crash — The processing failed due to a crawler crash. This is a more severe error compared to the Processing Error type. When it occurs, Intella will also reject all items that are related to crashed item (e.g. PST file and all of the emails that it contains). More details about why the crawler crashed can usually be found in a hs_err_pid_XYZ.log file which is located in the case logs folder (one file per crash). Crawler crashes will not affect other items and the case integrity.

When an item has multiple exceptions, it will occupy several rows in the table.

During indexing Intella tries to prevent processing of duplicate items (detected by their MD5 hash), as their contents will also be the same. Therefore, an item may occur only once in the exceptions report, even though there can be many copies throughout the case.

All items that produced an exception during indexing can easily be found using the Exception Items category in the Features facet, with subcategories for the warning codes.

The XLSX variant of the exception report additionally holds the following information:

  • Number of exceptions per source, subdivided by the warning codes.

  • Overall statistics for the warning codes.

  • Source-level errors, e.g. broken PST files.

Besides holding more information, the XLSX variant is also better able to handle non-ASCII characters.

11.10. Removing sources

To remove one or more sources, choose them in the sources list and click the Remove button beneath the list.

Source removal is an expensive operation. When multiple sources are to be removed, it is recommended to remove them all at the same time, as the total time required will be less than when the sources are removed one-by-one.

Sources can be added again after removal, by following the normal "Add Source" procedure.

Removing a source will remove:

  • The data, metadata, OCR results and (load file) images associated with the removed items, except for those that are still associated with item duplicates originating from other sources.

  • Any redactions and comments associated with the removed items.

  • All references to the removed items in tags, flags, batches, export sets, custodian sets and near-duplicate groups.

What remains after source removal are:

  • Refences to the source and the evidence items contained in the logs files.

  • References to the numeric item IDs in the event log.

  • MD5 hashes of item locations.

  • Metadata extracted by the email threading procedure, such as Message-ID headers and Conversation Index properties.

These artifacts are typically not visible to the end user but could be obtained by reverse engineering of the case files. Please consider this when handing over a case with removed sources to an opposing party.

11.11. Restoring annotations

If may occur that a case will no longer open. Possible causes are unexpected power failures or the incorrect handling of external or network drives, as this can result in the files in the case folder getting damaged to the point where normal handling of the case becomes impossible.

When this has happened to your case, it may still be possible to extract the annotations (tags, flags and comments) from the broken case and import them into a backup copy of the case, so that the results of your work on the case are also restored.

To restore the annotations, create a copy of the backed-up case (ideally the back-up has been made right after indexing), open the copy and select File > Restore Annotations…​ in the main menu. Next, select the "audits" subfolder of the case that contains file named "events.log". The annotations from the broken case will then be imported into the current case copy.

It is important to consider the following:

  • The original case must be indexed with Intella version 1.6.3 or later.

  • Annotations can be restored only from a copy of the same base case and only if both case copies have not been re-indexed.

  • Any annotations that exist in the importing case will be removed before importing. The copy of the current "events.log" file is stored in the audits folder as events.log.old. You can use this copy to restore the state in case this removal was not intended.

12. Optical Character Recognition (OCR)

Cases often contain images with human-readable text in them, e.g. web page screenshots. These images can be embedded in documents, e.g. a scanned or faxed document is packaged as a PDF containing TIFF images, or a chart is embedded as a picture in a Word document.

The techniques for identifying the text in such images (embedded or not) is called Optical Character Recognition, commonly abbreviated to OCR. Application of such OCR techniques can make the textual contents of these images available for keyword search.

Some modern scanners already apply OCR techniques during scanning and add the extracted text to the PDF. If this is the case, Intella will pick up the text automatically during indexing. Often this machine-accessible text is missing though, or it contains too many recognition errors to be useful for keyword searching. Also, loose images do not come with such text at all.

To overcome this, Intella offers embedded and external OCR options, letting you improve your case index.

Note that there can be some limitations with OCR processing:

  • OCR quality can be affected by several factors, including the quality of the original item, the way the original item is structured, the quality setting used in Intella, etc.

  • The OCR text may not be displayed in the same order as shown in the Preview tab.

  • Certain characters may not be OCRed correctly. E.g. a '1' may be read as an 'l', and '0' may be read as a 'O'.

  • Handwritten documents, or handwritten comments in documents, may not be OCRed.

12.1. Starting OCR

OCR is available only for item’s original content. Other types of content such as load file images are not supported at the moment. This may be improved in future releases.

Intella’s OCR support is a post-processing step, performed manually by the case admin after indexing has completed, or as an automated post-processing task. In the future, we may make this part of the indexing process.

To OCR a collection of search results, you can use the following procedure:

  1. Use Ctrl-click or Shift-click to select multiple items in the Details pane, using the table or thumbnails view. Alternatively, right-click and choose Select All to select all items in the list.

  2. Right-click and choose Process > OCR…​

This opens the OCR Wizard. This wizard lets you choose the OCR method and its settings.

Note that the “OCR Candidates” task condition can be used to automate OCR-ing. See the OCR section in the Post-processing section of the Sources chapter for more information on running OCR as a post-processing task.

You can also use the OCR button in the Previewer to OCR the current item. This uses the embedded ABBYY FineReader engine.

12.2. OCR methods

Intella currently supports three OCR methods:

  • ABBYY FineReader (embedded)
    This method processes items using an OCR engine that is embedded into Intella. The method is fully automatic and does not require any additional software, systems, or licenses.

  • ABBYY Recognition/FineReader Server
    This method consists of sending the files to an ABBYY Recognition Server or ABBYY FineReader Server for processing, automatically incorporating the received results into the case. This method is fully automatic and requires a licensed and configured instance of ABBYY Recognition Server or ABBYY FineReader Server available over the network.

  • External OCR tool
    This method consists of exporting the items as loose files, processing them with the user’s preferred OCR software, and importing the OCRed files back into the case.

12.3. Using ABBYY FineReader (embedded)

This method utilizes the ABBYY FineReader engine that is embedded into Intella. Use this method to get high-quality OCR results fully automated and without requiring any additional software, systems, or licenses.

As stated in the Intella Software License Agreement, the use of the embedded OCR functionality must be in conjunction with the supply of results for eDiscovery and services that are normally related with the Intella software. Please see the licenses\intella-license.rtf bundled with Intella for more information.

Steps to OCR selected items with the embedded ABBYY FineReader:

  • Select the desired items and open the OCR Wizard, as described above.

  • Specify the desired profile, balancing between speed and quality:

    • Accuracy – OCR-ing may take a longer time, but will likely yield a higher quality output.

    • Speed – OCR-ing may be faster, but produce output of a lower quality.

  • Specify the languages that are used in the items. Note that adding more languages will make the process slower.

  • Specify the number of workers. For best performance, it should match the number of logical CPU cores in your machine.

  • Specify the output format: Plain Text or PDF. When the PDF format is selected, Intella will store both the OCRed text and a searchable PDF version of the document.

  • Use the “Detect page orientation” option to automatically rotate an image if its text orientation differs from the standard view of the image.

  • Use the “Correct inverted images” option to detect whether an image is inverted (white text against black background).

  • Use the “Skip OCRed items” checkbox to skip items that have already been OCRed before. Otherwise, Intella will replace any existing OCR-ed text.

Click the “OCR” button to start the OCR process.

The following document formats are supported:

  • PDF

  • MS Office document (Word, Excel, PowerPoint, both old and new)

  • OpenDocument document (Text, Spreadsheet, Presentation)

  • XPS

The following image formats are supported:

  • BMP

  • DjVu

  • GIF

  • HEIC/HEIF

  • JBIG2

  • JPEG

  • JPEG-2000 (JP2)

  • PCX / DCX (DCX has not been tested)

  • PNG

  • TIFF

The embedded OCR engine uses the location of the "Temp" folder specified in the Intella preferences. By default, it is set to the Windows user folder for temporary files, which is typically located on the C: drive. A custom location can be entered in the preferences, see File > Preferences > General > Temp folder > Custom.

12.4. Using an external OCR tool

To OCR the selected items with an external OCR tool, you initially only need to specify an export folder. Once you click the Export button, Intella will export the items in their original format to the folder. Every file will be named after the MD5 of the item – note that this means that unique items are only exported once!

Next you can use any OCR tool to process the exported files.

To import the OCRed files back to Intella, the tool and its configuration should comply with the following requirements:

  • The OCR tool must be able to create a single OCRed file for each input file. Put these files in a separate folder.

  • The file name of the OCR output must match the original file name, but it may have a different file extension, per the file type produced by the OCR tool. For example, if the original file name is 6345b60187d08be573133376d7543c54.tif, then the OCRed file name can be 6345b60187d08be573133376d7543c54.txt.

  • The OCRed file format must be of one of the Intella supported formats, e.g. plain text, PDF, MS Office, etc.

After you have OCRed the files, select “File → Import OCRed files…​” in the main menu. Next, specify the folder where the OCR output is located. Use the “Skip OCRed items” checkbox to skip items that have already been OCRed before. Uncheck the “Skip OCRed items” to replace any existing OCRed text with the new one. The “Import as” option can be used to specify the format for the OCRed files, otherwise Intella will try to detect it automatically. Click on the Next button to scan the chosen folder for importable files and report problematic files. Click the “Start OCR Import” button to start importing the files.

Intella will analyze every file in the specified folder, extract the text and link it to the original item and all its copies. The imported OCR text can be found under a separate OCR tab in the Previewer.

12.5. Using ABBYY Recognition/FineReader Server

When you have access to an ABBYY Recognition Server, or its successor ABBYY FineReader Server, you can utilize it to OCR selected items in the case fully automatically.

ABBYY Recognition Server 3.5 or 4.0, or ABBYY FineReader Server 14, should be used.

Steps to OCR selected items with ABBYY Recognition/FineReader Server:

  • Select the desired items and open the OCR Wizard, as described above.

  • Specify the server’s IP address. The “Service URL” field will be populated automatically based on the entered IP address.

  • If you know that your server uses a different URL, you can override it by checking the “Use custom service URL” check box.

  • Specify the workflow name that should be used. Alternatively, you can press the “Get list from server” button to select a value from all available workflows on that server.

  • Specify the number of workers to let the Recognition/FineReader Server process more than one document at a time. The optimal number of workers depends on the Server capabilities (in particular the number of CPU cores on the server) and is also restricted by the server’s license (the number of CPU cores allowed to be used by the Recognition Server license).

  • Use the “Skip OCRed items” checkbox to skip items that have already been OCRed before. Otherwise, Intella will replace any existing OCR text.

  • Click the “OCR” button to start the OCR process.

The selected documents are will now be send to the Recognition/FineReader Server. The results that it sends back will be processed automatically, similar to how the external method works.

Please make sure that your ABBYY Recognition/FineReader Server is configured correctly:

  • The “Web Service” component is installed and configured properly.

  • You can open the Service URL in a browser and it shows the following text:
    “RSSoapService. The following operations are supported…​”.

  • The latest versions of FineReader Server come with a demo app that can be used to test whether the web service component is installed and configured correctly. Open the following web page and enter the fields to test the API:

  • A separate document should be generated for each input file.

  • The output format is a format that Intella can index.

  • The following parameters need to be set correctly in the following file
    (suggested parameters allow for processing files up to 30 MB):

C:\Program Files (x86)\ABBYY Recognition Server 3.5\RecognitionWS\web.config

Parameters:

<?xml version="`1.0`" encoding="`UTF-8`"?>
<configuration>
<system.web>
<httpRuntime maxRequestLength="`409600`" />
</system.web>
<system.webServer>
<security>
<requestFiltering>
<requestLimits maxAllowedContentLength="`300000000`" />
</requestFiltering>
</security>
</system.webServer>
</configuration>
If the FineReader Server is configured to use a non-standard port such as 8080, do NOT include the port in the "Server IP address" field. Instead, click on the "Use custom service URL" checkbox and add the port directly after the IP address. For example: Server IP Address: 192.168.0.1 and Service URL: http://192.168.0.1:8080/FineReaderServer14/WebService.asmx.

12.6. Reviewing OCRed items

To find all items in a case that have been OCRed, you can use the OCRed category in the Features facet. This attribute is also reflected in the Details table in the OCRed column. When an OCRed item is previewed, this will be shown as an additional property in the Properties tab.

When importing OCRed documents, Intella will extract the text, add it to the index, and store a searchable (original view) version of the document. The text can be found in the OCR tab of the previewer. The original view can be found in the OCR Preview tab. Note that the original content of the item will not be replaced. See the Exporting section for more details about exporting OCRed text and original view.

When converting a case made with Intella 2.0.x or older to the 2.1.x format, the OCR-ed text will be transferred as if it were part of the original document text, i.e. placed in the Contents tab rather than the OCR tab.

13. Insight view

The Insight tab contains several information panels that together give a concise overview of the information inside the case, revealing suspect behavior, and giving rise to follow-up investigative questions.

The information is extracted from a variety of sources, such as emails and documents, web browser histories, Windows registries and more.

Clicking on entries like a document type or custodian name in the Insight tab will add a relevant search for that item category to the Cluster Map in the Search view. The main window will then automatically switch to the Search view as well.

The entire tab can be exported to HTML by clicking on the Export button in the top right corner.

13.1. Evidence

The Evidence section shows important global statistics regarding your data. A detailed description of each category can be found in the section explaining the Features facet.

13.2. Types

The Types section shows a breakdown of the different types of files and other items in the case. It shows the same hierarchical structure as the Type facet in the Search tab.

13.3. Custodians

The Custodians section shows the list of custodians in the case, if any, together with the number of items that are assigned to them. A pie chart showing these amounts is shown to the right of the table.

For detailed information on how to define custodians see the section titled “Custodians”.

13.4. GDPR

The GDPR section gives an overview of privacy-sensitive information encountered in the case. Examples of such information are person names, email addresses, phone numbers and other communication handles, credit card numbers, etc. Such information is important from a GDPR compliancy perspective, or similar legal frameworks in use around the world.

For each category of personally identifiable information (PII), the number of values found is listed. These values can be exported to a CSV or XLS file. Furthermore, the number of items that contain at least one of these values is listed. This amount is further split up in Documents, Emails, and Other categories.

The PII categories are split into two groups, based on whether the PII was found in the document/email body or in the metadata.

Double-click on a table row to switch to the Search tab and see the items involved in that category.

Some categories are determined during indexing, yet some other categories may require Content Analysis to be run first. To launch the Content Analysis procedure from the Insight tab, click the hyperlink at the bottom of the GDPR panel.

13.5. Internet Artifacts

The Internet Artifacts section contains information about web browser activity, based on the browser histories detected in the evidence data.

All major browsers are supported: MS Internet Explorer/Edge, Mozilla Firefox, Google Chrome and Apple Safari.

The top chart shows the list of encountered browser histories, listing the following information:

  • The path of the browser history in the evidence data.

  • The type of browser, represented by the browser’s desktop icon.

  • The number of visited URLs in the browser history, both as a number and as a bar showing the amount relative to the total amount of visited URLs in the entire case.

  • The last used date of the browser history, i.e. the last time a new URL was added or a visit count was updated. Note that manual deletions of URLs in the history by the end user are not considered when determining the last used dates; it is merely indicative of when the regular day-to-day usage of that browser ended.

At the very top of this list is a row that represents the total amount of visited URLs in the case, regardless of location and web browser type.

Beneath the list of browser histories there is a breakdown of the visited URLs:

  • The “Top 100 visited URLs” table shows the most visited URLs, with for each URL the number of visits as indicated by the browser history.

  • The “Top 100 visited domains” table shows the most visited domains, together with the sum of the visit counts of all URLs in that domain. Subdomains are treated as independent domains.

  • The panels “Social media”, “Cloud storage”, “Webmail” and “Productivity” show the number of visits that belong to some commonly used websites, such as Facebook and Twitter for social media, DropBox and OneDrive for cloud storage, Gmail, and Yahoo Mail for webmail, etc.

By default, this breakdown covers all visited URLs in the case. By clicking on a row in the list of browser histories one can narrow down on the visited URLs in that browser history. The blue URL count bar indicates the selected browser.

The categories and domains that are checked can be configured by editing the common-websites.xml file in the [CASEDIR]\prefs folder.
During the development of this functionality we observed that the semantics of a “visited URL” may differ between browsers, possibly even between browser versions. In some cases, it indicates that the user explicitly visited a URL by entering it in the browser’s address bar or by clicking a link. In other cases, all resources loaded as a consequence of displaying that page may also be registered as “visited”, even resources from other domains, without making any distinction between the explicitly entered or clicked URLs on the one hand and the other resources on the other hand. One should therefore carefully look at the operation of a specific browser before drawing any final conclusions.

13.6. Timeline

The Timeline shows the timestamps of all items in the case over the years of months. This not only gives a rough overview of events over time, but can also be used to find data anomalies, e.g. unexpected peaks or gaps in the volume of emails, which for example may be caused by an incomplete capture of evidence files, bugs in the custodian’s software, default values entered by client software and actions of malicious custodians (resetting date fields, deleting information).

To the right of the chart are all date fields that Intella currently supports. Each date field shows the number of items that have that date field set. Date fields that do not occur in this case are disabled. (De)selecting one of the checkboxes changes the timeline to include or exclude the counts for that date field.

This update may take some time, depending on the case size and whether a local or remote case is used. The resulting counts are cached so that afterwards the user can toggle that checkbox and see the chart change instantly.

The chart can alternatively show months or years.

The Timeline’s time axis only shows dates between January 1, 1969 and two years from “now”. This is to prevent obviously incorrect dates that have been extracted from corrupt files from spoiling the graph.

13.7. Identities

The Identities section consists of three tables with various types of identities, which may be representing users or other entities.

The User accounts table shows a list of user accounts extracted from the evidence data. These can be:

  • Windows user accounts, extracted from Windows registry hives.

  • Skype user accounts, extracted from Skype databases. These are the database’s local account, not the entire contacts list of that account.

  • Pidgin user accounts. Again, these are the local accounts, not the entire contact list.

  • User accounts in cellphone reports as produced by Cellebrite UFED, Micro Systemation XRY and the Oxygen Forensic suite. See the documentation of the respective product for details on the correct interpretation of such information.

The “Origin” column in this table shows either a machine name extracted from a Windows registry or the location of the evidence file that the account was extracted from.

The Top 10 email addresses table shows the 10 email addresses with the highest number of emails in the case. Both the raw and deduplicated counts are shown. The top 10 is based on the raw counts.

The Top 10 host names table shows the host names that have the most emails associated with them. These are essentially the host names that show up when you expand the “All Senders and Receivers” branch in the Email Address facet. Both the raw and deduplicated counts are shown. The top 10 is based on the raw counts.

13.8. Notable Registry Artifacts

You may click on the "Calculate Notable Registry Artifacts" button to begin the analysis process (it might take significant time on larger cases). After the analysis is complete, the Notable Registry Artficats (NRA) section will appear. This section gives insight into the most important artifacts extracted from the Windows registry hives of the investigated machines/operating systems.

A case may contain evidence files (usually in the form of disk images) that relate to multiple operating systems (OSes), simply because multiple machines may be involved, but also because a machine may have multiple operating systems installed. Hence the artifacts are grouped by OS, labeled by the “Computer Name” that was extracted from the registry, and further subdivided in several categories.

The following artifact types are currently extracted and reported:

  • Basic OS information

  • OS time zones

  • OS user accounts

  • Network interfaces

  • Network connections

  • USB mass storage devices that have been connected

  • Recently used files

  • Shellbags

  • Typed URLs registered by web browsers using the registry

A “registry artifact” is a logical concept in Intella that is modeled as an atomic item in the case and that holds important information typically used in digital forensic investigations. This information is specially selected for this purpose by experienced forensic experts. While the properties of a registry artifact may be scattered across different registry hives and backups of these hives, Intella will unify them into a coherent item.

The NRA section is divided into two parts. On the left-hand side, labeled “Overview”, the tree organizing the registry artifacts is shown. The first level nodes represent OSes labeled with the “Computer Name” extracted from the registry. One lever deeper we find sub-nodes for the various registry categories (e.g. “User Accounts”), followed by leaf nodes representing the actual artifacts (e.g. a specific User Account).

One can select a leaf node in this tree, which will show the properties of that registry artifact in the Details view on the right-hand side.

Double-clicking on a leaf node opens the registry artifact item in the Previewer. This shows additional information such as the location of the item and allows for browsing to nearby items in the item hierarchy using the Previewer’s Tree tab. One can also right-click on a leaf node and select “Preview” from the context menu.

Right-clicking on a category node (e.g. a “User Accounts” node) shows a context menu with a Search option. This launches a search for all User Accounts in the Search view. Note that this searches for all user accounts, not just the ones in the currently explored OS.

Besides the regular registry hives, the Windows registry maintains backup files in the form of so-called “RegBack” files. Intella will process these files as well and display the extracted data in the NRA section. Values coming from such backup registry hives are marked with a “RegBack” label and are only displayed when they differ from the corresponding values in the current files. Not doing so would greatly increase the amount of redundant registry information.

13.8.1. Supported registry hives

Intella will process the following registry hives:

Registry Hive Name Location

SYSTEM

Windows/System32/config/SYSTEM

SYSTEM (RegBack)

Windows/System32/config/RegBack/SYSTEM

Windows/repair/SYSTEM

NTUSER.DAT

Found under folder Users/<user id> or Documents and Settings

SOFTWARE

Windows/System32/config/SOFTWARE

SOFTWARE (RegBack)

Windows/System32/config/RegBack/SOFTWARE

Windows/repair/SOFTWARE

SAM

Windows/System32/config/SAM

SAM (RegBack)

Windows/System32/config/RegBack/SAM

Windows/repair/SAM

Registry artifacts can be extracted from disk images and folders only if all relevant files are in the proper folders, e.g. Windows\System32\config\SYSTEM.
Support for Windows XP and older is limited.

13.9. Devices

The Devices section contains a list of all USB mass storage devices that have been connected to the suspect machines. This information is taken from the Notable Registry Artifacts section. It provides the ability to quickly oversee and sort all devices found in the case.

13.10. Networks

The Networks section contains a list of wired and wireless networks that a suspect machine has been connected to. This information is taken from the Notable Registry Artifacts section and from cellphone reports. It provides the ability to quickly oversee and sort all networks found in the case.

13.11. Significant Words

You may click on the "Calculate Significant Words" button to begin the analysis process (it might take significant time on larger cases). After the analysis is complete, the Significant Words panel will appear. It visualizes important words encountered in the item texts in the case, based on a statistical model of term relevance. The bigger the font of a word, the higher the relevance that word may have for the data set at hand.

These results are purely suggestive: though they are based on commonly used information retrieval techniques, they only look at the evidence data. They do not take the investigative research questions into account, or any investigative results such as items tagged as “relevant”.

The Paragraphs section shows statistics on the paragraphs that Intella has registered, when the Analyze Paragraphs setting was set on the source(s) in the case. It lists the number of unique and duplicate paragraphs, both as raw numbers and as percentages. Furthermore, the Paragraphs marked as Seen or Unseen are counted. Finally, the number of Documents, Emails, and Other item types with unique content (i.e. a paragraph that does not occur in any other item) is listed. These groups can be clicked, which shows these item sets in the Search tab.

13.12. Workflow

The Workflow section lists additional tasks that one might consider after the initial indexing is done. These tasks can further refine the case index quality and kick-start the investigation and analysis phases.

Additional Processing category:

  • The Process OCR candidates link opens the OCR wizard and will process all items that fall into the default OCR Candidates category: all empty PDF documents and all TIFFs and JPEGs that are not embedded in documents.

  • The Export encrypted items link opens the Export wizard for all items that are encrypted but have not been decrypted.
    Export encrypted items list exports the metadata of these items to a CSV file.

  • The Export unprocessed items link opens the Export wizard for all items that fall into the “Extraction Unsupported” category in the Features facet.
    Export unprocessed items list exports the metadata of these items to a CSV file.

  • The Export exception items link opens the Export wizard for all items that fall into the “Exception Items” category in the Features facet.
    Export exception items list exports the metadata of these items to a CSV file.

Search & Analysis category:

  • The Generate thumbnails link initiates the thumbnail generation process for all items in the case. Doing so will speed up the performance of the Thumbnails results view and the Previewer’s Thumbnails tab.

  • The Run content analysis link initiates the content analysis procedure for all items in the case. This detects person, organization and location names used in the item texts and reports them in the Content Analysis facets.

  • The Run email thread analysis link initiates the email thread analysis procedure for all email items in the case, as explained in the Email Threading section.

  • Add keyword list adds a keyword list to the case, for use in the Keyword Lists facet or Keywords tab in the Statistics view.

  • Add MD5 list adds an MD5 or message hash list, for use in the MD5 and Message Hash facet.

  • Add saved search adds a saved search obtained from another case to this case, for use in the Saved Searches facet and Keywords tab in the Statistics view.

  • Add task adds a post-processing task (e.g. running a keyword list and tagging the results), to be used during post-processing or on-demand via the Tasks option in the File menu.

Report category:

  • The Export n events link opens the Export Event Log window, letting one export all or selected case events. These events capture all actions that have taken place inside the case since the start of its existence, such as source creation and indexing, searching, tagging and exporting. These events include the user that took the action and the time the action took place.

  • Open log folder opens the folder where the case’s log files are stored in Windows Explorer.

To search for text, enter a query in the Search panel and click the Search button.

For query syntax rules, refer to the “Search query syntax” section below.

When the Search button is clicked and no keyword query has been entered, all items in the case will be returned.

Due to technical limitations a search on the “Comments” field cannot be combined with a search on other fields.

14.1. Search options

With the search options panel, you can limit keyword searching to specific item parts or attributes:

image

  • Text

  • Title / Subject

  • Summary & Description

  • Path (= folder and file name)

  • File name

  • Message Headers

  • Raw Data (e.g. low-level data from PST files,
    MS Office documents, vCards)

  • Comments

  • Authors & E-mail Addresses

  • Each of the From, Sender To, Cc and Bcc fields separately

  • Export IDs

To see the search options, click the Options button under the search text field. The options box will be displayed as a popup menu below the button.

Select the options for properties that you want to include in your search and deselect those you want to exclude. Your selected search options will be stored and used for future searches until you change them.

The Options box also has a checkbox for setting whether the excluded paragraphs should be considered. By default, this is turned on. Uncheck this checkbox to search the entire document text again.

image As a reminder, the Options button will show a yellow triangle when not all options are selected.
To prevent application instability, the maximum length of a single search query is limited to 16,000 characters.

To hide the options box, click the Options button again. If you have made any changes, the icon on the Options button will change to a yellow warning sign as a reminder that you have changed options that will affect your searches.

Click the arrows in the Search button to start an Include or Exclude search, rather than a regular search. See the “Including and excluding facet values” section for more information.

14.2. Search query syntax

In the text field of the Search panel you can use special query syntax to perform complex multi-term queries and use other advanced capabilities.

You can also see the list below by clicking on the question mark button in the Search panel.

14.2.1. Lowercase vs. uppercase

Keyword searches work in a case-insensitive manner: during indexing all characters are lowercased, as are the characters in a keyword query.

This means that the query “john” will match with “john”, “John” and “JOHN”.

14.2.2. Use of multiple terms (AND/OR operators)

By default, a query containing multiple terms matches with items that contain all terms anywhere in the item. For example, searching for:

john johnson

returns all items that contain both “john” and “johnson.” There is no need to add an AND (or “&&”) as searches are performed as such already, however doing so will not negatively affect your search.

If you want to find items containing at least one term but not necessarily both, use one of the following queries:

john OR johnson
john || johnson

14.2.3. Minus sign (NOT operator)

The NOT operator excludes items that contain the term after NOT:

john NOT johnson
john -johnson

Both queries return items that contain the word “john” and not the word “johnson.”

john -"john goes home"

This returns all items with “john” in it, excluding items that contain the phrase “john goes home.”

The NOT operator cannot be used with a single term. For example, the following queries will return no results:

NOT john
NOT "john johnson"

14.2.4. Single and multiple character wildcard searches

To perform a single character wildcard search you can use the “?” symbol. To perform a multiple character wildcard search you can use the “*” symbol.

To search for “next” or “nest,” use:

ne?t

To search for “text”, “texts” or “texting” use:

text*

The “?” wildcard matches with exactly one character. The “*” wildcard matches zero or more characters.

To search for a certain phrase (a list of words appearing right after each other and in that order), enter the phrase within full quotes in the search field:

"john goes home"

will match with the text “John goes home after work” but will not match the text “John goes back home after work.”

Phrase searches also support the use of nested wildcards, e.g.

"john* goes home"

will match both “John goes home” and “Johnny goes home”.

Intella supports finding items based on words or phrases that are within a specified maximum distance from each other in the items text. This is a generalization of a phrase search.

To do a proximity search you place a tilde (“~”) symbol at the end of a phrase, followed by the maximum word distance:

"desktop application"~10

returns items with these two words in it at a maximum of 10 words distance.

It is possible to mix individual words, wildcards and phrases in proximity queries. The phrases must be enclosed in single quotes (' ') in this case:

"'desktop application' 'user manual'"~10

Nested proximity searches are also possible:

"'desktop application'~2 'user manual'~4"~10
Nested phrase and proximity queries are always use single quotes. Using regular double quotes for them will cause a syntax error. Only one level of nesting is possible.

14.2.7. Grouping

You can use parentheses to control how your Boolean queries are evaluated:

(desktop OR server) AND application

retrieves all items that contain “desktop” and/or “server,” as well as the term “application.”

Intella supports fuzzy queries, i.e., queries that roughly match the entered terms. For a fuzzy search, you use the tilde (“~”) symbol at the end of a single term:

roam~

returns items containing terms like “foam,” “roams,” “room,” etc.

The required similarity can be controlled with an optional numeric parameter. The value is between 0 and 1, with a value closer to 1 resulting in only terms with a higher similarity matching the specified term. The parameter is specified like this:

roam~0.8

The default value of this parameter is 0.5.

Intella’s Keyword Search searches in document texts, titles, paths, etc. By default, all these types of text are searched through. You can override this globally by deselecting some of the fields in the Options, or for an individual search by entering the field name in your search.

title:intella

returns all items that contain the word “intella” in their title.

The following field names are available:

  • text - searches in the item text

  • title - searches in titles and subjects

  • path - searches in file and folder names and locations

  • filename - searches in file names only

  • summary - searches in descriptions, metadata keywords, etc.

  • agent – searches in authors, contributors and email senders and receivers

  • from – searches in email From fields

  • sender – searches in email Sender fields

  • to – searches in email To fields

  • cc – searches in email Cc fields

  • bcc – searches in email Bcc fields

  • headers - searches in the raw email headers

  • rawdata - searches in raw document metadata

  • comment - searches in all comments made by reviewer(s)

  • export - searches in the export IDs of the items that are part of any export set

The summary field can contain a lot of metadata fields:

  • Comments (as originating from the evidence files)

  • Template

  • Organization

  • Location

  • Contact note

  • Generators

  • Keywords

  • Password

  • Certificate

  • Message ID

  • Document ID

  • Native ID

You can mix the use of various fields in a single query:

intella agent:john

searches for all items containing the word “intella” (in one of the fields selected in the Options) that have “john” in their author metadata or email senders and receivers.

14.2.10. Regular expressions

Keyword queries can also be expressed using regular expressions. Be aware that these regular expressions are evaluated on the terms index, not on the entire document text as a single string of characters. Your search expressions should therefore take the tokenization of the text into account.

To search for a regular expression, put it between "/" slash characters:

/.?ext.*/

The result will match tokens like “next”, “text”, “texts”, “texting” and so on.

For regular expressions evaluated on the raw document text, see the section on the Content Analysis facet.

14.2.11. Tokenization and Special characters

Tokenization underlies the keyword search functionality in Intella. It is the process of dividing texts into primitive searchable fragments, known as "tokens" or "terms". Each token makes a separate entry in the text index, pointing to all items containing this token. Keyword search works by finding matches between the tokens in the user’s query and in the index. Therefore, for effective keyword search, it is vital to have a basic understanding of how tokenization works in Intella.

Tokenization employs different algorithms, but in the most common case it is simply splitting the text around specific characters known as "token delimiters". These delimiters include spaces, punctuation symbols, and other non-alphabetic characters, to produce tokens close to the natural language words.

A side effect of this method is that it is impossible to search for words together with the token delimiters. If these characters are met in the user query, they play their delimiting role, thus being handled the same as simple spaces. This is rarely a problem, although it should be taken into account when doing a keyword search.

To search for exact text fragments, including all punctuation and special characters, the Content Analysis functions can be used (see the section on the Content Analysis facet for details).
A list of all search tokens, generated for an item, can be seen in the "Words" tab of the Previewer window.

There is no specific support for the handling of diacritics. E.g., characters like é and ç will be indexed and displayed, but these characters will not match with 'a' and 'c' in full-text queries. A workaround can be to replace such characters with the '?' wildcard.

The following characters have special meaning in the query syntax and may cause an error message if not used in accordance to the syntax rules:

+ - && || ! ( ) { } [ ] ^ " ~ * ? : / \

To prevent the syntax errors, these characters need to be escaped by the preceding \ character. Please note that if the character is classified as a token delimiter, then escaping it in the query will not make it searchable.

15. Using facets

image

Besides keyword searching, the indexed items can be browsed by facets, which represent specific item properties. Every facet organizes the items into groups (possibly hierarchical) depending on a specific item property.

Selecting a facet in the Facet panel will give you a list of all values of the selected facet in the lower part of the panel. In the example on the right, the Type facet has a list of file types as values.

To search for items that match a facet value, select the facet value, and click the Search button.

When search results are displayed in the Results panel, and items in that set are associated with a facet value, that value will be highlighted in bold blue text in the facet’s value list. This indicates that that value occurs in the current search results and can be used to further drill-down in those results. Furthermore, the amount of items in the result set that have that value will be shown, followed by the total amount of items in the case that have that value. This feature is called facet highlighting.

To export facet information, (1) select a facet, (2) open the context menu - right mouse click - on the facet values, and (3) select Export values…. This will open the Export values dialog. Choose a file name and folder and save the export file. The CSV file will contain the facet values (e.g. file types, email addresses, folder names), their total counts in the case, and their currently shown counts, which represents the overlap with the currently shown search results.

15.1. Available facets

15.1.1. Saved Searches

The Saved Searches is a list of previous sets of searches that the user has stored.

image

When there are search results displayed in the Cluster Map and the Searches list, the Save button beneath the Searches list will be enabled. When the user clicks this button, a dialog opens that lets the user enter a name for the saved search. A default name will be suggested based on the current searches. After clicking on the OK button, the chosen name will appear in the list in the Saved Searches facet.

Click on the name of the saved search and then on the Restore button to bring the Cluster Map and the Searches list back into the state it had when the Save option was used.

The “Replace current results” checkbox controls what happens with the currently displayed searches when you restore a saved search. When turned on, the Cluster Map and Searches list will be emptied first. When turned off, the contents of the saved search will be appended to them.

The “Combine queries” checkbox can be used to combine the result sets of all parts of the saved search into a single result set. This is for example useful when the various parts conceptually are meant to find the same set of items, just in a technically different way. Example are different complex Boolean queries, which could have been combined into a single Boolean OR query but that the user prefers to keep separate in the saved search definition.

Saved searches can be shared across cases. To transfer a saved search, right-click on the saved search in the list and select “Export search…”. The search is then exported as an XML file can then be imported into any other case by right-clicking in this list and selecting “Import searches…”.

Saved searches are grouped by the user who made them. Depending on the Intella version used to create the case, a “Default searches” branch may also be present with pre-defined saved searches.

15.1.2. Features

The Features facet allows you to identify items that fall in certain special purpose categories:

  • Encrypted: all items that are encrypted. Example: password-protected PDF documents. If you select Encrypted and click the search button, you will be shown all items that are encrypted.

Sometimes files inside an encrypted ZIP file are visible without entering a password, but a password still needs to be entered to extract the file. Such files cannot be exported with Intella if the password has not been provided prior to indexing. In this case both the ZIP file and its encrypted entries will be marked as Encrypted, so searching for all encrypted items and exporting those will capture the parent ZIP file.
  • Decrypted: all items in the Encrypted category that Intella could decrypt using the specified access credentials.

  • Unread: all emails, SMS/MMS messages, chat messages, and conversations that are marked as “unread” in the source file. Note that this status is not related to previewing in Intella.

This property is only available for PST, OST and EDB emails, and some cellphone dumps. If the Unread property is not set, it could mean that either the item was not read or that the property is not available for this item. Some tools allow the user to reset a message’s unread status, so even when the flag is set, it cannot be said with certainty that the message has not been read.
  • Empty documents: all items that have no text while text was expected. Example: a PDF file with only images.

  • Has Duplicates: all items that have a copy in the case, i.e. an item with the same MD5 or message hash.

  • Has Shadow Copies: all items that have another version located in a shadow copy volume.

  • Has Geolocation: Indicates whether the item has geolocation information.

  • Downloaded from Internet: Indicates items that may have been downloaded from the Internet. Intella determines such items by looking at the Zone.Identifier alternate stream in NTFS file systems. Where possible, Intella will extract the URL the file was downloaded from. This URL can hen be found in the Raw Data tab.

  • OCRed: indicates whether the item has been OCRed after indexing. See the separate chapter on OCRing of documents and images.

  • Has Imported Text: indicates whether the item has text imported via the “-importText” command-line option.

  • Content Analysed: all items for which the Content Analysis procedure has been applied.

  • Images Analysed: all items for which the Image Analysis procedure has been applied.

  • Exception items: all items that experienced processing errors during indexing. This has six subcategories that match the warning codes in the exception report:

    • Unprocessable items: the data cannot be processed because it is corrupt, malformed or not understood by the processor. Retrying will most likely result in the same result.

    • I/O errors: the processing failed due to I/O errors. The processing might succeed in a repeated processing attempt.

    • Decryption failures: the data cannot be processed because it is encrypted and a matching decryption key is not available. The processing might succeed in a repeated processing attempt when the required decryption key is supplied.

    • Timeout errors: the processing took too long and was aborted. See more details on how to configure crawler timeout in "Memory, crawler count and timeout settings" chapter.

    • Out of memory errors: the processing failed due to a lack of memory.

    • Processing errors: the processing failed due to a problem/bug in the processor. The description should contain the stack trace.

    • Truncated text: the extracted text was not processed entirely. See the “Exceptions report” section for more details.

    • Crawler crash: the processing failed due to a crawler crash. This is a more severe error compared to the Processing Error type. When it occurs, Intella will also reject all items that are related to crashed item (e.g. PST file and all of the emails that it contains). More details about why the crawler crashed can usually be found in a hs_err_pid_XYZ.log file which is located in the case logs folder (one file per crash). Crawler crashes will not affect other items and the case integrity.

  • Extraction unsupported: all items that are larger than zero bytes, whose type could be identified by Intella, are not encrypted, but for which Intella does not support content extraction. An example would be AutoCAD files: we detect this image type but do not support extraction any content out of it.

  • Text Fragments Extracted: indicates whether heuristic string extraction has been applied on a (typically unrecognized or unsupported) binary item.

  • Irrelevant: all items that fall into one of the “Irrelevant Items” categories and that themselves are often considered to be of little relevance to a review. See the Preferences section for details on this automatic classification.

  • Threaded: all items that have been subjected to email threading processing and that were subsequently assigned to a thread (see the Email Thread facet). Subtypes:

    • Inclusive: all email items marked as inclusive.

    • Non-Inclusive: all email items marked as non-inclusive.

    • Missing Email Referent: Indicates that the threading process has detected that the email item is a reply to another email or a forwarded email, but the email that was replied to or that has been forwarded is not available in the case.

  • Recovered: all items that were deleted from a PST, NSF, EDB, disk image, cellphone report, cloud source or volume shadow copy and that Intella could still (partially) recover. The items recovered from PST, NSF and EDB files are the items that appear in the artificial “<RECOVERED>” and “<ORPHAN ITEMS>” folders of these files in the Location facet. The items recovered from volume shadow copies are located in the artificial "<Volume Shadow Copies>" folder of the parent volume in the Location facet. The Recovered branch in the Features facet has the following sub-branches, based on the recovery type and the container type:

    • Recovered from PST.

    • Orphan from EDB.

    • Orphan from NSF.

    • Orphan from PST.

    • Recovered from cellphone.

    • Recovered file metadata from disk images.

    • Recovered entire file content from disk images.

    • Recovered partial file content from disk images.

    • Recovered from cloud source.

    • Recovered from volume shadow copy.

  • Attached: all items that are attached to an email, conversation, or document. Only the direct attachments are reported; any items nested in these attachments are not classified as Attachment. Furthermore, items that are classified as Embedded Image are not classified as Attachment, and vice versa.

  • Has attachments: all emails, documents and user activities that have other items attached to them. Note that it does not include embedded images.

  • Embedded images: all inlined images in emails, documents, spreadsheets, or presentations.
    Formally, an image is classified as “embedded image” when it is displayed as part of the native rendering of its direct parent. This happens when the parent is displayed in the Preview tab of the Previewer and when the parent is exported to PDF in “Original view” mode. An embedded image is thus already “visible elsewhere” when the original view is used, which (depending on your policy) may be reason not to export this image item separately as well, saving exporting and/or reviewing time.

  • Tagged: all items that are tagged.

  • Flagged: all items that are flagged.

  • Commented: all items that have a comment.

  • Previewed: all items that have been opened in Intella’s previewer.

  • Opened: all items that have been opened in their native application.

  • Exported: all items that have been exported.

  • Redaction: all items that have been subject to one of the redaction procedures. See the section on Redaction for more information.

    • Redacted: all items that have one or more parts blacked out due to redactions. Items on which the Redact function has been used but in which no parts have been marked as redacted are not included in this category.

    • Queued for Redaction: all items that have their “Queued for Redaction” checkbox selected. These will turn to “Redacted” once the user performs the “Process Redaction Queue” function on them.

    • Missing keyword hits: all items that had a redaction issue when Process Redaction Queue was invoked.

  • Batched: all items that have been assigned to a coding batch. This can happen when a case is reviewed in Intella Connect, our web-based companion product.

  • Top-level parents: all top-level parent items in the case.

  • Item stubs: when items are exported to a target case without their parents in the export set as well, their parents will be represented by stubs.

  • W4 Delta: new items found by Intella in imported W4 cases.

  • Analyzed for Near-Duplicates: all items that have been analyzed in the last Near-Duplicate Analysis procedure.

  • Has Near-Duplicates: all items that have been positively assessed and included in near-duplicate groups by the last Near-Duplicate Analysis procedure.

  • Has Event Note:: all items that have a note assigned to one of its events

  • All items: all items (non-deduplicated) in the entire case.

In cases in which multiple users have worked, i.e. shared cases, the Previewed, Opened, Exported, Commented, Tagged, and Flagged nodes shown in the Facet panel will have sub-nodes, one node for each user.

To export the categories and their counts to a CSV file, right-click anywhere in the facet’s values area and select “Export values…”.

15.1.3. Tags

Tags are labels defined by the user to group individual items. Typically used tags in an example are for example “relevant”, “not relevant” and “legally privileged”.

Tags are added to items by right-clicking in the Results table or the Cluster Map and choosing the Tags > Add Tags… option. Tags can also be added in the Previewer. The exact procedure is described in other sections of this manual.

To search for all items with a certain tag, select the tag from the Tags list and click the Search button below the list.

When tags have been added by different users in the same case, the Tags facet panel will have a drop-down list at the bottom, listing the names of all reviewers that have been active in this case. You can use this list to filter the tags list for taggings made by a selected reviewer only. Select the “all” option to show taggings from all users.

If the same tag has been used by different reviewers, their names and the numbers of tagged items are displayed in the tag statistics line (in the parentheses after the tag name). The list of reviewers can include “You” to indicate taggings made by the current case user. If no reviewer name is mentioned, it means that all taggings with this tag are made by the current user only.

The tags can be organized into a hierarchical system by the creation of sub-tags within an existing (parent) tag group. To create a sub-tag, select an existing tag in the Tags facet and choose “Create new tag inside …​” in the context menu. In the dialog box, enter the name and (optionally) the description of the new tag.

To rename a tag or change the tag description, select the tag in the facet and choose “Edit…​” in the context menu.

When a tag is renamed, all items associated with this tag will be assigned the new tag name automatically. However, some operations that depend on specific tag names (such as indexing tasks with the Tag condition, see 10.5.1) may need to be corrected manually.

To delete a tag, select it in the facet and choose “Delete…​” in the context menu.

To export the tags and their counts to a CSV file, right-click anywhere in the facet’s values area and select “Export values…”.

15.1.4. Identities

The Identities facet makes it possible to search for all items related to an identity, as defined in the Identities tab.

When searching for an identity, it queries for all items that have any of the identity’s aliases as sender/receiver/caller/callee/etc. Effectively, it gathers all messages in which the identity is a participant through one of its aliases.

To export the identities and their counts to a CSV file, right-click anywhere in the facet’s values area and select “Export values…”.

15.1.5. Custodians

Custodians are assigned to items to indicate the owner from whom an evidence item was obtained. The “Custodians” facet lists all custodian names in the current case and allows searching for all items with a certain attribute value.

Custodian name attributes are assigned to items either automatically (see the section on custodian name post-processing) or manually in the Details panel. To assign a custodian to items selected in the Details panel, use the “Set Custodian…” option in the right-click menu. To remove custodian information from selected items, choose the “Clear Custodian…” option.

To change a custodian’s name, select it in the list and choose “Edit custodian name…” in the right-click menu.

To delete a custodian from the case and clear the custodian attribute in all associated items, select the value in the facet panel and choose “Delete” in the right-click menu.

To export the custodians and their counts to a CSV file, right-click anywhere in the facet’s values area and select “Export values…”.

Custodians in Compound Cases

Custodian names defined and assigned to items in the sub-cases are available in the “Custodians” facet of a Compound Case. If two or more sub-cases define the same custodian name, this name appears in the facet as a single value. The search result of this facet value will be a union of all items assigned to this custodian name in all sub-cases.

Sub-case custodians are read-only in the Compound Cases: it is not possible to edit or delete them, as well as remove the item assignments made in the sub-cases. However, it is possible to reuse the sub-case custodians for new assignments in the Compound Case. These assignments will be specific to the Compound Case and will not affect the sub-cases.

The Compound Cases can also specify the local custodian names that can be assigned, unassigned, edited, or deleted just like custodians in a regular case.

15.1.6. Location

This facet represents the folder structure inside your sources. Select a folder and click Search to find all items in that folder.

When “Search subfolders” is selected, the selected folder, all items in that folder, and all items nested in subfolders will be returned, i.e. all items in that entire sub-tree.

When “Search subfolders” is not selected, only the items nested in that folder will be returned. Items nested in subfolders will not be returned, nor will the selected folder itself be returned.

When your case consists of a single indexed folder, then the Location tree will show a single root representing this folder. Selecting this root node and clicking Search with “Search subfolders” switched on will therefore return all items in your case.

When your case consists of multiple mail files that have been added separately, e.g. by using the PST and NSF source types in the New Source wizard, then each of these files will be represented by a separate top-level node in the Location tree.

To export the subfolders and their counts of a given location node to a CSV file, right-click on that node and select “Export values…”.

15.1.7. Email Address

This facet represents the names and/or email addresses of persons involved in sending and receiving emails. The names are grouped in the following categories:

  • From

  • Sender

  • To

  • Cc

  • Bcc

  • Addresses in Text

  • All Senders (From, Sender)

  • All Receivers (To, Cc, Bcc)

  • All Senders and Receivers

  • All Addresses

The first five categories list email addresses found in the corresponding message headers. Most emails typically only have a From header, not a Sender. The Sender header is often used in the context of mailing lists. When a list server forwards a mail sent to a mailing list to all subscribers of that mailing list, the message send out to the subscribers usually has a From header representing the conceptual sender (the author of the message) and a Sender header representing the list server sending the message to the subscriber on behalf of the author.

The “All Senders”, “All Receivers” and “All Senders and Receivers” categories group addresses into specific sender or recipient roles, abstracting from the specific header that was used.

The “Addresses in Text” category lists email addresses that are mentioned in message and document bodies.

“All Addresses” group together all other categories and thus contains all email addresses found anywhere in either message headers or textual content.

Sorting and grouping
The contacts can be sorted alphabetically by email addresses (the default order), by the contact name associated with them or by the number of items associated with this contact. To change the sorting method, right-click anywhere in the facet and choose the desired sorting method from the “Organize” menu.

The addresses can optionally be grouped by the host name used in the email address. To enable or disable grouping, select the “Group by host name” option in the “Organize” section of the context menu. Enabling this option adds another level of nodes to the tree, representing the host names.

Filtering on text
To quickly find specific email addresses, contact names or host names, it is possible to filter the facet content to only display the values that contain a specific substring. To filter the contacts in a specific category, expand the tree branch and click on the image button below the tree. In the text field that appears enter the text. The tree will be filtered to show only those contacts whose contact name or email address matches the entered text.

To cancel filtering and hide the text field, click the filter button again or type Escape.

Filtering on presence in the current search results
To display only the highlighted email addresses, i.e. the addresses that occur in the currently visible or selected search results, click on the image button. To return to displaying all addresses, just click this button again.

This type of filtering is removed automatically when a different branch is expanded, the selection in the facet or Cluster Map changes or when the sorting or grouping mode changes.

This filter can be used in combination with the text filter.

Exporting email addresses
To export the email addresses and their counts of a given Email Address facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.8. Phone Number

This facet lists phone numbers observed in phone calls, SMS and MMS messages extracted from cellphone reports. Furthermore, this includes phone numbers listed in PST contacts and vCard files.

The “incoming” and “outgoing” branches are specific to phone calls and SMS/MMS messages. The “All Phone Numbers” branch combines all the above contexts.

Depending on the type of evidence files and their contents, the phone numbers may or may not have a name associated with them.

This facet also supports the filtering options described in the Email Address section.

To export the phone numbers and their counts of a given Phone Number facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.9. Chat Account

This facet lists chat accounts used to send or receive chat messages, such as Skype and WhatsApp account IDs. Phone numbers used for SMS and MMS messages are also included in this facet.

Depending on the type of evidence files and their contents, the chat accounts may or may not have a human-readable name associated with them.

This facet also supports the filtering options described in the Email Address section.

To export the chat account names and their counts of a given Chat Account facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.10. Recipient Count

This facet lets the user search on recipient count ranges by entering the type and the number of recipients (minimum and maximum). The following recipient types are supported:

  • All Recipients: all email, chat, and cellphone recipients.

  • Visible Recipients: visible email, chat, and cellphone recipients (To, Cc).

  • Blind Recipients: blind carbon copy email recipients (Bcc).

15.1.11. Date

This facet lets the user search on date ranges by entering a From and To date. Please note that the date entered in the To field is considered part of the date range.

Besides start and end dates, Intella lets the user control which date attribute(s) are used:

  • Sent (e.g. all email items)

  • Received (e.g. all email items)

  • File Last Modified (e.g. file items)

  • File Last Accessed (e.g. file items)

  • File Created (e.g. file items)

  • Content Created (e.g. file items and email items from PST files)

  • Content Last Modified (e.g. file items and email items from PST files)

  • Primary Date

  • Family Date

  • Last Printed (e.g. documents)

  • Called (e.g. phone calls)

  • Start Date (e.g. meetings)

  • End Date (e.g. meetings)

  • Due Date (e.g. tasks)

A checkbox is provided for easy (de)selection of all attributes at once.

The Date facet will only show the types of dates that occur in the evidence data of the current case.

Furthermore, it is possible to narrow the search to only specific days or specific hours. This makes it possible to e.g. search for items sent outside of regular office hours.

Note that the Preferences dialog has a setting that controls how dates are displayed: by selecting a geographic region, all dates will be displayed in a manner commonly used in that region.

15.1.12. Type

This facet represents the file types (Microsoft Word, PDF, JPEG, etc.), organized into categories like Documents, Spreadsheets, etc. To refine your query with a specific file type, select a type from the list and click “Search”.

Note that you can search for both specific document types like PNG Images, but also for the entire Image category.

Empty (zero byte) files are classified as “Empty files” in the “Others” branch, regardless of their file extension.

To export the types and their counts to a CSV file, right-click anywhere in the facet’s values area and select “Export values…”.

15.1.13. Author

This facet represents the name(s) of the person(s) involved in the creation of documents. The names are grouped into two categories, as is done in most office formats:

  • Creator

  • Contributor

To refine your query by a specific creator or contributor name, select the name and click the Search button.

This facet also supports the filtering options described in the Email Address section.

To export the author names and their counts of a given Author facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.14. Content Analysis

The Content Analysis facet allows you to search items based on specific types of entities that have been found in the textual content of these items.

The top three categories are populated automatically during indexing and are available immediately afterwards:

  • Credit Card Numbers – suspected numbers of the major world-wide credit card systems (Visa, MasterCard, American Express and others).

  • Social Security Numbers – suspected SSN numbers issued by the United States Social Security Administration.

  • Phone Numbers – suspected phone numbers.

For credit card numbers, the algorithm looks for a sequence of digits matching the patterns of most common credit card systems (e.g. Visa, MasterCard, American Express, Diners Club). This digit sequence can be arbitrarily mixed with space and dash characters. Additionally, the Luhn checksum is tested for this sequence to make sure that this is a valid credit card number.

The other categories in this facet are empty by default. To populate them, a user needs to perform the automatic content analysis procedure on a selected set of items. Please see the “Content Analysis” section for instructions.

This facet also supports the filtering options that are available in the Email Address facet.

To export the entities and their counts of a given Content Analysis facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.15. Image Analysis

image

The Image Analysis facet allows you to search items based on the results of the Image Analysis performed on these items.

The categories in this facet are empty by default. To populate them, a user needs to perform the automatic Image Analysis procedure on a selected set of items. Please see the “Image Analysis” section for instructions.

The facet tree contains three branches:

  • Skin tone (sub-categorized as Weak, Medium and Strong based on the presence of human skin colors)

  • Image categories (general types of images, such as Documents, IDs, Photos, etc.)

  • Detected objects (the objects detected by the Image Analysis algorithm, such as Persons, Vehicles, etc.)

Image analysis provides probabilistic confidence score estimations for each result, expressed as a floating point number between 0.0 and 1.0. The “Min. confidence” slider specifies the value of a threshold filter applied to the results in this facet. The image items evaluated with lesser confidence scores than the current threshold will not appear in the search results of this facet. It is possible to execute multiple searches with the same facet value but different confidence thresholds.

The “Min. confidence” setting does not affect the values in the "Skin tone" category.

This facet also supports the filtering options that are available in the Email Address facet.

To export the entities and their counts of a given Image Analysis facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.16. Email Thread

In the Email Thread facet you can search for emails based on the email thread identified by the email threading procedure. To populate this facet, a user needs to perform the email threading procedure on a selected set of items. Please see the Email Threading section for instructions.

Be default, all threads containing only a single email are hidden from view, as they can greatly increase the length of the list and are typically of little use. To include these threads in the list, uncheck the “Hide threads with one email” checkbox.

To export the email thread names and their counts to a CSV file, right-click anywhere in the facet area and select “Export values…”.

15.1.17. Near-Duplicates

This facet lists all item groups identified by the last near-duplicates analysis. To populate this facet, a user needs to perform near-duplicate analysis on a selected set of items. Please see the "Near-duplicates Analysis" section for instructions.

The names of near-duplicate groups are derived from the titles of their master items. Searching for a group produces a set of items that includes the master item and its near-duplicates that have similarity scores larger than or equal to the threshold specified for near-duplicates analysis.

The groups can be sorted alphabetically by name (the default order) or by size (the number of associated near-duplicate items). To change the sorting method, right-click anywhere in the facet and choose the desired sorting method from the “Organize” menu.

To export the near-duplicate group names and their counts to a CSV file, right-click anywhere in the facet area and select “Export values…”.

15.1.18. Keyword Lists

In the Keyword List facet, you can load a keyword list, for automating searching with sets of previously determined queries.

The most basic keyword list is a text file in UTF-8 encoding that contains one search term per line. Once loaded, all the search terms found in the keyword list are shown in the Keyword Lists facet. They are now available for searching: just select one or more queries, or select the name of the keyword list, and click the Search button to search with these queries.

When the “Combine queries” checkbox is selected, and you have multiple queries selected, they will be combined into one query, effectively creating a single Boolean OR query. The matching items are then returned as a single set of results. When the checkbox is deselected, the selected queries will be evaluated separately, resulting in as many result sets as there are selected queries in the list. This may cause the Cluster Map to turn to Sets mode to handle a large amount of result sets.

Keyword lists can also use more advanced queries. The complete keyword search query syntax is supported here, e.g. wildcards, Boolean operators and field names can be used.

Besides searching, keyword lists can also be used to tag items. To do this, select the keyword list and click the Auto-tag button. A window will open that lists the queries in the first column and the proposed tag in the second column. When you click Apply, each query in the first column will be evaluated separately and have its results tagged with the proposed tag.

By default, the proposed tags are the queries itself. You can change this interactively in the table by clicking on the proposed tag and entering a new value. Alternatively, a keyword list can take the form of a CSV file in which the first column specifies the query and subsequent columns specify the tags. Use slashes to denote hierarchical tags. If a line has only one column, the proposed tag will default to the query text itself.

An example keyword list with associated tags could look like this:

confidential,Responsiveness/Privileged,Sensitivity/High
patent,Sensitivity/Medium

This will tag all items containing the term “confidential” with the tag “Privileged”, nested beneath the tag “Responsiveness”, and as “High” in the context of the “Sensitivity” parent tag. Furthermore, all items containing the term “patent” will he tagged as “Medium” in the context of the “Sensitivity” parent tag.

As keyword lists are essentially CSV files, it is not recommended to use commas in queries, because they result in a different interpretation of the keyword list. If a comma in a query is required, you can wrap the entire query in quotes.

The tags specified in the CSV file will be mapped to or result in the creation of top-level tags.

15.1.19. MD5 and Message Hash

Intella can calculate MD5 and message hashes to check the uniqueness of files and messages. If two files have the same MD5 hash, Intella considers them to be duplicates. Similarly, two emails or SMS messages with the same message hash are duplicates. With the MD5 and Message Hash facet you can:

  1. Find items with a specific MD5 or message hash and

  2. Find items that match with a list of MD5 and message hashes.

Specific MD5 or message hash
You can use Intella to search for files that have a specific MD5 or message hash. To do so, enter the hash (32 hexadecimal digits) in the field and click the Search button.

List of MD5 or message hashes
The hash list feature allows you to search the entire case for MD5 and message hash values from an imported list. Create a text file (.txt) with one hash value per line. Use the Add… button in the MD5 Hash facet to add the list. Select the imported text file in the panel and click the Search button below the panel. The items that match with the MD5 or message hashes in the imported list will be returned as a single set of results (one cluster).

Structured vs Legacy message hash
In Intella 2.2.2 a more flexible algorithm for calculating message hashes has been introduced: structured message hashes. Cases that have been created with Intella 2.2.2 or newer will use the structured message hashes by default. Cases that have been created with older versions will keep using the old algorithm until the case is fully re-indexed. That re-index is required to calculate the Body Hash, one of the four components of structured message hashes, for applicable items. You can configure the algorithm for message hashes from the Preferences window.

Structured message hash
Intella’s structured message hash exists of four components: Header, Recipients, Body, and Attachments. By default, the calculated message hash will be based on all four components, but you can deselect any of these to make deduplication of message items less strict. For example, when the Recipients component is deselected, an email with a Bcc header will be considered as a duplicate of an email without that header (assuming all other components are equal).

For email items, the following data is included in the four components of a structured message hash:

  • Header – The sender, subject and sent date.

  • Recipients – The To, Cc and Bcc header values.

  • Body – The email’s text body.

  • Attachments – The combined MD5 hashes of all email attachments.

All upper case/lower case differences of textual data is ignored, and for the email body all whitespace and formatting characters (Unicode categories C and Z) are ignored too. The sent date is rounded down to full minutes. For attachments that are embedded emails, the structured message hash of that email is used, instead of the MD5 hash.

When deduplicating a set of items, Intella will select the item that has the lowest item ID for each set of duplicates. This item may be missing specific details that are present in duplicates. This effect becomes more likely when a less strict message hash configuration is used.

Legacy message hash
The message hash is calculated by calculating the MD5 hash of a list of concatenated item properties. For emails the following properties are used:

  • From, Sender, To, Cc and Bcc headers.

  • Subject header.

  • Date header.

  • Email body.

  • All other MIME parts (attachments, nested messages, signatures, etc.).

For SMS, MMS, and other types of chat messages such as Skype and WhatsApp messages, the following parts are used:

  • The sender information.

  • The receiver information.

  • The textual content of the message.

When certain headers/properties occur multiple times, all occurrences are used.

A difference between email message hashes and chat message hashes is that the hashing procedure for emails will simply skip missing values, whereas for chat messages all fields need to be present to calculate a hash.

These message hash computation methods have the benefit that they are source-agnostic: a specific email message always gets the same message hash, regardless of whether it is stored in e.g. a PST, NSF, Mbox or EML file. Message hashes can therefore find duplicates across a variety of mail formats and be used to deduplicate such a diverse set of mail formats.

When one of the copies has a minor difference, the email will get a different hash and be treated as different from the other occurrences. A good example is a bcc-ed email, as the bcc is only known by the sender and the recipient listed in the Bcc header. Therefore, these two copies will be seen as identical to each other but different from the copies received by the recipients listed in the To and Cc headers. Another example is an archived email which has one or more attachments removed: it will be seen as different from all copies that still have the full list of attachments.

Install a free tool such as MD5 Calculator by BullZip to calculate the MD5 hash of a file. You can then search for this calculated hash in Intella to determine if duplicate files have been indexed.
Use the “Export table as CSV” option in the Details table to export all MD5 and message hashes of a selected set of results to a CSV file.

15.1.20. Item ID Lists

In the Item ID Lists facet, you can load a list of item IDs, to automate the searching with sets of previously determined item IDs, e.g. obtained by exporting the Details table to a CSV file.

An item ID list is a text file in UTF-8 encoding that contains one item ID per line.

Once loaded into the case, you can select the list name and click Search. The result will be a single result set consisting of the items with the specified IDs. Invalid item IDs will be skipped.

15.1.21. Language

This facet shows a list of languages that are automatically detected in your items.

To refine your query with a specific language, select the language from the list and click the Search button.

When Intella cannot determine the language of an item, e.g. because the text is too short or mixes multiple languages, then the item will be classified as “Unidentified”. When language detection is not applicable to the item’s file type, e.g. images, then the item is classified as “Not Applicable”.

Language detection is applicable for the following media types (see the "Types" facet):

  • All types in the Documents → Word Processing category

  • All types in the Documents → Presentations category

  • All types in the Documents → Spreadsheets category

  • In the Documents → Other documents category: Plain Text Document, HTML Document, XHTML Document, Comma-separated Values File

  • Communication → E-mail → Email Message

  • Others → Source code → XML Document

To export the detected languages and their counts to a CSV file, right-click anywhere in the facet area and select “Export values…”.

15.1.22. Size

This facet groups items based on their byte size.

To refine your query with a specific size range, select a value from the list and click the Search button.

To export the size ranges and their counts to a CSV file, right-click anywhere in the facet area and select “Export values…”.

15.1.23. Duration

This facet reflects the duration of phone calls listed in a cellphone report, grouped into meaningful categories.

To export the duration ranges and their counts to a CSV file, right-click anywhere in the facet area and select “Export values…”.

15.1.24. Device Identifier

This facet groups items from cellphones by the IMEI and IMSI identifiers associated with these items. Please consult the documentation of the forensic cellphone toolkit provider for more information on what these numbers mean.

This facet also supports the filtering options described in the Email Address section.

To export the device identifiers and their counts of a given Device Identifier facet branch to a CSV file, right-click on the node of that branch and select “Export values…”.

15.1.25. Export Sets

All export sets that have been defined during exporting are listed in this facet. Searching for the set returns all items that have been exported as part of that export set.

To export the export set names and their counts to a CSV file, right-click on the node of that branch and select “Export values…”.

15.2. Requiring and excluding facet values

image

Facet values can be required and excluded. This allows for filtering items on facet values without these values appearing as individual result sets in the Cluster Map visualization.

To require or exclude items based on a facet value, select the value, and click on the arrows in the Search button. This will reveal a drop-down menu with the Require and Exclude options.

15.2.1. Requiring a facet value

Requiring a facet value means that only those search results will be shown that also match with the chosen required facet value.

image

For example, see the image on the right. The “Enron” search term resulted in 2,752,284 items, but after applying the E-mail category with its 1,668,416 items as a requirement filter, only 1,239,282 items remain.

When multiple required filters are specified, a drop-down list will appear in the Required list:

desktop search required multiple

Depending on the option selected here, the results will differ:

  • Any – the results belong to at least one of the required sets. Filtering uses the union of all required sets.

  • All – the results belong to all of the required sets. Filtering uses the intersection of all required sets.

The "any" option can be of use when filtering on e.g. document types or custodians, where the sets tend not to overlap. The "all" option can be of use when filtering on criteria from multiple facets, where it effectively does a drill-down search without all searches being visualized in the Cluster Map.

15.2.2. Excluding a facet value

Excluding a facet value means that only those search results will be shown that do not match with the chosen excluded facet value.

Example: The user selects the facet value “PDF Document” and excludes this facet value with the drop-down menu of the Search button in the facet panel. The searches panel in the Cluster Map shows that “PDF Document” is excluded. As long as this exclusion remains, all result sets and clusters will not hold any PDF Documents. Empty clusters will be filtered out.

15.2.3. Limitations

Excludes are often used to filter out privileged items before exporting a set of items, e.g. by tagging items that match the privilege criteria with a tag called “Privileged”. This tag can then be used to filter the privileged items from the results. This methodology has a few limitations that one needs to be aware of:

  1. The required and excluded sets are applied on the result list shown in the Details view. The original, unfiltered results are still present in the Cluster Map, where they can be selected for review in the Review tab via the right-click menu. In this case, the unfiltered set is reviewed.

  2. The items that are filtered out may still be reached through traversal of the item hierarchy (e.g. in the Previewer’s Tree tab) or through other search methods.

  3. When exporting an email to e.g. Original Format or PST format, it is exported with all its attachments embedded in it. The same applies to a Word document: it is exported intact, i.e. with all embedded items. Therefore, when an attachment is tagged as “privileged” and “privileged” is excluded from all results, but the email holding the attachment is in the set of items to export, the privileged attachment will still end up in the exported items. The solution is to also tag both the parent email and its attachment as “privileged”. The tagging preferences can be configured so that all parent items and the items nested in them automatically inherit a tag when a tag is applied to a set of items.

When filtering privileged information with the intent to export the remaining information, we recommend that you verify the results by indexing the exported results as a separate case and checking that there are no items matching your criteria for privileged items.

16. Cluster Map

The Cluster Map shows search results in a graphical manner, grouping items by the queries that they match. This chapter will help you understand how this visualization works.

image

16.1. Understanding a Cluster Map

The figure above shows a graph with two labels and three clusters. The larger, colored spheres are called clusters. They represent groups of items such as emails and files. The queries entered by the user are shown as labels and are used to organize the map.

Every cluster is connected to one or more labels. In this Cluster Map, we see that the user has evaluated two keyword searches: one for the word “buy” and one for the word “sell”. The Cluster Map shows these two result sets, using the search terms as their labels:

  • “buy” returned 128 items and is represented by the red edges.

  • “sell” returned 67 items and is represented by the blue edges.

The colored edges connect the clusters of items to their search terms, indicating that these items are returned by that search term. For example, this Cluster Map shows that there are 16 items that were returned by both the “sell” and “buy” queries, 51 items that contain “sell” but not “buy”, and 112 items that contain “buy” but not “sell”.

It is important to understand that the set of results for “buy” are split across two clusters: one that also matches “sell” and another that only matches “buy”. The same split happens for the “sell” results.

When a third keyword search for “money” is added, the graph changes as follows on our data set:

image

In the middle is a single cluster of 6 items that is connected to all three labels. This represents the 6 items that match all three search terms. There are three clusters of 9, 10 and 20 items, each connecting to two labels but not a third. They represent the items that match two out of the three search terms. Finally, three large clusters at the periphery represent all items that only match the search term that it is connected to.

A Cluster Map can always draw a reasonable picture of up to three search terms: the above map shows the maximum complexity that such a graph may have. Beyond three search terms the graph may become too complex and cluttered to be meaningful. That is why the Cluster Map has a second visualization mode called Sets. This mode can be chosen by clicking on the Sets mode in the toolbar. When the user enters more than seven queries, the Cluster Map will automatically switch to that mode.

In Sets mode, the three result sets are visualized like this:

image

Here, each result set is depicted as a single rounded square shape with the label and number of items on top. The size of the square is related to the number of items in the set: bigger means more items. Furthermore, all sets are grouped by their order of magnitude indicated on the left – in this case all result sets are of the same order of magnitude. The overlap between sets is no longer visualized until the user selects one of the sets.

Sets mode can scale to a much larger amount of result sets. The following image is a visualization of 16 result sets, divided among four different orders of magnitude. Adjacent groups get alternating colors for better separation. Note that the visual size of the result sets, indicating the number of items in each set, is only comparable within the group.

image

16.2. Manipulating Cluster Maps

image

The result sets created with the current query are listed in the box at the top right corner of the Cluster Map panel. To remove a result set from the Cluster Map, click on the remove icon (red X) in the list.

To clear the Cluster Map - remove all result sets - and start a new search, click the Clear button in the terms list.

If the Cluster Map regeneration takes too long, you can stop the process by clicking the Stop button.

To view and open the individual items in a cluster or result set, first click on the cluster or label. This will list the items in that set in the Details view below. From there the items can be opened by a single or double click, depending on the currently selected view mode of the Details view.

16.3. Options

When the Cluster Map is in Clusters mode, the Filters button in the toolbar will be enabled. When this toggle button is selected, the graph is filtered to show only the clusters with the most connections. These could be seen as the most relevant result clusters. This filtering has no equivalent in Sets mode and therefore is disabled in that mode.

The last button in the toolbar indicates whether the graph should be shown at normal size (with scrollbars if necessary) or be scaled to fit in the visible space. For Clusters mode, the fit to size mode makes the most sense. For Sets mode, showing at normal size is often preferable, especially when dealing with lots of result sets (tens or more).

The current visualization can be exported as a transparent, 24-bit PNG image. To do so, choose the “Cluster Map…” option in the Export menu.

17. Histogram

The Histogram shows how a set of search results is spread over time. This tells the user when certain communications or other activities took place.

Another important use case of this visualization is to find anomalies in the data. Any gaps in the chart may indicate shortcomings in the data collection process, e.g. due to a device or disk that should have been included. However, it can also indicate custodians intentionally or unintentionally withholding data, e.g. by deleting emails prior to the collection.

The image below shows the results of a keyword search for “invoice”, grouped by year:

image

This histogram immediately tells you that the timestamps of the items matching “invoice” range from 1980 to 2016, with the majority between 2000 and 2009, and peaking in 2007.

The date range that Intella looks for ranges from 1969 to the current year plus two years. This will filter out bogus dates that are far in the past or future. Large, real-life data sets will often show items on specific dates like January 1st, 1970, January 1st, 1980 or similar “round” dates. These are typically caused by default date values used in some applications. Future versions may make it possible to filter out such dates.

Initially the Histogram will show the items grouped by year. It is possible to toggle between years and months by using the supplied toggle buttons.

17.1. Date attributes

The date attribute used to create the chart is configurable. By default, the Family Date is used. This typically gives a good sense of the “when” of an evidence set, without dates in older email attachments and files giving a warped sense of the relevant dates.

To use a different date, click on the button showing Family Date (if the default has not been altered yet) and use the checkboxes in the popup to indicate the desired date attribute(s).

Multiple date attributes may be used, e.g. “Sent” and “Received”. In this case it may occur that an item is represented in multiple bars in the histogram, because one date attribute may be in one bar’s date interval and another date attribute falls into a different bar’s date interval. To get a sense of the volume of the data, it is best to use a single date attribute.

17.2. Selections

To see the items in a specific year or month, simply click on the bar representing that year or month. This will list the items in the Details view below.

To see all items in a range of years or months, drag the mouse cursor across the chart. This will show a marker in the background, indicating the selected date range. All bars overlapping with that date range will be selected.

image

18. Geolocation

The Geolocation view shows the (estimated) locations of all search results that have geolocation information on the world map.

image

This chapter will help you understand how this visualization works.

18.1. Basics

Currently, geolocation data is extracted from the following sources:

  • Images – GPS coordinates in the EXIF metadata.

  • Cellphone reports – available information depending on the device model, extraction utility and extraction method.

  • Emails – through geolocation lookup of the sender IP.

  • Google Maps URLs – e.g. from browser histories and bookmarks.

Using this information, a set of search results can be mapped to a set of geographic coordinates, roughly representing the “where” of the found items.

Any items that do not have any geolocation information associated with them are omitted in this view.

Showing each item’s estimated location on the map would make the view very cluttered. Items laying in the same area are therefore grouped into clusters, shown as a blue circle in the screenshot above. The number in a cluster represents the number of items whose geolocation falls in that area.

When zooming in, the geographic size of what constitutes the “same area” will be reduced, resulting in clusters getting split up into smaller clusters. Zooming out of the map consolidates clusters into fewer and larger clusters again. This cluster management allows the user to inspect specific locations in detail.

          Zoom in                   Zoom out

image image

The clustering is determined by imposing an invisible grid on the map and bundling all items in a grid cell into a cluster. When a grid cell contains only a single item, an icon will be placed on the map instead, representing that single item. For image items, this icon will be a thumbnail of the image. This gives the user a quick overview of the images located in a specific area. For all other items, the item’s file type icon will be shown.

image image

18.2. Interaction

Zooming can be done using the control buttons in the top-right toolbar or by using the mouse wheel.

To pan (move sideways) in a zoomed map, move the mouse while holding down the left mouse button.

To inspect the content of clusters, the user can select:

  • A single cluster, by clicking on it.

  • Multiple clusters, by clicking on them while holding down the CTRL key.

  • Multiple clusters, by dragging in the map while holding down the right mouse button.

The contents of the selected cluster(s) will be displayed in the Details view below the Geolocation view.

The view also responds to facet selections. The image below shows the Geolocation view, showing all images in a specific case. Selecting the “500 KB – 1 MB” category in the Size facet has turned all cluster discs into pie charts. The solid, dark blue areas in the pie charts represent the items that match this Size facet category. This way, the user can quickly identify the geographic spread of the matching items in the map without having to change the set of queries. This works for every available facet that supports value selections.

image

In the figure below, the map is shown after an “Include search” is added using the same Size facet query:

image

18.3. Resources

Intella may need two resources to make the most out of the Geolocation visualization.

18.3.1. Tile server

By default, Intella uses tiles (images containing parts of the map) that are embedded in Intella to construct the world map. This makes it possible to use the Geolocation view without any configuration and without requiring an Internet connection to download these tiles.

Due to the enormous size of a complete tile set covering all zoom levels of the entire world map, the embedded tile set is limited to the first 6 zoom levels. As a rule of thumb, this usually shows the major cities in most countries, but it will not let you zoom in to see where in the city an item is located.

To zoom in beyond that zoom level, a connection to a tile server is needed. This can be a public tile server or one located in your network. See the Preferences section on how to configure a tile server.

A tile server may not only let you zoom in and create more fine-grained maps, it can also let you apply a different map rendering, e.g. a map containing elevation data, infrastructural information, etc.

18.3.2. IP geolocation database

To determine the geolocation of emails, Intella uses the chronologically first IP address in the Received email headers (i.e. the one nearest to the bottom of the SMTP headers). Next, a geolocation lookup of that IP address is done using MaxMind’s GeoIP2 or GeoLite2 database. These databases are not distributed with Intella and therefore one needs to be installed manually.

See the Preferences section on how to acquire and install an IP Geolocation database.

18.4. Caveats

While the Geolocation view can quickly give a unique and insightful overview of a data set, there are some aspects of geolocation visualization to be aware of. Geolocation data is approximative by nature and manual verification of the findings will always be required. This is not an Intella limitation; it is inherent to the complexity and unreliability of the systems producing the geolocation information. Make sure that you are fully aware of these aspects and their consequences before relying on the findings.

18.4.1. GPS coordinates

GPS coordinates, such as obtained from the EXIF metadata of images or location-bound items extracted from cellphones, are usually quite accurate. However, they are subject to the limitations of GPS:

  • In the best-case scenario, the accuracy is typically in the range of several meters. The accuracy can be lower or coordinates can even be completely wrong when the GPS hardware cannot receive a good signal (e.g. in the direct vicinity of buildings), due to hardware limitations of the GPS device (the theoretical maximum precision possible varies between devices) or simply due to bugs and hardware faults in the device.

  • The same applies to comparable satellite-based navigation systems such as GLONASS.

  • Geolocation coordinates may also have been determined using other techniques, e.g. based on geolocation information about nearby Wi-Fi networks and cell towers.

  • Some devices combine several of these techniques to improve accuracy and coverage. Therefore, what is commonly referred to as “GPS coordinates” may not have been established through GPS at all.

  • Coordinates may have been edited after the fact by a custodian using an image metadata editor. A set of different images with the precise same coordinates may point in that direction. This may be harmless, e.g. to fill in the coordinates of images taken with a camera that does not have GPS functionality.

18.4.2. IP geolocation

The determination of an email’s geolocation by using its sender’s IP address is imprecise by nature, typically even more so than GPS coordinates. First, the determined Source IP address may be incorrect due to several reasons:

  • Some email servers mask such IP addresses. Instead, it may in fact be the second IP address of the transport path that is being used.

  • A web email client (e.g. Gmail used through a web browser) may have been used to send the email.

  • The IP address may have been spoofed.

  • The IP address may not reflect the sender’s location due to the use of a VPN, Tor, etc.

Second, IP geolocation databases are typically never 100% accurate and the accuracy varies by region. See MaxMind’s website for statistical information on their accuracy. Reasons for this imprecision are:

  • The geolocation of an IP address may change over time.

    Take this into account when indexing an older data set!
  • Some IP addresses may only be linked to a larger area like a city or even a complete country, yet the precise coordinates may give a false sense of GPS-style precision.

  • The techniques behind the collection process for creating this database introduces a certain amount of imprecision.

18.4.3. Tile servers

Using a public tile server may reveal the locations that are being investigated to the tile server provider and anyone monitoring the traffic to that server, based on the tile requests embedded in the retrieved URLs.

To use a public tile server, you need to ensure that you comply with the tile server’s usage policy. This is your responsibility, not Vound’s.

18.5. Attribution

We are grateful for obtaining the data we have used for the embedded tiles generation from the OpenStreetMap project, © OpenStreetMap contributors. See http://www.openstreetmap.org/copyright for more information on this project.

The tile set is made available under the Open Database License: http://opendatacommons.org/licenses/odbl/1.0/. Any rights in individual contents of the database are licensed under the Database Contents License: http://opendatacommons.org/licenses/dbcl/1.0/.

19. Social Graph

The Social Graph is another visualization of search results, showing who participated in the emails, phone calls and instant messages in the search results.

image

19.1. Basics

The social graph is revealed by clicking on the Social Graph button in the Results toolbar. Next, just enter any type of query and the results will be displayed as a social graph. When switching from a populated Cluster Map to the Social Graph, the graph will start loading immediately with these results.

image

When multiple searches have been evaluated, the graph is based on the union of all search results, with the Includes and Excludes applied. In other words, the social graph is based on the same items that are visible in the Cluster Map at the same moment.

To see the conversation items in this result set that relate to a specific contact, i.e. that have that contact as sender or recipient, click on the node representing that contact. To see the items in this result set that have been sent between two contacts, click on the edge between those nodes. In both cases the Details panel below the Social Graph will display these items.

note that the Timeline view is a natural fit to display the items associated with a node or edge in the Social Graph. All items are sorted by their natural date, and you can easily see the participants involved in the individual items.

When a person sends a message to several people, this will result in several edges in the graph. Therefore, you may encounter the same item several times when browsing the graph and selecting edges.

19.2. Interaction

The toolbar at the top left offers four buttons for managing the zoom level of the graph:

  • Zoom in.

  • Zoom out.

  • Reset zoom level to the default value.

  • Change the zoom level to make the graph fit the available screen space.

The fifth button shows or hides all node labels. When set to “hide”, only the labels of selected nodes and their connected nodes are displayed.

Finally, the sixth button collapses the toolbar and the Searches panel, revealing any graph structures beneath it. Click on the button that appears in the top-right corner to expand these panels again.

The lower parts of the toolbar are used to control the information shown in the graph.

The Edges filter filters out edges, and nodes with no remaining edges, based on identities. See the Identities section for a description of how to model and use identities. The Edges filter has three possible values:

  • All – leaves all edges in the graph.

  • At least one Identity – filters out edges that are not connected to an Identity.

  • Only Identities – only preserves edges that have Identity nodes on both ends.

The Node labels drop-down list controls what should be shown as node labels:

  • Name – Show only the contact name; use the address (e.g. the email address) if there is no contact name.

  • Network – Show only the address; use the contact name if there is no address.

  • Name and Network – Show both the contact name and address.

By default, only contact names are shown, as these are typically shorter than (email) addresses and lead to less cluttered displays.

The following mouse operations are supported:

  • Drag a node to improve readability.

  • Click on a node to highlight that node and the nodes connected to it.

  • Use Ctrl-clicking to select and highlight multiple nodes.

  • Hold down the right mouse button while dragging to scroll (pan) the graph.

  • Right-click on a node to add that email address to an Identity.

  • Right-click on a node or edge to review the corresponding items in the Review tab.

The graph can be exported to a PNG file by using Export Social Graph…​

19.3. Limitations

The graph displays a warning when your result set contains more than 700 unique emails, as this may take considerable time to create. Future versions will address this in various ways.

As the Edge filter is applied during rendering, not during graph construction and layout calculation, the chosen filter has no impact on the calculation speed. This may also be addressed in a future version.

20. Details panel

To inspect the contents of the visualization, the user can select a cluster or result set by clicking on it. Its contents will be displayed in the “Details” panel below the map. This panel contains a list of the items that can be presented in four modes:

  • Table view

  • List view

  • Thumbnails view

  • Timeline view

20.1. Table view

image

The table view displays the results as a table in which each row represents a single item and the columns represent the attributes such as title, date, location etc.

The set of attributes to display can be customized with “Toggle visible table columns” button - the right button of the Details Panel Control.

Click on a table column header to sort the table by specific item attributes.

Toggle the table row icon to control the presence of separators between rows that have different values in the column that is being sorted on, thereby grouping rows that have the same value. Note that not all columns support this separator.

20.1.1. Adding and removing columns

With the “Toggle visible table columns” button in the Details toolbar you can add and remove columns in the table, by (de)selecting column names in the popup that shows when you click the button.

To add or remove columns in the table, select or deselect the checkboxes in the “Available columns” list. Selected columns are indicated in the “Selected columns” list in the center of the window. You also can remove columns using the red remove buttons in this list, or by selecting one or more columns and pressing the Delete key.

image

To add or remove an entire column group, (de)select the group header checkbox in the “Available columns” list.

To add all available columns, select the “Check / uncheck all” checkbox. Unselect this checkbox to clear the selected columns list.

To search for a specific column or column group by name, type the first few letters of the name in the “Filter column names…” field. The “Available columns” list will be filtered to show only the matching columns and groups. Clear the search field to remove the filtering in the list.

Click “Ok” to confirm the changes and close the Column Chooser window. Press “Cancel” to close the window box without making any changes in the table.

For contacts, e.g. email senders and receivers, this window lets the user choose whether to display the contact name, the email address or both. The chosen setting will affect the table sorting when the involved columns are used to sort the table.

The contents of the date columns can be adjusted to show their time zones:

  • When set to “Always”, an explicit time zone always accompanies each date and time value.

  • When set to “For different sources”, time zones are only shown when items from different sources are being shown in the table.

  • “Do not show” suppresses all time zones.

20.1.2. Available columns

General columns:

  • All Locations: The locations of all duplicate items in the case (including this item).

  • Certificate: The certificate with which an encrypted item could be decrypted.

  • Contact name: The name of a contact encountered in a PST file or as a vCard file.

  • Content Analysed: Shows whether the item has been subjected to analysis by at least one of the Content Analysis categories.

  • Custom ID: The Custom ID of the item, assigned to it by a Generate Custom IDs task.

  • Custom Family ID: The Custom Family ID of the item, assigned to it by a Generate Custom IDs task.

  • Decrypted: Shows if an item is encrypted and Intella has decrypted it.

  • Direct Child IDs: The item IDs of the direct children of this item.

  • Direct Parent ID: The ID of the item’s direct parent item.

  • Document ID: The ID as imported from a load file. This ID is maintained for cross-reference purposes.

  • Duplicate Locations: The locations of all duplicate items in the case (excluding this item).

  • Duplicates: Shows the number of duplicates of an item within the case.

  • Embedded Image: Indicates whether the item is an embedded image extracted from an email, Microsoft Office document, XPS document, or PDF document. See the Features facet section for a precise definition of this category.

  • Encrypted: Shows if an item is encrypted.

  • Exception: Shows if an item had one or more issues indexing properly.

  • File Name: The name of a file in the file system, in an archive or used as an attachment name.

  • Geolocation: The geolocation (longitude, latitude) of an item, if any.

  • Has Geolocation: Indicates whether the item has geolocation information associated with it.

  • Item ID: The ID used internally in Intella’s database to refer to this item.

  • Language: The language of the item’s text. The language field is left blank when the language cannot be detected automatically. When the language could not be determined, e.g. because the text is too short or mixes various languages, the value shown will be “unidentified”. Item types that inherently do not have a language, e.g. images or archives, show the “not applicable” value.

  • Location: Name of the location in the original evidence data where the item is stored. For example, an email in a PST file would have a location that would start with the folder and file name of the PST file, followed by the mail folder path inside that PST file.

  • MIME type: The type of an item per the MIME standard.

  • Native ID: The native ID of the item. Currently only HCL/IBM Notes UNID (Universal Notes ID) are listed here. This column may be used for other native ID types in the future.

  • Near-Duplicate Group: The name of the near-duplicate group that the item belongs to.

  • Near-Duplicate Master Item: The ID of the master item of the near-duplicate group that the item belongs to.

  • Near-Duplicate Score: The similarity score of the item in its near-duplicate group.

  • Near-Duplicates: The number of near-duplicates of this item (other items in the near-duplicate group that the item belongs to).

  • BegAttach / Parent ID: The ID of a parent document (or first ID in the family) as imported from a load file. This ID is maintained for cross-reference purposes.

  • EndAttach: The last ID in the family as imported from a load file. This ID is maintained for cross-reference purposes.

  • Password: The password with which an encrypted item could be decrypted.

  • Recovered: Indicates whether the item has been recovered. See the Features facet section for the definition of the Recovered status.

  • Size: The item’s size in bytes.

  • Source: The name of the Intella source that holds the item. Typically, this is the root folder name or the name of the mail container file (e.g. PST or NSF file).

  • Source Path: The path to the evidence, e.g. the PST or NSF file, or the root folder of a Folder source. This helps reviewing items when dealing with a lot of evidence files – the name of the evidence file and the derived source name may not hold enough information to easily discern the origin of the information.

  • Subject: The subject of an email or document item – note that some document formats can have both a title and a subject.

  • Title: The title of a document item.

  • Text Snippet: Text summary containing at max first 1000 characters of item’s content. This column is especially usable for reviewing Chat message item types as it makes it possible to examine communication from different channels side by side in the Details view.

  • Top-Level Parent: Indicates whether the item is a top-level parent.

  • Type: The item’s human-readable type, e.g. “MS PowerPoint Document” or “Email Message.”

  • URI: Uniform Resource Identifier, the identifier used internally by Intella for the item in addition to the Item ID.

Email-specific columns:

  • All Receivers: The combined list of To, Cc and Bcc agents.

  • All Senders: The combined list of From and Sender agents.

  • Attached: Whether this item is an attachment to an email, conversation, or document.

  • Attachments: Shows the file names of an email’s attachments.

  • Bcc: The addresses in the Bcc header.

  • Bcc Count: The total number of unique blind carbon copy email recipients (Bcc).

  • Cc: The addresses in the Cc header.

  • Conversation Index: The value of the ConversationIndex field of the item (the PR_CONVERSATION_INDEX MAPI property commonly used in PST, EDB and MSG files), or the value of the Thread-Index header.

  • Email Thread ID: When the item has been subjected to email thread analysis, this shows the ID assigned to the email thread in which the item has been placed.

  • Email Thread Name: When the item has been subjected to email thread analysis, this shows the thread name assigned to the email thread in which the item has been placed. Often this is the “root” of the subject line that is common between the emails in the thread.

  • Email Thread Node Count: When the item has been subjected to email thread analysis, this shows the number of nodes in the email thread in which the item has been placed.

  • From: The addresses in the From header.

  • Has Attachments: Emails that are marked as having attachments.

  • Has Internet Headers: Emails that have regular SMTP headers. When this is not the case, information about e.g. the sender, receiver and dates may still be obtained from other fields, depending on the source format.

  • Inclusive: When the item has been subjected to email thread analysis, this shows whether the item has been marked as inclusive.

  • Message Hash: Shows the Message Hash for emails and SMS messages. This hash is used for deduplicating emails and SMS messages in a manner that works across different mail formats and phone data source types.

  • Message ID: Shows the Message ID extracted from email messages.

  • Missing Email Referent: When the item has been subjected to email thread analysis, this flag indicates that the threading process has detected that the email item is a reply to another email or a forwarded email, but the email that was replied to or that has been forwarded is not available in the case.

  • Non-Inclusive: When the item has been subjected to email thread analysis, this shows whether the item has been marked as non-inclusive.

  • Recipient Count: The total number of unique email, chat, and cellphone recipients.

  • Sender: The addresses in the Sender header.

  • Source IP: the determined source IP of the email.

  • Threaded: Shows whether the item has been subjected to email thread analysis.

  • To: The addresses in the To header.

  • Unread: Shows if an email item was unread at the time of indexing.

  • Visible Recipient Count: The total number of unique visible email, chat, and cellphone recipients (To, Cc).

Cellphone-specific columns:

  • All Phone Numbers: phone numbers relevant to a phone call, regardless of whether it is an incoming or outgoing call, combined with phone numbers found in contacts.

  • Chat Accounts: all instant messaging accounts (Skype, WhatsApp, but also SMS and MMS phone numbers) that have been used to send or receive a chat message.

  • Chat Protocol: all chat protocols used in a message or conversation (e.g. SMS, MMS, Skype, Teams).

  • Chat Receivers: all instant messaging accounts used to receive a chat message.

  • Chat Senders: all instant messaging accounts used to send a chat message.

  • Conversation ID: ID associated with conversation as found in the evidence data.

  • Duration: how long the phone call took.

  • IMEI: The International Mobile Station Equipment Identity (IMEI) number of the phone from which the item was obtained.

  • IMSI: The International Mobile Subscriber Identity (IMSI) associated with the item.

  • Incoming Phone Numbers: phone numbers used for incoming phone calls.

  • Intella Conversation ID: uniquely generated Conversation ID (changes on each indexing attempt)

  • Message Count: shows the number of messages in conversation items, i.e. items that bundle all messages between two or more participants in a single day.

  • Outgoing Phone Numbers: phone numbers used for outgoing phone calls.

File- and document-specific columns:

  • Contributor: The name(s) of the contributor(s) of a document. These are typically authors that edited exiting documents.

  • Creator: The name(s) of the creator(s) of a document item. These are typically the initial authors of a document.

  • Empty document: Shows that the item has no text while text was expected. Example: a PDF file that contains only images.

  • File Extension: the file extension of a file, e.g. “doc”, “pdf”.

  • Irrelevant: Indicates whether the item is classified as “Irrelevant”. See the Preferences section for the definition of the “Irrelevant” category.

  • MD5 Hash: The MD5 hash that uniquely identifies the item.

  • OCRed: Shows whether an OCR method has been applied on this file.

  • Page Count: the number of pages of the item, as reported by the metadata found in the original evidence item. I.e., this is not a verified value, and is only available for certain document formats that support such a metadata attribute.

  • Shadow Copies: the number of other versions of the item located in volume shadow copies.

Columns containing dates:

  • Called: The date a phone call was made.

  • Content Created: The date that the content was created, per the document metadata.

  • Content Last Modified: The date that the content of the item was last modified, per the document-internal last modified date.

  • Due: The due date of a task.

  • End Date: The end date of an appointment, task or journal item.

  • Family Date: The family date of the item. Family dates build on primary dates and take the item hierarchy into account. The family date of an item is defined as the primary date of its top-level parent, i.e. all items in an item family have the same family date. Sorting on Family Date sorts by this date, but also puts attachments and nested items right behind their parent. This is strictly enforced, i.e. two item families with the same family date are not intertwined. This makes it possible to review items in chronological order while maintaining a sense of their context. Certain types of items are skipped when determining the family root, namely all folders, mail containers, disk images, load files and cellphone reports.

  • File Created: The date a file was made, according to the file system.

  • File Last Accessed: The date a file was last accessed, according to the file system.

  • File Last Modified: The date of the last time the file was modified, according to the file system.

  • Last Printed: The date a document was last printed, according to the document-internal metadata.

  • Primary Date: The date that is the best match for the given item. Default or user-defined rules are used to pick the most appropriate date attribute based on the item’s type.

  • Received: The date the item was received.

  • Sent: The date the item was sent.

  • Start Date: the start date of an appointment, task or journal item.

  • Visited: The last visited date of an item obtained from a browser history or Windows registry.

Review-related columns:

  • All Custodians: The custodians of all duplicate items in the case (including this item).

  • Batches: Shows the batches an item has been put it. This can happen when a case is reviewed in Intella Connect, our web-based companion product.

  • Comments: Shows if an item has comments. When this is the case, a yellow note icon is shown in the table. Hover over the icon to see a tooltip with the comments attached to the item.

  • Custodian: shows the name of the custodian associated with this item.

  • Duplicate Custodians: The custodians of all duplicate items in the case (excluding this item).

  • Exported: Shows if an item has been exported.

  • Flagged: Shows a column at the left side of the table that indicates if an item is flagged. Click the checkbox if you want to flag an item.

  • Opened: Shows if an item has been opened in its native application.

  • Previewed: Shows if an item has been opened in the previewer.

  • Queued for redaction: Shows if an item has been queued for mass redaction.

  • Redacted: Indicates whether the item has been redacted.

  • Tags: Shows the tags connected to an item.

Analysis columns:

  • The columns in this group represent built-in and custom Content Analysis categories. See the “Content analysis” section for more information on their meaning.

    Note that the Location branch in the Content Analysis facet corresponds with the “Geographical Location” column in the table. This naming difference is to distinguish it from the other Location column indicating the location of the evidence items.

Tag groups (optional):

  • These columns are created for every top-level tag with sub-tags. When selected, the corresponding column shows the tags within that part of the tag tree. The column will be named after the top-level tag.

    For example, when a tag named “Relevancy” has been created with subtags “Relevant”, “Non-Relevant” and “Privileged”, the tag group in the column chooser would be called “Relevancy”. Enabling it would add a column named “Relevancy” to the table, with the “Relevant”, “Non-Relevant” and “Privileged” tags as possible values for those items that have been tagged as such.

Export (optional):

  • When items have been exported using the export set functionality, a column will be made available for every export set, holding the export IDs within that export set.

Custom columns (optional):

  • Custom columns are created either during load file import or in the Custom Columns dialog. See the “Custom Columns” section for more details.

20.1.3. Reorganizing table columns

The columns can be reorganized by dragging a column header to a different location in the table.

Alternatively, you can use the Column Chooser window to reorder the columns:

  • To move a column to a different location, drag and drop it in the “Selected columns” list. You can use this method for multiple columns selected in the list.

  • To move a column to one position up or down in the “Selected columns” list (left or right in the table), use the “Move Up” and “Move Down” buttons. This also can be applied to multiple selected columns

20.1.4. Table presets

When clicking the “Ok” button, all changes made to the table configuration are stored in the current table preset, selected in the Column Chooser window.

Every case has its own set of presets. When a new case is created, it contains two presets with predefined column sets for regular table display (“Default”) and for CSV exporting (see the “Export to a CSV file” section). You can create as many presets as needed and switch between them using the drop-down list in the Column Chooser.

To save the current state of the Column Chooser as a new preset, click the “Save As…” button and enter the preset name. You can enter an existing preset name to overwrite it.

To undo all changes made in currently opened Column Chooser window and revert it to the original state of a selected preset, click the “Reset” button.

To remove an existing preset, select it in the drop-down list and click the “Delete” button. This operation is available only if there are two or more presets defined in the case.

20.1.5. Sorting the list

By clicking on a column header, the search results will be sorted alphabetically, numerically, or chronologically, depending on the type of information shown in that column. By clicking the header once more, the sort order will be reversed. Clicking one more time will remove the sorting, letting the results be displayed in their original order.

Sorting on the Family Date column is implemented as a compound sorting on two columns. Items are first sorted by the Family Date itself and next by the Hierarchy criterion. This process is transparent to the user and results in attachments and embedded items always getting placed directly after their parent item, which can greatly simplify the review of the items.

Sorting by multiple columns can be achieved by holding the Ctrl button while clicking on the column names. Any additional clicked column will be added to the list of sorting criterions. When two items cannot be sorted using the values from the first column (because the values are identical), the second column will be used, and so on.

Besides clicking on column headers, you can alter the sorting with the “Sort table” button. This opens a dialog that lets you select the sorting columns and the sort order per column (ascending/A-Z or descending/Z-A). This dialog lets you use all the columns available, regardless of whether the column is currently present in the table.

Furthermore, this dialog offers a sort criterion called “Hierarchy”, which is not available as a table column. Sorting on this criterion puts the items in hierarchical order, e.g. an email is directly followed by its attachments.

20.1.6. Showing a conversation

Right-clicking a message item and selecting the Show > Conversation option will display a new result set in the Cluster Map, showing all messages that are part of the conversation. This includes replies and forwarded messages.

The messages in a conversation set are determined by matching keywords in their subject lines and by inspecting values in the “In-Reply-To” and “References” email headers. More specifically:

  1. The algorithm takes the item’s subject and reduces this to the “base subject” by stripping all prefixes like “Re:”, “Fwd:”. It supports common prefixes for several languages.

  2. Next, it determines the set of Message IDs mentioned in the item’s “Message-ID”, “In-Reply-To” and “References” email headers.

  3. It does a Boolean AND query for the words in the base subject, restricting the search to the “title” field.

  4. It narrows this set down to all items that have at least one of the Message IDs in the determined set in their headers, i.e. regardless of the specific header name it is associated with.

Due to how this method is implemented, Show Conversation may find a different set of items than the Email threading method. For example, single thread emails that have the same subject are typically returned using Show Conversation. A future Intella release may unify these two functionalities.

20.1.7. Showing the family items

To determine all family items of a set of selected items, select all relevant items in the Details table, right-click on one of them and click the Show > Family menu item. This will add a new result set in the Searches panel containing all family items of the selected items.

The family of an item is defined as its top-level parent and all descendants of that parent in the item hierarchy, including folders. The definition of a family used by the “Show family” option is the same as used in the Keywords tab and the Family Date attribute.

20.1.8. Showing the unique families

This operation finds the families with their top-level items deduplicated. This is identical to performing the following steps on a given set of items:

  • Find all top-level parents of the items.

  • Deduplicate these parents.

  • Extend the resulting set with their direct and indirect children.

To determine all items belonging to unique families of selected items, select two or more items in the Details table, right-click on one of them and click the Show > Unique Families…​ menu item.

In the dialog box, you can configure the following options:

  • Deduplicate by custodian: If selected, the top-level parents of families are deduplicated for each custodian separately, thus allowing duplicates in different custodian sets.

  • Include folders: If selected, folder items are included in the produced families.

Clicking on the OK button will add a new result set in the Searches panel.

20.1.9. Showing the child items

To determine all items nested in an item, right-click on the item and select Preview > Item. Next, switch to the Tree tab to see the full hierarchy, including all child items.

To determine the children of a set of selected items, select all relevant items in the Details table, right-click on one of them and click the Show > Children option. This will open a dialog that asks you what children to put in the result set, as child items may also again contain child items.

20.1.10. Showing the parent items

Right‑click an email attachment and select the option Preview > Parent E-mail to view the email message that contains the selected item. This feature looks up the parent item recursively until it reaches an email item.

To determine the parent of a set of selected items, select all relevant items in the Details table, right-click on one of them and click the Show > Parents… option. This will open a dialog that asks you whether to produce the top-level or direct parents, and what to do with items that have no parent.

See the search preferences for settings related to how the top-level and direct parents are determined.

20.1.11. Showing native ID duplicates

To determine all items that have the same Native ID as a specific item, right-click on the item and select Show > Native ID Duplicates. See the “Adding and removing columns” for the definition of the Native ID column.

20.1.12. Showing shadow copies

To determine all items that are other versions of a specific item extracted from volume shadow copies, right-click on the item and select Show > Shadow Copies.

20.1.13. Showing statistics

To show the statistics of a set of items, select the items in the table, right-click on the selection and choose the option Show > Statistics. The following statistics can be shown:

  • Cumulative file size. This estimates the storage space required to export these items in their native format.

  • Total number of document pages. Note that this only includes items that have a page count in their metadata (e.g. PDF and Word documents). Further note that this metadata may be unreliable.

  • Number of OCRed items.

  • Number of redacted items.

  • Number of encrypted items.

20.2. List view

The List view displays the results in a form similar to conventional web search engines. Select the third button in the Details toolbar to switch to this view.

image

For each item, the title and other important metadata will be displayed, as well as a fragment of the document text, if any text has been extracted from this item. When Intella currently is displaying keyword search results, the selected text fragment will show the keyword matches and their context.

The title is normally displayed in a light green color; dark green indicates that the item has been previewed before by the current user.

If the item has any tags applied to it, these will be shown on the right as blue labels. To flag an item, use the checkbox on the left.

Items can be selected by clicking, Ctrl-clicking and right-clicking. Right-clicking on any item reveals the same popup as used in the Table view.

20.3. Thumbnails view

The Thumbnails view displays the thumbnails of the image and video items detected within a selected cluster.

By default, the thumbnails include images embedded in email bodies, email attachments and images inside documents. Select "Hide Embedded Images" option to show only direct selected items.

Use the zoom slider or "Ctrl+mouse wheel" action to change the thumbnail size.

Hover over the thumbnails with your mouse cursor to see a summary of the data connected to the image. You can flag an image with the checkbox below the thumbnail.

When the mouse cursor is hovered over a thumbnail, the small zoom button becomes visible. Click on that button to activate the quick image preview in a pop-up panel. Click anywhere to close the quick preview panel.

When you double-click a thumbnail, the image will open in the previewer.

image

The Thumbnails view will work a lot smoother when you let it pre-generate the thumbnail representation of all images in the case in advance. This can be done by selecting “Generate Thumbnails” from the File menu.

20.4. Timeline view

The Timeline view shows a chronological representation of email communications, phone calls and SMS/MMS messages.

image

The left pane shows the senders and receivers, i.e. email addresses or phone numbers, with their communication plotted chronologically. Every edge in the timeline view represents a communication and points to the receiver of that communication.

The node color represents the role a contact (i.e. an email address or phone number) has in a communication, e.g. sender or caller. Click the Legend button to see an explanation of all node colors that can occur.

When displaying emails, it may occur that an email appears to have two senders. That happens when the email has both a From and a Sender header. As in most circumstances the From header is of primary interest, the visualization of the Sender headers is by default disabled. It can be enabled by clicking on the Options button and checking the “Display the Sender header in addition to the From header” checkbox.

When you click an arrow, the arrow, the connected arrows, and the connected squares will be highlighted. When you double click an arrow, the email will show in a preview window. TIP: Export a timeline by choosing Export > Timeline… from the menu. The timeline will be saved as a PNG image.

20.5. Deduplication, irrelevant and non-inclusive items

With the Deduplicate button, duplicates are removed from the search results. This is based on the MD5 and message hashes of the results: when two items have an MD5 or message hash in common, only one of them is shown.

If one or more custodians are defined in the current case, the Deduplicate button has two deduplication options:

  • Global: Default behaviour; all items in the case are deduplicated against each other.

  • By Custodian: Deduplicate each custodian’s item set separately. Duplicate items belonging to different custodian sets will all be shown.

When deduplicating a set of items, Intella will select the item that has the lowest item ID for each set of duplicates. This item may be missing specific details that are present in duplicates. This effect becomes more likely when a less strict message hash configuration is used.

Similarly, the Hide Irrelevant button removes all items marked as Irrelevant during indexing. See the Preferences section for information on the Irrelevant Items category.

Finally, the Hide Non-Inclusive button filters out items marked as non-inclusive by the email thread analysis.

When used in the Thumbnails view, which shows both the images in the selected results as well as any images nested in those results, the result is filtered. In other words: first the set of images in the item set is determined, then it is extended with the set of nested images, and finally the deduplication and irrelevant item filters are applied on this combined set.

21. Previewing results

desktop previewing items

21.1. Overview of the Previewer

When you double-click an item, it will display in a separate Previewer window - unless you specify in the Preferences that the file should be opened in its native application instead.

The Previewer allows you to inspect, flag, and tag the item, to explore its relations with other items, and to export the item for later use.

The Previewer will show several tabs, presenting differ aspects of the item, such as Contents, Preview, Headers, Raw Data, Properties, Attachments, etc. The set of tabs will differ from item to item, depending on the type of item that you selected and what information is available for that item.

21.2. The Toolbar

The toolbar on the right of the window contains options for producing and annotating the current item, as well as navigating to other items and starting new searches that use this item as a starting point.

At the top is a panel with buttons for producing the current item in several formats:

  • Export
    This button opens the “Export result as” dialog. Enter a name and location if you want to store the item. This exports the item in its original format.

  • Print Tab
    This button opens a print dialog that shows the contents of the selected tabs (Contents, Headers, Thumbnails, etc.) of the item. Click the print button on the lower right to print the item. Alternatively, the print output can also be saved as a PDF document.

  • Print Report
    This button opens a print dialog that shows the native rendering of the item with a minimal amount of metadata. If the item has attachments, you are asked if these should also be printed. Click the print button on the lower right to print the item. Alternatively, the print output can also be saved as a PDF document.

  • Open in Application
    This button opens the item using the computer’s default application (e.g. a PDF file would be opened with Adobe Acrobat Reader if that is the default PDF viewer on your computer).

  • Open Containing Folder
    This button is enabled for items that represent files in the file system and provides quick access to it. When clicked, Windows Explorer will open, show the file’s folder and select the file in the folder.

The next panel lets one iterate over all items in the Details view from which the Previewer was launched:

  • Previous and Next buttons
    Go to the next or previous item in a list. Alternatively, you can also use the keyboard shortcuts Alt+right-arrow to go to the next item, and Alt+left-arrow to go to the previous item.

This functionality is not available when the Previewer was launched by clicking in the Cluster Map, from the Tree tab of another Previewer, etc.

The next two panels are for annotating the current item:

  • Tag button
    Opens the tag space where you can add new tags to your case and select a tag from a list of existing tags.

  • Quick tag buttons
    You can assign a tag to a quick tag button, or remove an existing tag. If no tag is pinned to a Quick tag button, it is randomly associated with one of the recently used tags by default.

  • “Go to next item after tagging” check box
    When this check box is selected, clicking the quick tag buttons will switch the Previewer to the next item in the list (if there is one).

  • Flagged
    Select this check box to flag the previewed item. You might want to flag an item for organizational reasons. For example, to keep track of the items that you have reviewed in the case.

The next panel holds actions for navigating to and searching for related items:

  • Preview Parent
    Use this button to open the parent item in a previewer window. A parent item contains one or more items. Example: Pictures found in a Microsoft Word document are separate items in Intella. The Word document is the parent item for these pictures. The same is true for items found in archive file, such as a ZIP file: The archive file is the parent item for these items.

  • Preview Parent Mail
    Use this button to open the parent email item in a previewer window. A parent email item contains one or more items. Example: A picture attached to an email is a separate item in Intella. The email is the parent for the picture. This button is visible only when one of the parents of previewed item is email.

  • Preview Parent Conversation
    Use this button to open the parent conversation item in a previewer window. A parent conversation contains one or more message. Example: A chat message which is part of conversation is a separate item in Intella. The conversation is the parent for the chat message. This button is visible only when one of the parents of previewed item is conversation.

  • Show Family
    Use this button to search for all items in the same family as the current item.

  • Show Children
    Use this button to search for and display the children associated with the item being viewed in the previewer. When selected, a search result with the associated children of the selected items will be available in the Cluster Map panel. The label of the cluster will be “Children of [file name]” or “Children of [subject].”

An example of a child item would be an attachment of an email. Intella views emails and attachments as separate items. The attachment would be the child of the parent email.

Child items can have child items of their own. Depending on the option that you select, the Show Children shows either only the directly nested children or all children in the tree.

  • Show Conversation
    Based on the Subject of an email and certain other email headers, Intella can find items that are part of a conversation. Click the button Show Conversation to show all these items in the Cluster Map panel.

The label of this cluster will be “Conv: [email subject].” The email subject is the email subject of the item in the previewer.

  • Show Duplicates
    When an item has duplicates in the case, click Show duplicates to display these duplicates in the Cluster Map. The label of this cluster will be “Duplicates of [file name]” or “Duplicates of [subject]”.

  • Smart Search
    Smart search lets one search for items that are similar to a selected item.

    It determines a set of keywords in the selected item that have a high information value. Typically, these are keywords that occur often in the selected document but are not common words across the case or in any of the supported languages, which makes them representative for the content of the selected document.

    image

    Using the Smart Search dialog one can then find other documents that share these keywords and therefore have a good statistical chance of being related to the selected document.
    A slider is provided that the user can use to set a threshold: the lower the threshold, the more documents are returned but at the cost of less relevance to the set of keywords.
    Checkboxes are provided to control which item fields should be used when determining the set of keywords. This way one can restrict the search for similar items to e.g. the document or message body only.

The next panel controls redaction:

  • Redact
    When this button is clicked, a PDF is generated for the current item and shown in the Redaction tab. See the section on Redaction for more details.

  • Queue for Redaction
    When clicked, the current item with its currently highlighted keyword search hits are stored in the redaction queue. This queue can be processed later, resulting in the redaction PDFs being generated and the visual areas where these hits appear in this PDF being determined. This allows for the quick review of keyword hits, without having to wait for the redaction PDFs to be generated on demand when one moves from item to item.

  • OCR
    This button will OCR the current item, using the embedded ABBYY FineReader Engine. This button is disabled when the OCR engine does not support this file type. See the “Using FineReader Engine” section for more details.

Finally, the last panel contains options related to the Contents tab:

  • Hide seen paragraphs
    When selected, paragraphs that have been marked as Seen by the user are removed from the text, only leaving an “eye” icon in the left margin as an indication that a paragraph has been removed there. Click on the eye to bring back the text.

  • Colorize paragraphs
    When selected, paragraphs marked as Seen by the user are displayed as grayed out text.

  • Highlight Content Analysis results
    When selected, entities from Content Analysis categories are emphasized with color highlighting.

21.3. Tabs

The tabs show the various aspects of the current item. The set of tabs shown for an item can differ from item to item, depending on the item type and which information that item holds.

When moving from one item to the next using the Next and Previous buttons, the current tab will stay selected – if that tab is also available for the next or previous item.

When a specific tab is never used in a case, its visibility can be toggled using the Previewer’s View menu. The benefit of this is a less crowded user interface and shorter loading time.

Keyword matches
When the current item has any keyword matches, the tabs containing one or more of the keywords change their appearance:

  • The tab name will show with a bold blue font and contain a number indicating the number of hits.

  • When the tab contains text (not metadata properties), like the document text or email headers, it will get a status bar at the bottom listing the found keywords and providing buttons to jump from one match to another.

  • When the tab contains text and has a scrollbar, the location of the keyword matches will be marked in the scrollbar using yellow indicators.

Hit highlighting in the Preview and Redaction tabs may be overzealous in highlighting the matching terms when using phrase or proximity queries. Generally, all occurrences of the individual terms are highlighted, not just the text parts that resulted in these query matches. E.g. the query "big car" will result in all occurrences of "big" and "car" being highlighted. This is a limitation of the technologies used to render these tabs. This limitation is not present in the other tabs that support hit highlighting (Contents, Headers, Raw Data, Properties and Comments).

Next, we explain which tabs can occur.

21.3.1. Contents

This tab shows the body of an item, e.g. the message in an email or the text inside a Word document. The Contents shows a limited set of stylistic elements such as bold, italic, and underlined text, tables and lists. However, text is always drawn as black text on a white background, as to reveal all extracted text. For a native rendering of the item use the Preview tab (when available).

If the item text is too long, it is truncated in the previewer for performance purposes. Click on the “Show full text” button to view the complete item text. Note though that there is also a limit on the maximum amount of text that is subjected to full-text indexing. See the note on the “ItemTextMaxCharCount” setting in the Source Types section.

When the item is an image, this tab will show the image’s content. An extra toolbar is then provided, allowing for zooming, rotating, and flipping the image. If the image has extracted text, it will be shown in a separate tab called “Extracted Text”, next to the Contents tab.

When the item is a video, this tab contains the thumbnails of static frames extracted from the video content.

When an item is encrypted and could not be decrypted, the Contents tab will show an image of a lock, to explain why no text could be shown.

Handling paragraphs
When the “Analyze paragraphs” option was selected during source creation, extra UI elements will be shown in the left margin. These UI elements indicate the start and end of the paragraphs that Intella has detected. They can be used to collapse and expand the paragraph. The UI elements are omitted for very short paragraphs (typically one-liners).

Furthermore, a popup menu will be shown when the user right-clicks on a paragraph, offering the following options:

  • Mark the paragraph as Seen, or back to Unseen. This grays out all occurrences of this paragraph in all items, facilitating the review of large amounts of long and overlapping documents such as email threads with lots of quoted paragraphs.

  • Mark all paragraphs above or below the current paragraph as Seen or Unseen.

  • Search for all items in which this paragraph occurs. All items that contain the selected paragraph will be returned, ignoring small variances such as white spaces.

  • Mark the paragraph for exclusion from keyword search. This can be used to suppress information present in lots of items but with little relevance to the investigation, such as email signatures and legal disclaimers. Consequently, keyword queries containing terms such as “confidential” and “legal” are more likely to return meaningful results.

Detected Objects

When the "Image Analysis" was executed on this item, this item is an image and objects have been detected in it, then extra UI elements will be shown in the image itself. These UI elements indicate the detected objects in this image with a rectangle at coordinates where the object was found and a description of the object. Transparency of the rectangle and label depend on the detected object’s confidence score.

Detected objects

The objects that have been searched for in Search tab will be highlighted with different color (see "Persons" on the picture above).

The “Min. confidence” slider on the Previewer toolbar specifies the value of a threshold filter applied to the detected objects on the current image preview. The objects evaluated with lesser confidence scores than the current threshold will not be highlighted in the preview.

Highlighting of detected objects can be turned off completely by unchecking the "Highlight Content Analysis results" checkbox.

21.3.2. OCR and OCR Preview

When the item has been OCR-ed by any of the supported methods, the OCR text will be shown in an OCR tab. The searchable version of the document will be shown in an OCR Preview tab.

Cases that were originally made with Intella 2.0.x versions or older and that have been transformed to a 2.1.x case or newer may still show the OCR text as part of the document text in the Contents tab.

21.3.3. Preview

This tab shows the item as if it was opened in its native application. The Preview tab is only shown when the format of the current item is supported, and the Contents tab is not already showing it in its native form. The following file formats are supported:

  • Emails (when the email contains an HTML body; excluding MSG emails)

  • Legacy MS Office formats (doc, xls, ppt)

  • New MS Office formats (docx, xlsx, pptx)

  • RTF

  • HTML

  • PDF

  • XPS

  • CSV and TSV files

  • WordPerfect

  • Open Office (Writer, Calc, Impress)

When previewing emails, only images that are already bundled with the email are shown. Any images that a mail client would load from a web server are shown as static icons. When there are any such missing images, a “Show external images” button appears. Clicking this button will load the images from the servers and show them embedded in the email representation. Note that loading these images may constitute a violation of investigation policies.

21.3.4. Headers

This tab shows the complete header of the email item. This tab is only shown when you open an email item.

21.3.5. Raw Data

The content of this tab depends on the item type. For example, in case of PST emails the low-level information obtained from the PST is listed here. This typically includes the transport headers (shown on the Headers tab) and the email body, but also a lot more. In case of vCard files the raw vCard contents is displayed here.

All this information is also searched through when using a keyword search. This may lead to additional hits based on information in obscure areas that Intella does not process any further.

21.3.6. Properties

This tab shows a list of properties connected to the item. Examples are Size, MIME Type, Creator and Character Set. The list of properties shown depends on the type of the item and what data is available in that item.

To copy all the text to the clipboard click Copy all.

Hover over the question marks at the right-hand side with your mouse and see a short definition of each property.

21.3.7. Attachments

This tab lists the attachments of an item.

When you double-click an attachment, or select it and click View, it will be opened in new Previewer window.

Attachments will be reported also in the case of conversations despite they are not representing direct children of Conversation item but are instead attachments of Message items consisting previewed conversation.

21.3.8. Thumbnails

This tab shows thumbnails of the images (jpg, png, gif etc.) attached to an item or embedded in a document, e.g. the images embedded in a MS Word document.

Select the checkbox below the image to flag a thumbnail.

When you double-click a thumbnail, the image will be opened in a new previewer window.

Thumbnails will be reported also in the case of conversations despite they are not representing direct children of Conversation item but are instead attachments of Message items consisting previewed conversation.

21.3.9. Tree

This tab shows the location of the reviewed item in the item hierarchy (entire path from root to descendants), as well as all its child items.

The file names and subjects are clickable. You can also right-click and choose to either select all above or select all below, or simply select items manually, to assign them to a tag.

The Tree tab can optionally display Custom ID and Custom Family ID columns. That can be useful to determine an item’s role in the item family. The presence of these columns can be toggled in the Preferences window.

21.3.10. Email Thread

This tab visualizes the email thread in which the currently previewed email is located. A blue border indicates the current email.

Each type of icon in this visualization has a special meaning. To see a basic explanation of the icons, click the Legend icon. The icons have the following meaning:

  • Inclusive Email – this email is part of the set of emails that a reviewer should read, in order to read everything in the thread.

  • Non-Inclusive Email – all content of this email is also present in at least one of the replies or forwards.

  • Missing Email – indicates that the existence of an email could be derived from references found in other emails, though the email itself could not be found in the case.

  • Duplicate Emails – indicates that one or more duplicates exist of this email.

  • Reply – indicates that the email was sent as a “Reply” to another email.

  • Reply All – indicates that the email was sent as a “Reply all” to another email.

  • Forward – indicates that the email contains a forwarded email.

  • Attachment – indicates that the email has one or more attachments.

The user can double-click on the nodes in the visualization. This opens that email in a separate Previewer. When the node represents a set of duplicates, one of these duplicates is opened.

To tag all items represented in the visualization, click the Tag Thread button.

To export the graph as a PNG image, click the Export button.

21.3.11. Entries

This tab shows the list of items found in an archive file, e.g. a ZIP or RAR file.

When you double-click an item in the list or select it and click View, it will be opened in a new Previewer window. However, when the entry is a sub-folder inside the archive, its content will be opened in the same 'Entries' tab. Double-click the '..' entry at the top of the list to return to the parent folder.

21.3.12. Comments

This tab lists the reviewer comments attached to the item. Every comment has an author name and time stamp, and the option to Edit or Delete the comment.

Note that this is not related to the comments such as found in the MS Word document metadata.

21.3.13. Words

The Words tab lists all words/terms extracted from this item, together with the following information:

  • The search field the term belongs to: text, title, path, etc.

  • The frequency of the word in this document and document field.

  • The number of documents having this term in the same field.

This list can be used to diagnose why a certain document is or is not returned by a certain query.

The list can be exported as a CSV file by right-clicking anywhere in the table. Right-clicking also lets you evaluate a query with the right-clicked term.

21.3.14. Actions

This tab shows the list of actions performed on the item. The action’s date and the user that triggered the action are shown in the list. Actions listed are:

  • Previewed – the item was opened in the previewer.

  • Opened – the item was opened in its native application.

  • Exported – the item was exported.

  • Tagged with – the item was tagged with the specified tag.

  • Flagged – the item was flagged.

  • Commented – the item was commented.

  • OCRed – the item has text content imported from OCR.

  • Redacted – the item was redacted.

  • Imported text – additional text was imported to the item via the “-importText” command-line option.

21.3.15. Geolocation

This tab shows the item’s geographic location on the world map. It is only present when the item has geolocation information (Longitude and Latitude properties) associated with it.

See the section on the Geolocation view for the correct interpretation of this information, including its caveats.

21.3.16. Redaction

This tab is only visible after the Redact button in the toolbar has been clicked (see above), or when a redaction PDF has been generated for this item through some other means. See the section on Redaction for a detailed explanation of the functionality in this tab.

21.3.17. Analysis

This tab is visible if one or more entities from Content Analysis or Image Analysis categories are identified in this item’s content. The tab lists all entities found in the item content, organized by category.

Click the Search button to query for other items containing the selected entity.

21.3.18. Near-Duplicates

This tab is only visible for items included in a near-duplicate groups (see the "Near-duplicates Analysis" section), except for the group’s master items and their exact duplicates in those groups.

The tab visualizes the differences between the text content of the current item and the master item in its near-duplicate group. Information about the near-duplicate group (name, master item ID, and the current item score) is visible on the top panel.

Different text blocks (paragraphs) are marked with red and green colors, indicating occurrences specific to the current and to the master items, respectively. Visibility of the different blocks is controlled with two checkboxes ("Occurs only in this item" and "Occurs only in the Master item"). The regular black-on-white text represents the text blocks that the two items have in common.

The Near-Duplicates tab uses simplified text formatting with the most of the text styling stripped out. Therefore, the view may differ from what one can see in the Contents and Preview tabs.

22. Reviewing results

Users can open a Review tab on a specific set of items. This top-level tab combines the functionality of the Details view and the Previewer. On the left of the tab is the item list, either as a Table, List, Thumbnails, or Timeline view. On the right of the tab is an embedded Previewer. Selecting items in the list on the left displays them in the Previewer on the right. This makes a quick and efficient review of a list of items possible within a single window.

image

To open a Review tab, the user can:

  • Right-click on a cluster or result set in the Cluster Map and select Review > Review n items.

  • Right-click on a result set in the Searches list and select Review > Review n items.

  • Right-click on a selected item range in the Table, List or Thumbnails view and select Review > Review n items.

It is possible to have an unlimited number of Review tabs. Note that it is even possible to open a new Review tab by selecting items in an existing Review tab, right-clicking and selecting the Review option. Likewise, it is also possible to open a Previewer by double-clicking on an item in a Review tab.

A Review tab can be closed by clicking on the Close icon next to the tab’s name.

Currently, the set of Review tabs is not persisted: closing Intella will close all Review tabs.

The embedded Previewer features all features of the standalone Previewer, i.e. one can tag, flag, comment, redact, etc. Clicking on any of the search links, e.g. Show Conversation, will let Intella switch automatically to the Search tab.

23. Identities

The Identities functionality lets one build an “address book” of the persons of interest in a case. An identity bundles the communication aliases used by a person, such as email addresses, phone numbers and chat accounts, into a single unit. The identity is given a Full Name and can be annotated with other properties. This information is used to enhance the querying and display of items in other parts of the user interface.

The Identities tab lists the currently defined identities and offers various options for adding and editing identities. This way, investigators can enter the information they know about suspects and other people involved in a case.

image

The Identities facet makes it possible to query for all items linked to an identity. An identity query combines the results of the queries for the individual email addresses, phone numbers and chat accounts into a single item set. The result is a holistic view of the communication of that person, regardless of the media and aliases used for that communication.

image

In case of email addresses, an Identity query also finds items where the email address occurs in the item text. It therefore casts a wider net than merely looking at senders and receivers.

Other facets and displays that use identities to improve their content and presentation:

  • The Email address facet bundles the email addresses of an identity into a single node.

  • The Chat Accounts and Phone Number facets do the same with their values.

  • The Social Graph groups nodes representing aliases of the same identity into a single node.

The result is both a simpler and more accurate presentation of the information. Future Intella releases will extend this to other parts as well, such as the Timeline and Email Thread views.

This unit of information is called an “identity” rather than a “person”, even though in practice it often will correspond on-one-one with a person. This was chosen because linking an identity to an actual person is still an important investigative step to make. For example, people may use pseudonyms to hide their identity, use different names in different languages, or may use different names to indicate different cultural roles. Also, the same or similar name may not necessarily imply the same person. How identities relate to persons and what their aliases are is therefore for the investigator to determine.

23.1. Adding identities

The simplest way of creating a new identity is by clicking the Create Identity button on the left side of the Identities tab. A form opens in the middle where the information of the new identity can be entered.

The Full Name will be used to display the identity in the displays mentioned above. Optionally, a Role, Organization and Note can be added.

Beneath these fields are three lists of identity aliases:

  • Email addresses

  • Chat accounts

  • Phone numbers

  • Authors

For a manually created identity, these lists will initially be empty. There are two ways of populating the lists:

  • Click the Add button above the respective list. A dialog will open that lets the user add several aliases at once, one on each line. The values can also be copied from the Clipboard.

image

  • Click the Suggestions button above the respective list. A dialog will open that shows suggestions derived from the Full Name and any aliases that have already been added manually. For example, after adding “John Smith” as the Full Name, email addresses with “john” and “smith” or similar names may be shown.

The suggestions are sorted by their estimated relevancy, so usually the best matches are located at the beginning of the list. Please note that these are only suggestions; manual verification of the addresses is always required.

After editing the identity, click Save Identity to make the additions or changes persistent. This will also update the information in the left side of the Identity Details box with the following statistics:

  • Emails – the number of emails that this identity is involved in. This includes occurrences of the email address facet in the document text.

  • Chats – the number of chats items that this identity is involved in.

  • Phone calls – the number of phone calls that this identity is involved in.

  • Case coverage – the number of all items associated with this identity, relative to the total number of items in the case.

Each of these statistics can be clicked, resulting in a query for these items being launched in the Search tab.

Once an identity has been saved, it can also be assigned an avatar image. Click on the circle with the person icon and select an image in the file chooser that opens. To remove the avatar image, right-click on it and select “Remove avatar”.

To remove an identity from the Identities list, right-click on it and select Remove. After confirming the deletion, the Identity will be permanently removed. There is no way to undo this operation.

23.2. Identity suggestions

image

As an alternative to the manual definition of identities, Intella offers a suggestions mechanism. Using patterns found in the evidence data such as similarly looking email addresses, it will suggest identities. These can be reviewed by the user and added to the Identities list, or merged into an existing identity. Only explicitly added or merged identities will be used in the other displays.

To see Intella’s identity suggestions, click the Calculate Identity Suggestions button in the lower right corner of the Identity tab. This operation may take some time depending on the case size, typically seconds to minutes. After completion, the Suggestions list on the right will be populated with the top 40 suggested identities derived from the evidence data. The suggested identities are sorted by the number of items that they cover, with the identity covering the largest number of items being at the top of the list.

Click on a suggested identity to expand that node and see the aliases associated with that identity.

Right-click on the empty space in the Suggestions list to open the Organize menu. This lets one sort the Suggestions list by item count (the default) or by name.

A suggested identity can be added to the Identities list by selecting it and clicking the Create Identity button beneath the Suggestions list. One can also drag identities from the Suggestions to the Identities list.

When the Suggestions list uncovers additional aliases for an existing identity, the existing and suggested identities can also be merged. To do this, select the related identities in both lists and click the Add to Identity button. Alternatively, the suggested identity can be dragged into the aliases lists of the currently displayed identity. All aliases of the suggested identity will then be merged with the existing lists of aliases of the current identity. Be aware that, once the user clicks the Save Identity button, there is no way to undo this operation easily: one would have to manually select the unwanted addresses and remove them.

23.3. Facets

The Identities facet makes it possible to search for all items related to an identity. When searching for an identity, it queries for all items that have any of the identity’s aliases as sender/receiver/caller/callee/etc. Effectively, it gathers all messages in which the identity is a participant through one of its aliases.

Several other facets use the identity information to optimize their content and presentation.

The Email Address facet uses identities to organize the email addresses in each of its branches. Without any identities defined in the case, each branch holds a flat list of email addresses sorted alphabetically. Once identities are defined, the list will group the email addresses in that branch by the identities they belong to. All addresses of the same identity are bundled in a node labeled with the identity’s Full Name. The email addresses of that identity are placed beneath it as sub-nodes. The identity nodes are always placed at the start of the list, before the addresses that are not associated with an identity.

Both the identity nodes and their email address sub-nodes can be used as query parameters. When searching using an identity node, the query effectively searches for all email addresses listed as sub-nodes in the tree. Note that this does not search for other types of aliases such as phone numbers and chat accounts; that is what the Identity facet is for.

The Chat Account, Phone Number and Author facets optimize their content and presentation in a way that is conceptually similar to what the Email Address facet does.

Each of these facets offers right-click menus for editing identities, e.g. when the user spots an email address that clearly does not belong to the identity that it is placed under.

23.4. Social Graph

The Social graph uses identities to bundle address nodes together that belong to the same identity. This can greatly improve the graph’s readability.

To simplify the graph even further, an Edges filter is available with three options:

  • All – shows all edges, regardless of what types of nodes they connect.

  • At least one Identity – shows edges that are connected to at least one identity.

  • Only identities – shows only edges that connect identities.

This can have an enormous effect on the graph complexity. Consider this original social graph:

image

Restricting the graph to only those edges that connect at least one Identity results in this graph:

image

Further restricting it to edges connecting only identities results in this image:

image

The Social graph can also be used to edit identities on the fly. Right-click on a non-identity node (e.g. an email address node) and select “Add to Identity”. The dialog that opens can be used to pick an existing identity to assign the address to. The graph will be updated accordingly.

Note that changes to the identities made in the facets or in the Identity tab do NOT lead to an automatic update of the Social Graph.

23.5. Caveats

Saved searches involving identities are not portable.

The identity functionality is a recent addition to Intella. We anticipate making refinements and extensions to this functionality in future releases and welcome your feedback.

24. Chat messages

24.1. Overview

This section describes the processing and rendering of chat messages, and how it differs from the way other artifacts are handled.

Let’s look at an example of how chat conversations are processed. Suppose that we have a chat message database, holding a conversation called “Main Chat” that spans over three years. In such a case Intella will create artificial Conversation items, based on the Indexing Options that control how such Conversation items are to be constructed. Let’s say that these were set like this:

  • Present chat messages as: Conversations and Messages

  • Split chat conversations: Per year

  • Limit number of messages per conversation: 100

image

Each of these three Conversations items will contain chat messages that were sent in the same calendar year. The start and end dates of conversations will be set to the sent date of the first and last chat messages respectively.

Furthermore, children Chat Message items will be produced for each individual chat message in this conversation. The conversation item will contain the message texts of all its child Chat Message items.

When the maximum number of messages per conversation item is reached, the conversation will be split further into additional conversation items.

Constructing conversations items out of the individual chat messages has following benefits:

  • The presentation of Conversation items inside the previewer makes reviewing chat data effective and efficient, as it mimics how a chat client will display the chat messages.

  • Having Conversation items make it possible to use AND and OR search operators and proximity queries when searching for text across chat message boundaries.

Producing separate Chat Message items has the following benefits:

  • Chat Message items can be individually tagged, flagged, and exported.

  • Chat Message items can be listed inside the Details view, making it possible to see how they relate to other item events in the case. For example, website visits, emails, phone calls, etc. that took place right before or after the moment that chat message was sent or received.

The Text snippet column can be especially useful when reviewing Chat Message items inside the Details view, as it shows the first 1,000 characters of each item.

The Present chat messages as indexing option controls whether: 1. Both Conversation items and Chat message items will be produced 2. Only Conversation items will be produced 3. Only Chat message items will be produced

The Split chat conversation Indexing Option controls how chat messages are bundled into Conversation items:

  • Per Hour - conversations are split by calendar hour

  • Per 12 Hours - conversations are split at noon and midnight

  • Per Day - conversations are split by calendar day

  • Per Week - conversations are split by calendar week

  • Per Month - conversations are split by calendar month

  • Per Year - conversations are split by calendar year

The maximum number of messages bundled in a single Conversation item can be controlled through the Limit the number of messages per conversation Indexing Option . The maximum value is 1,000 messages per conversation.

Altering these values will affect reviewing and exporting to PDF at the later stage. A reasonable default setting is “Per day” splitting, capped by a maximum of 100 chat messages per conversation item. When exporting such a Conversation item, the exported document will contain at most 100 chat messages, thereby not producing unnecessarily large PDF documents. Reviewing such Conversations in the Previewer is also more straightforward, as the reviewer is not overwhelmed with many messages inside the previewed conversation.

Note that (re-)indexing of the chat data is needed to let changes in these options take effect.

All attachments associated with Chat Message items will also be reported as attachments in the Conversation item. The number of attachments can thus be large if there are many Chat Message items with attachments present in the data.

The Number of recipients property of the Conversation item will be set as [number of all participants – 1] to make it consistent with how the number of participants is calculated in the case of emails: the sender is not included in the recipient count.

In chat conversations extracted from a Cellebrite phone dump, the amount of participants is derived from the entire conversation (all days) and then applied to all daily parts. This is different from Skype chat, which may have a different amount of participants per day.

24.2. Previewing

One can preview both the Conversation items and the Chat Message items nested in them. In the case of previewing a Conversation item, the whole conversation thread will be rendered, with links to the preceding and succeeding Conversation items. When previewing a Chat Message item, only that single message will be rendered.

The Raw Data tab will contain the raw data based on which the conversation and message preview representation was constructed. The data that is shown here depends on the type of evidence data, e.g., a Skype SQLite database, a Cellebrite UFDR report, etc.

24.2.1. Previewing of Conversation items

When a Conversation item is opened in the Previewer, there are a number of differences with how other item types are displayed:

image

  • A checkbox is rendered in front of each chat message. This makes it possible to flag the corresponding Chat Message item straight from within the conversation view. This is useful if only specific messages in the conversation are to be exported.

  • Double-clicking on the Chat Message body (the gray area) will open the corresponding Chat Message item in a separate Previewer window.

  • The attachments of all Chat Message items contained in the conversation will be reported in the Attachment tab of the Conversation item as well.

  • Thumbnails of all Chat Message attachments contained in the conversation will be reported in the Thumbnails tab.

  • Additional info shown in the Contents tab:

    • Start Date: indicates the date of the first chat message covered in this Conversation item.

    • End Date: indicates the date of the last chat message covered in this Conversation item (i.e., not necessarily the end date of the entire conversation).

    • Chat Accounts: shows all chat account that were participating in this particular conversation

When an attachment is an image and its binary content is present, it will be rendered as an inline image in the Conversation item’s preview for easier review.

There are a few additional conversation-related properties reported in the Properties tab: Number of recipients, Number of visible recipients, Protocol, Messages count

More information about these can be found by hovering the mouse over the question mark icon next to the property.

24.2.2. Previewing Chat Message items

The below image shows how a Chat Message item is previewed:

image

Note that there is no checkbox in front of the message text, as Chat Message items can be flagged by using the Flagged checkbox in the previewer’s toolbar on the left. Another reason for this is to make a visual distinction between Conversation items and Chat Message items.

If you want to preview the Conversation item that this Chat Message item corresponds to, you can use the “Preview Parent Conversation” action in Previewer, or navigate to it through the Tree tab.

In the case of Chat Messages items, the Properties tab contains the following chat message-related information: Recipients Count, Visible Recipient Count, Chat protocol

24.3. Exporting of Chat conversation and Chat Message items

One should be aware that there are two ways of exporting styled chat messages / conversations.

  • Export as PDF

  • Export as Report

24.3.1. Export as PDF

When exporting as PDF, the Conversation item or Chat Message item will be exported as it is rendered in the Previewer.

In the case of Conversation items, the whole conversation fragment covered by this Conversation item will be exported. There is no way to export only specific chat messages this way, but you will be able to redact it as you would with any other item.

image

In the case of Chat Message items, each individual chat message gets exported as a PDF.

image

The overall process of exporting to PDF is not explained here, as it is identical to exporting any other type of item to PDF.

24.3.2. Export as Report

The main difference of exporting as an Item Report, compared to the PDF export type, is that one can export either the whole conversation or just particular chat messages in it.

To export all messages in a specific conversation, one needs to select the Conversation item and export it using

Report export type. Make sure that Display as: Conversation is used in the Report – Sections step. A report created this way will contain all messages in that conversation.

To export specific messages in a conversation, you just need to select the desired Chat Message items and use the same export options as above. Intella will export the related conversation but restricted to the messages present in the export set:

image

When Indicate gaps of omitted messages is checked at the Report-sections step, Intella will add the following information the Report:

  • [x skipped messages], inserted between two chat messages, and indicating how many messages are present in the conversation in between these two chat messages that were not included in the report. This message is omitted if there are no chat messages in between the two exported messages.

  • [Total: y skipped messages], indicating the total number of skipped messages in the related conversation.

25. Tagging

Tagging is the process where you connect a descriptive word to an item or a group of items. For example, one of your items is a PDF document containing valuable information. You decide to tag the item with the word “Important.” Tagging helps you to organize results, for example by separating important and unimportant information.

Tagging can be done in several ways in Intella. This chapter gives you an overview of the possibilities:

  • Tagging in the main window

  • Tagging in the previewer

  • Letting other items inherit tags automatically

  • Pin a tag to a button

  • See all tagged items

  • Searching with tags

  • Deleting a tag

  • Tagging in Compound cases

When entering a tag name, one cannot use the slash character ('/') as that is commonly used in various places to denote hierarchical tags.

25.1. Tagging in the main window

25.1.1. Adding or removing tags

To add tags:

image

  1. Select one or more items from the table, the thumbnail view or the timeline.

  2. Open the context menu (right mouse click), and select “Add or edit tags”

  3. In the “Edit tags of x items” dialog you can select already defined tags, or define a new tag with optional description and tag color. When you click OK, the marked tags will be linked to the selected items.

The “Edit Tags” dialog can also be used to remove tags from selected items. After unmarking the checkboxes and clicking OK, the tags will no longer be connected to selected items.

In general, when selecting more than one item, there are three symbols that can be displayed in the checkbox next to a tag name:

  • An empty checkbox means that none of the selected items are connected to the tag.

  • A minus sign in a checkbox means that some of the selected items, but not all of them, are connected to the tag.

  • A marked checkbox means that all of the selected items are connected to the tag.

The Edit Tags menu option is also available in the Cluster Map: right-click on a cluster or label to open a popup menu with this and other options.

When you start typing the name of a new tag, the list of tags is filtered to show existing tags whose name contain the entered text. This can be used to check whether the intended tag already exists or to quickly navigate to the tag in a long list of tags.

When creating a new tag, a parent tag can be specified. Parent tags can be used to logically group tags, e.g. grouping custodian names, reviewers, locations, or priorities.

Parent tags can also be used to tag items. For example, when you have tags called Europe and Asia with subtags representing specific countries, you can choose whether to tag an item with a continent or a country.

The tag color is used as a background color for the tag name in many places where it appears, including the Tags facet and the Previewer. The default tag color is white, which doesn’t stand out from the background, but you can select a different color by clicking on the square button under “Tag Color” to open the color chooser. When adding a new tag, if you select the tag’s parent before choosing the custom color, the tag color will be automatically set to the parent tag’s color.

25.2. Tagging in the previewer

If you want to tag or remove a tag in the previewer, please take the following steps:

image

  1. Open the previewer

  2. Click the Tag button to open the tag space

  3. Enter a new tag or select an existing tag. To remove a tag (to remove the connection between an item and a tag) just deselect the tag from the list.

Three, six or nine tags can be shown as button in the previewer. When a tag is listed as a button, clicking the button results in the tag being assigned to the current item. You can set the desired amount of these quick-tag buttons in the File > Preferences > Results tab > Previewer section.

You can also use Ctrl+1, Ctrl+2, Ctrl+3, etc. to quick-tag an item. The numbers correspond with the button positions.

When the “Go to next item after tagging” toggle button is selected, the previewer will automatically switch to the next item in the list.

25.3. Automatic tag inheritance

When tagging items, the policy of your investigation may be that some related items should be tagged as well. One use case is when tagging items as irrelevant: all nested items may then be considered as irrelevant as well. Another use is tagging items as privileged; depending on your policy, this may then be extended to all other items within the same mail as well.

Intella offers mechanisms that let these additional tags to be set automatically. For more information, see the section on tagging preferences.

25.4. Pin a tag to a button

In File > Preferences > Tagging tab > Previewer section you can select the number of quick tag buttons: three, six or nine. The default value is three quick tag buttons.

You can pin a tag to a button and keyboard shortcut (Ctrl+1, Ctrl+2, Ctrl+3) with the following steps:

image

  1. Select Tags in the facet panel

  2. Right-click on a tag in the list to open the context menu.

  3. Select “Pin tag to button” and select a number from the submenu.

Now you can use the buttons in the previewer and the keyboard shortcuts to tag an item.

Tags that are pinned to a button are marked with a small blue pin in both the Tag facet and previewer.

To unpin a tag from a button, select 'Unpin tag' in the context menu of Tags.

25.5. See all tagged items

To get an overview of all items that are tagged in your case, please take the following steps:

  1. Select Features in the facet panel.

  2. Select Tagged from the list and click Search

Now you can see all the items that have a tag in the Cluster Map panel.

25.6. Searching with tags

To search with tags, please take the following steps:

  1. Select Tags in the facet panel.

  2. Select a tag and click Search.

Now you can see the items that have the selected tag in the Cluster Map panel.

When querying for a parent tag, the result set will contain all items tagged with that tag or with any of its child tags.

25.7. Editing a tag

To edit a tag, please take the following steps:

  1. Select Tags in the facet panel.

  2. Right-click on a tag in the list.

  3. Use the dialog that opens to either:

    • Change the tag name.

    • Relocate the tag by choosing a different parent tag.

    • Alter the tag’s description.

    • Change the tag color.

25.8. Deleting a tag

To delete a tag from your case, please take the following steps:

  1. Select Tags in the facet panel.

  2. Right-click on a tag in the list.

  3. Select “Delete” and confirm.

Now this tag is no longer in your case.

When you delete a parent tag and confirming the operation, the tag and all its child tags are removed.

25.9. Tagging in Compound cases

Tags assigned to items of Compound cases are stored separately from the sub-cases and do not affect their data. If a sub-case is opened as a single case, these tags will not be visible.

Tags, assigned to items of the sub-cases are available in the Compound case. They can be found in case-specific sections of the Tags facet panel below the local (Compound case-specific) tags. The sub-case tags are available read-only: you can view the tags and query items assigned to them in the sub-cases, but not edit or delete them. Tagging items in the Compound case with the sub-case tags is also not possible: the tags have to be imported to the Compound case before that.

Queries made with the sub-case tags cannot be restored using the Saved Search mechanism. You can import the tags to the Compound case before using them in Saved Searches.
Queries made with the sub-case tags will render a sub-case name in parentheses, to make them easier distinguishable from local and imported tags.

25.9.1. Importing sub-case tags

To import one or more sub-case tags to the Compound case:

  1. Select Tags in the facet panel

  2. Right-click on a tag in the list to open the context menu.

  3. Select "Import sub-case tags"

The tags are copied to the Compound case with their content, i.e. the associated items.

If a tag with an existing name is imported, you will see a dialog window to either create a new tag in the Compound case or merge the imported tag with the existing one:

  • Select the "Import tag with new name" option to create a new tag with another name. Enter a new tag name in the text field.

  • Select the "Merge with existing tag" option if you want to merge the content of two tags, i.e. to assign the items associated with the sub-case tag to the Compound case tag of the same name. Select "Repeat this for all name conflicts" to remember this decision for the rest of the imported tags.

After selecting an option, press the "Import" button to proceed. Press "'Skip'" to not import the current tag.

25.10. Annotations History

To revert a tagging or other type of annotation, open the Annotations History dialog by selecting File > Annotations History…​ This dialog represents a chronological list of all changes in tags, flags, comments, and custodians made on the case.

image

In the top panel, it is possible to set the start and end dates of the events to show. To include only events of specific case users, click the Select button and choose their names in the pop-up list. Click the Refresh button to update the events list.

To revert a specific change, select the annotation in the list and click the “Undo selected action” button. The reverted changes are marked with a strikethrough font. To restore a previously reverted change, click “Redo selected action”.

Note that some actions cannot be reverted directly. For instance, the creation of a tag cannot be undone without undoing or removing all associated item taggings that depend on it beforehand. The actions that cannot be undone at the moment are grayed out in the list.

26. Keyword Statistics

The Keywords tab gives detailed statistics about the keywords in a keyword list. The workflow is as simple as choosing a keyword list, specifying several calculation options, and clicking Calculate. This will produce a table showing the keyword list and several statistics for every keyword query in the list.

image

The nature of the information shown here potentially goes beyond what can be established in the Search tab.

26.1. Configuration

All controls for configuring the calculation are placed on the left side of the tab. The options are divided into four groups:

  1. The keyword list to use.

  2. The filter(s) to be applied.

  3. The document fields to search in.

  4. The statistics to be calculated.

At the top, the user can choose a previously uploaded keyword list or add one here. This uses the same collection of keyword lists as the Keyword Lists facet in the Search tab. Any list added in the facet can be used here and vice versa.

The second panel allows the search results to be filtered. When a saved search is chosen as a filter, the saved search is evaluated and its result is intersected with the result of each keyword search. For example, when searching for the keyword "letter" and filtering using a saved search for OCRed items, requiring PDF documents, and excluding a custodian, this would result in an item set containing OCR-ed items that contain the keyword "letter", are PDF documents, and that does not come from that custodian.

Although we call this functionality “keyword statistics”, the user can use the complete full-text search syntax here: wildcards, Boolean operators, phrase queries etc. are all available. Field-specific searches are also possible. When used in a query, these overrule the field settings set in the third panel.

The third panel offers the available search fields. These are the same as offered in the Search tab. By default, all fields are searched, but the user can choose to restrict searches to e.g. the document text, email headers, etc. Any combination of fields can be used.

The last panel offers five checkboxes that determine what information the table will contain:

  • The Items option adds columns indicating:

    • the number of items containing the keyword,

    • the corresponding percentage of items,

    • the deduplicated number of items, and

    • the exclusive items of a keyword, i.e. the deduplicated amount of items that are not returned by any of the other keywords. It shows how many extra items are returned when a keyword is added, or how many items are lost when it is removed. This can e.g. be used to measure the impact of a search keyword on the length of the review process.

  • The Hits option counts the number of occurrences of the search term in the texts.
    For example, when a keyword produces a document that contains the keyword 3 times and another document that contains the keyword 5 times, this column will show 8.
    The hits are counted across all the selected search fields, but only on the deduplicated items.

If you use keyword lists with advanced search query syntax, please be aware that hits counting is supported for a limited set of query types.
  • The Custodians option adds a column for every custodian in the case. Each custodian column indicates how many of the matching items originate from that custodian.

  • The Families option adds two columns: “Families” and “Family items”. As described elsewhere in this manual, a family is an item set consisting of a top-level item (e.g. a mail in a PST file) and all its nested items (e.g. attachments, embedded images, archive entries). See the Preferences window for how families are determined in the case. The meaning of the two columns is then as follows:

    • The Families column shows in how many families the keyword occurs. For example, if a mail and two of its attachments all contain the keyword, that counts as a single family.

    • The Family Items column shows the total number of items that are contained in these families. This may (and usually will) include items that do not contain the keyword at all; they just belong to a family that has a hit in one of its other items.
      In cases where you are not directly exporting search results but rather their top-level parents (i.e. the default setting when exporting to PST), this will tell you how much of the case is conceptually being exported this way. This may give an indication of how well a certain search filters items in a case.

26.2. Calculation

When the Calculate button is clicked, Intella will populate the table row by row.

The time required for the calculation is dependent on several factors, including the size of the keyword list, the hardware, the chosen search options and the storage location and size of the case. While most options can benefit from indices that make the calculation fast regardless of case size, the Hits option will have a considerable impact on the search speed.

The progress of the calculation will be shown in the status panel above the table.

During calculation, the Calculate button will change into a Stop button, allowing for manually terminating the process.

When clicking Calculate again, the previous results will be discarded and the table will be populated from scratch, using the (possibly changed) configuration options.

26.3. Results

The table order is the same as the order in the keyword list.

The last row shows the total amounts for each column. For the Exclusive and Hits columns, this is equal to the sum of the values in that column. For the other columns, the total amounts are calculated from the union of the item sets. Note that they are typically not the sum of all rows, as items may occur in multiple result sets.

Each table header also shows the total value that can appear in that column on that case. These maximums are not yet filtered by any saved searches that may have been selected. Only the totals at the bottom of the table take the saved searches into account.

The table can be sorted by clicking on one of the headers. This lets you find the queries with the most or the fewest item counts, the queries that involved the most families, etc. By default, the table order is the same as the order in the keyword list.

One can double-click on a row in the table and see the matching items from that result set in the Details view in the Search tab. It is also possible to select multi rows using Ctrl-clicking or Shift-clicking. Then, right-click in the table and select “Query”, to query for all these rows. A dialog will open, asking whether to evaluate the rows as separate queries or combine them into a single Boolean OR query.

To query the exclusive items of single or multiple keywords, select single or multiple rows using Ctrl-clicking or Shift-clicking, right-click on the selected rows and select “Query Exclusive”. The exclusive query will be shown in the Search tab. Note that the results table needs to be in deduplicating mode to see the exact same set of items as in the Keywords tab’s table.

26.4. Exporting

The table contents can be exported by clicking on the Export button in the top right corner. This opens a dialog that lets the user choose the desired format(s) and, for the PDF and DOCX formats, a description to add to the report.

  • CSV and XLSX - exports the table as a comma-separated values file or as a table in an Excel spreadsheet.

  • PDF and DOCX - creates a document containing a keyword statistics report.

Keyword statistics report
This reports builds on the information in the results table.

Each page has a header composed of the case name, the keyword list’s name, the date and time the report was created, and an optional description that was entered in the export dialog.

The page following the cover page contains an overview in the form of bar chart, showing how the keyword list compares to all items in the case. It contains the following values:

  • All items that contain any of the keywords in the list.

  • The deduplicated amount of all items that contain any of the keywords.

  • The deduplicated amount of all items that contain any of the keywords, extended with their family items.

  • All items without any hits.

The following pages contain bar charts showing the Deduplicated and Exclusive values for each keyword in the list.

The last pages in the report contain a table showing for each keyword:

  • The number of items containing the keyword.

  • The deduplicated number of items.

  • The number of items containing the keyword, extended with their family items.

  • The deduplicated amount of items not returned by any of the other keywords.

27. Redaction

Redaction is the process of concealing part of an item’s text, graphics and/or metadata, to ensure that that information is not disclosed to a specific audience.

A typical use case of redaction is the concealing of legally privileged information in information that is produced for an opposing party in an eDiscovery matter, e.g. because of attorney-client privilege. Other scenarios are hiding person names, birth dates, social security numbers, credit card numbers, etc. due to privacy laws or when they are not relevant to the matter at hand.

27.1. Workflow

When redacting an item, Intella first creates a temporary PDF representation of the item and then lets the user mark the sensitive areas in it. This PDF and the added redactions are stored in the case. The original evidence item is not changed, nor is any information removed from the Intella case. At any time, the redaction marks can be reviewed, edited, and removed.

Only when the item is exported to the final PDF or to a load file, are the redactions “burned in”: all pages in the temporary PDF are converted to images in which the sensitive part is replaced with pixels in the chosen redaction color. The result is a PDF that is guaranteed not to contain the sensitive information.

Redaction affects the results of the regular PDF export and the PDFs and TIFFs that are created as part of a load file. For the sake of brevity, the remainder of this section will only refer to exported PDFs when both are meant.

Creation of a redaction PDF can be a time-consuming process, depending on the item’s file type and complexity. This can get in the way of a smooth and efficient review. To alleviate this, one can choose to pre-generate the redaction PDFs, or to queue the items for redaction.

27.2. Redacting an item

Items are redacted by opening them in the Previewer and clicking on the Redact button in the toolbar. This adds a tab called “Redaction”. The Redaction tab contains a PDF rendering of the item and offers various controls for adding and editing redactions. As the PDF is generated on demand, the tab may take some time to appear, depending on the type and complexity of the item. The item is now ready to be redacted.

To redact a part of the content, simply select the rectangular area in the rendered item that needs to be hidden. The selected area will now be covered with a colored rectangle. You can repeat this step to conceal additional parts of the item. The redactions are stored automatically; no manual save action is needed.

The rectangle is semi-transparent so that the reviewer can still see what content has been redacted without having to move it. In the final exported document, the rectangle will be a solid color.

To move or resize a redaction mark, simply place the mouse pointer above the redaction rectangle. When placed in the middle, the mouse cursor changes to a four-arrowed cross and the rectangle can then be moved by holding the mouse button and dragging the mouse. When placed on a corner, the mouse cursor changes into an arrow and the rectangle can then be resized by holding the mouse button and dragging the mouse.

To set or edit redaction text, select a redaction, click the T button and enter the desired overlay text in the dialog that opens. This text will have a priority over the text defined by the redaction template (if any).

To remove a redaction, select it and click Delete Redaction. To remove all redactions of this item, click the Clear Redactions button.

When you close and reopen the item, the Previewer will immediately show the Redaction tab again with all previously made redactions, as the PDF is cached. Only when no redactions are added will the PDF be discarded.

Redacted items can easily be found using the Redacted category in the Features facet.

27.3. Redacting keyword search hits and arbitrary words

A common redaction method is to search for a company or organization name and to review and optionally redact the search hits. Intella can assist with this process: when the Redaction tab is viewed while Intella’s search interface shows one or more keyword queries, the keyword search hits will be highlighted in the Redaction tab and can be redacted with the click of a button.

Note that this highlighting works best on single term queries. It does not work reliably or even at all for more advanced queries such as phrase searches, wildcard queries, etc.

The currently used keyword(s) will be shown in a text field beneath the item content and can be changed. Use the arrow keys to move from one keyword hit to another. Click the Redact button to redact the currently highlighted occurrence, or click the Redact All button to redact all occurrences in the current item.

It is also possible to search for arbitrary words and redact their occurrences directly in the Redaction tab. Type one or more words into the text field and press the Search button. If this word or a sequence of words is found in the item content, the hits will be highlighted and can be redacted using Redact and Redact All buttons.

Click the configuration button next to the text field to set the following search options:

  • Case sensitive: If selected, the search will be performed in a case-sensitive manner.

  • Match whole words only: If selected, only whole word matches will be considered. Deselect this option to also allow for partial word matches.

Please see the subsection on Caveats below when using the Redact All button.

27.4. Redaction templates

The appearance of every redaction mark, such as its color and overlay text, is contolled by its redaction template.

Every new case has a predefined redaction template named "Default". One can define additional templates to make a visual distinction between various types of redactions. For example, one could create redaction templates to visually distinguish legally privileged information from privacy-sensitive information and business-sensitive information. It is possible to create as many custom redaction templates as needed.

To set the redaction template that is to be used for new redactions, choose it in the template chooser drop-down box in the Redaction tab.

To change the redaction template of an existing redaction mark, select the redaction mark in the Redaction editor, then choose the desired template in the drop-down box.

To find all items that contain redactions that use a specific redaction template, see the "Redaction Templates" node in the Features facet.

To create a new template, or edit or remove existing templates, press the gear icon next to the template chooser. This opens the template editor dialog. Here, one can change the properties of the new or selected template:

  • Template name.

  • Color of the redaction marker.

  • Optional text to is to be displayed inside the redaction marks, if an individual redaction text for a redaction mark is not defined.

  • Color and maximum font size for this redaction text.

  • The "Draw border around white redactions" option can be used to render a thin black border around redaction marks that use a white background color.

To create a new template, press the "New…​" button and change the template properties as needed.

To edit an existing redaction template, choose it in the "Template" drop-down box and then click the gear icon.

Editing a redaction template will affect the rendering of all redactions previously created with that template.

To delete a template, choose it in the drop-down box and press “Delete”. All redactions associated with the deleted template will be reassigned to the default template. The default template cannot be deleted.

27.5. Exporting

When exporting an item to PDF, Intella will by default use the redacted version if there is one. More specifically, it will convert the temporary PDF into a final PDF that contains only images. These images will have the redactions “burned in”, so that the sensitive content is concealed permanently. Exported load files containing PDFs or TIFFs will undergo a similar process.

The result of this last conversion step is a PDF that has no regular machine-processable text. To verify this, simply open the PDF in a PDF reader like Acrobat and try to select the text. That makes this redaction method very safe (as opposed to removing the sensitive text from the source file) as all information is in plain sight; there is e.g. no hidden metadata that could still leak the sensitive information. The downside is that the PDFs can have a large file size as all text is represented as images, and that they would need to be OCR-ed to make the non-concealed text accessible again for text selection, keyword search, etc.

As the final PDF is derived from the temporary PDF, the PDF export settings entered in the Export dialog will only have any effect on the non-redacted items in the export set.

The redaction toolbar in the Previewer also has an Export button, to export the current item as a redacted PDF. This PDF will be the same as when it is exported as part of a collection of items to PDF, i.e. all pages will be converted to images with their redacted parts showing as colored rectangles. This option is useful when only a few redacted documents are necessary or to verify the redaction export.

When a redaction PDF has been generated for an item, but no redactions have been added to it, this item will be exported as if no redaction PDF was ever generated. In other words, generating redaction PDFs without adding redactions to them has no effect on the export output.

27.6. Redacting duplicates

Typically, a lot of items will have duplicates in a case. It may be desirable to let all redactions of an item also be associated with all duplicates of that item. To let Intella do this automatically, click the “Auto-redact duplicate items” checkbox in the Redaction tab’s toolbar.

The determination of duplicates for this functionality differs from the general notion of duplicates throughout the Intella application:

  • Message hashes are not involved, i.e. only MD5 hashes are checked.

  • The Image MD5 attribute of load file images is also involved.

The checkbox may therefore not be available for emails that have only duplicates identified by message hashes, even if the “Show Duplicates” function can find these duplicates.

The checkbox can be checked or unchecked at any time during the redaction of an item. The redactions are associated with the duplicates once the user closes the Previewer window or navigates to a different item.

Note that duplicates that are redacted separately may contain different data if they were created using different redaction profiles. The functionality for automatically redacting duplicates was intended to work with redaction profiles that do not include instance-specific metadata.

27.7. Pre-generating redaction PDFs

Redaction PDFs can also be pre-generated. This will result in the Redaction tab initializing a lot faster, as it can immediately load this PDF from disk.

To pre-generate redaction PDFs, select the items in the Table, right-click, and choose Redactions > Pre-generate redaction PDF.

The downside of pre-generating redaction PDFs is that you may end up generating PDFs for items that will turn out not to need any redactions. For a large case, the cost of generating all PDFs up-front may be prohibitively time-consuming.

27.8. Queueing items for redaction

Only languages that use a Latin script are supported.

Items may also be queued for redaction. This workflow starts with gathering the items to redact via a keyword search, e.g. using a person name. The user reviews the search hits using the Content and Preview tabs, which are generally quick to load. When the user determines that redaction is appropriate for the current item’s keyword search hits, the user can click the “Queue for Redaction” checkbox in the Previewer.

Clicking this checkbox stores the item in the so-called redaction queue, together with its currently highlighted search terms. The user can let Intella process this redaction queue later, at a time when the case is not being worked on. Intella will then generate the redaction PDFs for the queued items and determines the visual areas where these hits appear in the PDF.

To process the redaction queue and generate the redaction PDFs with their redactions, select the items in the Table, right-click, and choose Redactions > Process Redaction Queue…

The benefit of queuing items for redaction is that no redaction PDFs are generated before or during the actual review. This means that reviewing for the purposes of redaction can start right after the case has been indexed. Additionally, redaction PDFs will only be generated for those items that need to be redacted, potentially saving a lot of processing time.

The downside of queueing items for redaction is that an additional manual review of these PDFs is still needed afterwards, as the visual output of the PDF rendering may be different from what is shown in the Contents and Preview tabs.

When the queue is processed and Intella detects for an item that there is a difference between the hit count in the redaction PDF versus the hit it had when the “Queue for Redaction” checkbox was selected, it will put these items in the “Missing keyword hits” in the Features facet. It is strongly recommended to review the redaction PDFs of these items afterwards.

27.9. Removing redactions

It may turn out to be necessary to remove all redactions, e.g. because of a change in policies. To do this, select the items in the table, right-click and select Redactions > Remove Redactions…

Note that it is currently not possible to revert this operation.

27.10. Redaction profiles

When the Redact button in the Previewer is clicked, a PDF that is generated will consist of a limited set of content and metadata properties. For example, emails will show their most important headers (email sender and recipients, subject and sent/received dates) on the first page, followed by the email body. The full SMTP headers of the email are printed on one or more separate pages, followed by the list and content of the email’s attachments.

When this default set of content and metadata properties is not suitable for a specific case, or when different settings are desired for different types of items or different audiences, the user can define one or more redaction profiles for the case. Such a profile defines the set of content and metadata properties to be used in the redacted PDF. When a redaction profile is defined and the reviewer subsequently clicks the Redact button in the Previewer, Intella will ask which redaction profile to use for this item and generate the PDF accordingly.

To define a redaction profile, click the “Configure redaction profiles…” button in the Export menu of the main window and choose Create in the next dialog. The window that opens allows the reviewer to enter a profile name and select which content and metadata properties should be used when this redaction profile is chosen. For a detailed description of the available properties see the section on exporting to PDF.

27.11. Caveats

As the purpose of redaction is to conceal sensitive information, it is vital that the reviewer takes notice of the following caveats on the redaction functionality.

First, there are a few issues to be aware of when using keyword hit highlighting to control the redactions. When highlighting the search hits in a PDF, the highlighted area may not exactly cover the responsive text in the PDF. The redaction rectangle then needs to be manually moved and resized. Whether this happens depends on the fonts used in the PDF: PDFs that Intella has generated using texts from its own databases are fine (e.g. pages with email bodies and headers), but text in existing evidence PDFs or in Word documents that are converted to PDF may be a different story. We have no control over the font characteristics used in those documents and therefore cannot guarantee correct placement of the redaction rectangle.

Another important aspect is that hit highlighting may not find all occurrences of the text that is searched for. For example, words that are misspelled, use a spelling variation, or are hyphenated may not be found. Texts inside graphics will also not be found. Note that OCR software that is used to combat this can also introduce spelling errors. Finally, tables and graphs may require extra attention.

When creating a redaction PDF for an item, the PDF is only associated with that specific item, unless the “Auto-redact duplicate items” checkbox is selected.

Be aware that redaction profiles may insert instance-specific information such as the item’s location. When that item has duplicates and auto-redacting of duplicates is used, all duplicates will still be associated with the same redaction PDF. All items will then show the item-specific metadata of one of the duplicates.

When using the “Queue for Redaction” functionality, be aware that the generated PDF may render information differently than the Contents and Preview tabs, causing differences in detected keyword hits. Therefore, a manual review of the redaction PDFs and the determined redaction rectangles is always recommended, to ensure that no sensitive information is revealed.

28. Exporting

Intella supports several exporting formats, each focusing on a different use case.

When exporting a single result or a list of results, the maximum length of a file name is 120 characters on operating system older than Windows 10. There is no limitation on the file name length when using Windows 10 and newer.

28.1. Exporting a single result

To export a single result, right-click on a row in the Details table (or on the item in any of the other views) and select Export > Item… in the content menu. Alternatively, select an item by clicking on it and choose Export > Result…​ in the menu bar.

A file chooser will open that lets you specify the folder and file name. Click “Save” to export the result to that file. The mouse cursor will show a “busy” icon while the exporting is taking place.

The result will be saved in its original format, i.e. a Word document attached to a mail gets saved as a Word file. All mails from mail sources (PST/OST/NSF/DBX/MBX/Mbox files and IMAP servers) are exported as EML files. Evidence files that are already in EML, EMLX or MSG format as exported as such. Contacts will be stored in vCard format. Calendar items from PST files will be stored in iCal format.

When an item has been OCRed, decrypted, or has a load file image associated with it, an additional dialog may be shown. This dialog will allow the user to select which variant of the content should be exported.

28.2. Exporting a list of results

To export a collection of search results at a time, you can use the following procedure:

  • Use Ctrl-click or Shift-click to select multiple items in the Details pane, using the table or thumbnails view.

  • Alternatively, right-click and choose Select All to select all items in the list.

  • Right-click and choose Export > Selection, or choose Results List… from the Export menu.

This opens the Export Wizard. This wizard lets you choose the export format and its settings and the export process.

28.2.1. Export formats

The first wizard page lets you choose an export format:

  • Original format exports a file into its original format, i.e. a Word document attached to an email is saved as a Word file. All emails from mail sources (e.g. a PST or NSF file) are exported as EML files. Emails that are already in EML, EMLX or MSG format are exported as such. All contact items from PST/OST files are exported as vCard (.vcf) files. All calendar items from PST sources are exported as iCalendar (.ics or. ical) files. The exported files can be opened with the program that your system has associated with the file extension used. Folders may be created during exporting whose name ends with _files. This is necessary to separate e.g. an archive from the items that were extracted from that archive, as both may be exported at the same time.

  • PDF converts every item into a PDF document, containing the content of the original item and a configurable set of properties.

  • PST lets you export items to a MS Outlook PST file. The main purpose of this option is to use the PST file as a carrier for transport of emails, but other item types are supported as well. The receiver can open the PST file in Microsoft Outlook or process it in another forensic application.

  • Load file will export the items in a format that can be imported into Summation, Concordance, Ringtail and Relativity.

  • Relativity will export items directly to a Relativity server, i.e. without the use of an intermediary load file.

  • i2 Analyst’s Notebook/iBase exports the results in a format that can easily be digested with i2’s Analyst’s Notebook and iBase applications. All metadata of all items, all attachments and all email bodies can be imported into these tools, allowing rapid social network analysis and all other analytical abilities of these applications on email and cellphone evidence data.

  • Report produces a nicely formatted report, containing a configurable set of item properties and optionally the extracted text of the item. The report is composed of sections that can be defined and further customized by the user.

Only one format can be chosen per export run.

28.2.2. Destination folder

The chosen destination folder will contain all exported items, including all export reports (see below). You will get a warning when this folder is not empty. Though Intella tries not to overwrite any files in the specified folder, we recommend specifying an empty folder to be sure.

For every selected format, a subfolder will be created that holds the files of that export format. All export reports will be placed in the top folder.

When exporting several sets to the same destination folder, the subfolders with produced files will be merged, but earlier produced files will not be overwritten. Each export run will have its own set of export reports.

28.2.3. Export templates

The configuration entered in the Export window or sub-windows like the load file field chooser will automatically be restored the next time the Export window is opened. No manual action is necessary to achieve that.

The current configuration can also be stored as a user-named template in the last wizard sheet. In the first sheet, all stored templates are listed in a drop-down list. Selecting one restores the state of the Export wizard to the one stored in the selected template. Export templates are typically stored in:

C:\Users\<USERNAME>\AppData\Roaming\Vound\Intella\export-templates

You can easily access this folder through the Help > Open Export Templates Folder menu item.

Export templates are stored outside of the case data folder. This makes all templates automatically available across all cases on the same machine and user account. To use templates with other user accounts or on other machines, just copy the XML file named after the template to the export-templates folder on that account or machine.

When you click next, the wizard will let you configure the format-specific options.

28.2.4. Suppressing irrelevant items

You can use the “Suppress irrelevant items” checkbox to automatically exclude all items from the export that have been classified as “Irrelevant” during indexing. See the Preferences section for a definition of irrelevant items. The number of irrelevant items in the current item set will be shown in parentheses.

28.2.5. Export sets

When a set of items is exported, they can optionally be added to an export set. This is a named set that captures information about the export.

When a specific item is about to be exported, the file name and number is recorded in the export set. Furthermore, the current export settings are stored as part of a set. When the export set is later selected again when exporting another set of items, this will affect that export run in the following ways:

  • All export settings such as the chosen export format, file naming and numbering schemes, etc. will all be the same as in the first export run. In other words, the export set works like an export template.

  • File numbering continues where it left off, rather than starting at 00000001 again.

  • Items that have been exported before with this export set selected will get the same name and number as the previous time(s) they were exported.

When an export set is specified, the resulting export ID (typically based on subject, file name and/or consecutive number) can be made visible in the Details column by adding an Export column that corresponds to the export set.

The Export IDs can also be searched for using keyword search and keyword list search.

28.2.6. Preferred content type options

The options in this sheet allow to select the preferred content type for the original format items, when multiple variants are available. Intella will export the first available content in the order specified in the table. The following content types are available:

  • Original – The original content of the item, as obtained from the evidence data.

  • Decrypted – The decrypted content is available for items that have been decrypted during indexing.

  • OCRed – The OCRed version of the original document. Note that this content type is not available for items where the OCRed content was imported as plain text.

  • Load file image – The image associated with an item imported from a load file. Note that the load file image is always exported in PDF format.

28.2.7. PDF file options

The first wizard sheet on PDF options lets you decide whether to export to individual PDF files, one for every selected item, or to export all items into one single concatenated PDF file.

When exporting to a concatenated PDF, the resulting PDF can optionally be split in chunks of a given size. This is recommended for performance and stability reasons.

28.2.8. File naming and numbering (original format, PDF, load files)

This wizard sheet consists of three sections:

  • File naming defines how to compose an exported file name (original format, PDF) or page (load file export).

  • File numbering defines how exported files are numbered.

  • File grouping defines how exported files are grouped into folders.

File naming
By default, exported files will be named using the original evidence file’s name or the subject of an email. Alternatively, you can choose to number the files using consecutive numbers. These options can also be combined: a number followed by the file name or subject.

Load file naming offers more elaborate numbering style, whose parts can be further configured in the File Numbering section.

When using a numbering style, you can also define a prefix. Anything you type here will be added to the beginning of the filename. E.g. the prefix export- will result in the first email being named export-00000001.eml, when you combine it with consecutive numbering.

The Custom ID option can be used to use custom IDs generated by the Generate Custom IDs task. When page numbering is used, each page will be numbered using the custom ID plus a page number suffix. The "Number of digits for page" option determines the number format that will be used to number pages.

Using "Advanced" mode, you can define a file name template that will be a base for exported file name. The template may include the following fields:

  • %num% – A counter value will be added. You can also define leading zeroes in the counter using the following format: %000num%. The number of zeroes defines the number of digits used in the counter. The default number format for the counter is to use 8 digits.

  • %page% – A page number that starts at 1 for each document.

  • %group1%, %group2% – Group counters used with load file export only. See the "Export as a load file" section for details.

  • Any Intella column identifier surrounded by the % symbol, like %md5%.

  • %Best_Title% – One of the following fields: File name, Subject, Title, Contact Name or "Untitled".

To insert any field in the template you can either type it manually or select the field from the drop-down list and press "Add field".

File numbering
Using the "Start at" option you can define the number to start counting with. By default, exporting will start counting at 1. A typical reason to use a different start number is when you want to combine the exported results with another set of already exported files. Numbers are always 8 digits long.

“Folder”, “Page rollover” and “Box” are only relevant when using load file naming.

When exporting to PDF, the “Number pages” option can be used to number individual pages instead of files. The numbering then works the same way as when exporting to a load file.

File grouping
Select the option “All in one folder” to put all exported files in one folder.

Select the option “Keep location structure” to preserve the original folder structure that the items have in the evidence files. A folder will be created for every source, in which the original folder structure of that source (as shown in the Location facet) will be recreated.

File name examples
On the right side of the window you can see a live preview of how the exported file names would look based on the current settings, using items from your current item set as examples.

28.2.9. PDF rendering options (PDF, load files)

When exporting a load file, this sheet is called “PDF or image rendering options”

The options in this sheet only apply to non-redacted items; the exporting of redacted items is governed by the “Redacted items” sheet.

For all types of items, you can indicate whether to include a basic item header, properties, raw data, body and comments in the PDF:

  • The item header is shown at the top, above a black line, and shows the email subject or file name.

  • The properties include typical metadata attributes such as titles, authors, all dates, hashes, sizes, etc. By default, all properties are included, but you can remove some of them in the “Select properties…​” dialog.

  • The raw data varies between item types. For example, for PSTs the low-level information obtained from the PST is listed here and for vCards the actual content of the file is listed. This field may reveal properties that Intella does not recognize and are therefore not to be found in the Properties section.

  • The main properties above body include main email properties such as Subject, From, To, Sent, Received, etc. By default, all properties are included, but you can remove some of them and change the order of the remaining elements by clicking the “Configure…” button.

  • The comments refer to the ones made by Intella user(s) in the Comments tab in the Previewer. They are not to be confused with comments that can be made in, for example, a Word document. These are part of the Properties section. Note that the reviewer comments may include sensitive information such as evidence file names, investigator insights, etc.

Furthermore, the item’s content can be exported in its original format, as the extracted text, or both. The following file formats can be exported in their original view:

  • Emails with a HTML body.

  • MS Office (doc, docx, xls, xlsx, ppt, pptx)

  • Open Office (Writer, Calc, Impress)

  • WordPerfect

  • RTF

  • HTML

  • PDF

When you select “Original view”, you will also be able to define a list of item types that should be skipped for this. You can use this to e.g. prevent native view generation of spreadsheets, which often are hard to read in PDF form. An optional placeholder text can be added to make clear that original view generation has been skipped on purposes for this item. Select the “Append file type to placeholder text” option to add the type of a skipped file to the end of the placeholder text. For example, this could look like:

Document rendering skipped (Microsoft Excel 97-2003 Workbook)

When you also select the “Export skipped item as native file” option during load file export, the resulting load file will contain the corresponding native file. By selecting “Also skip extracting text” you can skip generating the extracted text as well. This includes extracted text added to the resulting PDF and extracted text exported as a separate file as part of a load file.

The “Configure 'Original view'” button allows to configure which content type needs to be included in the Original view of the item. See the “Preferred content type options” section for more information about the available content types.

If you uncheck “Include item metadata”, the resulting PDF will not contain any additional information except for the actual item content (in its original format and/or as extracted text), the document title/subject and the headers and footers defined in the next sheet. Most of the options on this sheet will then be disabled.

For emails, the following additional information can optionally be included:

  • A checkbox is provided, controlling whether the HTML or plain text body is preferred. This option is only available when the “Content as” setting is set to a value that involves original view generation, i.e. anything other than “Extracted text”.

  • The full email headers.

  • A list of all attachments, as a separate page. The file name, type and size of each attachment will be listed.

  • The contents of the attachments. The original view (described below) will always be selected by default, with the extracted text used as a fallback.

For loose files and attachments that are not emails, the following additional information can optionally be included:

  • The “OCRed text for images” checkbox controls whether to include the OCRed text when the file is an image.

  • List all embedded images and attachments.

It is possible not to include the lines that separate the headers and footers from the content by unchecking the “Draw header and footer line separators” checkbox. Section names such as “Image”, “Original view”, “Extracted text” etc. can also be excluded from the resulting PDF by unchecking “Include section names”.

28.2.10. PST options

Enter a file name to use for the generated PST.

Enter a display and folder name. After opening the exported PST file in MS Outlook, you will see the names you entered. They help you to locate the PST file and its contents in MS Outlook. When a folder name is not specified, the items will be exported directly to the PST root folder.

Select the option “Keep location structure” to preserve the original folder structure during the export.

The resulting file can optionally be split into chunks of a given size. This is highly recommended for larger result sets that would make the PST grow beyond the default suggested file size, as Outlook may become unstable with very large PST files. The produced files will have a file size that is close to the specified maximum file size (usually smaller). The export report will list for every item to which PST it was added.

Item types that can be exported directly to a PST file
Besides emails, the following item types can be exported directly to a PST file:

  • Contacts

  • Calendar items:

    • Appointments

    • Meetings

    • Meeting requests

  • Tasks

  • Journal entries

  • Notes

  • Distribution lists

  • Teams chat messages extracted from another PST file

Limitations:

  • iCal recurrence rules (RRULE property) are not exported.

  • PST Distribution lists are exported, but their list members are not.

These limitations may be removed in a future release.

Please note that non-email items will be exported to a regular PST folder under the Mail section, so not in e.g. the Contact section.

How to export other item types to a PST file
Items such as Word and PDF documents cannot be exported directly to a PST file. As such items may be attached to an email, Intella can be configured to export the parent email instead.

You can choose to either include the top-level email parent or the direct email parent. An example would be an attachment contained within an email message within another email message. With the top-level parent selected all parent items of the attachment (both emails) would be included in the PST, one nested within the other. The second option exports the nested email to the PST. You can also choose to simply skip non-email attachments.

Although this option only mentions parent emails, it also applies to e.g. PDF files attached to a meeting request or any of the other exportable items. In this case, enabling this option will export the meeting request instead. This option may therefore be renamed in the future.

Files in a folder source lack a parent email and therefore cannot be exported to a PST file, except for mail files like EML, EMLX and MSG files, or files of the types listed above.
For stability reasons, items which size exceeds 5MB excluding attachments will not be exported to a PST. It can be changed via "ExportPstSkipLargeItemsEnabled=False" property in the case.prefs file. The property "ExportPstMaxItemMetadataSizeMB" can be used to change the upper limit. For example, to increase the limit from 5 to 10MB use "ExportPstMaxItemMetadataSizeMB=10".

How to export attached emails
The last setting controls what happens with emails that are selected for export and that also happen to be attachments. These are typically forwarded messages. Such emails can technically be exported to a PST without any restrictions, but the investigation policy may require that the parent email is exported instead, to completely preserve the context in which this email was found. That can be done by choosing the Replace with its top-level parent email option. Alternatively, use the Export attached email option to export the attached email directly to the PST.

28.2.11. iBase and Analyst’s Notebook options

The Analyst’s Notebook and iBase export does not provide any configuration options.

Templates, import specifications and instructions are provided for Analyst’s Notebook and iBase. Please contact support@vound-software.com for more information.

28.2.12. Load file options

You can select one of the following load file formats:

  • Summation.

  • Concordance.

  • Relativity.

  • Ringtail.

  • Comma Separated Values file.

Each load file export consists of several parts:

  • The main load file, containing the selected fields.

  • Native files, representing the items in their original format.

  • Image files, containing metadata and content as configured in the “PDF or image rendering options” sheet.

  • Text files that contain the extracted text.

The first part is mandatory; the others can be turned off.

The main load file name can be changed using the “File name” text field. It is also possible to specify the main file encoding when the Summation format is selected.

The "Export native chat content as PDF" option can be used to export native content of chat conversations and messages in PDF format instead of plain text. This was primarily designed for direct Relativity export due to certain limitations of that platform.

By selecting “Use custom date/time formats” you can override the date and time format used in the load file. Please see this document for the date/time format syntax details: http://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html

The Size column can be optionally exported in kilobytes, megabytes or gigabytes instead of bytes, by using the “Size unit” option.

To control the quality of the exported images, one can use the “Image DPI” parameter. It defines the number of dots (pixels) per inch. A higher DPI setting results in higher quality images, but these will take more time to produce and consume more disk space.

It is also possible to adjust the TIFF compression type. Note that the image will be converted into black-and-white variant if one of the “Group Fax Encoding” compression type is selected.

The “Also include PDF version of images” option can be used to additionally export PDF version of images if the image format is different from PDF. The PDFs will be exported to the folder specified in the Folder option.

If the “Opticon Page Count field contains number of pages of entire document” option is turned on, in the Opticon file the 7th field of the first page record (NEW_DOC_BREAK=Y) will contain the total number of pages of the entire document. The 7th field of all the other page records will contain an empty value. If the option is turned off, the 7th field will contain the number of pages of the current page record only. That means it will always be “1” for any single page format like JPEG or PNG.

The extracted text can be configured by clicking the “Configure” button. In the “Configure extracted text” dialog you can choose which components to include and change their order. Different components can be configured for emails (including email-like items: instant messages, conversations and phone calls) and files. The Properties component can be configured by clicking the “Configure” button below the table. You can choose which properties to include.

When you need to embed the extracted text directly into the load file itself (the DII, DAT or CSV file) instead of exporting it into a separate file, you can use the checkbox “Embed extracted text into load file”. A custom field of type EXTRACTED_TEXT should be used to insert the text as a field in this case.

When exporting to Summation the checkbox “Include Summation control list file (.LST)” can be used to generate a plain text file that lists all document IDs along with the extracted text files. The “OCR Base” field controls the prefix used for the extracted text files.

The “Exclude content” option can be used to completely exclude the items tagged with a specified tag. For every excluded item, only the metadata will be added to the load file. The text and the images will contain the text specified in the “Placeholder text” field. Native files will also not be generated for such items.

Export sets must be used when one needs to be able to create an overlay load file later. Adding the load file to an export set during the initial export will create an Export ID field that can be used as a Document ID later in overlays.

Numbering with load files
The numbering used for load files differs from the other export formats. When exporting to a load file, every exported page has its own unique number. The number of the first page is usually used as a number of the document. Please note that pages are numbered only if image files are included in the export.

On the “Headers and footers” sheet you may choose a special field PAGE_NAME which is available only with load file export. This will put the current page name as it was configured on the “Naming and numbering” sheet.

Another difference is that by default all export files are grouped into folders and optionally boxes. The “Page rollover” option defines a maximum number of pages that a folder can contain. The maximum number of folders in a box is fixed to 999 (at the moment, it can be changed via an export template XML file only). Additionally, you can set a starting number for the page (“Start at”), folder and box.

By default, the page counter starts over when switching to the next folder, so the first page in the next folder will have the number “1”. This approach can be changed when using the “Continue page numbers from previous folder” option. When it is selected, the page counter will continue page numbering from the last page of the previous folder. In other words, page numbers will be unique among the entire export set.

Additionally, the “Advanced” numbering mode can be selected when exporting to a load file. In this case, you will be able to set a custom file name template. Please see the file naming and numbering section for details. Note that %num% means a page number, not a document number in this case. Also, there are two new fields that can be used:

  • %group1% – folder counter

  • %group2% – box counter

You can also use the %000group1% syntax to define the number of leading zeroes in the counter (similar to the %000num% syntax). Thus, the default load file numbering schemes can be expressed using the following templates:

  • PREFIX.%group2%.%group1%.%num% = Prefix, Box, Folder, Page

  • PREFIX.%group1%.%num% = Prefix, Folder, Page

  • %group2%.%group1%.%num% = Box, Folder, Page

When using the “Advanced” mode it is important to set a file grouping: “All in one folder” or “Load file mode”. When the “Load file” grouping mode is selected, the exported files will be grouped by folders and, optionally, boxes in exactly the same way as it is described above.

Field chooser
The “Field chooser” sheet contains a table of the fields that will be included in the load file. By default, the starting set of fields depends on the selected load file format.

The “Name” and “Comment” columns in this table are used only for managing the fields within Intella and are not included in the load file. The “Label” column value is used as a column label in the load file. The “Type” column can be one of the following:

  • SUMMATION – It can be used only with Summation load file format and cannot be modified.

  • RINGTAIL – It can be used only with Ringtail load file format and cannot be modified.

  • CUSTOM – User-created field. It can be used with any load file format.

You can include an additional custom field by pressing the “Add custom field…​” button. Next, enter the name, label, and comment. Select one of the following types:

  • Fixed value – Fixed value as specified in the “Value” field.

  • Intella column – One of the Intella columns.

  • Best title – One of the following Intella columns: File name, Subject, Title, Contact name or “Untitled”.

  • Start record ID – Name of the first page of the document.

  • End record ID – Name of the last page of the document.

  • BEGATTACH - First record in family – Name of the first page of the first document in the current family (“parent-child” group).

  • ENDATTACH - Last record in family – Name of the last page of the last document in the current family (“parent-child” group).

  • Parent record ID – Name of the first page of the parent document.

  • ATTACH_RANGE: BEGATTACH-ENDATTACH – Names of the first page of the first document and the last page of the last document in the current family (“parent-child” group). In other words, the BEGATTACH and ENDATTACH fields separated by a hyphen.

  • Number of pages in record – number of pages of the record (document).

  • Path to native file – relative path of the original format of the document to the base folder.

  • Path to image file – relative path of the first image of the document to the base folder.

  • Path to text file – relative path of the extracted text file of the document to the base folder.

  • Extracted text – extracted text directly embedded in the load file body. See the “Embed extracted text into load file” option described above.

  • Email Internet headers – full Internet headers of the email.

  • Attachment IDs – The list of attachment IDs.

  • Number of attached documents in family – The number of attachments that belong to the current family (“parent-child” group).

  • Is email – “True” if the document is email, “False” otherwise.

  • File extension – The file extension of the document.

  • Direct parent ID – ID of the document’s direct parent.

  • Direct children IDs – The list of IDs of the document’s direct children.

  • ID of first email attachment – Name of the first page of the first attachment document in the current family (“parent-child” group). Empty if there are no attachments in the current group. Used for emails only.

  • ID of last email attachment – Name of the last page of the last attachment document in the current family (“parent-child” group). Empty if there are no attachments in the current group. Used for emails only.

  • Has extracted or OCRed text – “True” if the item has any extracted or OCRed text, “False” otherwise.

  • Raw data – One of the raw data fields. Use the “Value” option to specify the name of the raw data field that is to be exported.

  • Duplicate Locations (excl original item) – The locations of all duplicate items in the case, excluding the item itself.

  • All locations (incl all duplicates) – The locations of all duplicate items in the case, including the item itself.

  • Duplicate custodians (excl original item) – The custodians of all duplicate items in the case, excluding the item itself.

  • All custodians (incl all duplicates) – The custodians of all duplicate items in the case, including the item itself.

When exporting to a load file, all documents are grouped by their parent-child relationship. For example, an email and its attachments form a single group. The columns “RECORD_ID_GROUP_BEGIN” and “RECORD_ID_GROUP_END” denote the start and end page numbers of such a group.

When adding a date column as a custom field, it is possible to choose the way how the date is formatted: show date only, show time only, show full date and time, timezone offset, or timezone name. Note that you can add the same date field more than once and use different formatting options. For example, you can add two custom fields: DATE_SENT (“Sent” column, show date only) and TIME_SENT (“Sent” column, show time only).

Click the “Select default fields” button to select only those fields that are part of the default field set for the selected load file format.

28.2.13. Relativity options

This functionality was tested with Relativity 8.2. Other supported versions are 9.7 - 10.3.

Intella can export items directly into a Relativity database, i.e. without the need to manually handle load files. Note that this functionality requires Microsoft .NET and the Relativity SDK to be installed. See the Installation section for further details.

On the “Relativity options” page you can specify a service URL, user name and password. Please ask your Relativity administrator for the correct settings. The Relativity service URL usually looks like this: https://host/relativitywebapi. You should use the same service URL as you use in Relativity Desktop Client.

Click the “Get list from server” button to get a list of Relativity workspaces. Select the workspace you want to export the items to. You should also choose an identity field which is used as a key field in the selected workspace (it’s usually “Control Number”).

The rest of the settings are the same as you can use during an export to a load file. You can include natives, images, and texts.

Please note that when you use a field chooser, you can choose an existing field from the selected workspace. The field editor will also show a little warning icon near the field label if you enter an incorrect field name.

Current limitations:

  • The Overwrite mode is currently fixed to “Append”. An option may be added in a future release.

  • To export a folder structure, the “Location” field should be added to the list.

  • To export natives, the field FILE_PATH should be added to the list.

  • To export texts, the field FILE_TEXT should be added to the list.

  • Items are exported to the workspace root folder. An option may be added in a future release.

28.2.14. Headers and Footers (PDF, load files)

You can set headers and footers for the generated PDFs and images. For each corner you can select one of the following fields to display:

  • EMPTY – Nothing will be displayed.

  • EXPORTED_FILE_NAME – A file name as it was configured on the "File naming and numbering" sheet.

  • PAGE_NAME – A page name as it was configured on the "File naming and numbering" sheet. Note: this option will work only with load file export. For other export types this will be replaced with EXPORTED_FILE_NAME.

  • PAGE_NUMBER – A page number within the current document starting from 1. By default, the format is "Page XYZ", where XYZ is the current page number. It is possible to change the default format. To do that, edit the "pageNumberFormat" field in the template XML. We might add a UI option for that in a future version.

  • BEST_TITLE – This is one of the following fields: File name, Subject, Title, Contact name or "Untitled".

  • DESIGNATION – one or more textual labels, e.g. "Confidential" or "For Attorneys’ Eyes Only". Tags control the presence of the labels. After selecting DESIGNATION, click on the gear icon next to the field chooser to specify the controlling tag and the text that should be shown on items that have that tag. It is possible to specify multiple tags here. If an item has multiple tags, the designation will be a comma-separated list of the corresponding designation labels.

  • Any Intella column – This will be exactly the same value as it is displayed in the result table.

Also you can type any static text instead of selecting one of the fields.

28.2.15. Report – Title page

This wizard sheet controls what the cover page of the report will look like. The options are divided into two sections:

  • Report Title
    This is where one can specify the title of the report and choose whether the Vound and Intella logos should be included. One can also choose custom logos to be shown.

  • Custom Fields
    This part controls the addition of custom fields that will appear on the cover page. In order to add a new field, press the Add field button and enter the values to show as the field name and field value. To remove a custom field, press the Remove button next to the field that needs to be removed.

28.2.16. Report – Sections

In this wizard sheet, the sections that make up most of the report can be defined. Each section is defined through a row in the table. The table columns have the following meaning:

  • # – Sequence number of the section.

  • Title – Title of the section.

  • Category – The type or tag whose items will be shown in this section.

  • Display – The type of layout that will be used for this section.

  • Items – Indicates the number of items that will be shown in this category.

  • Sort by – The item attribute used to sort the items.

  • Orientation – the page orientation that is to be used for this section.

To add a new section, click the Add section button to the right of the table. A dialog will be shown where you can select either a type or a tag of items. The items corresponding with that type or tag will appear in the newly defined section. To simply list all items in the case, one can choose the top-level type labeled All items.

To change the type or tag of an existing section, click the Change category button.

After clicking the Ok button, a new section will be added to the Sections table. It will be automatically selected. The selected section can now be further configured in the Section details section:

  • Title – This can be any text. The title will be displayed at the top of the section.

  • Description – a description of the section. This will be rendered beneath the title in the report.

  • Display as – the type of layout that will be used for this section. Possible layouts are:

    • List

    • Table

    • Image Gallery

  • Sort by – the property that will be used for ordering of items.

  • Page Orientation – the orientation of the pages in this section: either Portrait or Landscape.

  • Thumbnail columns – The number of thumbnail columns, when the Image Gallery layout is used.

  • Include original format files – This controls whether the original item should also be exported to its original format. Note that items that are exported in this way will be linked from the report.

  • Include extracted text – This controls whether the extracted text of the item should be included in the report.

  • Selected columns – The columns that you wish to appear in the section. Note that the initial set of suggested columns is based on the selected Section Category type. To edit this list, click on the Edit columns button. You will be presented with a column selection dialog.

The ordering of the sections can be modified by dragging sections in the Sections table.

28.2.17. Report – Other Settings

Setting that can be configured on this wizard sheet are:

Headers and footers

  • Show report title in footer – Controls whether the report title should be shown in the page footer.

  • Show page number in footer – Controls whether the page number should be shown in the page footer.

Table of contents

  • Include table of contents – Controls whether a Table of Contents should be included in the report.

Summary

  • Include summary section – Controls whether a summary section should be included into the report.

    • Include sources summary – Controls whether basic source information summary should be included.

    • Include types summary – Controls whether basic item type information summary should be included.

Errors

  • Include error report (CSV) – Controls whether an error report in CSV format should be produced.

Output format
The item report can be exported to these two formats:

  • Portable Document Format (PDF)

  • Microsoft Word Document (DOCX)

28.2.18. Redacted items

This wizard sheet controls how Redacted items are to be handled when they are part of the set of items to export. The options available depend on the chosen export format.

When exporting to Original format or PDF:

  • When the option “Use redacted images when available” is selected, any redacted item will be exported in its redacted form.

    Note that for Original format export a PDF will then be generated, rather than the item being exported in its original file format.

  • The "Ignore redaction color" option renders all redaction marks as either white (if this color is specified in the redaction’s template) or black (for all other colors).

When exporting to Original format, PST or i2 iBase/ANB:

  • When the option “Suppress redacted items” is selected, then any redacted item will be skipped.

When exporting to Load file or Relativity:

  • When the option “Use redacted images when available” is selected, then the image will be exported in its redacted form.

  • When the option “Suppress natives for redacted items” is selected, then exporting of the native file will be skipped when the item has been redacted.

  • When the option “Also suppress natives for family items” is selected, the natives of items in the same family as a redacted item will also be skipped.

  • When the option “Suppress text for redacted items” is selected, then exporting of the extracted text will be skipped when the item has been redacted. The text can optionally be replaced with the specified placeholder text.

28.2.19. Reports and ordering

desktop exporting reports

You can indicate whether you want to create an export report for this export. The report can be formatted as a PDF, RTF, CSV and/or HTML file.

For PDF, RTF, and HTML reports you can also add a comment that will be displayed on the first page of the report.

Export reports link the original files to the exported files, by listing identifying information about the original item (e.g. source evidence file, MD5 hash) and linking to the exported file. Also the export report may contain information that is lost during export, such as the evidence file’s last modification date; like any copy, the export file has the date of export as its last modification date.

If the export of a specific result resulted in errors, you will be notified with an error message in the application. You can find the error notifications at the end of the PDF and RTF report or in the last column of the CSV report.

Additionally, you can specify in what order the items are to be exported:

  • Use Intella default – Items will be sorted first by Hierarchy, followed by Sent date, followed by File Name.

  • Use Family Date – Sorts items by Family Date before exporting.

  • Use Custom ID – Sort items by Custom ID before exporting.

  • Use current table row order – Uses the current order used in the Details table.

When exporting items to a load file with the current table row order option, the order needs to be load file compatible. Otherwise, incorrect families might be produced.
The Custom ID sort order is recommended to use with Custom ID numbering only. Otherwise, it might produce unexpected results in terms of parent-child groups in the exported load file.

28.2.20. Skipped items

The exporting progress user interface may report skipped items. These relate to the fact that not all items are inherently exportable to the chosen export format(s). Examples are:

  • A file inside an encrypted ZIP file may be known to Intella but it cannot be exported to Original Format if Intella could not decrypt the ZIP file. Exporting to PDF is possible though, with the information that is known.

  • When using the default PST export settings, Intella will try to replace non-exportable items with their parent email. If there is no parent email, the item is skipped.

  • Folder results are always skipped.

All skipped items are listed in the export report.

28.3. Exporting to a CSV file

You can export a results list to a comma separated value (CSV) file. A CSV file contains all information listed in the table. CSV files can be opened in a spreadsheet application such as Microsoft Excel and can be processed through scripting, which opens up new analytical abilities. This functionality can also be used to generate MD5 lists.

To export the table to a CSV file:

  1. Select the results in the table that you want to export to a CSV file.
    You can use the Select All option in the right-click menu to easily select all rows.

  2. click on the selected files and click Export > CSV File…​

  3. In the Column Chooser panel, select all columns that you want to include in the CSV file. See the section about the Details panel for more information on how to operate this panel.

  4. Use the “Include raw data fields” option to include arbitrary raw data fields. A comma-separated list (e.g. PR_MESSAGE_CLASS, PR_MESSAGE_FLAGS) can be used to include more than one field in the CSV export. Note that this list is case-sensitive.

  5. Select “Truncate length of text in cell to 32,000 characters” option to limit the length of values in individual cells, if necessary.

  6. Select “Preserve line breaks” option to keep line breaks in multi-line values (such as comments).

  7. Give the CSV file a name and select Export.

The selected columns are stored in the current table column preset. You can switch between presets and create new and delete presets, as described in the “Table column presets” section. Note that editing the preset currently used in the Details table would result in the visible columns being changed when the CSV export window is closed.

By default, Intella uses the comma character to separate cells and uses a double quote character to escape values containing commas or other special characters. These can be changed using the Column delimiter, Quote character and Escape character drop-down boxes.

The contents of the Senders and Receivers columns are configurable to show either the contact name(s), the email address(es), or both.

The maximum text length of a value inside a cell can optionally be trimmed to 32,000 characters. This is often necessary when one wants to open the CSV file in MS Excel. When opening a CSV with longer texts in Excel, these long texts are typically broken up and roll over to the next row, breaking the table structure.

The CSV format is not a formal standard; different applications may have different conventions on how to separate cells and escape special characters. Intella uses the comma character to separate cells and uses a double quote character to escape values containing commas or other special characters.
To import such files in MS Excel 2010, select Data From Text in the ribbon. Next, select the file in the file chooser. In the wizard that opens next, choose “Delimited”. Set the Delimiters option to “Comma” and set the Text Qualifier to the " character. Click Finish.
To import such files in MS Excel 365, select Data Get Data From File From Text/CSV. Select the CSV file in the file chooser. In the dialog that opens next, set the Delimiter to “Comma”. Click Load.

28.4. Exporting the result counts

The number of hits per search query can be exported by right-clicking in the Searches list in the upper-right corner and selecting “Export queries”. This produces a CSV file with the following columns:

  • Facet – e.g. Type or Keyword Search.

  • Result – the textual representation of the search, e.g. the entered search terms or selected facet values.

  • Total Count – the total number of items that matched this query.

  • Count after Includes and Excludes – the number of items that were retained after applying the Includes and Excludes (if any) to the original set.

28.5. Exporting to an Intella case

Items in a case can be exported to another case. We refer to these as the source case and target case, respectively. This functionality supports several use cases:

  • Exporting selected items to a new (and empty) case, e.g. to filter privileged information or irrelevant items, or to divide the work among reviewers in such a way that each reviewer only has access to their own assigned subset.

  • Merging of two or more cases to create one unified case, or to import the work done on a previously exported case back into the main case.

When items are exported to another case, Intella will add the related sources to the target case, and the items will be added to those new sources. However, when the target case is a copy of the source case, or if items from the source case were exported to the target case before, then the target case can already contain these sources. In such cases Intella will add the exported items to the existing sources. Any items that already exist in the target case will not be added again. Subject to the selected export options, the associated data like tags and comments will be copied though.

Exporting items to a case will increase the registered case size of the target case. Intella calculates the increase based on the size of the selected items and their (recursive) parent items, as far as these contributed to the case size of the source case. For example, exporting a set of email items from a PST file that was stored in a crawled file system folder will increase the size of the target case with the size of that PST file. When the to-be-exported items were indexed with Intella 2.1.1 or older, the required information is not available though, and Intella will add the full size of the source case to the target case instead. If this is problematic, for example due to case size limits of the Intella edition that is being used (see the “Intella editions” section), then you can consider re-indexing the source case with a newer Intella version before attempting to export the items.

The size of the target case after export cannot exceed the limit imposed by your current Intella license.
An Intella case or source that was created by exporting items to it cannot be re-indexed. This is done to prevent the filtered out items from being restored if the original evidence files are available. Sources that cannot be re-indexed will be marked with "exported" label.

To export a set of items to another case, use the following procedure:

  • Run a query that contains all the to-be-exported items.

  • Select the to-be-exported item in the Details view.

  • Right-click on the selected items and choose the Export > Intella Case… option from the context menu.

This opens the “Export to Intella case” wizard:

desktop exporting case

To export items to an existing case, simply select it in the list. * To export to a new Intella case, use the “Create…” button to create the case (see the “Creating a new case” section). When the case has been created, it will appear in the cases list. * To export to an existing Intella case that is not shown in the list, add it using the “Add…” button, as described in the section titled “Opening an existing case not in the list”.

Once the target case has been selected, click the “Next” button to move to the exporting options page:

desktop exporting case options

This page allows you to specify which information should be included in the export, along with the item content:

  • Passwords and authentication tokens that are stored in source configurations (IMAP, iCloud, etc.). When these are not included in the export, the source cannot be re-indexed in the target case.

  • Indexing tasks defined in source configurations.

  • Tags that are referenced by exported items, or the complete tag hierarchy defined in the source case.

  • Custodians that are referenced by exported items, or all custodians defined in the source case.

  • Flags.

  • Comments.

  • Redactions.

  • Excluded paragraphs.

  • OCRed content, with an option to overwrite existing OCR content in the target case for exported items that already exist in the target case.

  • Imported text, with an option to overwrite existing imported text in the target case for exported items that already exist in the target case.

  • Content Analysis results.

To start exporting to another case, press “Export” button. This completes the exporting wizard and will start the export.

The case export currently only includes the crawled data and the options shown above. Also, the timestamps of exported annotations are not yet preserved; their timestamps will be the time of export. This will be extended and improved in future releases.

When exporting items to an Intella case, Intella will export only the items that are in the current selection. If you want to export emails with their attachments, you must include the emails and all their attachments in the selection for exporting.

This functionality therefore allows for specific items to be excluded. E.g., if an email has an attachment and that attachment is privileged (should not be included in the export), the email can be exported without the attachment by simply exporting only the email itself. Note that the binary file associated with that email will still contain the attachment in binary form though! This is therefore not a secure way of filtering out all privileged information.

When items are exported without their parents, their parents will still be represented in the target case by item stubs. These stubs are necessary to show the context of the exported item. An item stub contains a minimal set of metadata of the original parent item, such as its name, location and type.

Intella will record the start and end of the export process in the event logs of both the source case and the target case. Besides the user and timestamp, these events record:

  • The Case ID of the case that the items were exported to/imported from.

  • The Case name of the case that the items were exported to/imported from.

  • A unique Event ID that can be used to identify corresponding import and export events in source and target cases.

28.6. Exporting the social graph data

Intella can export the social graph of a collection of emails by selecting the items in the Details view, right-clicking on one of the selected items and choosing Export > Social Graph…​ This procedure creates a graph data file where all nodes represent contacts and all edges represent the fact that mails have been sent between those two contacts.

This is different from exporting social graph image, which is covered in the Social Graph chapter.

The edges are weighted, with the weight representing the number of mails that have been sent from one contact to another. The edges are directed to differentiate mails from A to B to those sent from B to A.

The graph can be exported into one of the following formats:

  • A CSV file containing three columns: the sender, the receiver, and the number of mails.

  • A GML (Graph Modeling Language) file containing that same information.

  • A GraphML file containing that same information.

A CSV file can be very practical because it can be viewed and edited in spreadsheets and it is easy to write scripts that can process them. Be aware though that CSV is a very informal standard. Different tools may have different rules on how to encode special characters. Some tools that can process CSV graph files may require that the third column be removed.

GML and GraphML are formats specifically designed for specifying graph structures. They can be processed in free tools such as Gephi and NodeXL as well as several commercial applications. As GraphML is based on XML, it offers the best solution for dealing with foreign character sets.

28.7. Exporting the event log

The case event log keeps the history of all actions performed by all users of this case, such as adding new sources, individual searches, taggings, exports and so on. The event log records can be exported to CSV or MS Excel (.xlsx) format for auditing purposes.

To export the event log, select “Export > Event Log…​” in the main menu.

desktop exporting event log

In the Export Event Log window, specify the format (CSV or XLSX), name and path of the CSV file to export to.

In the “Export events between…​” part you can choose the start and end dates of the events to export. By default, the dates range is automatically set to cover all events.

The options in the “Select event types” part allow for specifying the type of events that should be exported.

Furthermore, you also can narrow down the events to cover specific users only, by checking the “Only include events of these reviewers” option and selecting the user(s) in the pop-up box.

The “Export IDs of items associated with the events” option control how item identifiers are exported:

  • Do not export item IDs – disables exporting of item identifiers.

  • Export item IDs to separate text files – identifiers are exported (one per line) to separate text files in the same folder. The files are named per the numbers indicated in the first column of the exported event log records (1.txt, 2.txt, …​). The file format is compatible with the Item ID Lists facet, i.e. this facet can be used to locate these items using the produced files.

  • Export item IDs within the log file – identifiers are included into the last column of the exported event log file. Note that this can produce a significantly larger file.

The “Export item IDs within the log file” option may cause problems when previewing an exported file with long identifier lists in MS Excel. This is related to limitations on individual cell lengths imposed by Excel and does not indicate that the event log file is malformed or corrupted. Using the option for separate text files allows to overcome this issue.

The exported event log file contains the chronologically ordered event log records, one record per line. The records include the following columns:

  • Time – date and time of the event.

  • Event Type ID – a number indicating the type of the event. This can be used for sorting event records by type, for example in MS Excel.

  • User – name of the user responsible for this event.

  • Message – human-readable description of the event.

  • Item IDs – a space-separated list of item identifiers associated with this event (only when the “Export item IDs within the log file” option is selected).

29. Command-line support

Intella supports the use of command-line arguments for tasks such as opening or creating a case and indexing a folder of evidence files.

29.1. Executables and licenses

You can choose between two different executables:

  • Intella.exe will always open the Intella main window with the (new or existing) case open. Use this for e.g. shortcuts or other scripts that should launch Intella’s main user interface.

  • IntellaCmd.exe will only show feedback on the command-line. Use this for automating case creation and indexing.

For IntellaCmd.exe, an Intella Professional, Node, Connect or Investigator license is required.

Intella.exe can be used with all licenses.

29.2. Command-line arguments

Opening a case
To open a specific case the following arguments can be used:

Intella.exe -user <user> -case <case location>

This can be used for both local and remote cases.

If the case folder doesn’t exist, the case will be created automatically at this place with default options. You can also use abbreviated argument names, like this:

Intella.exe -u <user> -c <case location>

Creating a new case
The following instruction creates a new case at a specific location with specific user, name, and description:

Intella.exe -user <user> -case <case location> -caseName <name> -caseDescription <description>

When a case already exists at that location, it is simply opened, and the specified case name and description arguments are not used. In other words, there is no difference in syntax between creating a new case and opening an existing case.

The abbreviated form of the above instruction would be:

Intella.exe -u <user> -c <case location> -cn <name> -cn <description>

Indexing a folder
The following instruction can be used for open or create a new case and index a folder with evidence files:

Intella.exe -user <user> -case <case location> -evidence <evidence location> -sourceName <name>

or in the abbreviated form:

Intella.exe -u <user> -c <case location> -e <evidence location> -sn <name>

Indexing options can be specified using the following arguments:

  • -e, -evidence <evidence location> - this option will accept either file or folder, for example to add disk image as evidence, specify the first part of the disk image.

  • -tz, -sourceTimezone [TZ] - The time zone of the new source (example: -sourceTimezone CET).

  • -ima, -indexMailArchives [true|false] - Index mails and files in mail archives (default: true).

  • -ia, -indexArchives [true|false] - Index files inside archives, such as ZIP and RAR files (default: true).

  • -ie, -indexEmbedded [true|false] - Extract images embedded in emails, MS Office, and PDF documents (default: true).

  • -icm, -indexChatMessages [true|false] - Index chat messages (default: true).

  • -scc, -splitChatConversations [day|week|month|year] - Split chat conversations by day|week|month|year (default: day).

  • -pcma, -presentChatMessagesAs [conversationsAndMessages|conversationsOnly|messagesOnly] - Present chat messages as conversationsAndMessages|conversationsOnly|messagesOnly (default: conversationsAndMessages).

  • -mmpc, -maxMessagesPerConversation [number] - Maximum number of messages per chat conversation (default: unlimited).

  • -iu, -indexUnstructured [true|false] - Heuristically extract text fragments from unsupported and unrecognized file types (default: true).

  • -rd, -recoverDeleted [true|false] - Recover deleted emails, files, and Notes deletion stubs from the supported file types (default: false).

  • -cef, -cacheEvidenceFiles [true|false] - Copy all evidence files into the case folder (default: false).

  • -ap, -analyzeParagraphs [true|false] - Enable paragraph analysis (default: false).

  • -geo, -determineEmailSenderIpGeolocation [true|false] - Enable the determination of the geographic location of an email’s sender IP (default: false).

  • -tf, -taskFile [File] - Specify a .json task file to run after indexing completes, containing e.g. keyword of hash list searches and tagging or exporting the results. Use -runTaskFile to run tasks on an existing case.

  • -nc, -newCase - Convert the case to a new version when available. The '-newCaseName' parameter is required. The confirmation dialog will not be shown.

  • -ncn, -newCaseName - The name of the new case if the case can be upgraded and the '-newCase' parameter was specified.

Following options can only be used with IntellaCmd.exe.

  • -ct, -caseType [compound|regular] Type of the case default is regular.

  • -asc, -addSubCase <sub case location> Specifies a path to a sub case to be added to the compound case when it’s created. If the case already exists then it adds the sub case to the compound case. The option is ignored if a regular case is opened or created

  • -latz, -listAllTimezones - Prints to standard output list all available timezones which are allowed to be used in conjunction with -tz parameter.

  • -sc, -sourceCustodian <custodian> - Custodian to be associated to the source

  • -ict, -importCaseTemplate <path to ICT file> - specifies a path to the case template to be used when creating a new case.

  • -stf, -sourceTypeFilter <comma separated list> specifies a comma separated list of media types or categories that need to be included or excluded depending on the value of the -sourceTypeFilterMode.

  • -stff, -sourceTypeFilterFile <plain text file> same as sourceTypeFilter but specifies a plain text file with a list of media types or categories. Media types are separated by new lines, type per line.

  • -stfm, -sourceTypeFilterMode [include|exclude] specifies how sourceTypeFilter should work. If the mode is set to "include" that means IntellaCmd will only index the types specified in the text file. If the mode is set to "exclude", IntellaCmd will index all types except for the ones specified in the text file.

  • -secs, -sourceEnableCrawlerScript - Enables a crawler script for this source.

  • -scst, -sourceCrawlerScriptType [python|groovy|jar] - Mandatory if sourceEnableCrawlerScript present

  • -scsf, -sourceCrawlerScriptFile <path to script file> - Path to a script file (python, groovy or jar) (mandatory)

  • -scsesl, -sourceCrawlerScriptEnableScriptLog - Enables the script log feature (optional, default: false)

  • -erf, -exceptionReportFile <path to CSV> - When used with one of the indexing options (indexAddedSource, indexNewData or reindexAll), tells Intella to generate CSV file with exceptions.

  • -fef, -fatalErrorsFile <path to CSV> - When used with one of the indexing options (indexAddedSource, indexNewData or reindexAll), tells Intella to generate a CSV file with the fatal errors that are shown in the Errors tab.

  • -addSourcesFromJson - Bulk add data sources defined in JSON data sources configuration file.

The following format should be used to define data sources for "addSourcesFromJson" parameter:
{
	"sources": [
    {
      "name": "Pst archive",
      "path": "d:\\Data\\evidence\\some-evidence.pst",
      "timezone": "Europe/London",
	  "indexMailArchives" : true,
 	  "indexChatMessages" : true,
	  "processingMode" : "CONVERSATIONS_AND_MESSAGES",
	  "splitMode" : "PER_DAY",
	  "numberMessagesPerConversation" : 100,
	  "indexArchives" : true,
	  "indexEmbeddedImages" : true,
	  "indexDatabases" : true,
	  "indexWindowsRegistry" : false,
      "indexWindowsEventLog" : false,
	  "indexBrowserHistory" : true,
	  "recoverDeleted" : false,
	  "extractFullText" : false,
	  "indexVolumeShadowCopies" : false,
	  "cacheEvidenceFiles" : false,
	  "analyseParagraphs" : false,
	  "emailsGeolocationEnabled" : false,
	  "custodian" : "Custodian 1 "

    },
    {
      "name": "Compressed aff4",
      "path": "d:\\Data\\aff4-working\\APFS - AFF4 - Compressed.aff4",
      "timezone": "Europe/Paris",
	  "indexMailArchives" : true,
 	  "indexChatMessages" : true,
	  "processingMode" : "CONVERSATIONS_AND_MESSAGES",
	  "splitMode" : "PER_DAY",
	  "numberMessagesPerConversation" : 100,
	  "indexArchives" : true,
	  "indexEmbeddedImages" : true,
	  "indexDatabases" : true,
	  "indexWindowsRegistry" : false,
      "indexWindowsEventLog" : false,
	  "indexBrowserHistory" : true,
	  "recoverDeleted" : false,
	  "extractFullText" : false,
	  "indexVolumeShadowCopies" : false,
	  "cacheEvidenceFiles" : false,
	  "analyseParagraphs" : false,
	  "emailsGeolocationEnabled" : false,
	  "custodian" : "John Doe"
    }

	]
}

Importing a load file
The following options can be used to import a load file: * -ilf, -importLoadFile [true]false] - Indicates whether to interpret the evidence data as a load file. * -of, -opticonFile [File] - Import images from the specified Opticon file. * -lft, -loadFileTemplate [File] - The import template file to use.

This command can only be used with IntellaCmd.exe.

Add hash filters
* -shf, -sourceHashFilter <list1,list2> - Specifies a comma separated list of hash lists if they have already been added to Intella before. * -ihFCSV, -importHashFilterCsvFile <Hash filter csv file> - Imports the specified CSV file as a hash filter to Intella. Works with additional parameters "importHashFilterName","importHashFilterFirstRowIsHeader","importHashFilterMd5Column" * -ihfn, -importHashFilterName <name> - Name specifies the name of the hash filter to be created. * -ihFih, -importHashFilterFirstRowIsHeader [true, false default false] -Specifies if the first row in the file should be treated is a header. * -ihFMd5No, -importHashFilterMd5Column [number] - Specifies if the MD5 column number in the CSV hash file. The column is 1 based: 1, 2, 3…​

Running tasks
Use the -runTaskFile option to run a .json task file on an existing case. This command can only be used with IntellaCmd.exe.

Note that the -taskFile option is used for running tasks at the end of indexing.

Converting a case
The following instruction can be used for converting an old case (2.1.x or newer) to the current case format:

Intella.exe -user <user> -case <case location> -newCase <new case location> -newCaseName <new case name>

Enhancing the case data through external means

The following commands can only be used with IntellaCmd.exe.

The following options are available to export and import data in a case. This can be used for externally analyzing item content and using that to enhance the case.

  • -et, -exportText - Export the extracted texts to a folder. The options -matchQuery, -savedSearch,
    -deduplicate and -exportDir can be used to control this operation. The resulting files will be named based on their item ID, e.g. 123.txt.

  • -eb, -exportBinaries - Export the item binaries in their original format to a folder. The options
    -matchQuery, -savedSearch, -deduplicate and -exportDir can be used to control this operation. The resulting files will be named based on their item ID, e.g. 123.jpg.

  • -ed, -exportDir [Folder] - When used with the -exportText argument, the folder where the resulting text files will be exported to.

  • -mq, -matchQuery [String] - Can be used to limit the exported items to those that match a specific keyword query.

  • -ss, -savedSearch [File] - Can be used to limit the exported items to those that match the specified saved search. The argument is the path to an XML file holding the saved search. Such a file can be exported from the Saved Searches facet. This allows for using other facets, such as the Date and Type facets, and to combine queries.

  • -dedup, -deduplicate [true|false] - When used with the -exportText command, deduplicate the items before exporting (default: false).

  • -itext, -importText [Folder] - Import text files from the specified folder. The file names must be based on item IDs such as 123.txt, where 123 is an item ID. The files must be in UTF-8 format. A use case for this is item texts that have been processed by a service such as Google Translate. The texts will be shown in an "`Imported Text” tab in the Previewer. The imported texts are subject to keyword search, content analysis, exporting, etc.

  • -itags, -importTags [File] - Import tags from the specified CSV file. The first column must contain an item ID. The remaining columns must contain tags that will be associated with the item.

  • -idup, -importToDuplicates [true|false] - Controls whether the imported texts and tags should only be applied to the specified item IDs, or also to all duplicates.

  • -ei, -exportItems - Export items to a directory. Options '-matchQuery', '-savedSearch', '-deduplicate', '-exportDir', '-exportTemplateFile', '-exportTemplateName' can be used to control this operation.

  • -etn, -exportTemplateName <template name> - Name of the export template that should be used for the export. It must be an already existing export template located in the system folder.

  • -etf, -exportTemplateFile <path to XML> - Path to the export template XML that should be used for the export. It must be an already existing export template located in the system folder.

  • -eel, -exportEventLog <export file location> - Export all events from the case to a CSV file specified by this location.

Listing the defined sources
Use the -exportSourceList or -esl option to export the list of sources in a case, including their configuration, to an XML file:

IntellaCmd.exe -u <user> -c <case location> -exportSourceList <file>

Adding passwords and certificates to key store

The following commands can only be used with IntellaCmd.exe.

The following options are available to add passwords and certificates to a case, so that it is possible to index encrypted data. The general usage is:

IntellaCmd.exe -u <user> -c <case folder> -keystore <operation> <parameters>

For example, this command will add password "123456" to the case key store:

IntellaCmd.exe -u <user> -c <case folder> -keystore -addPassword 123456

Intella will create a new case if the specified case location doesn’t exist.

The supported operations are:

  • -addPassword [password] - adds a single password to the key store.

  • -addPasswordList [password list file] - adds all passwords from a text file (UTF-8).

  • -addNotesIdFile [ID file] - adds a single Nodes ID file that is not protected by password. Use -validateNotesIdFile option to also validate the ID file before adding it.

  • -addNotesIdFilePw [ID file] [password] - adds a single Nodes ID file protected by password. Use -validateNotesIdFile option to also validate the ID file before adding it.

  • -addX509Certificate [certificate file] - adds a X.509 certificate that can be used to decrypt S/MIME emails.

  • -addPgpCertificate [certificate file] - adds a PGP certificate that can be used to decrypt PGP/MIME emails.

  • -addBitLockerKey [key] - adds a single BitLocker recovery key (password).

  • -addBitLockerRecoveryFile [BEK file] - adds a BitLocker recovery file (BEK).

  • -addKeystoreFromJson [JSON file] - adds keystore elements defined in JSON input file to the keystore.

The JSON input file for -addKeystoreFromJson parameter should have following format:
{
	{
	"keys" : [
	{
		"type" : "password",
		"data" : {
				"__comment_1__": "This is comment which is not going to be processed",
			"password" : "password_value",
			"encrypted" : false
		}

	},
	{
		"type" : "passwordList",
		"data" : {
			"file" : "d:\\tmp\\keystore.data\\passwords.txt"
		}
	},
	{
		"type" : "notesIdFile",
		"data" : {
			"password" : "123",
			"file" : "d:\\tmp\\keystore.data\\user1.passwd-123.id",
			"validate" : true
		}
	},
	{
		"type" : "X509Certificate",
		"data" : {
			"password" : "",
			"file" : "d:\\tmp\\keystore.data\\cert.p12"
		}
	},
	{
		"type" : "PGPCertificate",
		"data" : {
			"password" : "password",
			"file" : "d:\\tmp\\keystore.data\\cert-pgp-sec-pass-password.asc"
		}
	},
	{
		"type" : "BitlockerRecoveryFile",
		"data" : {
			"__comment_1__": "We do not support encrypted key data so password is not necessary",
			"file" : "d:\\tmp\\keystore.data\\5BF6E030-7379-42CC-9D92-B2DB81520A28.BEK"
		}
	},
	{
		"type" : "BitlockerRecoveryKey",
		"data" : {
			"__comment_1__": "We do not support encrypted key data so password is not necessary",
			"key" : "181621-236489-662046-063536-387695-066913-184910-400015"
		}
	}

	]
}

Headless mode
It is possible to run Intella in a non-interactive (headless) mode. IntellaCmd.exe should then be used instead of Intella.exe, for example:

IntellaCmd.exe -user <user> -case <case location> -evidence <evidence location> -sourceName <name>

The following commands are supported with IntellaCmd.exe only:

  • -exportSourceList

  • -exportText

  • -exportBinaries

  • -importTags

  • -importText

  • -importLoadFile

  • -runTaskFile

  • -keystore <operation>

Logging
The desired log level can be specified using the -log argument, for example:

Intella.exe -user <user> .... -log DEBUG

Valid options for the log level are:

  • ERROR

  • WARN

  • INFO

  • DEBUG

Both Intella.exe and IntellaCmd.exe support specifying the log level.

Selecting an appropriate license
In some cases, it may be useful to let Intella select a license without showing the interactive license selection dialog. The following options can be used:

  • -autoSelectFullLicense - Automatically select the first available license, ignoring trial and expired licenses. Intella will fail to start if no such licenses were found.

  • -autoSelectFullProcessingLicense - Automatically select the first available license that allows for case creation, ignoring trial and expired licenses. Intella will fail to start if no such licenses are available.

  • -autoSelectLicenseType <type> - Automatically select the first available license of the given type, ignoring trial and expired licenses. Intella will fail to start if no such licenses are found. The following types can be used:

    • I10 - Intella 10

    • I100 - Intella 100

    • I250 - Intella 250

    • VIEWER - Intella Viewer

    • PROFESSIONAL - Intella Professional

    • NODE - Intella Node (applicable only with IntellaCmd.exe)

  • -keyID - Only the licenses from the given key ID will be used. This option may be useful when you have more than one dongle. This parameter can be combined with any of the other three license selection parameters.

Be aware that trailing backslashes in paths may give issues when the path is placed between quotes. Either try to remove the trailing backslash, or escape it: "C:\My Docs\" becomes "C:\My Docs\\"

29.3. Example: invoking Google Cloud AI services

This section describes an example batch script, showing how Google Cloud’s AI services can be used to extend an Intella case with entity extraction and content categorization. A Windows batch script is used to index evidence data, submit specific items to Google Cloud for analysis, and import the results back as tags in the case.

The approach described here is not restricted to just these two types of case data enrichment, or to only Google Cloud’s offerings. Examples of other types of document enrichment that are made possible this way are:

  • Document translation.

  • Sentiment analysis.

  • Key phrase extraction.

  • Summarization.

  • Document classification, using classification models specifically trained for a case or task (i.e. predictive coding).

Other types of media may also be analyzed, e.g.:

  • Audio and video files:

    • Spoken language detection.

    • Transcription.

  • Image and video files:

    • Object detection, e.g. vehicles, weapons, drugs, people.

    • Topical detection, e.g. logo’s, celebrities, landmarks.

    • Face detection and recognition.

    • Explicit content detection.

    • Finding similar media on the web.

Such services are currently available from several providers, including Google, Amazon, IBM, Microsoft, and others.

Preparation
Ensure that Intella 2.2 or above is installed. The method outlined below relies on command line instructions that have been added in Intella 2.2.

Prepare access to Google Cloud:

  • Visit https://www.google.com and register for a Google account.

  • Visit https://cloud.google.com and click Try GCP Free to let the account access the Google Cloud Platform. Fill in the required details. At the time of writing, Google offers USD$ 300 of free resource usage in the first 12 months.

  • Click on Home button (top left corner) > Getting started > Create an empty project. Enter an arbitrary number, e.g. "Intella Test" and click Create.

  • Click on the project chooser dropdown in the top navigation bar and select the newly created project. The project may display an animated icon, indicating that the project is still being initialized. One can only select the project once it has been initialized.

  • In the APIs box, click on Go to APIs overview > Enable APIs and services. You may also be greeted by the Getting started page, then just click on APIs.

  • Search for Cloud Natural Language API. Select it and click the Enable button.

  • Optional step: Click on Home > APIs and Services > Credentials. Click the Create credentials button and select API key. A dialog opens that shows the key. Copy this key to a text file. This key is not needed for our example script, as the gcloud utility handles authentication and authorization transparently. It is needed for other services such as Google Translate, which are not supported by the gcloud utility. Google Translate does expose a REST web service, in which this key will be used.

  • Alternative optional step: one can also consider registering a service account. Service accounts make it possible to access files stored on Google Cloud Storage, allow for more advanced access restrictions and auditing, and are not as easy to leak as API keys. Setup and use of service accounts is beyond the scope of this tutorial.

  • Visit https://cloud.google.com/sdk and click the Install the Cloud SDK link.

  • Download and run the installer. At the end of the installation wizard, select the option to run the init procedure. When asked, enter/select your Google account, and select the project defined earlier. This will set the default authentication and project for the gcloud command; no access credentials and project identifier will be needed when running entity recognition or text classification.

Step by step description of the batch script
At the end of this section, two Windows batch scripts are listed; one for running entity extraction on a case, and one for text classification. This section provides a step-by-step explanation of the entity extraction script. The text classification script is almost identical.

First, we declare some variables that make the script easier to read and manage. Also, it alleviates a current limitation in Intella’s command line parameters, namely that all files paths specified as parameters to IntellaCmd.exe need to be absolute.

:: Prepare state
set INTELLA="C:\Program Files\Vound\Intella 2.2\IntellaCmd.exe"

set BASE="C:\Intella GCP Test"

set CASE=%BASE%\Case +
set EVIDENCE=%BASE%\Evidence +
set EXPORT=%BASE%\Export +
set ENTITIES=%BASE%\Entities +
set TAGS=%BASE%\Tags

To facilitate experimenting, we clean up all folders produced by this script. This will have to be adjusted for a production environment, as it makes it very easy to lose work this way.

:: Clean up from previous runs
rmdir /Q /S %CASE% +
rmdir /Q /S %EXPORT% +
rmdir /Q /S %ENTITIES% +
rmdir /Q /S %TAGS%

First, we index all evidence files into a new case:

:: Index the evidence files into a new case
%INTELLA% -user John -case %CASE% -caseName "Google Cloud Entity Extraction Test" -evidence %EVIDENCE% -sourceName "Emails" -indexAddedSource

Next, all items are exported as plain text files:

:: Export the bodies of all items as text files
%INTELLA% -user John -case %CASE% -exportText -exportDir %EXPORT%

This produces a collection of text files. The file name is the numeric ID of the item in the case.

All data is now ready for analysis by Google Cloud’s entity analysis service. We loop over all exported files and invoke gcloud to upload the text and extract all entities in it. gcloud will automatically use the account credentials and the selected project that was entered during the init procedure of the Cloud SDK installer. For each item, the output is written to a file in the Entities folder, with the same name as the original text file.

:: Use Google Cloud entity analysis on all text files.
:: […]
mkdir %ENTITIES%

for %%f in (Export\*.txt) do (
   echo %%f
   call gcloud ml language analyze-entities --content-file=%%f
      --flatten="entities[]" --format="csv[no-heading,separator='/',
      terminator=','](entities.type,entities.name)" >> %ENTITIES%\%%~nf.txt
)

The standard gcloud output format is a JSON document. As this is hard to process in a Windows batch script, we set the format to CSV, restricted to listing the entity type (e.g. "ORGANIZATION", "PERSON") and the entity name (e.g. "Enron", "John Doe"). Normally, this produces a two column CSV file, with the type in the first column and the name in the second, and with "`name” and "`type” headers in the first row. By changing the comma to a slash, changing the line terminator to a comma, and by suppressing the heading, we produce a single line of comma-separated entities, like this:

ORGANIZATION/Enron,PERSON/John doe,

This is already very close to the hierarchical tag format that Intella will import at the end of the script.

Next, we concatenate all text files into a single CSV file and add the file base name (i.e. the item ID) as the first column in that file.

:: Concatenate all the entities to a file that can be imported as a tag file.
:: […]
mkdir %TAGS%

for %%f in (Entities\*.txt) do (
   if %%~zf gtr 2 (
      echo | set /p="%%~nf," >> %TAGS%\tags.csv
      type %%f >> %TAGS%\tags.csv
      echo.>> %TAGS%\tags.csv
   )
)

The line may now look like this:

53,ORGANIZATION/Enron,PERSON/John doe,

A few tricks are needed to make this work in a Windows batch script. The echo command normally ends its output with a newline. To prevent the first echo invocation to put a newline after the item ID and comma, we let the set command handle the output of the item ID and the comma. The type command does not produce a newline, hence the need for a second echo call after it. The test for files greater than 2 bytes is there because some items do not produce any entities. In that case, the entity file will contain the text "/,", which when added to the CSV file will result in an error; redundant commas are accepted but redundant slashes are not.

Finally, we are ready to import the found entities as tags into the case:

:: Import the entities as tags into the case +
%INTELLA% -user John -case %CASE% -importTags %TAGS%\tags.csv

Script considerations
The Windows batch script demonstrates how information in an Intella case can be enriched using an external service such as Google Cloud and how the results can be imported into the Intella case. The only requirement is the installation of the Google Cloud SDK and some online project setup.

Still, the script has many shortcomings that one likely needs to address for production usage. These may be tackled in a future version of the script.

  • The script does not escape any special characters. For example, in the entity extraction output we have seen entity names containing quote characters. This produces some interpretation issues with the generated tags file.

  • Email bodies also contain metadata of other emails, such as senders, receivers, and subjects. These may affect the output of certain analysis techniques. Ideally, these headers would be filtered out before the text is being submitted for analysis.

  • The gcloud command-line utility does not (yet) support invoking Google’s text translation service, or several of the other mentioned services. For that, a different approach is needed, e.g. utilizing curl and parsing the JSON output that comes back.

  • The Windows batch scripting language is weak and brittle. Windows PowerShell may be a better alternative, with built-in facilities for invoking REST web services and parsing JSON output, as well as better options for text processing.

  • The Google Cloud entity recognition and text classification output contain salience and confidence numbers respectively. These can be used to suppress the weaker entities and categories, or to put them in separate categories.

  • In the Windows Task Manager we observed that memory consumption of the cmd.exe process rose to hundreds of MBs during the concatenation of the entities files into a single CSV file. In this case, a collection of only 4000 files were processed, with a resulting CSV file of only 5 MB.

  • Both scripts upload the same data to Google Cloud. For realistic case sizes, the data is ideally moved to the cloud in one batch upload, after which several analysis types can be applied without requiring the data to be uploaded each time. Data storage in the cloud also comes at a price though. Another benefit is speed; the script is single-threaded, i.e. all gcloud invocations run purely sequentially, and the analysis of a single email (incl. upload and download) takes about 1-3 seconds each. The script makes no use of horizontal scalability typically provided by cloud technologies.

  • We present a single script to emphasize that everything can be fully automated and to encourage experimenting. In real-life situations, one may want to introduce some breakpoints to perform manual inspection of the intermediary results, e.g. review the quality of the exported items before mass-analysis in the cloud or do a sanity check on the structure of the tags file before importing it into a case.

Full entity extraction script
The script below can also be found in the Intella installation folder, see scripts\RunEntityExtraction.bat.

:: Prepare state
set INTELLA="C:\Program Files\Vound\Intella 2.2\IntellaCmd.exe"

set BASE="C:\Intella GCP Test"

set CASE=%BASE%\Case
set EVIDENCE=%BASE%\Evidence
set EXPORT=%BASE%\Export
set ENTITIES=%BASE%\Entities
set TAGS=%BASE%\Tags

:: Clean up from previous runs
rmdir /Q /S %CASE%
rmdir /Q /S %EXPORT%
rmdir /Q /S %ENTITIES%
rmdir /Q /S %TAGS%

:: Index the evidence files into a new case
%INTELLA% -user John -case %CASE% -caseName "Google Cloud Entity Extraction Test" -evidence %EVIDENCE% -sourceName "Emails"

:: Export the bodies of all items as text files
%INTELLA% -user John -case %CASE% -exportText -exportDir %EXPORT%

:: Use Google Cloud entity analysis on all text files.
:: The default CSV structure is a two-column setup, with the entity type (e.g. "PEOPLE")
:: in the first column and the entity name (e.g. "John Doe") in the second column.
:: By changing the default separator from ',' to '/' and the default terminator from '\n'
:: to ',' , we put all entities on one line as hierarchical tags.
mkdir %ENTITIES%

for %%f in (Export\*.txt) do (
   echo %%f
   call gcloud ml language analyze-entities --content-file=%%f --flatten=""
      --format="csv[no-heading,separator='/',terminator=','](entities.type,entities.name)"
      >> %ENTITIES%\%%~nf.txt +
)

:: Concatenate all the entities to a file that can be imported as a tag file.
:: Such a file starts with an item ID, followed by a comma, followed by a comma-separated
:: list of tags.
:: gcloud does not escape quotes, so we still need to do that somehow
mkdir %TAGS%

:: The use of set is necessary in order to prevent the first echo from adding a newline
:: The test for files greater than 2 bytes is to skip items that produced no entities.
:: Their content typically equals "/," which would break the CSV structure.
for %%f in (Entities\*.txt) do (
   if %%~zf gtr 2 (
      echo | set /p="%%~nf," >> %TAGS%\tags.csv
      type %%f >> %TAGS%\tags.csv
      echo.>> %TAGS%\tags.csv
   )
)

:: Import the entities as tags into the case
%INTELLA% -user John -case %CASE% -importTags %TAGS%\tags.csv

Full text classification script
The script below can also be found in the Intella installation folder, see scripts\RunClassification.bat.

:: Prepare state
set INTELLA="C:\Program Files\Vound\Intella 2.2\IntellaCmd.exe"

set BASE="C:\Intella GCP Test"

set CASE=%BASE%\Case
set EVIDENCE=%BASE%\Evidence
set EXPORT=%BASE%\Export
set CLASSIFICATIONS=%BASE%\Classifications
set TAGS=%BASE%\Tags

:: Clean up from previous runs
rmdir /Q /S %CASE%
rmdir /Q /S %EXPORT%
rmdir /Q /S %CLASSIFICATIONS%
rmdir /Q /S %TAGS%

:: Index the evidence files into a new case
%INTELLA% -user John -case %CASE% -caseName "Google Cloud Classification Test" -evidence %EVIDENCE% -sourceName "Emails"

:: Export the bodies of all items as text files
%INTELLA% -user John -case %CASE% -exportText -exportDir %EXPORT%

:: Use Google Cloud text classification on all text files.
:: We change the default terminator from '\n' to ',' so that it conveniently
:: puts all classifications on one line.
mkdir %CLASSIFICATIONS%

for %%f in (Export\*.txt) do (
   echo %%f
   call gcloud ml language classify-text --content-file=%%f --flatten="categories[]"
      --format="csv[no-heading,terminator=','](categories.name)" >> %CLASSIFICATIONS%\%%~nf.txt
)

:: Concatenate all the classifications to a file that can be imported as a tag file.
:: Such a file starts with an item ID, followed by a comma, followed by a
:: comma-separated list of tags.
mkdir %TAGS%

:: The use of set is necessary in order to prevent the first echo from adding a
:: newline.
for %%f in (Classifications\*.txt) do (
   echo | set /p="%%~nf," >> %TAGS%\tags.csv
   type %%f >> %TAGS%\tags.csv
   echo.>> %TAGS%\tags.csv
)

:: Import the classifications as tags into the case
%INTELLA% -user John -case %CASE% -importTags %TAGS%\tags.csv

29.4. Example: using relative paths with command line

IntellaCmd.exe does not support relative paths directly. However, it is possible to use relative paths with the help of a batch script. Below is an example of a script that would index a file in the current directory:

setlocal

set INTELLA_PATH=c:\Program Files\Vound\Intella 2.6.0.3
set CASES_FOLDER=c:\cases
set CURRENT_DIR=%cd%
set EVIDENCE=%1

"%INTELLA_PATH%\IntellaCmd.exe" -u user -c "%CASES_FOLDER%\%EVIDENCE%" -e "%CURRENT_DIR%\%EVIDENCE%" -ias

endlocal

The script can be used like this:

> intella-index.bat test.pst

30. General Checklist

It is advisable to follow a repeatable set of procedures to ensure quality control and repeatable results. Exactly what procedures are needed will vary from case to case and even from organization to organization. We have created a short checklist of procedures that you should look to follow. It is by no means an exhaustive list, but it is designed to assist your organization to develop their own checklist.

30.1. Before you begin

Quality control
Create your own checklist of tasks to ensure consistent and reputable results. The checklist should ensure that no data or results are missed. The checklist should cover all processes from early planning to final archiving.

Appropriate system setup
When scoping a processing system, you should ensure that your system meets the minimum hardware specs for the version of Intella that you are using and the data that is to be processed. The recommended hardware and software specs are shown elsewhere in this user manual and on our online forum: http://community.vound-software.com/.

Check that you have enough disk space on the system. This includes the Case drive, Evidence drive, and Optimization drive. Note that compressed items with be expanded during the indexing process. You will need to allow for more free space, over and above the size of the evidence to be indexed.

There are several memory settings that can be configured in Intella. Depending on the data that is being indexed, you may need to set Intella to use more memory and crawlers for the indexing phase. We have created a forum post that explains the memory settings for Intella. This should be reviewed, and the memory setting should be configured to work with the type of data that is being indexed and the type of hardware your system has.

The Case, Evidence, and Optimization drives should be local drives within your processing system. Indexing from/to USB attached drives, or network locations is not recommended. Any brief dropouts or I/O issues when indexing from USB or network drives can cause data to not index properly, or your case can become corrupt.

Version and case Control
Check for new versions of Intella before starting a new case. The “Check for updates” feature in Intella’s Preferences will alert the user that a new update is available. If you have this feature turned off, you should check for updates on our website before proceeding with creating a case and indexing data.

Note that Intella updates that involve the first two digits of the version number (e.g. the 2 or the 1 in the 2.1.0.3 release) require the dongle to be updated as well. If the customer has at least 60 days left on their maintenance agreement (MA) at the date of release, they can use the “Check for updates” feature in the Dongle Manager to update their dongle.

Note that some antivirus and security applications can cause the dongle licensing service to not work properly. We recommend that you white-list the Intella and Sentinel files in these security applications to minimize dongle and license issues.

Evidence data can contain corrupt data that Intella may not be able to process completely. It is important to have a plan to identify and mitigate corrupt data.

Through our customer support channel, we have seen several instances where the evidence data was corrupt or malformed, and it could not be indexed fully. Examples are PST files (often created by 3rd party tools) where the PST does not conform to the official specifications. When you encounter this type of data corruption, please contact our support team as they may be able to help.

30.2. Creating a case

Encryption
After you have created your case, and before you index any data, make sure you have considered encrypted items that may be in the dataset. Not considering encrypted items may lead to data being missed during searching and analysis.

The user should use the Key store feature (File > Key store) to load any known passwords, ID files, certificates and keys to be used when encrypted items are encountered during indexing.

Adding data to a case and processing
Verify that the type of evidence data is supported by Intella. A list of supported data types is shown in this user manual. Check whether there are any documented limitations with the type of your evidence data in the user manual.

As noted above, we recommend that you copy the evidence data to a local drive within the processing system. Indexing data from USB-attached drives and network locations is not recommended.

30.3. Post processing tasks

Exceptions
One of the first tasks to complete after indexing, is to check the Exceptions Report for any processing issues. An Exceptions Report can be produced by choosing Sources > Exceptions Report. This produces a XLSX or CSV file that lists all items that had issues during indexing. This can range from minor issues such as date parsing problems, to file corruptions that affect an entire container and all its nested items.

These processing issues can also be seen in the Features facet, under Exception items.

Version and Case Control
Check that the case version matches or is compatible with the version of Intella you are planning to use in the review phase. Know that some cases will need case conversion if made with an older version of Intella. See the bundled Release Notes for more information.

Completeness of data
Check for items that do not contain machine-readable text and may need to be OCRed. Note that Empty documents (documents that have a non-zero size but do not contain text, e.g. scanned documents) will not be respondent to keyword search criteria unless they have been OCRed. These types are shown in the Features facet, under Empty Documents.

Other file types, e.g. image files such as TIFF and JPG, may also need to be OCRed. You should agree which types are subject to OCR processing with your client.

Data quality control
Encrypted items will not be respondent to keyword search criteria unless they can be decrypted by Intella using the specified credentials or when they have been decrypted prior to indexing. You should check which items are still encrypted after the indexing process. A list of the encrypted items can be exported from the Workflow section in the Insight tab. This list can be provided to the client, who should advise on which files should be decrypted with further processing (e.g. by using a dedicated decryption program).

Check that the data and header information appear to be correct for the indexed data. As mentioned above, some evidence data can be corrupt, and may consequently not index fully. Not checking that the indexing has completed properly could lead to data being missed.

30.4. Searching and analysis

Report and result quality
It is important to have your work peer-reviewed to ensure that the findings and number of results are correct. Simple user mistakes such as not selecting a tag, or not turning off an Include or Exclude search during your searching and analysis, can invalidate the results that you get. A peer review to show how you got the results will add to the accuracy of those results.

A peer review may also help identify whether evidence data is corrupt.

30.5. Export

Report and result quality
Any exports should be verified for correctness before handing them over to the client. The items that have been exported should be checked with the items in Intella which are tagged and ready for export, to confirm that the numbers match. The last part of the export report should also be reviewed to see whether there were any errors during the export process.

The format for the items being exported should be agreed with your client prior to exporting. This is more so with complex formats such as load files. We provide a separate load file check lists in this user manual to help with this process. You must be in full agreement with the client (client sign-off) as to all the settings and requirements for the load file, before running the production. In some cases, you may be required to produce a small test load file so that the client can ensure that all settings and deliverables are correct.

30.6. Backup

Continuity, Availability, Integrity
We advise running frequent backups for your cases. It only takes a drive to fail or a case to become corrupt, and you lose all your previous work. A corrupt or inaccessible case can be frustrating. But without a backup, it is difficult or impossible to get the work which has been done back.

We recommend creating a backup right after indexing completes, and before starting the review/investigation. In the event the case becomes corrupt during review, the annotations (tags, flags, comments, etc.) from the corrupt case can be imported into this backed up-case. For more information, see the Restoring Annotations section in this user manual.

Along with backups, it is just as important to periodically test the restore process for the backup. The last thing you want to discover is that the backup was not actually working properly.

When archiving old cases, consider archiving a copy of the Intella installer that created the case along with the data. This will help to speed up opening of that case in the future.

Ensure you have thorough notes on case specifics in the archive. This will assist you in the future. What data, what version, any issues or errors that may have occurred.

31. Load file checklist

To help battle the complex nature of load file exporting and importing, we provide checklists for use with Summation (formerly iBlaze) and Concordance / Relativity load files, as well as a general checklist for common issues and solutions with load files.

The Summation checklist contains general considerations that can also apply to the use of other load file formats.

The Concordance / Relativity checklist is specific to the standards used by the US Securities and Exchange Commission (SEC) and US Department of Justice (DoJ).

31.1. Load file diagnostics

This section is intended to identify common issues in the preparation and presentation of load files. It is addressed to individuals using Intella to produce load files for third party applications like Concordance, Relativity, Ringtail and/or Summation.

Load file purposes
It is useful to understand what the final recipient of the load files is going to do with them. Your client will use your load files:

  • To create document records in their database.

  • To upload field values for the records in their database.

  • To upload native files associated with the records in their database.

  • To upload image files associated with records in their database.

  • To upload the text associated with records in their database.

Load file errors
Essentially these come in two types:

  • Data-related

  • Presentation-related

Common mistakes – avoiding them
There are several common errors that can occur when producing load files. Most of them can be avoided by following some simple rules of planning:

  • Obtain a specification.

  • Carefully consider which columns will be needed, as adding lots or even all columns to the export settings can result in very large and unwieldy load files.

  • Run a test ahead of time with a small selection of the real data that you have put through the process.

  • Do not manually edit the load file after it has been produced – it is terribly easy to introduce further errors when doing this.

  • Use templates for your export settings.

  • Develop and use a checklist to quality control a sample from your export and match it to the specification provided.

Common mistakes – correcting them

  • Incorrect field names – If you have presented data to a client and they are complaining that the field names do not fit the requirements, repeat the export. When you come to specify the fields remember that you can customise the label for any field that you choose to export.

    The example below shows exporting of the value extracted by Intella from the Sent field as a date time value and exporting it to a field called “EM Sent Date” in the load file.

    desktop load file edit

  • Field type / data presentation mismatch – When there is a field type mismatch this will frequently be where the date export options have not been set. The presentation of date values required by the various review systems varies and you will need to consult your client to determine the right settings to use.

    These are set using the following page in the export process:

    desktop load file options

Detailed Troubleshooting
When you are trying to understand why a load file will not load, here are some suggestions as to how to proceed. Note that Concordance, Relativity, Ringtail, and Summation have different requirements in terms of data load files and each also provides differing levels of diagnostics and error messages to assist in troubleshooting. The following suggestions are aimed mainly at Concordance and Relativity but may also provide some assistance with Summation.

When following these steps, it is best practice to work from a copy of the load file.

Open the load file with a text editor and check the following:

  • What is the encoding and is it correct? (Summation only)
    If you are exporting data from Intella the only issues that can occur with encoding are for Summation. For all the other load file formats, the system defaults to using UTF-8, which your client should be able to use.

    If incorrect, regenerate the load file from Intella with updated settings.

  • Confirm that field names are correct
    These should appear in the first row, which should happen automatically for the formats where it might apply. Are field names spelt correctly?

    Best practice is to update the Intella settings (and template) and regenerate the load file to avoid introducing errors using manual editing.

  • Is the ID field the first field and is the field order correct?
    Having the ID field first is best practice and makes for easier diagnostics but isn’t necessary. The field order can be specified in the export process.

  • Are delimiters present and correct?
    It is highly unlikely that these will be wrong since Intella uses the default values and you cannot change them. If the problem appears to be with data appearing in the wrong field, it may be that there is a data problem, e.g. delimiters appearing in the data such that they are misinterpreted during the import process.

    Intella is like MS Excel in that it only inserts quotes as text qualifiers in csv files if they are needed. This can cause issues when using the CSV load file format with some of the review platforms.
  • Is it a data issue?
    In many ways data issues are the hardest to diagnose. The best technique is to start with any indication from the software as to where the issue occurs – references to error in line X can be helpful in this regard.

    If there is no indication of where the error might be, an approach is to edit the load file manually (always use a copy), to slice it up into chunks (vertically first and then if necessary horizontally) and try loading each chunk until you find the error.

    First of all, try loading just the header line and the first record. If this succeeds, your field names are right and the non-null fields in the first line have correct data types. You can then try loading the first half of the load file. If this works then the issue is somewhere in the second half of the file. Load the first half of that and so on.

    Using this approach you will usually be able to identify an individual record, or possibly a set of records, which will not load. At this point you slice the records up “horizontally” by test loading each field in turn, eliminating those fields that load until you find the culprit.

In practice, most difficulties arise from three sources – (1) failing to check the quality of the finished product against the specification, (2) somebody editing the load file after it was generated (and introducing errors) or (3) from delimiters being present in the data.

31.2. Summation

Due to the customization options available in Summation, it is often the case that no two clients will have the same load file specifications or requirements.

To ensure the most professional outcome when working with Summation load files, it is highly recommended that you (the Intella user) engage with the recipient of the load file (the client) at the beginning of any engagement or well before any deadline to produce a load file.

Our suggested workflow would be along the lines of the following steps:

  1. Ask the client to confirm that they require a Summation load file.

  2. Supply the checklist below to the client and ask them to complete it.

  3. Collect the completed checklist.

  4. Ask the client to confirm that they have added any extra fields they require to their Summation installation.

  5. Make any changes to the Intella Summation load file export options that are needed to comply with the client’s requirements.

  6. Create a load file from a test data set similar to the data set of this engagement.

  7. (*) Test the load file in your own Summation installation with the client’s configuration.

  8. Ask the client to verify that the sample load file imports correctly in their own Summation installation on all fields, OCR and so on.

  9. If not, make any corrections needed and repeat steps 5 to 8.

  10. Once it imports correctly, ask the client to sign-off on this format.

  11. Save the export options as a custom export template in Intella.

  12. Use the custom export template for producing the final load file(s).

(*) When creating a Summation load file as a part of your engagement, it is highly recommended that you have sufficient qualifications for Summation to understand and troubleshoot any issues that may arise.

Furthermore, it is also highly recommended that you have a copy of Summation in-house that you can use to test and improve the output of your work.

The next pages contain a sample checklist that you can use with your client.

31.2.1. Load File Engagement Checklist

Due to the customization options available in Summation it is often the case that no two clients will have the same load file specifications or requirements.

The information below provides you with a list of options to configure a Summation load file, as part of the proposed engagement.

Please complete the sections A to D and return the form to our litigation support team prior to this engagement.

Options Description Completed
YES | NO

Table selection

Option A

Summation E-tables or Stdtable

Option B

Additional Summation Fields

Document Rendering

Option C

Document numbering during rendering

Option D

Document exclusions

Dates

Option E

Date Selection

Option F

Other options

Option A: Summation E-tables or Stdtable
The following represents the standard Summation E-table or Stdtable offered in Intella. Please select the fields that you require for the production of the load file requested. To do this ticking the appropriate check boxes on the right-hand side of the table. If a particular field is not present, write it in at the bottom of the table and send us those details.

Token Field Description Required

@DOCID

DOCID

Document ID

Required

@ATTACH

ATTCHIDS

Document IDs of attached documents

@PARENTID

PARENTID

Parent document ID

Required

@MEDIA

MEDIA

Document category: eMail or eFile

Required

@FOLDERNAME

FOLDER

File or email location (e.g. Bob.pst/Top of Personal Folders/Inbox)

Required

@DOCTYPE

DOCTYPE

MIME type of document (e.g. application/pdf)

@DOCTITLE

DOCTITLE

Document title

@AUTHOR

AUTHOR

Name of the document’s author

@EDITEDBY*

EDITEDBY*

Other authors or contributors of the document

@DATECREATED

DATECRTD

Creation date of the document

@DATESAVED

DATESVD

Last modification date of the document

@SUBJECT

SUBJECT

Email subject

@FROM

FROM

Email sender

@TO

TO

Email recipient(s) (TO field)

@CC

CC

Email recipient(s) (CC field)

@BCC

BCC

Email recipient(s) (BCC field)

@DATERCVD

DATERCVD

Date that the document was received

@TIMERCVD

TIMERCVD

Time that the document was received

@DATESENT

DATESENT

Date that the document was sent

@TIMESENT

TIMESENT

Time that the document was sent

@INTMSGID

INTMSGID

Internet message ID

@READ

READ

Whether the email message was read (Y or N)

@HEADER

HEADER

Email message headers

@EMAIL-BODY

BODY

Email message body

@ATITLE

ATITLE

Name of the email attachment(s)

@PGCOUNT

PGCOUNT

Page Count

@HASHCODE

HASHCODE

MD5 hash

Required

@IID*

IID*

Intella item ID. Required to locate items in Intella.

Required

<Add custom…>

Option B: Additional Summation Fields
It is also required that you include two additional fields that are not listed as standard E-tables in Summation.

The additional fields are:

ETABLE Description

1

IID

Used to identify the item in Intella.

2

EDITEDBY

Taken form the Intella Authors and Contributors Facet and used to identify which user edited or created the document.

These fields and any additional custom fields that are not default in Summation need to be added using the Summation “Form Editor” before the load file is imported.

Option C: Document Numbering during Rendering

Option Description Value

Numbering scheme

How the documents and pages are numbered

ABC.001.00001

Starting number

The number of the first document

1

Starting folder

The starting folder number

1

Starting box

The starting box number

n/a

Page rollover

Maximum number of pages per folder

99999

Grouping scheme

☑ All files in one folder
☐ Prefix, Folder

Please identify the positions where you would like the DocID and numbering options to be displayed on the rendered images:

Positions Placement

desktop load file positions

Please identify at what position you would like the DocID to be located in the rendering:

Position 1: …​…​…​…​…​…​…​…​…​…​…​…​
Position 2: …​…​…​…​…​…​…​…​…​…​…​…​
Position 3: …​…​…​…​…​…​…​…​…​…​…​…​
Position 4: …​…​…​…​…​…​…​…​…​…​…​…​

Also indicate if you would like either to:

. Show the same DocID, for a particular document, on all pages of that document. . Increment DocIDs for each subsequent page of a document, e.g. first page ABC.001.00001, second page ABC.001.00002, etc. . Do not stamp the image files stamp with the DocID.

Option D: Document Exclusions
Some documents formats such as spread sheets do not render very well as images. On occasions, a single spread sheet may generate thousands of rendered image files that will have no value to the reviewer. It is recommended that you opt to exclude certain file types for final rendering.

File Extension Comment EXCLUDE
YES | NO

Spreadsheets

Recommended to exclude

CSV

Recommended to exclude

XLS

Recommended to exclude

XLSX

Recommended to exclude

Custom

Option E: Date Formatting
The date format in the load file needs to match the date format selected in the Summation’s default settings. Failure to do so may cause the day and month to be reported incorrectly in Summation when reviewing.

Although the dd/MM/yyyy date format is the standard in many countries, some clients prefer to use dd/MMM/yyyy. This format is preferred because there can be no mistake interpreting the date. For example, the date 4/5/2013 could be interpreted as either 4 May 2013 or 5 April 2013. Using the format dd/MMM/yyyy, the date will be displayed as 4/Apr/2013.

Option Description Value

Date format

How to format date only fields

☑ dd/MM/yyyy
☐ MM/dd/yyy
☐ dd/MMM/yyyy
☐ Other: …​…​…​…​…​…​…​.

Time format

How to format time only fields

HH:mm:ss

Date/time format

How to format full date/time fields

dd/MM/yyyy HH:mm:ss

Option F: Other options

Option Description Value

File encoding

UTF-8

Native files

Include native files?

☑ Yes / ☐ No

Image files

Include image files?

☑ Yes / ☐ No

Image format

☑ PDF
☐ TIFF
☐ PNG

Text files

Include extracted text?

☑ Yes / ☐ No

31.3. Concordance / Relativity

US Securities and Exchange Commission (SEC) standard
Please select the fields that you require for the production of the requested load file by ticking the corresponding checkboxes on the right-hand side of the table.

If a particular field is not present, add it at the bottom of the table and send us those details.

Field Description Required

FIRSTBATES

First Bates number of native file document/email

LASTBATES

Last Bates number of native file document/email

BEGATTACH

First Bates number of attachment range

ENDATTACH

Last Bates number of attachment range

PARENT_BATES

First Bates number of parent document/email

FROM

Email: Sender

TO

Email: Recipients (To, Cc, Bcc)

SUBJECT

Email: Subject

DATE_SENT

Email: Date the email was sent

TIME_SENT

LINK

Hyperlink to the email or native file document

MIME_TYPE

The content type of an email or native file document

AUTHOR

Author of the document

DATE_CREATED

Date the document was created

TIME_CREATED

Time the document was created

DATE_MOD

Date the document was last modified

TIME_MOD

Time the document was last modified

DATE_ACCESSD

Date the document was last accessed

TIME_ACCESSD

Time the document was last accessed

PRINTED_DATE

Date the document was last printed

FILE_SIZE

Size of native file document/email in bytes

PGCOUNT

Number of pages in native file document/email

PATH

Document location

INTMSGID

Email message ID

MD5HASH

MD5 hash

TEXT

Extracted text of the native file document/email

<Add custom…>

US Department of Justice (DoJ) standard
Please select the fields that you require for the production of the requested load file by ticking the corresponding check boxes on the right-hand side of the table.

If a particular field is not present, add it at the bottom of the table and send us those details.

Field Description Required

COMPANIES

Company submitting data

HASHMD5

Document MD5 hash value

BEGDOC#

Start Bates

ENDDOC#

End Bates

DOCID

Must equal the value appearing in the BEGDOC# field and be UNIQUE

NUMPAGES

Page count

PARENTID

Parent record’s BEGDOC#

FOLDERLABEL or FILEPATH

Document location

FROM

Email: Sender

TO

Email: Recipients (To, Cc, Bcc)

SUBJECT

Email: Subject

DATECREATED

Date electronic file was created

DATESENT

Date the email was sent

TIMESENT

Time email was sent

DATERECEIVED

Date email was received

TIMERECEIVED

Time email was received

HEADER

The internet header information for email sent through the internet

INTERNETMSGID

Globally unique identifier for a message which typically includes message ID and a domain name

DATESAVED

Date native file was last modified

DATEPRINTED

Date native file was printed

EAUTHOR

Author of the document

LAST AUTHOR

Last Saved By field value extracted from metadata of a native file

ESUBJECT

Document title

FILESIZE

File size in Bytes

FILENAME

File name of native file

APPLICATION

MIME type of document (e.g. application/pdf)

DOCLINK

File path location to the current native file location on the delivery medium

DATEAPPTSTART

Start date of calendar appointment

TIMEAPPTSTART

Start time of calendar appointment

DATEAPPTEND

End date of calendar appointment

TIMEAPPTEND

End time of calendar appointment

32. Preferences

To open the Preferences dialog, select the File > Preferences menu option.

To apply changes of the settings, click the Apply button. To apply changes and close the dialog box, click the OK button. The Cancel button will close the dialog box and discard all unapplied changes.

The specific settings per tab are explained below.

32.1. General

The Startup section controls how Intella behaves when a user opens a case. When the Check for availability of original evidence files option is on, Intella will check for the presence of evidence files at their original locations every time a case is opened. If any of the evidence files are missing, the user will be warned and directed to the Sources tab, where the missing evidence paths can be corrected.

General preferences

The Check for updates on start-up option lets Intella look online for new versions of the software during startup. This lookup will be done once in every 24 hours. New versions will be shown in the upper right corner of the application. A message will also be shown here when this option is turned off or when fetching the last version information has failed.

The Temp Folder controls where Intella stores its temporary files, e.g. for opening an item in its native application. By default, the used folder is inherited from the operating system, but it can be modified here, e.g. to accommodate a system with a small operating system drive or for performance or security reasons.

The Shutdown section controls how case backups are handled when the user closes the case. The three options control whether a backup of the case needs to be made when the case is closed, or whether this needs to be asked on every occasion. This setting is set for each case individually. The Backups folder is shared by all cases though. When a case is backed up, a copy of the entire case folder is made and placed in this folder. A previous backup is removed, if the backup has succeeded – note that this will have consequences for the disk space that needs to be available. The default location of this backup folder is next to the cases folder. We recommend changing this to a location that is located on a physical disk, so that disk malfunctions do not damage both the actual case and the backup.

When connecting to a shared case, then the backup is only created from local case folder, not the remote case. In order to have proper backup of case that is shared, it needs to be unshared first and then made a backup of the case on server from which it was shared.

32.2. Display and Locale

Display and Locale preferences

The Display splash screen while loading a case option controls whether a splash screen will be displayed after you have selected a case in the Case Manager for opening in Intella.

The Language selection option lets you select the display language used for Intella. The set of values in the list depends on which language profiles are detected in the "translations" subfolder, located in the folder where Intella is installed.

Intella checks online whether new language profiles are available for the current Intella version and the currently used language. When this is the case, a message is displayed in the upper right corner of the main window. Clicking on that message will open a web browser and download the new language profile. The Browse…​ button in this panel can then be used to install the new profile.

The Date format setting lets the user select how dates and times will be displayed. The dropdown menu allows for various formats selected by country. This setting is not dependent on the display languages and allows for all generally used formats, regardless of which language profiles are available.

The Draw border around white redactions option generates a thin black border around those redaction marks that have a white color. Such redactions would otherwise not be visible as redacted areas in documents with a white background.

Finally, the Page format lets you select which paper size to use when exporting to PDF or printing items. Available options are ISO A4 and US Letter.

32.3. Dates

Dates preferences

The Primary Date option controls how Primary Dates are determined for each item, based on a set of rules holding preferred attributes.

While processing the dates of all items, Intella will try to pick a matching date rule based on the item’s type and use it to determine the primary date attribute for that item.

  • First, it will first look for a rule that has the same MIME type as the item has, e.g. the MS Word MIME type.

  • When no such rule exists, it looks for a more general rule covering the type group that holds this MIME type, e.g. the Documents group. See the Type facet for how item types are grouped.

  • If no such rule exists either, it will fall back on the default rule to compute the Primary Date.

Each rule holds a prioritized list of all the date attributes that Intella supports. Once a primary date rule is selected for the item, the first date in this list that occurs in that item’s metadata is used to set the item’s primary date.

You can define many date rules for different MIME types or groups. You can add or remove rules from the set, but it must always contain the default rule. By clicking the Reorder dates button, you can change the priorities of the date attributes for the selected rules.

Because of the way rule selection works, the order of the rules does not affect the outcome. Only the order of attributes in a rule matters.

Note that the Primary Date settings also affect the Family Date attributes, as the Family Date of an item is defined as the Primary Date of its top-level parent.

When a change is made to the Primary Date settings, Intella will ask whether you want to rebuild the indices for those two dates. These indices are used for displaying and sorting the Primary Date and Family Date columns and for any Date facet searches on these attributes. Updating these indices can be a lengthy operation on large data sets. In case you wish to cancel this update operation, you can click the Cancel button in the progress dialog. This will revert your Primary Date settings back to the previous configuration and leave the indices unaltered. Note that it is not possible to alter the Primary Date settings without updating the relevant indices.

32.4. Message Hash

Message Hash preferences

You can switch between the legacy message hash that was used in Intella 2.2.1 and older, and the structured message hash that has been introduced in Intella 2.2.2.

The structured message hash allows you to use a less strict algorithm for deduplicating email, SMS and chat items, by deselecting components. For example, when "Include recipients" is deselected, an email with a Bcc header will be considered as a duplicate of an email without that header (assuming all other things are equal).

The structured message hash can only be used in cases created with Intella 2.2.2 or newer. Older cases will first need to be re-indexed with 2.2.2 or newer.

32.5. Irrelevant Items

Irrelevant Items preferences

Intella automatically classifies certain items as "Irrelevant" during indexing. These are items that are generally regarded as non-relevant from a review point of view. Note that this only applies to the item itself, not its child items. One can think of this category as containing those items that you would likely not review when they show up as part of a keyword search result, unless in certain types of deep forensic analysis.

This classification has no effect on the processing of the affected items, other than storing the classification. It can be used to suppress items during searching and exporting, e.g. by toggling the Hide Irrelevant button, using the Features facet category or setting the corresponding options during export. This reduces the time needed for reviewing and exporting. If such filtering is not desirable, all one has to do is leave these options to their default, unselected state.

The following items are classified as Irrelevant:

  • Folders – regardless of origin

  • Email containers – e.g., PST, NSF, Mbox, …​

  • Disk images – e.g., E01, L01, DD, …​

  • Cellphone reports – e.g., UFDR, XRY XML, …​

  • Archives – e.g., ZIP, RAR, …​

  • Executables – e.g., EXE, BAT, …​

  • Load files – e.g., DII, DAT, …​

  • Empty – i.e., zero-byte files

  • Embedded images – see the Features facet section for a definition of this category

Note that the flag is not automatically inherited by child items. Child items are only classified as Irrelevant if they match any of the criteria by themselves.

Currently, the criteria for classifying items as Irrelevant are hardcoded and fixed; the disabled checkboxes in the Irrelevant Items tab are only there to explain the process. We may make these options configurable in a future release.

Search preferences

The Enable Search History option allows you turn off the search history. The main use of this is when you do not wish these search terms to be recorded – be aware that they are still being added to the audit trail and may leave traces in the log file. This setting is also a workaround for character sets (e.g. Korean characters) that cannot be entered properly when the history functionality is active.

The Restore the queries that were shown last option results in the current queries being stored during shutdown and restores them the next time the case is opened.

The Show Children options allow you to specify what children are returned when you click on Show Children in the Previewer or in the search results popup menu. You can specify the level by including only directly nested children (direct children only) or directly and indirectly nested children (all children). When you select the Ask every time option, you will be prompted for the desired level every time you use Show Children.

The Show Parents options control what items are ignored when the top-level or direct parent is selected for an item. This operation affects what the Show Parents, Show Top-level Parents and Show Family functions produce, what items are tagged when the Also tag all other items nested in the same top-level item option is selected in the Tagging tab, and what items are shown under the Top-Level Parents feature facet and column. Note that changing any of the Show Parents options will trigger a database update that might take some time to complete, depending on the number of items in the case

32.7. Results

Results preferences

The Opening results option controls what happens when a result is double-clicked: open it in Intella’s internal Previewer, or in the native application registered with that file type.

The Following HTML links option relates to the links and externally linked images that can be found in HTML-based emails. Both can be dangerous to download automatically, e.g. because they can tip-off suspects that their emails are being read by another party. This panel lets you control how these link types are handled. By default, links are blocked and external images are not loaded automatically. This can be managed per individual email in the Previewer window or for all items at once in this preferences panel.

The Cluster Map options let you specify whether transitions on the Cluster Map should be animated and if so, how long that animation may take. You may want to disable animation if it causes performance problems on your system.

Furthermore, you can specify whether or not the Cluster Map should automatically be scaled when it does not fit inside the window. You can also change this option using the Cluster map toolbar button, or go to View > Cluster Map > Scale to fit window.

The Details View setting controls the amount of values that are shown or exported when using Content Analysis columns. The values after the threshold are filtered out and replaced with ellipsis. To export all values, simply set this threshold high enough for all values to be exported.

The Thumbnails View group controls the aspects of thumbnail generation:

  • Thumbnail minimum size specifies which thumbnails are shown based on the size of the original image in kilobytes. Images that are below this threshold are filtered out.

  • The three video thumbnail options select an algorithm for extracting static frames to compose thumbnails for video items:

    • Fixed number of frames: The specified number of frames in equal time intervals is extracted.

    • Any number of frames every X seconds: The frames are extracted in the specified time intervals. The total number of frames depends on the length of the video item. Use this option with caution, as it can make the process of thumbnail generation substantially slower.

    • First X frames every Y seconds: The fixed amount of frames in the specified time intervals is extracted and the rest of the video is skipped. Note that the actual number of extracted frames can be less than this limit, depending on the length of the video item.

Changing the settings in the Thumbnails View group will cause clearing the cache of previously generated thumbnails. This operation can also be performed separately using the Clear thumbnails cache now button.

The Previewer Window setting controls the maximum size of the files that will be presented in their native rendering in the Preview tab. Rendering of this tab may trigger a conversion from the document format at hand (e.g. an MS Word document) to PDF. This can take a long time for large and complex documents. By default, this limit is set to 10 MB.

Furthermore, the paragraph controls shown in the left margin of the Contents tab can be disabled using the Enable paragraph features checkbox. This only has an effect when the Analyze paragraph setting has been used during source creation.

32.8. Tagging

Tagging preferences

When tagging items, the policy of your investigation may be that some related items should be tagged as well, e.g. tagging items in a mail as privileged may require that all other items in that same mail are also tagged as privileged. The settings in this tab can make that happen automatically.

The three radio buttons specify how other items in the hierarchy need to be handled:

  • Only tag the selected item is self-explanatory.

  • Also tag all attached/nested items results in all attached or nested items being tagged with the same tag as well. This works recursively, i.e. all children in the hierarchy are tagged.

  • Also tag all other items nested in the same top-level item means that everything from the top-level item down to the most deeply nested child gets the tag.

In addition to these three settings, you can also specify that all duplicates should also be tagged. When this setting is enabled, all items in the case with the same MD5 or message hash will inherit the tag. Furthermore, their children or siblings may also be tagged automatically, according to the settings described above.

The top-level parent of an item is determined per the Show Parents settings in the Search preferences.
The settings described above can also be accessed directly from the dialog where tagging dialog. That will allow you to (temporarily) override the default settings that you specify here.

The Previewer setting controls the maximum number of quick tag buttons that are shown in the Previewer.

The Expand top-level tag groups in item properties setting controls how hierarchical tags are shown. By default, hierarchical tags will all show their hierarchical path, e.g. "Parent/Child1" and "Parent/Child2". When enabling this option, the parent tag will be shown as a properties line, followed by all child tags, e.g. "Parent: Child1 Child2". This can improve overview when items typically tend to have multiple tags in the same branch of the tag hierarchy.

32.9. HCL/IBM Notes

IBM Notes preferences

Click Validate to ensure that Intella can locate the HCL/IBM Notes program files on the system. The status is shown in the (non-editable) field.

If validation fails, click the Browse…​ button, select the path to the HCL/IBM Notes folder in the file chooser and click Apply.

During Notes validation Intella will check the HCL/IBM Notes version. Some versions are not recommended, see the Installation section. To enable use of such non-recommended HCL/IBM Notes versions, select the Enable using unsupported version of HCL/IBM Notes checkbox.

The default installation directories for HCL/IBM Notes is one of the following: C:\Program Files\IBM\Lotus\Notes
C:\Program Files\IBM\Notes
C:\Program Files (x86)\IBM\Lotus\Notes
C:\Program Files (x86)\IBM\Notes

32.10. Geolocation

Geolocation preferences

The Tile preferences section defines how the world map gets rendered in the Geolocation results view and the Previewer’s Geolocation tab.

Intella embeds a set of tiles for rendering this map. By default, this tile set is used. This embedded tile set enables use of the Geolocation views without requiring any configuration and/or network connection. The drawback of using this tile set is that the user can only zoom in six levels.

Another option is to integrate with a custom tile server. To enable use of such a server, select the Integrate with the tile server option. The Geolocation tab will then expand to offer additional settings.

Geolocation Tiles preferences

In the example shown here, MapBox’s tile server is used. You can use any tile server you wish by typing its address into the Tile server integration URL field. The format for the URL is dependent on the chosen tile server.

To use a public tile server, you need to ensure that you comply with the tile server’s usage policy. This is your responsibility, not Vound’s.

The Min. zoom option defines the desired minimum zoom level in the user interface. This should be in the range of supported zoom levels of the chosen tile server.

The Max. zoom option defines the desired maximum zoom level in the user interface. This should be in the range of supported zoom levels of the chosen tile server.

The Tile Size (pixels) option defines the size of a single square tile. This value should match the size of the tiles which are returned by the tile server.

The Reverse X tile numbering and Reverse Y tile numbering options should be used when the tile numbering order used by the tile server is reversed. Usually there will be no need to use these two options.

Using a public tile server may reveal the locations that are being investigated to the tile server provider and anyone monitoring the traffic to that server, based on the tile requests embedded in the retrieved URLs.
If the investigation system has no internet connection, a custom tile server can be set up on the local network. One way of how this can be achieved can be found at http://osm2vectortiles.org/docs/serve-raster-tiles-docker/. This is out of the scope of this manual and Vound’s technical support.

Email geolocation allows one to estimate the geographic location of an email’s sender using the sender IP address. This process takes place during indexing. See the Geolocation chapter for a description of the process and its caveats.

Determination of the geographic location of an IP address requires the presence of MaxMind’s GeoIP2 or GeoLite2 database. These databases associate IP addresses with geographic locations. The databases can be found here:

See the MaxMind website for a description of their differences, beyond price. Please note that when using either of these databases, you will need to register and generate a license key, as described here - https://blog.maxmind.com/2019/12/18/significant-changes-to-accessing-and-using-geolite2-databases/

The chosen database can be installed by placing it in the following folder:

C:\Users\[USER]\AppData\Roaming\Intella\ip-2-geo-db

Alternatively, when you are on an Internet-connected machine, you can let Intella download and install the GeoLite2 database automatically by putting your license key in the Your license key field and clicking the Download GeoLite2 database button. After clicking this button, a dialog opens that states that proceeding with this download implies that you agree with the GeoLite2 license terms. After clicking Proceed, the download will start. The download progress will be shown in the Status field. Once the download has completed successfully, a green validation message will be shown here.

To use the Email geolocation feature, check the Determine the geographic location of an email sender’s IP address option when adding a new source.

32.11. Advanced

Advanced preferences

Cellebrite

Select the Skip sections checkbox to make Intella skip listed sections of Cellebrite reports during indexing. List of sections that can be skipped include "taggedFiles, caseInformation, metadata, decodedData" in physical reports, and general sections (e.g. sms_message, contact, incoming_call, outgoing_call, missed_call, image, ringtone, audio, video, calendar, tasks, notes, entry) in logical reports.

Near-Duplicates

The Preferences window allows you to define the shingle size for both algorithms - word-based and character-based. A shingle represents a substring of a given length into which a text document will be split for detecting near-duplicates (N-D). For the word-based algorithm, which is recommended for Western languages, a shingle represents a sequence of words. By choosing the shingle size, we define how many words the text block will be composed of, serving as the basis for Near-Duplicate Analysis. For the character-based algorithm, which is recommended for Asian languages (CJKV), a shingle represents a sequence of language characters. When selecting the shingle size for this algorithm, we define how many characters (ideographs) the text block will consist of that will be used to compare text documents.

How to choose the size of the "shingle size"?

Generally, the shingle size should be selected in such a way that the probability of the occurrence of a given shingle in the analysed documents is low. This value is influenced by factors such as the amount of text in the analysed documents and the document languages. Therefore, to optimize near-duplicate analysis, two algorithms have been made, which are designed for two separate language groups. For the CJKV language group, the default shingle size is 2 (characters). This value allows you to divide the text in a way that provides a basic level of document comparison. For the second language group, the default shingle size is 3 (words). This value was chosen based on the average number of characters in English words, and is optimal when analysing large text documents. When predominantly processing emails, the shingle size value can often be scaled down to 2.

Image Analysis

  • Minimum image size - determines the minimum size of image it needs to have to be analysed. If very small images are providing false categorization or object detection, then it is recommended to set this value to the same or higher size than those images. This will prevent categorization or object detection to take place on such small images. An image is skipped if neither its width nor height exceed the threshold. That is a categorization process threshold - changing it only affect results of future categorization process (it must be redone if you want to filter out unnecessary results).

This setting get disabled when connecting to a case shared by Intella Connect or Intella Investigator.

Exporting

Select the Make email addresses compatible with PST format when exporting to PST checkbox to allow Intella export items with email addresses that might not be compatible with the PST format. That means that any email address that does not look like name@server (does not contain the @ symbol) will be modified by adding a "@server" suffix.

For example, a LDAP email address like /O=ORG/CN=RECIPIENT will become /O=ORG/CN=RECIPIENT@server.

33. Reading your log files

Each Intella case has its own log files: case-main-[date].log and case-warnings-[date].log. The main file contains all Intella’s case-specific log messages. The warnings file only contains those messages that have a WARN or ERROR message level. This will typically be a much shorter file and is therefore easier to handle, but it may miss critical contextual information needed for the proper diagnosis of errors.

Log files "roll over" every day at midnight. When a case is opened for the first time on a specific day, two new log files are made for that day. When you open the case another time on the same day, all log messages are appended to these files. When Intella still has the case open at midnight, it starts creating two new log files for the next day. When you keep track of when you indexed or exported information in the case, this split up will ease looking up the relevant log messages. This split up also prevents the log files from becoming too large to handle.

Your log files are always stored in the logs subfolder of the case folder. By searching your own log files for common error messages, you can often troubleshoot these common errors without need of contacting our support department.

To open very large log files, we suggest free tools such as Large Text File Viewer. This tool is optimized for viewing very large (> 1 GB) text files.

33.1. List of known errors

The following is a list of known error messages and their explanation and solution. Most of these errors have causes that are external to Intella.

java.lang.OutOfMemoryError

Explanation: Intella has run out of memory while processing a file or index. This error can also happen on machines with lots of RAM, as each of Intella’s processes has a maximum amount of memory it is allowed to use.
Solution: increase Intella’s memory settings. See the "Memory settings" chapter for details.
Caveat: some out of memory errors are caused by corrupt documents and Intella may safely recover from this, so that the integrity of the case is not harmed. When in doubt, please send the error message and the 50 lines surrounding it to our support department.

java.io.IOException: The device is not ready

Explanation: the USB disk or network drive is not available to Intella.
Solution: process the case on an internal disk to see if this solves the issue. The network/external disk or the connection to it may not be reliable.

Exception while processing stream of …​

Explanation: the file being indexed may be corrupt.
Solution: check the file in its native application.

java.io.IOException: An unexpected network error occurred

Explanation: your network is dropping connections.
Solution: move the case and evidence data to an internal drive and retry.

java.io.IOException: Access is denied

Explanation: you do not have the correct file or folder permissions.
Solution: ensure all permissions on the case files and evidence data are correct.

java.io.FileNotFoundException: X:\FILE_NAME_AND_PATH (Access is denied)

Explanation: the file or folder is missing. Perhaps you have moved the case or evidence folder?
Solution: verify that the original data is where it should be.

java.io.IOException: There is not enough space on the disk

Solution: Please check if you have sufficient drive space to continue. Note that heavily fragmented files can also cause such errors.

33.2. Background information

Log files are useful in diagnosing problems or errors that may arise while using Intella. Intella records log files during all normal program operation, including indexing and exporting. In the default logging mode, Intella records basic operational messages and errors. The log files are stored in the logs folder inside the case folder.

When using the default suggested case folder, the log files are typically stored here:

C:\Users\<USERNAME>\AppData\Roaming\Intella\cases\<CASE NAME>\logs

The log files are case-main-[date].log, case-warnings-[date].log and any other *.log files in the same directory. Note that the Help menu in Intella has an Open Log Folder option that takes you straight to this folder, regardless of where the actual log folder is located.

33.3. Log levels

When an issue arises and the log files do not contain sufficient information to resolve the issue, you may wish to increase the logging level from basic (INFO) to a more detailed level (typically DEBUG). Intella will then log actions and operations in more detail. Note that this will considerably increase the size of the resulting log files.

To increase the logging level. please do the following:

  1. Close your Intella case, remembering the case name and the location of the case folder.

  2. Assuming the default suggested case folder, open the following folder in Windows Explorer:
    C:\Users\<USERNAME>\AppData\Roaming\Intella\cases\<CASE NAME>\conf
    When your case is saved to a location other than the suggested location (e.g. a USB device), please use the path to that folder.

  3. Once in the conf folder, use a text editor (e.g. Notepad) to open the logback.xml file. This file is unique to every case. Locate the line containing
    <level value="INFO" />
    and change it to
    <level value="DEBUG" />

  4. Once the changes are made, save the file and close the text editor.

Now open your Intella case again. Logging will from now on report more detailed information.

33.4. Using the Log Analysis tool

For users convenience, Intella includes a Log Analysis tool which scans case log files for common errors (see the "List of known errors" above). With this tool, it is possible to make a quick overview of possible problems with the Intella case.

To run the Log Analysis tool, select Scan logs for errors item in the Help menu. It will scan all case-warnings-[date].log files stored in the logs folder inside the case folder and represent the results in a window:

image

  • To repeat scanning and refresh the results, click Rescan folder button

  • To scan another folder or file, specify the path into Log file or folder field (or use Browse button), then click Rescan folder. You can specify either the entire folder (to scan all case-warnings-[date].log files in that folder), or individual log file.

  • To scan the logs for crawler timeout errors, click Scan for timeouts button. This will scan all case-main-[date].log files in the folder and report all items that took more than one hour to process or item which processing never finished (timed out). This might help to identify a specific file that is reponsible for the timeout, so you can isolate or remove it.

  • To save the results to a text file, click Save As…​ button.

34. Menu, mouse, and keyboard shortcuts

34.1. Main Menus

Below is a description of all menu items in the main window. Not all options appear in all products.

34.1.1. File

Preferences

Open the Preferences window (see Preferences).

Key Store

Open the Key Store window, for viewing and editing decryption passwords, certificates, etc.

Annotations History

Open the Annotations History window (see Annotations History).

Restore Annotations…​

The user can restore the annotations from a copy of this case, e.g. when the working copy has been damaged beyond repair.

Import OCRed files…​

Import files that have been processed using an external OCR tool.

Import load file…​

Import an overlay load file (see Import overlay load file).

Generate Thumbnails (Ctrl+T)

Pre-generates all thumbnail images used in the Thumbnails view, speeding up its responsiveness.

Tasks

Opens a window that show all defined post-processing tasks and lets the user edit and launch them.

Excluded Paragraphs (Ctrl+Shift+F)

Opens a window that shows all paragraphs explicitly excluded from keyword search and let the user search for them or remove them from the list of excluded paragraphs.

Custom Columns

Open the Custom Columns window (see Custom Columns).

Close Case

Closes the current case and brings the user back to the Case Manager window.

Exit (Ctrl+Q)

Exit the application

34.1.2. View

Cluster Map Animate Changes

Turn cluster map animation on or off.

Cluster Map Scale to Fit Window

Turn cluster map size scaling on or off.

Details

Use the four sub-items to switch the Details panel to Table, List, Thumbnail or Timeline mode.

Preview Item…​ (CTRL+O)

Lets the user open a specific item. See the Item ID column in the Details table for these numbers.

Close All Previews (Ctrl+Shift+W)

Closes all open Previewer windows.

Full screen

Toggles full-screen mode.

34.1.3. Export

Cluster Map…​

Exports the current Cluster Map as a PNG image.

Social Graph…​

Exports the current Social Graph as a PNG image.

Timeline…​

Exports the current timeline as a PNG image.

Words…​

Export all words used in the indexed evidence files. When the results table shows a list of results, exporting of the words of only these items is also possible.

Result…​

Export a single result. This option is available when a single item is selected in the Detauks table.

Result List…​

Opens the export dialog to let you export the currently selected results.

Export to Intella Case…​

Opens the export dialog to let you export the selected items to a different or blank case.

Configure Redaction Profiles…​

Opens the dialog that lets you create and edit redaction profiles. See the section on redaction for more information.

Event Log…​

Opens the dialog that lets you export part of or the entire case event log.

34.1.4. Help

User Manual (F1)

Opens the bundled user manual (this document).

Release Notes

Opens the bundled release notes.

Forum

Opens the Intella forum in a web browser.

Dongle Manager

A shortcut to the separate Dongle Manager application, which is used to inspect and update the contents of your Intella dongle.

Open Log Folder

Opens the folder where Intella stores logging information.

Scan logs for errors

Runs the Log Analysis tool (see "Using the Log Analysis tool" section of the "Reading your log files" chapter)

Open Export Templates Folder

Opens the folder where the user-defined export templates are stored. These files are .xml files that can be shared and copied to other case folders.

About Intella <product edition>

Shows a dialog with three tabs. (1) The first tab contains the version number of Intella. (2) The second tab contains system information. (3) The third tab shows license information such as ID, type and restrictions.

34.2. Mouse actions

34.2.1. Table and thumbnail view

Click and drag

Select multiple items.

Ctrl+click

Select/deselect items.

Double click on item

Depending on the preferences, this opens the clicked item in Intella’s internal Previewer, the registered native application, or opens a dialog asking the user what to do.

Right click on item

Opens the popup or context menu.

34.2.2. Timeline

Click on email

Opens the email in the Previewer.

Double-click on email

Depending on the preferences, this opens the clicked item in Intella’s internal Previewer, the registered native application, or opens a dialog asking the user what to do.

Right click on email

Opens the popup or context menu on that email.

34.2.3. Cluster Map

Click on cluster or on label

Select a cluster or result set and shows its items in the Details panel below.

Click and drag

Move cluster to reorganize the Cluster Map.

Right click on cluster, label or on the selections panel

Open the popup or context menu on that item.

34.2.4. Social Graph

Click on a node

Select a node and show its items in the Details panel below.

Click on an edge

Select an edge and show its items in the Details panel below.

Click and drag

Move nodes to reorganize the graph.

Drag with right-mouse button pressed

Scroll (pan) the graph.

34.2.5. Histogram

Click and drag

Zoom in on a specific area in the chart.

Ctrl-click and drag

Pan (scroll) the chart.

Click and move up

Restore zoom level.

Mouse wheel

Zoom in and out of the chart.

34.3. Keyboard shortcuts

34.3.1. Main window

Ctrl+N

Add new source

Ctrl+R

Re-index all sources

Ctrl+O

Open a specific numbered item

Ctrl+Q

Exit the application

Ctrl+Shift+W

Closes all open preview windows

F1

Open Intella help file (requires PDF-viewer, like Adobe Acrobat)

Spacebar (in thumbnail view)

Flag selected item

Ctrl+A

Select all items or text

34.3.2. Previewer window

Alt+Right Arrow

Move to next item

Alt+Left Arrow

Move to previous item

Ctrl+C

Copy selected text

Ctrl+V

Paste copied text

Ctrl+A

Select all text

Ctrl+1, Ctrl+2, or Ctrl+3

Tag an item with the tag assigned to button 1, 2 or, 3 in the previewer

35. HASP dongle problem resolution

35.3. Installation problems

35.3.1. HASP dongle drivers do not install

desktop hasp purge

Problem: You are not able to install the HL Key (dongle) drivers.

Cause: Presence of older HASP HL key drivers installed on the machine.

Solution: Uninstall the older drivers.

  1. Click Start > Run or click the Windows key + R.

  2. Enter C:\Program Files\Vound\Intella\bin\haspdinst.exe -kp purge and click OK.

  3. Wait for message that operation was successful.

These steps uninstall ALL other HASP drivers. Make sure you have no other HASP dongle that requires an older driver. Install the latest driver.

35.3.2. HASP dongle not found

Problem: The following message is shown: HASP key not found (H0007).

Possible cause 1: The HASP dongle LED is not lit. The dongle is not connected or not properly connected to the USB port.

Solution:

  1. Disconnect, pause a few seconds, then reconnect. If the LED lights up, the application should be able to access the dongle. You may need to wait a few seconds for the dongle to be completely installed by the operating system.

  2. The required HASP HL key drivers are not installed. If you are running HASP SRM on a Windows platform, check for an entry for HASP SRM in the Device Manager utility. If there is no entry, you must install the drivers.

  3. Check if the USB port is functioning correctly. Disconnect all other USB devices from their respective ports. Connect the HASP dongle to a different USB port. Try using a different USB device in the port from which the dongle was not accessible to test if the port is working.

Possible cause 2: HASP License Manager Service is not running.

Solution:

  1. Check if the HASP License Manager Service is running by opening a Command Prompt (Start > All Programs > Accessories > Command Prompt).

  2. Enter: sc query hasplms

  3. Check the result. It should show something similar to this:

    SERVICE_NAME: hasplms
    TYPE : 10 WIN32_OWN_PROCESS
    STATE : 4 RUNNING
    (STOPPABLE,NOT_PAUSABLE ,IGNORES_SHUTDOWN)
    WIN32_EXIT_CODE : 0 (0x0)
    SERVICE_EXIT_CODE : 0 (0x0)
    CHECKPOINT : 0x0
    WAIT_HINT : 0x0

    When you see RUNNING it means that the hasplms service is running.

35.4. Hardware problems

35.4.1. No dongle detected

Problem: The computer does not detect the dongle. There are several potential causes, listed below.

Cause 1: Conflict with other USB devices
On occasion, the presence of other USB devices may cause problems with the HASP dongle.
Solution: Remove conflicting USB device/devices.

Cause 2: Incorrect device driver installed
The HASP dongle may not function if an incorrect version driver is installed.
Solution: see section Installation problems, HASP dongle drivers do not install.

Cause 3: USB port is defective or HASP dongle not properly inserted
Solution: Check that the LED light is lit on the dongle. If not, remove and reinsert. Wait for the operating system to detect the dongle. If it still does not light up, try another USB port or use a USB hub.

Cause 4: Faulty dongle
On rare occasions one may get a faulty dongle. The dongle neither lights nor is detected in Device Manager, even with proper driver installed. Request a replacement.

35.5. Firewall & anti-virus problems

35.5.1. Unable to access HASP SRM RunTime Environment (H0033)

Problem: The error message Unable to access HASP SRM RunTime Environment (H0033) might be caused by too restrictive firewall settings.

Possible causes:

  • C:\WINDOWS\system32\hasplms.exe is blocked by firewall or antivirus application.

  • Port 1947 is blocked by a firewall application.

  • HASP License Manager Service is stopped.

Preliminary test:

  1. Disable all antivirus and firewall applications. Note that some applications such as Norton, McAfee, and AVG have both antivirus and firewall settings that may need to be individually disabled.

  2. If the HASP License Manager Control Center does not appear in the browser at http://localhost:1947 then we know that the anti-virus or firewall application will have to be configured.

  3. If the Control Center still does not appear, check for other firewall or antivirus applications that may be running and disable them or turn them off.

Solution:

  1. Add C:\WINDOWS\system32\hasplms.exe in the Exception list of the antivirus and firewall application.

  2. Add port 1947 to the Exception list.

  3. Restart the HASP License Manager Service (Control Panel > Administrative Tools > Services).

An example of a firewall exception is shown in the image below.

You must perform an installation “Reinstall” of Intella as the antivirus software may have blocked components during the first install.

desktop hasp firewall

The following information has been adapted from the SafeNet Sentinel HASP knowledgebase:

Message: Unable to access HASP SRM RunTime Environment (H0033).

Problem: This error means that there is a communication error between the program and the local license manager. This error can be triggered by several causes, including (1) improper installation of the HASP RTE software, (2) personal firewall software blocking communication with the HASP LMS service, or (3) other software using the same port that the HASP License Manager uses (i.e. port 1947).

Solution: To troubleshooting the error follow the steps below until the cause for the error is found:

  1. Open a web browser and connect to http://localhost:1947.

    This is the HASP SRM Admin Control Center. If it’s possible to connect to this page, then the HASP SRM Runtime is installed properly. The problem lies elsewhere and you can disregard the rest of this document.

    If you get a message Page cannot be displayed, it’s possible that HASP SRM Runtime is not installed (go to step 2) or blocked (go to step 3 and 4).

  2. Go to Start > Run, enter services.msc and click OK.

    The list is alphabetical. Search for HASP License Manager in the table and then check if its status is "Started".

    If this entry is not listed, then the HASP SRM Runtime is not installed. Please reinstall it.

    If the status is not "Started", check the event log for entries relating to the HASP License Manager service that will give an error message and further diagnostic information.

  3. Check your personal firewall software. There are many types of personal firewall software including Norton Internet Security (the Firewall is one component of this software), ZoneAlarm and others.

    By default, most personal firewall software will request permission to allow access for the HASP License Manager the first time it is run. If access is allowed there should be no problems.

    If access is denied you will encounter communication problems. To resolve such problems either disable the firewall completely (Note: this option has risks. Please contact your firewall vendor for details) or create a rule or exception in the firewall to allow the HASP License Manager communication. If there is an option to create a rule/exception based on a port number, allow port 1947.

    As there are many personal firewall products on the market it is not possible to list all the ways to configure each piece of software here. Please contact your firewall vendor for details on how to create exceptions or rules as detailed above.

  4. Check that there aren’t any applications that use HASP registered port (Port 1947). If you find such a program, disable it and run the HASP application again.

35.6. Normal Operation

35.6.1. Dongle installation

desktop hasp dongles

Intella is shipped with the latest SafeNet HASP dongles. Intella is packaged with the SafeNet HASP RTE installer.

When correctly installed, the Windows Device Manager reports three items in the "Universal Serial Bus controllers" section:

  • SafeNet HASP HL Key

  • SafeNet HASP Key

  • SafeNet USB Key

image

image

When incorrectly or incompletely installed, warning icons appear on the device.

35.6.2. HASP License Manager Service

The HASP installer includes the HASP License Manager application that runs as a system service:

C:\WINDOWS\system32\hasplms.exe

The HASP License Manager Service hasplms.exe must be running to allow Intella to open. When this application is running, you should be able to load the HASP License Manager Admin Control Center by entering http://localhost:1947 in an internet browser.

image

HASP SL is the trial version license.

HASP HL will only show when the dongle is plugged in.

35.6.3. Windows system services

A good indication that the License Manager Service is running properly, is that the entry is flagged as “Started” in the table of Windows system services:

desktop hasp services