ICAIL
2011 Workshop on
Setting Standards for
Searching
Electronically Stored Information
in Discovery Proceedings
(DESI IV
Workshop)
June 6, 2011, University of Pittsburgh, Pittsburgh, PA, USA
Purpose
The DESI IV workshop is intended to provide a platform for discussion
of an open standard governing the elements of a state-of-the-art
search for electronic evidence in the context of civil discovery. The
dialog at the workshop might take several forms, ranging from a
straightforward discussion of how to measure and improve upon the
"quality" of existing search processes; to discussing the creation of
a national or international recognized standard on what constitutes a
"quality process" when undertaking e-discovery searches; to a more
ambitious discussion concerning creation of a type of standards
"authority" or certification entity that could certify compliance with
whatever standard emerges. Issues to be considered include the
potential benefits of a standard (e.g., reducing the need for
evidentiary hearings on the reasonableness of a process), versus its
potential costs (e.g., the risk of inhibiting innovation through early
convergence on rigid standards), and timelines.
Background
The Sedona Conference®, a leading legal think tank, recognized in
2009 that "the legal profession is at a crossroads," in terms of its
willingness to embrace automated, analytical and statistical
approaches in the area of e-discovery. However, while in the last
four years a cottage industry of published case law and commentaries
have come into existence recognizing that parties in litigation should
reach some kind of agreement on what constitutes a "search protocol,"
there is no widely agreed-upon set of standards or best practices for
how to conduct a reasonable e-discovery search for relevant evidence.
The Sedona Conference and others have, however, called out to industry
and academia to assist in the further development of standards of what
constitutes "best practices" in the area of performing searches of
electronic evidence.
In initiating a discussion about standards for what constitutes a
"quality" process in e-discovery search, the workshop will serve to
achieve the aim of bringing together academia and industry in the
development of standards in this area. A recent article on "Evaluation
of Information Retrieval in E-Discovery" in the journal
Artificial Intelligence and Law's special issue on E-Discovery
suggested that:
"One broad class of approaches that has gained currency in recent
years ... is known as 'process quality.' Essentially, ... the
important thing is that we agree on how each performer of E-discovery
services should design measures to gain insight into the quality of
the results achieved by their particular process. The design of their
process, and of their specific measures, is up to each performer. Of
course, some performers might benefit from economies of scale by
adopting measures designed by others, but because the measures must
fit the process and because process innovation should be not just
accommodated but encouraged, forced convergence on specific measures
can be counterproductive. So process quality approaches seek to
certify the way in which the measurement process is performed rather
than what specifically is measured."
Agenda
The full day workshop will be organized in four parts:
Part I: The workshop will begin with an overview of recent
developments in e-discovery search, including the needs of the legal
profession, recent case law, and recent work on evaluation design.
Part II will consist of a set of presentations selected to illustrate
the diversity of current research on e-discovery search processes
discussion of some approaches to standard-setting.
Part III will consist of a set of breakout sessions to engage the
workshop participants in brainstorming with respect to what process
quality standards for e-discovery search might entail. Each breakout
group will be asked to initially look at the problem from the
perspective of some specific standards-setting process. This
discussion will begin with informal discussions over lunch, and then
it will be the principal focus of the first session after lunch,
ending with brief reports from each group.
Part IV will be a facilitated session with panelists looking back over
the day and suggesting ways forward. The session will conclude with a
broader discussion of directions for further research, building from
two questions:
- What research questions should be explored, so as to contribute
to the development of process quality standards in the e-discovery
search area?
- Who, beyond those already in the room, do we need to engage with
to address the issues that we have identified?
A final schedule is now available.
Papers
The Workshop Proceedings is now
available. This contains all of the research and position papers in a
single PDF file.
Research Papers (peer reviewed)
Position Papers (not peer reviewed)
- Steve Akers, Jennifer Keadle Mason and Peter L. Mansmann, An Intelligent Approach to
E-Discovery
- Susan A. Ardisson, W. Scott Ardisson and Decker, bit-x-bit, LLC
- Cody Bennett, A Perfect Storm for
Pessimism: Converging Technologies, Cost and Standardization
- Bennet B. Borden, Monica McCarroll and Sam Strickland, Why Document Review is Broken
- Macyl A. Burke, Planning for Variation
and E-Discovery Costs
- David van Dijk, Hans Henseler and Maarten de Rijke, Semantic Search in E-Discovery
- Foster Gibbons, Best Practices in
Managed Document Review
- Chris Heckman, Searches Without
Borders
- Logan Herlinger and Jennifer Fiorentino, The Discovery Process Should Account
for Iterative Search Strategy
- Amanda Jones, Adaptable Search
Standards for Optimal Search Solutions
- Chris Knox and Scott Dawson, ISO 9001:
A Foundation for E-Discovery
- Sean M. McNee, Steve Antoch and Eddie O'Brien, A Call for Processing and Search Standards in
E-Discovery
- Eli Nelson, A False Dichotomy of
Relevance: The Difficulty of Evaluating the Accuracy of Discovery
Review Methods Using Binary Notions of Relevance
- Christopher H. Paskach and Michael J. Carter, Sampling -- The Key to Process Validation
- Jeremy Pickens, John Tredennick and Bruce KieferProcess Evaluation in eDiscovery as
Awareness of Alternatives
- Venkat Rangan, Discovery of Related
Terms in a Corpus using Reflective Random Indexing
- Howard Sklar, Using Built-In Sampling
to Overcome Defensibility Concerns with Computer-Expedited
Review
- Doug Stewart, Application of Simple
Random Sampling in eDiscovery
Important Dates
- Research papers due: April 1, 2011
- Position papers due: April 22, 2011
- Notification for research papers: April 22, 2011
- Preliminary Agenda posted: May 9, 2011
- Camera-ready research papers due: May 13, 2011
- Workshop: June 6, 2011
The Workshop will begin at 9:00 AM (with registration open from 8:30)
and it will end at 5:30 PM.
DESI History
At the beginning of the second decade of the 21st century, lawyers
continue to increasingly face the problem of how to effectively and
efficiently conduct searches for relevant documents across
increasingly complex, enterprise-wide collections within corporate and
institutional settings. Hundreds of millions of documents are now
routinely subject to searches across a wide spectrum of litigation and
investigatory contexts (e.g., the Lehman Brothers investigation
necessitated a review of 350 million pages or 3.5 petabytes of
material). In recognition of this situation, in 2006 the United
States adopted new rules governing civil litigation in federal
courts. These courts have recognized "electronically stored
information" (ESI) as a term of art, embracing all forms of electronic
documents made subject to the civil discovery process. Under the new
rules, opposing parties in federal court litigation now have an early
"meet and confer" duty to discuss a range of electronic discovery
("e-discovery") issues, including the continued storage, preservation,
and access to ESI in their respective physical and legal custodies.
DESI IV follows three successful prior DESI (Discovery of
Electronically Stored Information) Workshops: at ICAIL 2007 (DESI I, Palo Alto)
and ICAIL 2009 (DESI
III, Barcelona), and an intermediate workshop (DESI
II) sponsored by University College London in 2008. In DESI I, a
wide array of individuals came together for perhaps the first time to
foster engagement between e-discovery practitioners and a broad range
of research communities who can contribute to the development of new
technologies to support the e-discovery process. The DESI II and III
workshops broadened the scope of this discussion to include
comparisons of requirements between different national settings and
different legal contexts. DESI IV builds on these efforts with its
emphasis on standards-setting or benchmarking for e-discovery
searches.
Standards
There may well be many good ideas that we can draw on that have been
worked out in the context of existing standards-setting processes in
other fields, such as:
References
Much has been published on E-Discovery generally, so no list of
references could hope to be complete. Here are a few papers that we
know of that we believe would be useful as background reading for the
focus of this workshop. Please send recommended additions for this
list to oard@umd.edu.
- W. Andrews and D. Logan, Early
Case Assessment: E-Discovery Beyond Judges and Regulators Is About
Risks, Costs and Choices, January 27 (2010).
- W. Andrews, D. Logan, J. Bace and S. Childs, E-Discovery
SaaS and On-Premises Software Converge at Vendors as They Mature,
July 29 (2010).
- J. Bace, Communicating Metadata Issues in E-Discovery to the General
Counsel, November 10 (2010).
- J. Baron, Toward
a Federal Benchmarking Standard For Evaluating Information Retrieval
Products Used In E-Discovery, The Sedona Conference Journal,
6(1)237-246 (2005)
- J. Baron, Law in the Age of
Exabytes: Some Further Thoughts on 'Information Inflation' and Current
Issues in E-Discovery Search, Richmond Journal of Law and
Technology, 17(3), Spring (2011)
- M. Grossman, Reflections
from the Topic Authorities About the 2009 TREC Legal Track Interactive
Task (2010).
- M. Grossman and T. Sweeney, What
lawyers need to know about search tools, National Law Journal,
August (2010).
- M. Grossman and G. Cormack, Technology-Assisted
Review in E-Discovery Can Be More Effective and More Efficient Than
Exhaustive Manual Review, Richmond Journal of Law and Technology,
17(3), Spring (2011)
- B. Hedin, S. Tomlinson, J. Baron and D. Oard TREC
2009 Track Overview (2010).
- J. Krause, Human-Computer
Assisted Search in EDD, Law Technology News, December 20 (2010).
- D. Oard, J. Baron, B. Hedin, D. Lewis and S. Tomlinson, Evaluation
of Information Retrieval for E-Discovery, Artificial Intelligence
and Law 18(4)347-386 (2011).
- D. Oard, J. Baron and D. Lewis, Some
Lessons Learned to Date from the TREC Legal Track, 2006-2009,
February 24 (2010)
- H. Roitblat, A. Kershaw and P. Oot., Document
Categorization in Legal Electronic Discovery: Computer Classification
vs. Manual Review, Journal of the American Society for Information
Science and Technology, 61(1)70-80 (2010).
- The Sedona Conference®, Commentary
on Achieving Quality in E-Discovery (2009).
Organizing Committee
Jason R. Baron, National Archives and Records Administration, USA
Laura Ellsworth, Jones Day, USA
Dave Lewis, David D. Lewis Consulting, USA
Debra Logan, Gartner Research, UK
Douglas W. Oard, University of Maryland, USA
Archived Materials on Submissions
We invite both e-discovery stakeholders and practitioners from the
law, government, and industry, along with researchers on process
quality, information retrieval, human language technology,
human-computer interaction, artificial intelligence, and other fields
connected with e-discovery.
To help craft the program, we encourage the submission of research
papers and position papers on emerging best practices in e-discovery
search as well as papers discussing the efficacy of standards setting
in this area. Accepted position papers and accepted research papers
will be made available on the Workshop's Web page and distributed to
participants on the day of the event, and some speakers may be
selected from among those submitting position papers. See the Call for Submissions for submission details.
Last Update: June 4, 2011