Click here to download the full-text of the report.
Click here to download the worksheet.
In August 2023, the Canadian Association of Law Libraries (CALL) Executive Board approved a proposal for the creation of a new AI Standards Working Group (AI WG). The AI WG was created as a sub-committee of the Vendor Liaison Committee (VLC).
This Guide is governed by Creative Commons license CC BY-NC-SA 4.0.
The Committee members are:
This group is well-positioned to provide guidance on the topic, as they have decades of combined experience with applying electronic legal research tools in legal practice.
The WG consulted widely in preparing its recommendations. The consultations included members of CALL and legal professionals, vendor representatives and related professionals.
The goal of this Guide is to shape informed consumers who will set base standards and expectations that all vendors of AI legal research and writing (LRW) solutions will aim to achieve.
As many Law Societies in Canada 1 have stated, licensees who use artificial intelligence have ethical and professional responsibilities in so doing. Accordingly, this Guide was written to help users. Users include law librarians, judges, lawyers, administrators, paralegals, clerks, faculty, and students. The public and other users, such as legislators or policy makers may also benefit from the information provided in this Guide.
Users need to know what to look for, and what to ask for when:
AI applications designed to assist with LRW.
For those users who are judging a particular system for procurement or other purposes, we have also provided a box at the end of each section that allows users to score a system on those criteria and to provide other notes. A summary table of all scores is provided at the end of the document as well that allows users to provide weighted results. An excel spreadsheet has also been developed as a complementary tool to this Guide.
The content of this guide is provided for both AI users and vendors for information only. It is not intended as legal advice. Everyone uses AI products at their own risk and should undertake due diligence to ascertain any liability in using them. The authors of this guide make no warranty whatsoever with respect to the accuracy and reliability of information contained in this guide or any related content provided.
This guide is intended to complement other guidelines and best practices that may be provided in court practice directions, rules of professional conduct, legislation, standards, benchmarks and policies.
This guide does not endeavour to address all possible ethical issues pertaining to AI systems, such as concerns relating to the environment, human rights, etc. Users are encouraged to make their own inquiries.
“Mathematics and Computing. A procedure or set of rules used in calculation and problem-solving; a precisely defined set of mathematical or logical operations for the performance of a particular task.” 2
Artificial Intelligence“Information technology that performs tasks that would ordinarily require biological brainpower to accomplish, such as making sense of spoken language, learning behaviours or solving problems.” 3
Artificial Intelligence System (AI System) 4An Artificial Intelligence System is “any computing system using artificial intelligence algorithms.”5.
DataStructured and unstructured information that is collected, stored, and analyzed to support decision-making.
Labelled Data
“In machine learning, data labeling is the process of identifying raw data (images, text files, videos, etc.) and adding one or more meaningful and informative labels to provide context so that a machine learning model can learn from it. For example, labels might indicate whether a photo contains a bird or car, which words were uttered in an audio recording, or if an x-ray contains a tumor. Data labeling is required for a variety of use cases including computer vision, natural language processing, and speech recognition.” 6
Also known as “information barriers” or “firewalls”, are procedural safeguards used within organizations to prevent the exchange of confidential information between departments, teams, or individuals whose interactions could result in a conflict of interest or ethical breach. These barriers are designed to protect sensitive information, maintain client confidentiality, and uphold the integrity of professional practices by ensuring that only authorized personnel have access to certain information.
Foundation Models"Foundation models are artificial intelligence (AI) models trained on vast, immense datasets and can fulfill a broad range of general tasks. They serve as the base or building blocks for crafting more specialized applications."7
Generative AIGenerative AI, also known as generative artificial intelligence, is a subset of AI that uses advanced models to create new content such as text, images, videos, and audio based on user inputs. These models, often deep learning algorithms, learn patterns and structures from large datasets and use this knowledge to generate original content.
HallucinationA generative AI output “is considered hallucinated if it is either incorrect or misgrounded.” In other words, an AI system has hallucinated if it “makes a false statement or falsely asserts that a source supports a statement.”8
Large Language Model (LLM)“A deep learning algorithm that uses massive amounts of parameters and training data to understand and predict text. This generative artificial intelligence-based model can perform a variety of natural language processing tasks outside of simple text generation, including revising and translating content. LLMs aim to produce the most probable outcome of words for a given prompt.”9
Machine Learning (ML)Machine learning (ML) is a branch of artificial intelligence (AI) focused on enabling computers and machines to imitate the way that humans learn, to perform tasks autonomously, and to improve their performance and accuracy through experience and exposure to more data.10
Semi-supervised Learning
“Machine learning that makes use of both labelled and unlabelled data during training”. 11
Supervised Learning
“Machine learning that makes only use of labelled data during training.”12
Unsupervised Learning
“Uses machine learning algorithms to analyze and cluster unlabeled data sets. These algorithms discover hidden patterns in data without the need for human intervention (hence, they are ‘unsupervised’).”13
“Reinforcement learning (RL) is a machine learning (ML) technique that trains software to make decisions to achieve the most optimal results. It mimics the trial-and-error learning process that humans use to achieve their goals. Software actions that work towards [the] goal are reinforced, while actions that detract from the goal are ignored.”14
ModelAn AI model is a program that analyzes data to find patterns or make predictions.15
Natural Language Processing (NLP)“Natural language processing (NLP) is a subfield of computer science and artificial intelligence (AI) that uses machine learning to enable computers to understand and communicate with human language.”16
Prompt(See also “user input”).
OutputContent that is produced by an AI system in response to a prompt/input.
Retrieval-Augmented Generation (RAG)RAG is an artificial intelligence technique that can improve the quality of generative AI output by allowing large language models (LLMs) to access specialized data sets (e.g. primary or secondary legal databases) outside of its training data sources.
User InputInstructions or data provided by a user to an AI system which the system uses to produce an output. 17 (See also “prompt”.)
VendorA natural or legal person, public authority, agency or other body that sells or provides access to an AI system.18
All marketing and product documentation should be provided in plain language.
Marketing
Statements made in marketing should be true, accurate and transparent. Marketing should reflect realistic benefits and risks presented by the product.
When drawing on survey results to market the product, vendors should make available the questions asked, and methodology used.
Documentation for Prospective Users
Vendors should make available to all prospective users, documentation that sets out, in a transparent manner, all aspects of the AI system, as detailed in this Guide.
Documentation Designed to Assist Subscribed Users
Vendors should make available to all subscribed users, documentation that sets out, in a transparent manner, all aspects of the AI system, as detailed in this Guide.
Additionally, subscribed users should have access to tips for how to use the product.
All documentation and tips should be accessible from within the product (e.g. a “Help” page).
The documentation should outline the types of support available to a user, including, for example, access to live support from a human representative.
Marketing and Documentation Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Users should recognize that most, if not all, AI applications for LRW are built on foundation models trained on the open internet. LRW applications should be built upon a transparent collection of legal information sources. Vendors should clearly provide information about the following:
1. What data sources are used by the AI system? Examples include:
2. Is the scope of each dataset provided? The scope should minimally include, where relevant:
3. What data is NOT included when producing output? For example, is the AI using or not using certain subscribed content, and is unsubscribed content excluded?
4. Do different services offered within the product also specify the scope of data (as described above) upon which each service is trained and from which outputs are constructed?
5. Have data quality standards been established and documented?
BiasWhat, if anything is being done to mitigate bias originating from data sources? See also the section on Bias below.
Data Transparency Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
The data on which an AI system has been trained affects the accuracy, validity, and relevance of AI outputs. Vendors should disclose the data from which tools draw, and the quality control mechanisms deployed to assess outputs and prevent errors. If this information is not known for foundation models, the vendor should clearly disclose this. Information on these points should include:
1. On what data is the AI being trained? For example, is the training data limited to content from legal databases (if so, which ones?) or is it drawn from external resources or the open internet? Can the user find out all the sources used in training? What gaps have been identified in the training data that might contribute to hallucinations or anomalous outputs?
2. Has the user been alerted to the possibility of incomplete, irrelevant, or hallucinated outputs resulting from gaps in the specialized training data?
3. Are user inputs used as data to train the AI system? If so, is the data anonymized or otherwise stripped of identifying data? Does the vendor provide the option to opt out?
4. Are the AI system’s own outputs used to train itself?
5. Was data labelling done by humans, by another AI system, or a combination thereof? Describe any processes, standards, or policies relevant to the labelling process.
Bias
What, if anything, is being done to mitigate bias originating from the training data and the training processes? See also the section on Bias below.
Data Used in Training Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Vendors should provide a detailed overview of the AI system, including the components outlined below. If this information is not known for foundation models, the vendor should clearly disclose this.
Algorithms
What methods and algorithms are used to train and operate the AI system (e.g., natural language processing, retrieval augmented generation, type of learning (e.g., semi-supervised learning), transformer-based models, etc.)? Examples of related questions to ask:
1. How do the algorithms work?
a. How do the algorithms prioritize the content of datasets in producing outputs? For example:
i. Are court decisions prioritized based on precedential value, judicial treatment, jurisdiction and/or level, recency, citation frequency, etc?
ii. How does the AI system prioritize primary sources versus secondary sources, when secondary sources are included in the data?
b. How do the algorithms determine relevancy of sources in producing outputs? For example:
i. Can, and if so, how do they parse out the facts to determine whether and to what extent a specific case applies to a given legal situation?
ii. How do they interpret sources such as judicial decisions, legislation, and specific wording therein?
iii. How does proprietary editorialized content/classification/tagging affect relevancy?
iv. How are secondary sources prioritized, including those involving conflicting viewpoints by different authors and conflicting study outcomes?
2. How are the algorithms trained? For example:
a. Supervised, unsupervised, semi-supervised, or reinforcement learning, or a combination thereof?
b. What is the training process? What are the standards of review, how in-depth is the review, etc.?
c. To what extent and in what ways are humans involved? For example:
i. Are they lawyers? Are they subject matter experts? Are they data analysts? etc.
ii. What methods are used to minimize errors?
3. What third-party product(s)/AI is the AI system based on (e.g. GPT-5, DeepSeek) and to what extent is the third-party product trained on specialized/subject specific-data?
a. When and to what extent is the algorithm relying on third-party training versus local/specialized training?
b. Is the AI system based on more than one third-party product? How many and which ones?
c. Does it rely on one third-party product more than another? To what extent and in which situations?
Algorithm Transparency Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Output Factors
Vendors should be transparent about how the AI system’s outputs are tested, such as:
1. What constitutes adequate versus inadequate output?
a. What quality control measures are in place to ensure output is adequate?
b. How is adequacy of the output determined and who is responsible for the determination?
c. Does quality control occur continuously / how often are outputs tested?
d. How often are the quality control measures reviewed and updated?
e. What happens when an output is deemed inadequate?
f. When and how are algorithms/AI systems updated if outputs are found to be inadequate, and what steps are taken between the discovery of insufficient output and the completion of sufficient algorithm/AI system updates? For example:
i. How often are quality control checks performed, who performs them, and what level of expertise do they have?
ii. What level or type of problem triggers updates or revisions to the algorithms/AI system?
iii. What level or type of problem triggers a temporary suspension of the AI system, and when and how are users notified of current or past problems (including those reported by other users)?
g. What should a user do when an inadequate output is identified? (e.g., does the vendor solicit feedback from users?)
2. Are the testing instrument(s) and test results available? When did the testing take place? How representative of the current system configuration are the test results?
Output Factors Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Hallucinations
For generative AI systems:
1. How are hallucinations defined by the vendor? Are there types of hallucinations that are not acknowledged by the vendor? See for example our definition of hallucination above.
2. Does the vendor specify that hallucinations are possible? If so, which type(s)?
3. What percentage of output is hallucinatory? Does (and if so, how does) this percentage vary depending on the type of prompt/task that the AI is asked to perform? Does (and if so, how does) this percentage vary depending on the type of hallucination in question?
4. What controls are put in place to prevent hallucinations?
5. What should a user do when a hallucination is identified? (e.g., does the vendor solicit feedback from users on instances of hallucinations?).
Bias
1. What steps have been taken to create algorithms that reduce and eliminate bias and discrimination?
2. To what extent and how are humans involved in reducing and eliminating bias in algorithms, data, and outputs?
See also: the section on Bias below.
Hallucinations Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Are there instructions embedded in the product to advise users about how to interact with the AI system effectively (e.g. best practices for crafting prompts)?
2. Are training materials and documentation available in a variety of formats (e.g. videos, interactive tutorials, checklists, guides)?
3. Is jurisdiction clearly addressed in the interaction (e.g. through a specific prompt or pre- or post-filtering options)?
4. Does the AI system allow users to create reusable or shareable personas to their accounts?
5. Does the AI system allow users to create prompt libraries or otherwise save or reuse user inputs?
6. Does the AI system provide information about further cases or related questions to explore?
7. Does the AI system learn from user inputs (e.g. From user prompts, from user feedback to outputs, etc.)? If so, how?
User Inputs /Prompt and Prompt Engineering Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Credibility of Outputs
1. How does the AI system respond if it cannot produce an output? Does the AI indicate clearly that it cannot provide an output, does it provide an error message, or will it always produce an output?
2. Is a confidence indicator provided? If so, how is this calculated?
Attributing Generative AI Outputs19
Does the vendor provide sufficient information to cite AI generated outputs considering the requirements of law societies, court practice directions, etc. to provide attribution of AI generated outputs?
Referencing Sources
1. Are references consistently provided?20 If not, is there a reminder or indicator provided to encourage users to verify the output?
2. Are sources objectively recognizable and verifiable? See also: Algorithm Transparency – Bias
Outputs Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Do the training materials, terms of reference, system reminders or warnings include:
a. A notice that outputs must be verified by the user, because:
i. Outputs are not comprehensive or authoritative
ii. Outputs are dependent in part on the input provided by the user
iii. Outputs may differ with or without algorithm changes, data updates, or additional system learning.
b. A statement of professional responsibilities, including the vendor’s level of liability or role therein?
c. Whether users will be charged per search; if so, can organizations request that a paywall be displayed to prevent extraneous charges?
2. Does the vendor provide documentation demonstrating compliance with the laws and regulations that apply to this AI system?
Other Risk Management Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Does the vendor explain how user data will be used and protected, including:
a. What types of user data are collected? (e.g. prompts/ user inputs, IP addresses, personal information (name, organization, email address, microphone or camera, etc.)) If able to receive input from microphone or camera, can these be turned off?
b. What will user data be used for? (e.g. for sale to third parties, for internal metrics).
c. Who "owns" the user data once collected?
d. Does the AI system learn from user inputs? If so, is the data anonymized or otherwise stripped of identifying data? Does the vendor provide the option to opt out?
e. Who can access user inputs (e.g. vendor IT staff or third-party LLM developers)?
i. What restrictions exist on access to user data?
f. What is the data retention policy for user inputs and outputs?
g. What is the data retention policy for uploaded content?
h. Can the user request their data be removed, or is manual deletion available?
i. In what jurisdiction do the servers reside?
j. Through what jurisdictions does user data travel?
k. How is data received as feedback from users protected?
l. How does the vendor define data breach?
m. How does the vendor monitor for data breaches?
n. Has the risk of a data breach been assessed, and if so, how is the risk measured?21
o. How will users be notified of data breaches?
2. For AI systems that integrate with other applications or software, including but not limited to:
a. workplace productivity suites such as Microsoft Office or Google Workspace,
b. document management systems, or
c. billing and finance tools,
does the vendor explain whether and/or how ethical walls or other conflict or security mechanisms will carry over to ensure that users are only able to see content to which they should have access? How often will the ethical walls or other conflict or security mechanisms synchronise across the systems?
If ethical walls or other conflict or security mechanisms cannot be carried over, does the tool allow for custom security measures? How are these managed?
Privacy Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Is there a policy that governs the use of copyrighted data used by the AI system?
2. Is there a policy that governs user uploading of copyrighted content to the AI system?
3. What is the vendor doing to protect users from inadvertently breaching copyright when using content produced by the AI system?
4. Does the vendor confirm that copyrighted materials have been used with permission? If this information is not known for foundation models, the vendor should clearly disclose this.
5. Is citation or attribution made available to users along with the outputs?
6. Does the vendor allow authors and licensed content providers to opt out of having their content used by AI systems?
Copyright Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Does the vendor acknowledge the existence of bias in the AI system and the risk of perpetuating or compounding bias?
2. What is the vendor doing to highlight or mitigate bias in the AI system? If this information is not known for foundation models, the vendor should clearly disclose this. For example:
a. Does the vendor warn users that outputs may be based on biased training data or algorithms?
b. Is the vendor transparent about how sources are ranked or prioritized? Does the AI present alternative viewpoints in the authorities relied on and/or cited, particularly as it relates to secondary sources?
c. Does the vendor identify any steps taken or technological guardrails deployed to minimize algorithmic bias?
d. Does the vendor provide information on whether, and if so, how they include diverse groups in product development and testing?
e. When and how are algorithms/systems updated if outputs are found to be problematic, and what steps are taken between the discovery of problematic output and the completion of sufficient algorithm/system updates? For example:
-How often are quality control checks performed, who performs them, and what level of expertise do they have?
-What level, or type, of problem triggers updates or revisions to the algorithms/system?
-What level, or type, of problem triggers a temporary suspension of the AI system, and when and how are users notified of current or past problems (including those reported by other users)?
3. Does the vendor ensure that the AI system complies with all legislative and regulatory requirements and conforms to relevant international standards?
4. Does the AI system block some prompts? If so, under what circumstances?
Users should be particularly wary of bias in tools that purport to discern subjective characteristics of a document, for example, truthfulness, tone and sarcasm.
See also:
--Algorithmic Transparency – Bias
--Data Transparency – Bias
Bias Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
1. Does the vendor have a public-facing policy, for both authors and for commentary generated in-house, on the use of AI in generating, updating or editing commentary products?
2. Does the publisher state clearly how AI is used in generating, updating or editing commentary products?
3. In such situations, how is AI usage attributed? See for example: COAL-RJAL Editorial Group, Canadian Open Access Legal Citation Guide, Canadian Legal Information Institute, 2024 CanLIIDocs 830, https://canlii.ca/t/7nc6q section 8.
AI Generated Legal Commentary Score: 0 - 5 (0 - product doesn't meet the requirement at all. 5 - best possible score.)
Notes:
Criteria | Score | Weight |
---|---|---|
Marketing and Product Documentation | ||
Data Transparency | ||
Data Used in Training (Training Data) | ||
Algorithm Transparency | ||
User Inputs / Prompt and Prompt Engineering | ||
Outputs | ||
Other Risk Management | ||
Privacy | ||
Copyright | ||
Bias | ||
AI Generated Legal Commentary |
1 For example, the Law Societies of Alberta, BC, Manitoba, Ontario, Newfoundland and Labrador, Northwest Territories, Nova Scotia, Prince Edward Island, Quebec and Saskatchewan have all issued guidance on the use of AI in legal practice, and the Government of Canada has issued its Voluntary Code of Conduct on the Responsible Development and Management of Advanced Generative AI Systems (Innovation, Science and Economic Development Canada, September 2023), online: https://ised-isde.canada.ca/site/ised/en/voluntary-code-conduct-responsible-development-and-management-advanced-generative-ai-systems [perma.cc/EAF7-QP8R].
2 Oxford English Dictionary, (last accessed September 4, 2025) online: https://www.oed.com/dictionary/algorithm_n?tab=meaning_and_use&tl=true. [perma.cc/8FR9-P94S].
3 Government of Canada Treasury Board Secretariat, "Directive on Automated Decision Making", (last modified 2023-04-25), online: https://www.tbs-sct.canada.ca/pol/doc-eng.aspx?id=32592#appA. [perma.cc/NCS8-NHYJ].
4 As this document considers a range of AI products that rely on varied technologies, we use 'AI System' unless a topic addresses a specific type of AI model.
5 Definition derived from: Montréal Declaration on Responsible AI, (Université de Montréal, 2018), online: https://declarationmontreal-iaresponsable.com/wp-content/uploads/2023/04/UdeM_Decl-IA-Resp_LA-Declaration-ENG_WEB_09-07-19.pdf>. [perma.cc/WX8P-UD9E].
6 Amazon Web Services, “What is Data Labelling?”, (last accessed March 7, 2025), online: https://aws.amazon.com/what-is/data-labeling/#:~:text=In%20machine%20learning%2C%20data%20labeling,model%20can%20learn%20from%20it. [perma.cc/YDB9-W6DH].
7 Rina Diane Caballar, “What are foundation models?” (last accessed August 26, 2025), online: https://www.ibm.com/think/topics/foundation-models [perma.cc/8ET5-FUHF].
8 Varun Magesh et al., “Hallucination-Free? Assessing the Reliability of Leading Ai Legal Research Tools” (2025) J Empirical Legal Studies 1, online: https://doi.org/10.1111/jels.12413 [Stanford Study] at 8.
9 Britannica, (last accessed February 28, 2025), online: “large language model” (2025). https://www.britannica.com/topic/large-language-model. [perma.cc/R7G4-YYSV].
10 Adapted from IBM, “What is machine learning?”, (last accessed February 28, 2025), online: https://www.ibm.com/think/topics/machine-learning [perma.cc/J7FH-M3YQ] and UC Berkeley School of Information, “What is Machine Learning (ML)?” (June 26, 2020), online: https://ischoolonline.berkeley.edu/blog/what-is-machine-learning". [perma.cc/PST3-K73E].
11 ISO/IEC, Artificial Intelligence Concepts and Terminology, ISO/IEC 22989:2022(en) https://www.iso.org/obp/ui/#iso:std:iso-iec:22989:ed-1:v1:en. [perma.cc/3CRX-Y96Z].
12 Ibid.
13 IBM, “Supervised versus unsupervised learning: What’s the difference?”, (last accessed March 7, 2025), online: https://www.ibm.com/think/topics/supervised-vs-unsupervised-learning. [perma.cc/G2WL-39CR].
14 Amazon Web Services, “What is Reinforcement Learning?” (last accessed March 14, 2025), online: https://aws.amazon.com/what-is/reinforcement-learning [perma.cc/8N52-U3YR].
15 Adapted from IBM, “What is an AI Model?”, (last accessed March 7, 2025), online: https://www.ibm.com/think/topics/ai-model#:~:text=An%20AI%20model%20is%20a,they%27ve%20been%20programmed%20for [perma.cc/T6BK-39W9].
16 IBM, “What is NLP?”, (last accessed March 14, 2025), online: https://www.ibm.com/think/topics/natural-language-processing [perma.cc/5TBQ-W3LL].
17 Regulation of the European Parliament and of the Council of 13 June 2024 laying down harmonized rules on artificial intelligence and amending Regulations, (EC) No 300/2008, (EU) No 167/2013, (EU) No 169/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2014, PE/24/2024/REV/1, OJ L, 2024/1689, 12.7.2024, (In force 1 August 2024), online: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=OJ%3AL_202401689 [perma.cc/26Y7-HQ4L]. [EU Act].
18 Adapted from definition of “provider” in EU Act, ibid.
19 Some court practice directions and other guidelines are requiring clear attribution when AI is used to generate content. The recommended enquiries below are designed to assist with fulfilling these requirements. See for example, Practice Direction Re Use of Artificial Intelligence in Court Submissions, (Court of King's Bench of Manitoba, June 23, 2023), online: https://www.manitobacourts.mb.ca/site/assets/files/2045/practice_direction_-_use_of_artificial_intelligence_in_court_submissions.pdf [perma.cc/X4DY-595A] and Notice to the Parties and the Profession: The Use of Artificial Intelligence in Court Proceedings, (Federal Court of Canada, updated May 7, 2024), online: https://www.fct-cf.gc.ca/Content/assets/pdf/base/FC-Updated-AI-Notice-EN.pdf [perma.cc/F54B-N29G]. COAL-RJAL Editorial Group, Canadian Open Access Legal Citation Guide, Canadian Legal Information Institute, 2024 CanLIIDocs 830, online: https://canlii.ca/t/7nc6q, provides users with sample citations to use when citing AI generated outputs.
20 Some scholars refer to outputs that do not cite a source to support a statement as” ungrounded outputs”. See for example, Stanford Study supra note 8 at 6, 7 and 27.
21 See for example, Canada, Office of the Privacy Commissioner of Canada, ”Assess if a privacy breach poses a real risk of significant harm to an individual” (last accessed August 11, 2025), online: https://www.priv.gc.ca/en/privacy-topics/business-privacy/breaches-and-safeguards/privacy-breaches-at-your-business/rrosh-tool/ [perma.cc/RX5N-Z6NG].
The coding for this website was partially generated by Microsoft Co-Pilot™.
Last updated ADemers September 5, 2025.