Welcome to Inspire AI documentation portal. We welcome you as a beta user of this tool as we excitingly be working hard to improve the system and resolve bugs raised by you.
If you haven’t already, please fill in the invite form by clicking here. Doing so, will let us know that you are interested in testing our tool and we can create an account for you from the backend.
When you received your account credentials, head to ai.icubeutm.ca to start using Inspire AI. Please note that all of your data, credentials, chat history &… are being stored on a private server managed by ICUBEUTM. This means that your data will not be shared externally.
Text generation models
Text generation models, have been trained to understand natural and formal language. The inputs to these models are also referred to as “prompts”. Designing a prompt is essentially how you “program” a model, usually by providing instructions or some examples of how to successfully complete a task. Models can be used across a great variety of tasks including content or code generation, summarization, conversation, creative writing, and more.
Tokens
Text generation models process text in chunks called tokens. Tokens represent commonly occurring sequences of characters. For example, the string ” tokenization” is decomposed as ” token” and “ization”, while a short and common word like ” the” is represented as a single token. Note that in a sentence, the first token of each word typically starts with a space character. As a rough rule of thumb, 1 token is approximately 4 characters or 0.75 words for English text.
One limitation to keep in mind is that for a text generation model the prompt and the generated output combined must be no more than the model’s maximum context length. The maximum context lengths for each text generation and embeddings model can be found in the model section of the documentation.
This guide shares strategies and tactics for getting better results from large language models. The methods described here can sometimes be deployed in combination for greater effect. We encourage experimentation to find the methods that work best for you. In general, if you find that a model fails at a task and a more capable model is available, it’s often worth trying again with the more capable model.
Write clear instructions
These models can’t read your mind. If outputs are too long, ask for brief replies. If outputs are too simple, ask for expert-level writing. If you dislike the format, demonstrate the format you’d like to see. The less the model has to guess at what you want, the more likely you’ll get it.
Tactics:
Provide reference text
Language models can confidently invent fake answers, especially when asked about esoteric topics or for citations and URLs. In the same way that a sheet of notes can help a student do better on a test, providing reference text to these models can help in answering with fewer fabrications.
Tactics:
Split complex tasks into simpler subtasks
Just as it is good practice in software engineering to decompose a complex system into a set of modular components, the same is true of tasks submitted to a language model. Complex tasks tend to have higher error rates than simpler tasks. Furthermore, complex tasks can often be re-defined as a workflow of simpler tasks in which the outputs of earlier tasks are used to construct the inputs to later tasks.
Tactics:
Give the model time to “think”
If asked to multiply 17 by 28, you might not know it instantly, but can still work it out with time. Similarly, models make more reasoning errors when trying to answer right away, rather than taking time to work out an answer. Asking for a “chain of thought” before an answer can help the model reason its way toward correct answers more reliably.
Tactics:
Use external tools
Compensate for the weaknesses of the model by feeding it the outputs of other tools. For example, a text retrieval system (sometimes called RAG or retrieval augmented generation) can tell the model about relevant documents. If a task can be done more reliably or efficiently by a tool rather than by a language model, offload it to get the best of both.
Test changes systematically
Improving performance is easier if you can measure it. In some cases a modification to a prompt will achieve better performance on a few isolated examples but lead to worse overall performance on a more representative set of examples. Therefore to be sure that a change is net positive to performance it may be necessary to define a comprehensive test suite (also known an as an “eval”).
Inspire AI is powered by a diverse set of models with different capabilities and price points. You can also make customizations to our models for your specific use case by requesting a custom model via email to mohammad.tahvili@utoronto.ca
Here is a table comparing the main models we offer to each other:
Feature/Model | Llama 2 70B | Falcon 180B | ChatGPT 3.5 | ChatGPT 4.0 | Inspire AI | IDEFICS |
---|---|---|---|---|---|---|
Parameters | 70 billion | 180 billion | ~175 billion | ~175 billion | ~175 billion | 114.9 billion |
Max Context | 1024 tokens | 1024 tokens | 2048 tokens | 8192 tokens | 2048 tokens | 1024 tokens |
Modalities | Text-based | Text-based | Text-based | Text-based | Text-based | Text-based & Multi-modal |
Accuracy | High | Very High | High | Very High | High | High |
Complexity | High | Very High | High | Very High | High | High |
Speed | Medium | Medium | Fast | Fast | Slow | Medium |
Efficiency | Efficient | Efficient | Efficient | Efficient | Efficient | Medium |
Creativity | High | Very High | High | Very High | High | High |
Cost | Free | Free | $0.002 / 1K tokens | $0.03 / 1K tokens | $0.002 / 1K tokens | Free |
Keep track of changes to Inspire AI. This feature list is maintained in a best effort fashion and may not reflect all upcoming changes being made.
To start a conversation, find and press the button that represents a pen:
This will open up a blank page where you can select your preferred model and start a conversation. Finally, find the input field typically labeled as “Ask anything” or similar. Enter your query or message and press ‘Enter’ or click the Send button.
Locating Previous Conversations: On the Inspire AI interface, look for a section usually on the left side. Or on the phone, by pressing the menu icon which is on the top left corner.
Accessing a Conversation: Click on the desired conversation from the list to view the previous interactions.
Continuing a Conversation: If your query is a follow-up or related to the previous conversation, it’s best to continue in the same thread/conversation for context.
Starting a New Conversation: If your new query is unrelated to the previous topics or if you prefer a fresh start, begin a new conversation. This can be done by either navigating back to the home screen or selecting an option to start a new chat, if available.
Selecting a Conversation: Navigate to the conversation you wish to delete. You can do this by placing your mouse/cursor on the conversation from the sidebar.
Deleting the Conversation: Look for an option like ‘Delete’, ‘Remove’, or a trash bin icon. Click on it and confirm the deletion.
Accessing Model Options: Locate an options menu, when starting a new conversation. Usually titled “Current Model”:
Selecting a Different Model: Click on the Current Model options menu. This will open up a box showing all the models available. Scroll through the list to find what you are interested in. Select the one you are interested. Make sure there is a visible checkmark next to the model you are interested in. Finally press the “Apply” button to apply the model to your new conversation.
Checking the Model Information: The current model in use is typically displayed at the bottom of the chat window. There you can find a sample text like below:
Model: Llama 2 · Generated content may be inaccurate or false.
To analyze images, you need to insert the image in your prompt. At the moment we only have this feature active for one of the models called IDEFICS model.
Please first start a conversation and change the active model to IDEFICS. After changing the model, you will see a new button called “Upload Image” located at the top left corner of the chat window. By clicking on this button, you will be prompted to upload an image. After you have uploaded your desired images, please make sure to type in what you like the model to do with these images.
To analyze PDF documents, you need to insert the document in your prompt. At the moment you can do so by dragging and dropping the document into the chat window of any conversation.
Note: It is recommended to use ChatGPT 3.5 for this as the large context window will allow you to upload large PDF documents.
Information you provide:
Automatically collected information:
Introduction to Ethical AI:
Inspire AI is committed to the responsible development and use of AI technologies. We recognize the transformative potential of AI and its impact on society. Therefore, ethical considerations are at the forefront of our AI model development and deployment.
Understanding Risks and Limitations:
Large language models, like those used in Inspire AI, are advanced yet carry inherent risks. Testing predominantly in English does not encompass all possible scenarios, which means that outputs can be unpredictable and may sometimes be inaccurate, biased, or objectionable.
Bias and Fairness:
Significant research, including works like Sheng et al. (2021) and Bender et al. (2021), has highlighted bias and fairness issues in language models. Inspire AI models, derived from these, can inadvertently produce content with stereotypes or inaccuracies. We are actively working to mitigate these issues through diverse data sets and inclusive model development teams.
Red-Teaming for Quality Assurance:
Our Red-Teaming efforts focus on identifying instances where models may generate incorrect, biased, or offensive responses. This rigorous process, involving diverse scenarios and continuous updates, aims to refine our models responsibly.
Methodology:
Bias evaluation in Inspire AI is a multifaceted approach, primarily focusing on instruction-tuned model variants. Our methods include:
Intended for Research and Commercial Use:
Inspire AI models are designed for English language use in commercial and research environments.
Out-of-Scope Uses:
Sustainable and Responsible AI:
User’s Role in Ethical AI:
Transparency and Accountability:
Citing responses from an AI model, can be done in various formats depending on the citation style you are using. Below are templates for some of the most common citation styles:
In-text citation:
(Insert Model Name Here, 2023)
Reference list entry:
Insert Model Name Here. (2023). Title of the input query or topic. Retrieved from [URL of the AI platform]
In-text citation:
(Insert Model Name Here)
Works Cited entry:
“Insert Model Name Here.” Title of the input query or topic, 2023, [URL of the AI platform].
In-text citation:
(Insert Model Name Here 2023)
Bibliography entry:
Insert Model Name Here. 2023. “Title of the input query or topic.” Accessed [date]. [URL of the AI platform].
In-text citation:
(Insert Model Name Here 2023)
Reference list entry:
Insert Model Name Here, 2023. Title of the input query or topic. [Online] Available at: [URL of the AI platform] [Accessed Date].
——-
Note that the “Title of the input query or topic” should be replaced with a brief description of the content or the specific question asked. Also, if the AI model does not have a specific retrieval URL, you can omit the URL part and just mention the platform where the AI model operates.
Remember to always check with your specific citation guidelines, as there might be variations depending on the institution or publication.
If you have noticed that the platform is performing slower than usual, or if you are experiencing delays in receiving a response, please rest assured that we are working to address the issue. We currently use a 2GB memory, which may be a factor in the slower performance. However, we do have options to increase the memory and improve the overall functionality of the site. If you would like to inquire about increasing the memory or have any questions or concerns, please do not hesitate to reach out to us at the email address provided (mohammad.tahvili@utoronto.ca).
Do you see a bug that needs to be reported? Please fill in the form below so we can best understand the issue:
There are so many great ideas on where we can take this project. Feel free to share them with us by filling in the form below:
Indigenous business training program created by RedBird Circle Inc. in partnership with UofT Libraries, ICUBE UTM and The Bridge at UTSC
An experiential learning program that follows the school 3 semesters schedule, for anyone with a great idea who wants to make it happen.
A boutique-style remote program to support your business development and help you grow through one-on-one support. This program is open all year.
A student-led creative studio with intends to serve the prototyping and design needs of our ventures and small businesses in our community.
This 2-day retreat is designed to help you reflect on your journey, what drives you, all the hats you wear and what is next for you and your social enterprise.