<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="fr">
	<id>http://wiki.backprop.fr/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jboscher</id>
	<title>Wiki BackProp - Contributions [fr]</title>
	<link rel="self" type="application/atom+xml" href="http://wiki.backprop.fr/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jboscher"/>
	<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Sp%C3%A9cial:Contributions/Jboscher"/>
	<updated>2026-05-09T14:18:40Z</updated>
	<subtitle>Contributions</subtitle>
	<generator>MediaWiki 1.38.4</generator>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=AI_Agent&amp;diff=79</id>
		<title>AI Agent</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=AI_Agent&amp;diff=79"/>
		<updated>2023-09-19T09:33:16Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « AI agents are artificial entities that sense their environment, make decisions, and take actions.   The Rise and Potential of Large Language Model Based Agents: A Survey : https://arxiv.org/pdf/2309.07864.pdf »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;AI agents are artificial entities that sense their environment, make decisions, and take actions.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The Rise and Potential of Large Language Model Based Agents: A Survey : https://arxiv.org/pdf/2309.07864.pdf&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=SFT&amp;diff=78</id>
		<title>SFT</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=SFT&amp;diff=78"/>
		<updated>2023-08-01T12:10:19Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Supervised Fine-Tuning (SFT): Models are trained on a dataset of instructions and responses. It adjusts the weights in the LLM to minimize the difference between the generated answers and ground-truth responses, acting as labels.[1] &lt;br /&gt;
&lt;br /&gt;
However, in some cases, updating the knowledge of the model is not enough and you want to modify the behavior of the LLM. In these situations, you will need a supervised fine-tuning (SFT) dataset, which is a collection of prompts and their corresponding responses. SFT datasets can be manually curated by users or generated by other LLMs. Supervised fine-tuning is especially important for LLMs such as ChatGPT, which have been designed to follow user instructions and stay on a specific task across long stretches of text. This specific type of fine-tuning is also referred to as instruction fine-tuning [2] &lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32] Fine-Tune Your Own Llama 2 Model in a Colab Notebook&lt;br /&gt;
* [https://bdtechtalks.com/2023/07/10/llm-fine-tuning/] The complete guide to LLM fine-tuning&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=SFT&amp;diff=77</id>
		<title>SFT</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=SFT&amp;diff=77"/>
		<updated>2023-08-01T12:00:50Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Supervised Fine-Tuning (SFT): Models are trained on a dataset of instructions and responses. It adjusts the weights in the LLM to minimize the difference between the generated answers and ground-truth responses, acting as labels.  == Références ==  * [https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32] Fine-Tune Your Own Llama 2 Model in a Colab Notebook »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Supervised Fine-Tuning (SFT): Models are trained on a dataset of instructions and responses. It adjusts the weights in the LLM to minimize the difference between the generated answers and ground-truth responses, acting as labels.&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://towardsdatascience.com/fine-tune-your-own-llama-2-model-in-a-colab-notebook-df9823a04a32] Fine-Tune Your Own Llama 2 Model in a Colab Notebook&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=PEFT&amp;diff=76</id>
		<title>PEFT</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=PEFT&amp;diff=76"/>
		<updated>2023-07-27T09:34:26Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Thus, we, as a com- munity of researchers and engineers, need efficient ways to train on downstream task data.  Parameter-efficient fine-tuning, which we denote as PEFT, aims to resolve this problem by only training a small set of parameters which might be a subset of the existing model parameters or a set of newly added parameters.   https://arxiv.org/pdf/2303.15647.pdf »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Thus, we, as a com- munity of researchers and engineers, need efficient ways to train on downstream task data.&lt;br /&gt;
&lt;br /&gt;
Parameter-efficient fine-tuning, which we denote as PEFT, aims to resolve this problem by only training a small set of parameters which might be a subset of the existing model parameters or a set of newly added parameters. &lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/2303.15647.pdf&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Agent&amp;diff=75</id>
		<title>Agent</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Agent&amp;diff=75"/>
		<updated>2023-07-05T21:19:10Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In LangChain, agents are high-level components that use language models (LLMs) to determine which actions to take and in what order. An action can either be using a tool and observing its output or returning it to the user. Tools are functions that perform specific duties, such as Google Search, database lookups, or Python REPL.&lt;br /&gt;
Agents involve an LLM making decisions about which Actions to take, taking that Action, seeing an Observation, and repeating that until done.&lt;br /&gt;
&lt;br /&gt;
Several types of agents are available in LangChain:&lt;br /&gt;
* The zero-shot-react-description agent uses the ReAct framework to decide which tool to employ based purely on the tool&#039;s description. It necessitates a description of each tool.&lt;br /&gt;
* The react-docstore agent engages with a docstore through the ReAct framework. It needs two tools: a Search tool and a Lookup tool. The Search tool finds a document, and the Lookup tool searches for a term in the most recently discovered document.&lt;br /&gt;
* The self-ask-with-search agent employs a single tool named Intermediate Answer, which is capable of looking up factual responses to queries. It is identical to the original self-ask with the search paper, where a Google search API was provided as the tool.&lt;br /&gt;
* The conversational-react-description agent is designed for conversational situations. It uses the ReAct framework to select a tool and uses memory to remember past conversation interactions.&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Agent&amp;diff=74</id>
		<title>Agent</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Agent&amp;diff=74"/>
		<updated>2023-07-05T21:18:48Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « In LangChain, agents are high-level components that use language models (LLMs) to determine which actions to take and in what order. An action can either be using a tool and observing its output or returning it to the user. Tools are functions that perform specific duties, such as Google Search, database lookups, or Python REPL. Agents involve an LLM making decisions about which Actions to take, taking that Action, seeing an Observation, and repeating that until... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In LangChain, agents are high-level components that use language models (LLMs) to determine which actions to take and in what order. An action can either be using a tool and observing its output or returning it to the user. Tools are functions that perform specific duties, such as Google Search, database lookups, or Python REPL.&lt;br /&gt;
Agents involve an LLM making decisions about which Actions to take, taking that Action, seeing an Observation, and repeating that until done.&lt;br /&gt;
&lt;br /&gt;
Several types of agents are available in LangChain:&lt;br /&gt;
- The zero-shot-react-description agent uses the ReAct framework to decide which tool to employ based purely on the tool&#039;s description. It necessitates a description of each tool.&lt;br /&gt;
- The react-docstore agent engages with a docstore through the ReAct framework. It needs two tools: a Search tool and a Lookup tool. The Search tool finds a document, and the Lookup tool searches for a term in the most recently discovered document.&lt;br /&gt;
- The self-ask-with-search agent employs a single tool named Intermediate Answer, which is capable of looking up factual responses to queries. It is identical to the original self-ask with the search paper, where a Google search API was provided as the tool.&lt;br /&gt;
- The conversational-react-description agent is designed for conversational situations. It uses the ReAct framework to select a tool and uses memory to remember past conversation interactions.&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Temperature&amp;diff=73</id>
		<title>Temperature</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Temperature&amp;diff=73"/>
		<updated>2023-07-04T10:15:11Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple strategy might be to select the most likely word at every timestep.&lt;br /&gt;
&lt;br /&gt;
But this method can result in uninteresting and sometimes repetitive answers. On the contrary, if you randomly sample over the distribution returned by the model, you might get some unlikely responses.&lt;br /&gt;
&lt;br /&gt;
By controlling the degree of randomness, you can get more unexpected, and some might say creative, responses. Back to the model parameters, temperature is a number used to tune the degree of randomness.&lt;br /&gt;
&lt;br /&gt;
Low temperature: Means to select the words that are highly possible and more predictable. In this case, those are flowers and the other words that are located at the beginning ofthe list. This setting is generally better for tasks like q&amp;amp;a and summarization where you expect a more “predictable” answer with less variation. … High temperature: Means to select the words  that have low possibility and are more unusual. In this case, those are bugs and the other words that that are located at the end of the list. This setting is good if you want to generate more “creative” or unexpected content.&lt;br /&gt;
&lt;br /&gt;
Temperature is a number used to tune the degree of randomness. Low temperature means choosing the most likely and predictable words. For example, the word &amp;quot;flowers&amp;quot; in the sentence &amp;quot;The garden is full of beautiful__.&amp;quot; High temperature means choosing the words that have low possibility and are more unusual. For example, the word &amp;quot;bugs&amp;quot; in the sentence &amp;quot;The garden is full of beautiful__. &amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/381925&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=72</id>
		<title>Prompt design</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=72"/>
		<updated>2023-07-04T10:14:14Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The process of finding and designing the best input text to get the desired response back from the model is called prompt design, which often involves a lot of experimentation. &lt;br /&gt;
&lt;br /&gt;
Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best way to write a prompt. You may need to experiment with different structures, formats, and examples to see what works best for your use case. &lt;br /&gt;
&lt;br /&gt;
* Be concise&lt;br /&gt;
* Be specific and well-defined&lt;br /&gt;
* Ask one task at a time&lt;br /&gt;
* Turn generative tasks into classification tasks&lt;br /&gt;
* Improve response quality by including examples&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/383122&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Zero-shot&amp;diff=71</id>
		<title>Zero-shot</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Zero-shot&amp;diff=71"/>
		<updated>2023-07-04T10:13:18Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Zero-shot prompting - is a method where the LLM is given no additional data on the specifictask that it is being asked to perform. Instead, it is only given a prompt that describes the task. For example, if you want the LLM to answer a question, you just prompt &amp;quot;what is prompt design?&amp;quot;. One-shot prompting - is a method where the LLM is given a single example of the task that it is being asked to perform. For example, if you want the LLM to write a poem, you might provide a single example poem. and Few-shot prompting - is a method where the LLM is given a small number of examples of the task that it is being asked to perform. For example, if you want the LLM to write a news article, you might give it a few news articles to read. You can use the structured mode to design the few-shot prompting by providing a context and additional examples for the model to learn from. The structured prompt contains a few different components:&lt;br /&gt;
&lt;br /&gt;
First we have the context, which instructs how the model should respond. You can specify words the model can or cannot use, topics to focus on or avoid, or a particular response format. And the context applies each time you send a request to the model. Let’s say we want to use an LLM to answer questions based on some background text. In this case, a passage that describes changes in rainforest vegetation in the Amazon. We can paste in the background text as the context. Then, we add some examples of questions that could be answered from this passage Like what does LGM stand for? Or what did the analysis from the sediment deposits indicate? We’ll need to add in the corresponding answers to these questions, to demonstrate how we want the model to respond. Then, we can test out the prompt we’ve designed by sending a new question as input. And there you go, you’ve prototyped a q&amp;amp;a system based on background text in just a few minutes! Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best way to write a prompt. You may need to experiment with different structures, formats, and examples to see what works best for your use case&lt;br /&gt;
&lt;br /&gt;
- Zero-shot prompting: Provides one single command with no examples.&lt;br /&gt;
- One-shot prompting: Provides one example of the task.&lt;br /&gt;
- Few-shot prompting: Provides a few examples of the task often with the description of the context.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/383122&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=70</id>
		<title>Prompt design</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=70"/>
		<updated>2023-07-04T10:12:35Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The process of finding and designing the best input text to get the desired response back from the model is called prompt design, which often involves a lot of experimentation. &lt;br /&gt;
&lt;br /&gt;
Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best way to write a prompt. You may need to experiment with different structures, formats, and examples to see what works best for your use case. &lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/383122&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=69</id>
		<title>Prompt design</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt_design&amp;diff=69"/>
		<updated>2023-07-04T09:20:33Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best w... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best way to write a prompt. You may need to experiment with different structures, formats, and examples to see what works best for your use case. &lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/383122&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Zero-shot&amp;diff=68</id>
		<title>Zero-shot</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Zero-shot&amp;diff=68"/>
		<updated>2023-07-04T08:58:15Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Zero-shot prompting - is a method where the LLM is given no additional data on the specifictask that it is being asked to perform. Instead, it is only given a prompt that describes the task. For example, if you want the LLM to answer a question, you just prompt &amp;quot;what is prompt design?&amp;quot;. One-shot prompting - is a method where the LLM is given a single example of the task that it is being asked to perform. For example, if you want the LLM to write a poem, you might... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Zero-shot prompting - is a method where the LLM is given no additional data on the specifictask that it is being asked to perform. Instead, it is only given a prompt that describes the task. For example, if you want the LLM to answer a question, you just prompt &amp;quot;what is prompt design?&amp;quot;. One-shot prompting - is a method where the LLM is given a single example of the task that it is being asked to perform. For example, if you want the LLM to write a poem, you might provide a single example poem. and Few-shot prompting - is a method where the LLM is given a small number of examples of the task that it is being asked to perform. For example, if you want the LLM to write a news article, you might give it a few news articles to read. You can use the structured mode to design the few-shot prompting by providing a context and additional examples for the model to learn from. The structured prompt contains a few different components:&lt;br /&gt;
&lt;br /&gt;
First we have the context, which instructs how the model should respond. You can specify words the model can or cannot use, topics to focus on or avoid, or a particular response format. And the context applies each time you send a request to the model. Let’s say we want to use an LLM to answer questions based on some background text. In this case, a passage that describes changes in rainforest vegetation in the Amazon. We can paste in the background text as the context. Then, we add some examples of questions that could be answered from this passage Like what does LGM stand for? Or what did the analysis from the sediment deposits indicate? We’ll need to add in the corresponding answers to these questions, to demonstrate how we want the model to respond. Then, we can test out the prompt we’ve designed by sending a new question as input. And there you go, you’ve prototyped a q&amp;amp;a system based on background text in just a few minutes! Please note a few best practices around prompt design. Be concise Be specific and well-defined Ask one task at a time Turn generative tasks into classification tasks. For example, instead of asking what programming language to learn, ask if Python, Java, or C is a better fit for a beginner in programming. and Improve response quality by including examples. Adding instructions and a few examples tends to yield good results however there’s currently no one best way to write a prompt. You may need to experiment with different structures, formats, and examples to see what works best for your use case&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/383122&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Top_P&amp;diff=67</id>
		<title>Top P</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Top_P&amp;diff=67"/>
		<updated>2023-06-22T08:30:29Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple stra... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple strategy might be to select the most likely word at every timestep.&lt;br /&gt;
&lt;br /&gt;
But this method can result in uninteresting and sometimes repetitive answers. On the contrary, if you randomly sample over the distribution returned by the model, you might get some unlikely responses.&lt;br /&gt;
&lt;br /&gt;
By controlling the degree of randomness, you can get more unexpected, and some might say creative, responses. Back to the model parameters, temperature is a number used to tune the degree of randomness.&lt;br /&gt;
&lt;br /&gt;
Low temperature: Means to select the words that are highly possible and more predictable. In this case, those are flowers and the other words that are located at the beginning ofthe list. This setting is generally better for tasks like q&amp;amp;a and summarization where you expect a more “predictable” answer with less variation. … High temperature: Means to select the words that have low possibility and are more unusual. In this case, those are bugs and the other words that that are located at the end of the list. This setting is good if you want to generate more “creative” or unexpected content.&lt;br /&gt;
&lt;br /&gt;
In addition to adjusting the temperature, top K lets the model randomly return a word from the top K number of words in terms of possibility. For example, top 2 means you get a random word from the top 2 possible words including flowers and trees. This approach allows the other high-scoring word a chance of being selected. However, if the probability distribution of the words is highly skewed and you have one word that is very likely and everything else is very unlikely, this approach can result in some strange responses. The difficulty of selecting the best top-k value, leads to another popular approach that dynamically sets the size of the shortlist of words. &lt;br /&gt;
&lt;br /&gt;
Top P allows the model to randomly return a word from the top P probability of words. With top P, you choose from a set of words with the sum of the likelihoods not exceeding P. For example, p of 0.75 means you sample from a set of words that have a cumulative probability greater than 0.75. In this case, it includes three words: flowers, trees, and herbs.&lt;br /&gt;
&lt;br /&gt;
This way, the size of the set of words can dynamically increase and decrease according to the probability distribution of the next word on the list. In sum, Generative AI Studio provides a few model parameters for you to play with such as the model, temperature, top K, and top P. Note that, you are not required to adjust them constantly, especially top k and top p. Now let’s look at the second feature, which creates conversations.&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/381925&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Top_K&amp;diff=66</id>
		<title>Top K</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Top_K&amp;diff=66"/>
		<updated>2023-06-22T08:28:45Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple stra... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple strategy might be to select the most likely word at every timestep.&lt;br /&gt;
&lt;br /&gt;
But this method can result in uninteresting and sometimes repetitive answers. On the contrary, if you randomly sample over the distribution returned by the model, you might get some unlikely responses.&lt;br /&gt;
&lt;br /&gt;
By controlling the degree of randomness, you can get more unexpected, and some might say creative, responses. Back to the model parameters, temperature is a number used to tune the degree of randomness.&lt;br /&gt;
&lt;br /&gt;
Low temperature: Means to select the words that are highly possible and more predictable. In this case, those are flowers and the other words that are located at the beginning ofthe list. This setting is generally better for tasks like q&amp;amp;a and summarization where you expect a more “predictable” answer with less variation. … High temperature: Means to select the words that have low possibility and are more unusual. In this case, those are bugs and the other words that that are located at the end of the list. This setting is good if you want to generate more “creative” or unexpected content.&lt;br /&gt;
&lt;br /&gt;
In addition to adjusting the temperature, top K lets the model randomly return a word from the top K number of words in terms of possibility. For example, top 2 means you get a random word from the top 2 possible words including flowers and trees. This approach allows the other high-scoring word a chance of being selected. However, if the probability distribution of the words is highly skewed and you have one word that is very likely and everything else is very unlikely, this approach can result in some strange responses. The difficulty of selecting the best top-k value, leads to another popular approach thatdynamically sets the size of the shortlist of words. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/381925&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Temperature&amp;diff=65</id>
		<title>Temperature</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Temperature&amp;diff=65"/>
		<updated>2023-06-22T08:24:10Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple stra... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;First, there are different models you can choose from. Each model is tuned to perform well on specific tasks. You can also specify the temperature, top P, and top K. These parameters all adjust the randomness of responses by controlling how the output tokens are selected. When you send a prompt to the model, it produces an array of probabilities over the words that could come next. And from this array, we need some strategy to decide what to return. A simple strategy might be to select the most likely word at every timestep.&lt;br /&gt;
&lt;br /&gt;
But this method can result in uninteresting and sometimes repetitive answers. On the contrary, if you randomly sample over the distribution returned by the model, you might get some unlikely responses.&lt;br /&gt;
&lt;br /&gt;
By controlling the degree of randomness, you can get more unexpected, and some might say creative, responses. Back to the model parameters, temperature is a number used to tune the degree of randomness.&lt;br /&gt;
&lt;br /&gt;
Low temperature: Means to select the words that are highly possible and more predictable. In this case, those are flowers and the other words that are located at the beginning ofthe list. This setting is generally better for tasks like q&amp;amp;a and summarization where you expect a more “predictable” answer with less variation. … High temperature: Means to select the words  that have low possibility and are more unusual. In this case, those are bugs and the other words that that are located at the end of the list. This setting is good if you want to generate more “creative” or unexpected content.&lt;br /&gt;
&lt;br /&gt;
https://www.cloudskillsboost.google/course_sessions/3264154/video/381925&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=64</id>
		<title>LangChain</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=64"/>
		<updated>2023-05-11T04:03:50Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;LangChain est un framework qui sert à développer des applications qui s&#039;interfacent avec des LLM (Large Language Models).&lt;br /&gt;
&lt;br /&gt;
LangChain peut accéder à d&#039;autres sources de données que celles du LLM et peut interagir avec son environnement.&lt;br /&gt;
&lt;br /&gt;
LangChain makes the complicated parts of working &amp;amp; building with AI models easier. It helps do this in two ways:&lt;br /&gt;
&lt;br /&gt;
Integration - Bring external data, such as your files, other applications, and api data, to your LLMs&lt;br /&gt;
Agency - Allow your LLMs to interact with its environment via decision making. Use LLMs to help decide which action to take next [3]&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://python.langchain.com/en/latest/index.html] Documentation LangChain&lt;br /&gt;
* [https://www.youtube.com/watch?v=2xxziIWmaSA&amp;amp;list=PLqZXAkvF1bPNQER9mLmDbntNfSpzdDIU5&amp;amp;index=4] Cours LangChain&lt;br /&gt;
* [https://github.com/ia35/langchain-tutorials/blob/main/LangChain%20Cookbook%20Part%201%20-%20Fundamentals.ipynb] Notebook LangChain&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=63</id>
		<title>LangChain</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=63"/>
		<updated>2023-05-11T04:03:27Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;LangChain est un framework qui sert à développer des applications qui s&#039;interfacent avec des LLM (Large Language Models).&lt;br /&gt;
&lt;br /&gt;
LangChain peut accéder à d&#039;autres sources de données que celles du LLM et peut interagir avec son environnement.&lt;br /&gt;
&lt;br /&gt;
LangChain makes the complicated parts of working &amp;amp; building with AI models easier. It helps do this in two ways:&lt;br /&gt;
&lt;br /&gt;
Integration - Bring external data, such as your files, other applications, and api data, to your LLMs&lt;br /&gt;
Agency - Allow your LLMs to interact with it&#039;s environment via decision making. Use LLMs to help decide which action to take next&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://python.langchain.com/en/latest/index.html] Documentation LangChain&lt;br /&gt;
* [https://www.youtube.com/watch?v=2xxziIWmaSA&amp;amp;list=PLqZXAkvF1bPNQER9mLmDbntNfSpzdDIU5&amp;amp;index=4] Cours LangChain&lt;br /&gt;
* [https://github.com/ia35/langchain-tutorials/blob/main/LangChain%20Cookbook%20Part%201%20-%20Fundamentals.ipynb] Notebook LangChain&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=62</id>
		<title>LangChain</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=62"/>
		<updated>2023-05-11T03:45:02Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;LangChain est un framework qui sert à développer des applications qui s&#039;interfacent avec des LLM (Large Language Models).&lt;br /&gt;
&lt;br /&gt;
LangChain peut accéder à d&#039;autres sources de données que celles du LLM et peut interagit avec son environnement.&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://python.langchain.com/en/latest/index.html] Documentation LangChain&lt;br /&gt;
* [https://www.youtube.com/watch?v=2xxziIWmaSA&amp;amp;list=PLqZXAkvF1bPNQER9mLmDbntNfSpzdDIU5&amp;amp;index=4] Cours LangChain&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=61</id>
		<title>LangChain</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=LangChain&amp;diff=61"/>
		<updated>2023-05-11T03:30:09Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec «   == Références ==  * [https://python.langchain.com/en/latest/index.html] Documentation LangChain * [https://www.youtube.com/watch?v=2xxziIWmaSA&amp;amp;list=PLqZXAkvF1bPNQER9mLmDbntNfSpzdDIU5&amp;amp;index=4] Cours LangChain »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://python.langchain.com/en/latest/index.html] Documentation LangChain&lt;br /&gt;
* [https://www.youtube.com/watch?v=2xxziIWmaSA&amp;amp;list=PLqZXAkvF1bPNQER9mLmDbntNfSpzdDIU5&amp;amp;index=4] Cours LangChain&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=60</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=60"/>
		<updated>2023-05-02T12:43:18Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
Exemples:&lt;br /&gt;
&lt;br /&gt;
Un prompt basique, le LLM (Large Language Model) poursuit le texte commencé.&lt;br /&gt;
&lt;br /&gt;
Prompt: &lt;br /&gt;
Four score and seven years ago our&lt;br /&gt;
&lt;br /&gt;
Réponse:&lt;br /&gt;
fathers brought forth on this continent, a new nation, conceived in Liberty, and dedicated to the proposition that all men are created&lt;br /&gt;
&lt;br /&gt;
Les prompts peuvent être aussi des instructions simples ou complexes.&lt;br /&gt;
&lt;br /&gt;
Prompt:&lt;br /&gt;
Write an introduction for a weekly newsletter.&lt;br /&gt;
&lt;br /&gt;
Réponse:&lt;br /&gt;
Welcome to our weekly newsletter. This week we have a special focus on productivity. We&#039;ll be sharing tips and tricks on how to be productive in all areas of your life. We hope you enjoy and please share with your friends!&lt;br /&gt;
&lt;br /&gt;
ou dans un cas plus complexe : &lt;br /&gt;
&lt;br /&gt;
Prompt:&lt;br /&gt;
Write an introduction for a weekly newsletter to my company named Contoso. It should mention the great all hands meeting we had last week and thank the team for all their hard work over the tough past few months.&lt;br /&gt;
&lt;br /&gt;
Réponse:&lt;br /&gt;
Welcome to Contoso&#039;s weekly newsletter! We had a great all hands meeting last week and I want to thank the team for all their hard work over the tough past few months. We&#039;re looking forward to a bright future and I&#039;m confident that with our team&#039;s dedication and commitment, we will achieve great things. Thank you for being a part of our team and I look forward to working with you all in the coming weeks and months.&lt;br /&gt;
&lt;br /&gt;
On peut aussi écrire des prompts du type &amp;quot;Primary Content&amp;quot; où un texte est modifié.&lt;br /&gt;
&lt;br /&gt;
Prompt:&lt;br /&gt;
Can you please tell me how to get to the museum?&lt;br /&gt;
Translate to French:&lt;br /&gt;
&lt;br /&gt;
Réponse:&lt;br /&gt;
Pouvez-vous s&#039;il vous plaît me dire comment aller au musée?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
Ecrire un prompt, contrairement à ce qu&#039;on peut imaginer, n&#039;est pas chose facile. Il faut faire l&#039;analogie avec le piano. Obtenir des notes est facile, bien jouer est très compliqué.&lt;br /&gt;
&lt;br /&gt;
While these models are extremely powerful, their behavior is also very sensitive to the prompt. This makes prompt construction an important skill to develop.&lt;br /&gt;
Prompt construction can be difficult. In practice, the prompt acts to configure the model weights to complete the desired task, but it&#039;s more of an art than a science, often requiring experience and intuition to craft a successful prompt. [1]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;br /&gt;
* [https://www.promptingguide.ai/fr] Prompt Engineering Guide (fr)&lt;br /&gt;
* [https://www.youtube.com/watch?v=dOxUroR57xs] Prompt Engineering Overview&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions] Learn how to work with the ChatGPT and GPT-4 models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=59</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=59"/>
		<updated>2023-05-02T09:52:44Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
Ecrire un prompt, contrairement à ce qu&#039;on peut imaginer, n&#039;est pas chose facile. Il faut faire l&#039;analogie avec le piano. Obtenir des notes est facile, bien jouer est très compliqué.&lt;br /&gt;
&lt;br /&gt;
While these models are extremely powerful, their behavior is also very sensitive to the prompt. This makes prompt construction an important skill to develop.&lt;br /&gt;
Prompt construction can be difficult. In practice, the prompt acts to configure the model weights to complete the desired task, but it&#039;s more of an art than a science, often requiring experience and intuition to craft a successful prompt. [1]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;br /&gt;
* [https://www.promptingguide.ai/fr] Prompt Engineering Guide (fr)&lt;br /&gt;
* [https://www.youtube.com/watch?v=dOxUroR57xs] Prompt Engineering Overview&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions] Learn how to work with the ChatGPT and GPT-4 models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=58</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=58"/>
		<updated>2023-05-02T09:37:18Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
Ecrire un prompt, contrairement à ce qu&#039;on peut imaginer, n&#039;est pas chose facile. Il faut faire l&#039;analogie avec le piano. Obtenir des notes est facile, bien jouer est très compliqué.&lt;br /&gt;
&lt;br /&gt;
While these models are extremely powerful, their behavior is also very sensitive to the prompt. This makes prompt construction an important skill to develop.&lt;br /&gt;
Prompt construction can be difficult. In practice, the prompt acts to configure the model weights to complete the desired task, but it&#039;s more of an art than a science, often requiring experience and intuition to craft a successful prompt. [1]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;br /&gt;
* [https://www.promptingguide.ai/fr] Prompt Engineering Guide (fr)&lt;br /&gt;
* [https://www.youtube.com/watch?v=dOxUroR57xs] Prompt Engineering Overview&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=LLaMA&amp;diff=57</id>
		<title>LLaMA</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=LLaMA&amp;diff=57"/>
		<updated>2023-05-02T09:25:02Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « LLaMA (Large Language Model Meta AI) is a language model released by Meta (Facebook). It is Meta’s answer to OpenAI’s GPT models.  Like GPT, LLaMA is intended to be a general-purpose foundational model suitable for further fine-tuning. [1]  == Références ==  * [https://agi-sphere.com/llama-models/] A brief history of LLaMA models »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;LLaMA (Large Language Model Meta AI) is a language model released by Meta (Facebook). It is Meta’s answer to OpenAI’s GPT models.&lt;br /&gt;
&lt;br /&gt;
Like GPT, LLaMA is intended to be a general-purpose foundational model suitable for further fine-tuning. [1]&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://agi-sphere.com/llama-models/] A brief history of LLaMA models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=56</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=56"/>
		<updated>2023-05-02T08:56:45Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;br /&gt;
* [https://www.promptingguide.ai/fr] Prompt Engineering Guide (fr)&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Fichier:Cropped-cropped-LogoBackPropTrans-300x120-3.png&amp;diff=55</id>
		<title>Fichier:Cropped-cropped-LogoBackPropTrans-300x120-3.png</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Fichier:Cropped-cropped-LogoBackPropTrans-300x120-3.png&amp;diff=55"/>
		<updated>2023-05-01T08:02:22Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Logo BackProp transparent&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
Logo BackProp transparent&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=54</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=54"/>
		<updated>2023-05-01T07:36:49Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=53</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=53"/>
		<updated>2023-05-01T07:33:15Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Un prompt est une commande textuelle transmise afin d&#039;exécuter une action. Cette commande s&#039;adresse principalement aux LLM (Large Language Model) ou aux logiciels de génération d&#039;images entraînées pour comprendre ces commandes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs). Researchers use prompt engineering to improve the capacity of LLMs on a wide range of common and complex tasks such as question answering and arithmetic reasoning. Developers use prompt engineering to design robust and effective prompting techniques that interface with LLMs and other tools.[4]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=52</id>
		<title>Prompt</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Prompt&amp;diff=52"/>
		<updated>2023-05-01T07:29:28Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec «   == Références ==  * [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering * [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers * [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/prompt-engineering] Introduction to prompt engineering&lt;br /&gt;
* [https://www.deeplearning.ai/short-courses/chatgpt-prompt-engineering-for-developers/] ChatGPT Prompt Engineering for Developers&lt;br /&gt;
* [https://learn.microsoft.com/en-us/azure/cognitive-services/openai/concepts/advanced-prompt-engineering?pivots=programming-language-chat-completions] Prompt engineering techniques&lt;br /&gt;
* [https://github.com/dair-ai/Prompt-Engineering-Guide] Prompt Engineering Guide&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=51</id>
		<title>In-context learning</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=51"/>
		<updated>2023-04-27T16:40:37Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
[1] The in-context learning (ICL) ability is formally introduced by GPT-3 : assuming that the language model has been provided with a natural language instruction and/or several task demonstrations, it can generate the expected output for the test instances by completing the word sequence of input text, without requiring additional training or gradient update&lt;br /&gt;
&lt;br /&gt;
[[File:A comparative illustration of in-context learning (ICL) and chain-of-thought (CoT) prompting.jpg|500px]]&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=50</id>
		<title>In-context learning</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=50"/>
		<updated>2023-04-27T16:40:02Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
[1] The in-context learning (ICL) ability is formally introduced by GPT-3 : assuming that the language model has been provided with a natural language instruction and/or several task demonstrations, it can generate the expected output for the test instances by completing the word sequence of input text, without requiring additional training or gradient update&lt;br /&gt;
&lt;br /&gt;
[[File:Capture d’écran 2023-04-27 à 18.32.50.png|500px]]&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=49</id>
		<title>In-context learning</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=49"/>
		<updated>2023-04-27T16:38:44Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
[1] The in-context learning (ICL) ability is formally introduced by GPT-3 : assuming that the language model has been provided with a natural language instruction and/or several task demonstrations, it can generate the expected output for the test instances by completing the word sequence of input text, without requiring additional training or gradient update&lt;br /&gt;
&lt;br /&gt;
[[File:A comparative illustration of in-context learning (ICL) and chain-of-thought (CoT) prompting.jpg|500px]]&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Fichier:Capture_d%E2%80%99%C3%A9cran_2023-04-27_%C3%A0_18.32.50.png&amp;diff=48</id>
		<title>Fichier:Capture d’écran 2023-04-27 à 18.32.50.png</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Fichier:Capture_d%E2%80%99%C3%A9cran_2023-04-27_%C3%A0_18.32.50.png&amp;diff=48"/>
		<updated>2023-04-27T16:36:05Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : A comparative illustration of in-context learning (ICL) and chain-of-thought (CoT) prompting&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
A comparative illustration of in-context learning (ICL) and chain-of-thought (CoT) prompting&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=47</id>
		<title>In-context learning</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=In-context_learning&amp;diff=47"/>
		<updated>2023-04-27T16:35:13Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec «  [1] The in-context learning (ICL) ability is formally introduced by GPT-3 : assuming that the language model has been provided with a natural language instruction and/or several task demonstrations, it can generate the expected output for the test instances by completing the word sequence of input text, without requiring additional training or gradient update    == Références ==  * [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
[1] The in-context learning (ICL) ability is formally introduced by GPT-3 : assuming that the language model has been provided with a natural language instruction and/or several task demonstrations, it can generate the expected output for the test instances by completing the word sequence of input text, without requiring additional training or gradient update&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=46</id>
		<title>Emergent Abilities of Large Language Models</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=46"/>
		<updated>2023-04-27T16:23:12Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On entend par &amp;quot;Emergent Abilities of Large Language Models&amp;quot; une capacité présente dans un LLM qui ne se retrouve pas dans un modèle similaire mais plus petit. Ce qui veut dire aussi qu&#039;on ne peut pas prévoir (extrapoler) cette nouvelle capacité uniquement à partir de celles d&#039;un modèle plus petit.&lt;br /&gt;
&lt;br /&gt;
[1] We consider an ability to be emergent if it is not present in smaller models but is present in larger models.&lt;br /&gt;
&lt;br /&gt;
[1] We will consider the following general definition of emergence, adapted from Steinhardt (2022) and rooted in a 1972 essay called “More Is Different” by Nobel prize-winning physicist Philip Anderson&lt;br /&gt;
&lt;br /&gt;
[1] Emergence is when quantitative changes in a system result in qualitative changes in behavior.&lt;br /&gt;
&lt;br /&gt;
Les paramètres qui servent à changer d&#039;échelle pour les modèles sont : la taille du dataset, le nombre de paramètres du modèle et la puissance de calcul requise pour l&#039;entraîner (les deux derniers étant souvent corrélés).&lt;br /&gt;
&lt;br /&gt;
[2] Although scaling is mainly conducted in model size (with similar architectures and pre-training tasks), these large-sized PLMs display different behaviors from smaller PLMs (e.g., 330M-parameter BERT and 1.5B- parameter GPT-2) and show surprising abilities (called emergent abilities) in solving a series of complex tasks. &lt;br /&gt;
&lt;br /&gt;
[2] For example, GPT-3 can solve few-shot tasks through in-context learning, whereas GPT-2 cannot do well. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openreview.net/pdf?id=yzkSU5zdwD] Emergent Abilities of Large Language Models&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=45</id>
		<title>Emergent Abilities of Large Language Models</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=45"/>
		<updated>2023-04-27T15:58:23Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On entend par &amp;quot;Emergent Abilities of Large Language Models&amp;quot; une capacité présente dans un LLM qui ne se retrouve pas dans un modèle similaire mais plus petit. Ce qui veut dire aussi qu&#039;on ne peut pas prévoir (extrapoler) cette nouvelle capacité uniquement à partir de celles d&#039;un modèle plus petit.&lt;br /&gt;
&lt;br /&gt;
[1] We consider an ability to be emergent if it is not present in smaller models but is present in larger models.&lt;br /&gt;
&lt;br /&gt;
[1] We will consider the following general definition of emergence, adapted from Steinhardt (2022) and rooted in a 1972 essay called “More Is Different” by Nobel prize-winning physicist Philip Anderson&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[2] Although scaling is mainly conducted in model size (with similar architectures and pre-training tasks), these large-sized PLMs display different behaviors from smaller PLMs (e.g., 330M-parameter BERT and 1.5B- parameter GPT-2) and show surprising abilities (called emergent abilities) in solving a series of complex tasks. &lt;br /&gt;
&lt;br /&gt;
[2] For example, GPT-3 can solve few-shot tasks through in-context learning, whereas GPT-2 cannot do well. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openreview.net/pdf?id=yzkSU5zdwD] Emergent Abilities of Large Language Models&lt;br /&gt;
* [https://arxiv.org/pdf/2303.18223.pdf] A Survey of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=44</id>
		<title>Emergent Abilities of Large Language Models</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Emergent_Abilities_of_Large_Language_Models&amp;diff=44"/>
		<updated>2023-04-27T15:39:06Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « On entend par &amp;quot;Emergent Abilities of Large Language Models&amp;quot; une capacité présente dans un LLM qui ne se retrouve pas dans un modèle similaire mais plus petit. Ce qui veut dire aussi qu&amp;#039;on ne peut pas prévoir (extrapoler) cette nouvelle capacité uniquement à partir de celles d&amp;#039;un modèle plus petit.  [1] We consider an ability to be emergent if it is not present in smaller models but is present in larger models.   Although scaling is mainly conducted in mode... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;On entend par &amp;quot;Emergent Abilities of Large Language Models&amp;quot; une capacité présente dans un LLM qui ne se retrouve pas dans un modèle similaire mais plus petit. Ce qui veut dire aussi qu&#039;on ne peut pas prévoir (extrapoler) cette nouvelle capacité uniquement à partir de celles d&#039;un modèle plus petit.&lt;br /&gt;
&lt;br /&gt;
[1] We consider an ability to be emergent if it is not present in smaller models but is present in larger models.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although scaling is mainly conducted in model size (with similar architectures and pre-training tasks), these large-sized PLMs display different behaviors from smaller PLMs (e.g., 330M-parameter BERT and 1.5B- parameter GPT-2) and show surprising abilities (called emergent abilities) in solving a series of complex tasks. &lt;br /&gt;
&lt;br /&gt;
For example, GPT-3 can solve few-shot tasks through in-context learning, whereas GPT-2 cannot do well. &lt;br /&gt;
&lt;br /&gt;
Thus, the research community coins the term “large language models (LLM)”1 for these large-sized PLMs [32–35]. A remarkable application of LLMs is ChatGPT2 that adapts the LLMs from the GPT series for dialogue, which presents an amazing conversation ability with humans.&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openreview.net/pdf?id=yzkSU5zdwD] Emergent Abilities of Large Language Models&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Pre-trained_language_models&amp;diff=43</id>
		<title>Pre-trained language models</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Pre-trained_language_models&amp;diff=43"/>
		<updated>2023-04-27T14:53:16Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Les auteurs de &amp;quot;A Survey of Large Language Models&amp;quot; distinguent les &amp;quot;Pre-trained language models (PLM)&amp;quot; des Large language models (LLM).&lt;br /&gt;
&lt;br /&gt;
ELMo et BERT appartiendraient à la 1ère catégorie, un peu comme les pionniers des LLM.&lt;br /&gt;
 &lt;br /&gt;
As an early attempt, ELMo was proposed to capture context-aware word representations by first pre-training a bidirectional LSTM (biLSTM) network (instead of learning fixed word representations) and then fine-tuning the biLSTM network according to specific downstream tasks. &lt;br /&gt;
&lt;br /&gt;
Further, based on the highly parallelizable Transformer architecture [22] with self-attention mechanisms, BERT [23] was proposed by pre- training bidirectional language models with specially de- signed pre-training tasks on large-scale unlabeled corpora.&lt;br /&gt;
&lt;br /&gt;
La différence entre les PLM et les LLM c&#039;est l&#039;augmentation de taille des données et des modèles, qui a pour effet d&#039;améliorer significativement les performances sur les tâches.&lt;br /&gt;
&lt;br /&gt;
Researchers find that scaling PLM (e.g., scaling model size or data size) often leads to an improved model capacity on downstream tasks&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://arxiv.org/pdf/1802.05365.pdf] Deep contextualized word representations&lt;br /&gt;
* [https://aclanthology.org/N19-1423.pdf] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Pre-trained_language_models&amp;diff=42</id>
		<title>Pre-trained language models</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Pre-trained_language_models&amp;diff=42"/>
		<updated>2023-04-27T14:44:33Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Les auteurs de &amp;quot;A Survey of Large Language Models&amp;quot; distinguent les &amp;quot;Pre-trained language models (PLM)&amp;quot; des Large language models (LLM).  ELMo et BERT appartiendraient à la 1ère catégorie, un peu comme les pionniers des LLM.   As an early attempt, ELMo was proposed to capture context-aware word representations by first pre-training a bidirectional LSTM (biLSTM) network (instead of learning fixed word representations) and then fine-tuning the biLSTM network acco... »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Les auteurs de &amp;quot;A Survey of Large Language Models&amp;quot; distinguent les &amp;quot;Pre-trained language models (PLM)&amp;quot; des Large language models (LLM).&lt;br /&gt;
&lt;br /&gt;
ELMo et BERT appartiendraient à la 1ère catégorie, un peu comme les pionniers des LLM.&lt;br /&gt;
 &lt;br /&gt;
As an early attempt, ELMo was proposed to capture context-aware word representations by first pre-training a bidirectional LSTM (biLSTM) network (instead of learning fixed word representations) and then fine-tuning the biLSTM network according to specific downstream tasks. &lt;br /&gt;
&lt;br /&gt;
Further, based on the highly parallelizable Transformer architecture [22] with self-attention mechanisms, BERT [23] was proposed by pre- training bidirectional language models with specially de- signed pre-training tasks on large-scale unlabeled corpora.&lt;br /&gt;
&lt;br /&gt;
La différence entre les PLM et les LLM c&#039;est l&#039;augmentation de taille des données et des modèles, qui a pour effet d&#039;améliorer significativement les performances sur les tâches.&lt;br /&gt;
&lt;br /&gt;
Researchers find that scaling PLM (e.g., scaling model size or data size) often leads to an improved model capacity on downstream tasks&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Fichier:Cropped-LogoBackPropTrans-2.png&amp;diff=41</id>
		<title>Fichier:Cropped-LogoBackPropTrans-2.png</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Fichier:Cropped-LogoBackPropTrans-2.png&amp;diff=41"/>
		<updated>2023-04-27T13:31:55Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Logo BackProp&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
Logo BackProp&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Fichier:Shamentality_artificial_intelligence_beautiful_textures_and_ter_545b2bf2-aece-4268-b49e-78dee7f78862.png&amp;diff=40</id>
		<title>Fichier:Shamentality artificial intelligence beautiful textures and ter 545b2bf2-aece-4268-b49e-78dee7f78862.png</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Fichier:Shamentality_artificial_intelligence_beautiful_textures_and_ter_545b2bf2-aece-4268-b49e-78dee7f78862.png&amp;diff=40"/>
		<updated>2023-04-27T13:29:41Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : AI logo by MidJourney&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
AI logo by MidJourney&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=Reinforcement_Learning_from_Human_Feedback&amp;diff=39</id>
		<title>Reinforcement Learning from Human Feedback</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=Reinforcement_Learning_from_Human_Feedback&amp;diff=39"/>
		<updated>2023-01-22T21:59:21Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : Page créée avec « Reinforcement Learning from Human Feedback (RLHF)  https://huggingface.co/blog/rlhf »&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Reinforcement Learning from Human Feedback (RLHF)&lt;br /&gt;
&lt;br /&gt;
https://huggingface.co/blog/rlhf&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=38</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=38"/>
		<updated>2023-01-12T16:21:08Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg&lt;br /&gt;
&lt;br /&gt;
Voir le notebook à ce sujet : CLIP&lt;br /&gt;
https://colab.research.google.com/drive/1kr0xnHQZ7G-cvlFgl2DECmLUbpgvswcy?usp=sharing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=37</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=37"/>
		<updated>2023-01-10T16:26:48Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=36</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=36"/>
		<updated>2023-01-10T16:26:21Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:https://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=35</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=35"/>
		<updated>2023-01-10T16:21:23Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:https://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=34</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=34"/>
		<updated>2023-01-10T15:55:57Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=33</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=33"/>
		<updated>2023-01-10T15:55:17Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=32</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=32"/>
		<updated>2023-01-10T15:54:55Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:https://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=31</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=31"/>
		<updated>2023-01-10T15:54:14Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
... text text text&lt;br /&gt;
[[File:http://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
text text text ...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
	<entry>
		<id>http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=30</id>
		<title>CLIP</title>
		<link rel="alternate" type="text/html" href="http://wiki.backprop.fr/index.php?title=CLIP&amp;diff=30"/>
		<updated>2023-01-10T15:53:30Z</updated>

		<summary type="html">&lt;p&gt;Jboscher : &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CLIP pour (Contrastive Language–Image Pre-training) a été introduit par OpenAi en 2021.&lt;br /&gt;
&lt;br /&gt;
L&#039;objectif est de fournir un text (caption) pour une nouvelle image présentée lors de la phase d&#039;inférence.&lt;br /&gt;
&lt;br /&gt;
Cette méthode CLIP permet de contourner certains problèmes de l&#039;apprentissage supervisé, par exemple, ceux liés aux coûts des Datasets (voir [[ImageNet]]), à leur limitation de prédiction à l&#039;ensemble des catégories entrainées, et à leur faible performance (parfois).&lt;br /&gt;
&lt;br /&gt;
L&#039;apprentissage décrit par OpenAi s&#039;est fait avec 256 GPU durant 2 semaines.&lt;br /&gt;
Il semble que 400M de paires(caption/image) aient été utilisées pour l&#039;apprentissage. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;quot; CLIP is an open source, multi-modal, zero-shot model. Given an image and text descriptions, the model can predict the most relevant text description for that image, without optimizing for a particular task.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[[File:http://wiki.backprop.fr/images/1*ag6qUFmmXAr4E410Ll-eSQ.jpeg|caption]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Références ==&lt;br /&gt;
&lt;br /&gt;
* [https://openai.com/blog/clip/ CLIP: Connecting Text and Images]&lt;br /&gt;
* [https://www.kdnuggets.com/2021/03/beginners-guide-clip-model.html CLIP: A Beginner’s Guide to the CLIP Model]&lt;br /&gt;
* [https://github.com/openai/CLIP CLIP sur GitHub]&lt;br /&gt;
* [https://towardsdatascience.com/clip-the-most-influential-ai-model-from-openai-and-how-to-use-it-f8ee408958b1 CLIP: The Most Influential AI Model From OpenAI — And How To Use It]&lt;/div&gt;</summary>
		<author><name>Jboscher</name></author>
	</entry>
</feed>